SHARE
Now, we will focus on the Anyverse Sensor Simulation Pipeline, and later in the following chapters, we will dive deep into each stage.
Meet Anyverse’s camera sensor simulation pipeline
Let’s start with the basics: what does a generic camera sensor look like? It basically consists of three main blocks.

The light coming from the scene goes through the optical system. Then light reaches the sensor surface where it is converted into a voltage, and finally into a digital value. All this happens in the imaging sensor. The digital value coming from the imaging sensor is converted into a final RGB image in the image processor block.
It is very important to highlight the fact that light (energy) is simulated as an electromagnetic wave all the way across the rendering pipeline to its final digital values. For instance, light sources in Anyverse platform emit using a characteristic spectrum profile that depends on the type of light source (LED, incandescent bulb, sky, etc).
The materials’ behavior is also wavelength-dependent, this allows for typical effects like dispersion. This feature is key and absolutely necessary for an accurate camera sensor simulation.
Anyverse Camera Sensor Simulation eBook
Learn everything you need to know about our camera sensor simulation pipeline
Camera sensor pipeline detailed
Now is when the fun starts, let’s play with the light!
As we said, if the light is physically characterized we can see (and simulate) how it behaves interacting with the different components of the pipeline to create an image.
Typical camera mountings include some filters between the optics and the sensor surface. These filters work in the spectrum domain performing different transformations on the incoming light.
In the picture below you can see a common set of filters used.

The low-pass filter and the infrared cut-off filter remove part of the spectrum in the infrared range because our eyes are not sensitive to them. The color filter array (CFA) is another key component of the camera mounting. A light photodetector detects light intensity with no wavelength specificity, therefore color information can’t be separated. The CFA separates color information before reaching the sensor’s surface.
Don’t miss the next chapter
Now it’s time to go through and discover each stage of the Anyverse camera sensor simulation pipeline. Don’t miss the third chapter of this insights series to know more about the optical system, or put another way, what happens inside the camera when light hits the lens.
Read other chapters >>>
About Anyverse™
With Anyverse™, you can accurately simulate any camera sensor and help you decide which one will perform better with your perception system. No more complex and expensive experiments with real devices, thanks to our state-of-the-art photometric pipeline.
Need to know more?
Visit our website, anyverse.ai anytime, or our Linkedin, Instagram, and Twitter profiles.
SHARE