Cameras are a vital technology in scientific imaging, working as an interface between light and a digital image. There are a number of steps involved to convert photons of signal from your sample to the image you see on your computer monitor, each step has variables and factors that can change the ways in which images are generated. Bit depth is one of these variables, by understanding how it can affect your images and improve your experiments, you can perform more efficient and informed imaging research. The journey from light to an image is displayed in Fig.1:
This process is the same for all camera technologies, but changes in each of these steps can optimize the end result. Some important camera factors to consider before discussing bit depth are full well capacity and dynamic range.
Full Well Capacity and Dynamic Range
As photons are converted to electrons by a sensor pixel, the electrons are stored in the silicon substrate of the pixel, known as a ‘well’. In different camera models or different modes in the same camera, these pixels have a different full well capacity, which is the maximum number of electrons they can store and still display as an image. Some cameras offer full wells up to 80,000 electrons, meaning extremely bright samples can still be displayed, whereas some are much lower (100-1000), meaning they are suited to lower signal levels. Full well should be considered if your sample level is very high (brightfield imaging), if your sample can get significantly brighter and change over time, or if you are attempting to image bright and dim objects in one image. However, most fluorescent applications have low signal levels and are suited to a lower full well capacity.
The dynamic range of a camera is related to the full well capacity, and it describes the ratio between the highest and lowest signals that can be displayed. This is calculated by simply dividing the full well capacity by the read noise, as these represent the maximum and minimum readable signals respectively. Dynamic range helps to analyze the change in your sample, how do you know your signal has doubled if the sensor cannot capture it?
This article describes another factor of scientific cameras that affects both the full well capacity and the dynamic range, namely the camera sensor bit depth, as well as how to best match your application and signal level to a suitable bit depth.
What Is Bit Depth?
Camera sensors convert photons into electrons, and these electrons are amplified into a voltage, which is the ‘analog’ signal. This signal goes through an analog-to-digital converter (ADC) and is converted into a ‘digital’ signal in order to be displayed as an image on a computer. The majority of scientific cameras are monochrome, meaning that the digital signal comes in the form of a grey level, ranging from pure black to pure white. The more intense the analog signal, the whiter the grey level, meaning that fluorescent images are typically displayed as a grey-white signal on a dark black background.
Bit Depth and Scaling
Historical Bit Depths
Most scientific cameras operate at 8 bit or above, but different camera technologies have historically used different bit depths. The oldest camera technology, CCDs, used 8 bit in a different way. Classical 8 bit CCDs would scale to the camera full well instead of the signal, meaning that CCDs with around a 10,000 electron full well would be distributed amongst only 256 grey levels, giving you around 40-50 electrons per grey level. If a signal increased by 30 electrons, there would be no change in the grey level, meaning these 8 bit CCDs could not accurately represent the signal from the sample. This made it challenging to use these CCDs for fluorescence or low light imaging, and 12 bit CCDs became the standard from 2000-2010.
After CCDs came electron-multiplying CCDs (EMCCDs), which multiplied electron signals up to 1000x in order to increase the signal-to-noise ratio (SNR) for more sensitive imaging. Due to this multiplication signal levels were very high, meaning that EMCCDs were suited to a higher bit depth of 16 bit, spreading the signal over 65,000 grey levels. EMCCDs were well adopted by the scientific community due to the increased imaging speed and sensitivity compared to CCDs, and for a time the accepted bit depth for scientific imaging was 16 bit.
Due to the expense, EM decay, excess noise factor, low speeds, small field of view, and low resolutions, EMCCDs are being superseded by sCMOS sensors as a major imaging technology for the present day. As sCMOS sensors don’t use electron multiplication, received signals are not amplified and do not necessarily need the full range offered by 16 bit. As a result, sCMOS cameras can offer a more flexible range of bit depths.
Our Prime series of back-illuminated sCMOS cameras allow researchers to choose between 12 bit and 16 bit, with the majority using 12 bit due to the greater speed and most fluorescent imaging applications not needing a 16 bit range. Our newest camera, the Kinetix, goes one step further and offers 8, 12, and 16 bit modes. While early CCDs scaled 8 bit to the well, the Kinetix scales 8 bit to the signal, meaning that it can accurately detect and represent low light fluorescent signals at extremely high speeds and with half the file size. Our Prime and Kinetix cameras allow for flexible imaging with a bit depth that can be tailored to the signal level in the experiment. At this range of bit depths, data is being detected and presented, but it is vital for your imaging applications that your data is matched to depth. By changing from a 16 bit mode to 12 or 8 bit mode, there a many benefits.
Factors Affected By Bit Depth
With sCMOS cameras offering a range of bit depths for different applications, it is good to know what happens when bit depth is changed and the benefits to imaging.
At lower bit depths there are less possible grey levels to convert the signal to, this makes the conversion process much faster. With our sCMOS cameras, the 12 bit modes run at double the readout speed compared to 16 bit modes, which makes 12 bit essential when high speeds are required. For example, the Prime 95B can only access the full 80 fps at full frame (1.4 megapixels) when in 12 bit mode.
A further drop to 8 bit allows for extremely high speeds, for certain applications such as calcium/voltage imaging or when working with live, dynamic samples when speeds of >1000 frames per second are needed. The Kinetix operates at 500 fps at full frame (10 megapixels) when in 8 bit mode, allowing researchers to capture more than ever before. For high-speed applications, 16 bit mode is a limiter on the acquisition speed, especially as signal levels are likely to be low due to the low exposure times needed to achieve high-speed imaging.
In computer storage terms, 8 bits make up 1 byte, with storage in terms of megabytes, gigabytes, terabytes, etc. This means that 8 bit images take up one byte of data, and bit depths above 8 require more than one byte, with 12 and 16 bit images being scaled up to two bytes. This means that images taken with 12 or 16 bit have double the file size compared to 8 bit images. The Kinetix can not only acquire images at extremely high speeds in 8 bit mode, but data storage and transfer become far easier.
When taking high-speed or time-lapse acquisitions across multiple wavelengths or XY positions, experiments can accumulate vast numbers of images. Transferring or analyzing these large datasets takes up a large amount of computational processing power and time, at considerable inconvenience to the researcher. By dropping to 8 bit, images will be half the file size and become far more manageable. This results in easier data transfer, an issue which is becoming more and more important as camera speeds and sensor sizes continue to increase.
The dynamic range of a scientific camera is essentially the sensor’s ability to read very bright and very dark signals together. If a camera has a dynamic range of 4000 : 1 (ratio of highest to lowest detectable signals), it can only detect 4000 levels of signal, and using a 16 bit mode with 65,536 grey levels would be inefficient, this dynamic range is better suited to a 12 bit mode (4096 grey levels). Higher bit depth does not translate directly into a higher dynamic range, higher bit depths simply slice the signal more finely and do not affect the ratio between the highest and lowest brightnesses that a sensor can detect.
While bit depth dictates the number of available grey levels for the signal to be distributed amongst, the gain is the factor that determines how many electrons per grey level. If the gain is 1x, an increase in signal of 1 electron results in an increase in the grey levels. With sCMOS cameras, different bit depths often have access to different gain states. This is important for imaging as certain gain states have different read noise, making a selection of bit depth and gain state vital for low signal levels.
Signal Level Considerations
sCMOS cameras offer a range of different bit depths, but it is important to know which is best for your application. Every imaging experiment is unique, working with different samples, light levels, and speeds. All of these factors affect which bit depth is right for your application. The main piece of information to know is the signal level, here are some factors to consider:
- Brightfield or Fluorescence? Fluorescent signals are several magnitudes below brightfield signals (see Fig.4), meaning that most fluorescent imaging experiments have a lower signal level.
- Is your sample live or fixed? High-intensity light sources can induce phototoxicity and adverse effects on live samples, and there is a far smaller range of fluorescent probes/markers available for live samples when compared to fixed cells/tissues. Live samples typically produce lower signal levels.
- Are you imaging at high speed? In order to achieve a high acquisition framerate, low exposure times are needed. Only by exposing for 1 ms or less can speeds of 1000 fps (1kHz) be reached, this means the camera has very little time to collect the signal. High-speed imaging typically results in lower signal levels.
- Confocal or widefield? By their nature laser-scanning and spinning disk confocal systems use pinholes to block out-of-focus light and illuminate small sections of the sample, this light-rejection technique typically results in lower signal levels compared to widefield techniques, where the entire sample is illuminated.
- Visible light, UV, or infrared? Scientific cameras are typically most sensitive at wavelengths in the visible light spectrum. If your imaging involves ultraviolet, near-infrared, or infrared imaging, acquired signals will lower due to decreased camera quantum efficiency at these wavelengths.
Imaging fixed cells with a high-intensity laser and a widefield system would result in a far higher level of signal when compared to a high-speed spinning disk confocal system that images live cells or model organisms. By knowing your typical level of signal in your experiments, you can match this up to an optimal bit depth.
With fluorescent applications, the data should be matched to bit depth. The lower signal levels are better suited to 8 or 12 bit, by imaging low signals at 16 bit you are limiting imaging speed and unnecessarily increasing file size. Have a look at your application and signal level, and determine which bit depth is best for your imaging.
Bit depth is an important factor involved when imaging with scientific cameras. By changing how the signal is presented, imaging can be done more efficiently: lower bit depths allow for greater imaging speed, smaller file sizes, and more efficient data transfer, with no downsides. With fluorescent imaging delivering typically low signal levels, see how your imaging systems could adapt to lower bit depths.