Tag Archives: sensor

When is 4k really 4k, Bayer Sensors and resolution.

When is 4k really 4k, Bayer Sensors and resolution.

First lets clarify a couple of term. Resolution can be expressed two ways. It can be expressed as pixel resolution, ie how many individual pixels are there on the sensor. Or as TV lines or TVL/ph, or how many individual lines can I see. If you point a camera at a resolution chart, what you talking about is at what point can I no longer discern one black line from the next. TVL/ph is also the resolution normalised for the picture height, so aspect ratio does not confuse the equation. TVL/ph is a measure of the actual resolution of the camera system.  With video cameras TVL/ph is the normally quoted term, while  pixel resolution or pixel count is often quoted for film replacement cameras. I believe the TVL/ph term to be prefferable as it is a true measure of the visible resolution of the camera.
The term 4k started in film with the use af 4k digital intermediate files for post production and compositing. The exposed film is scanned using a single row scanner that is 4,096 pixels wide. Each line of the film is scanned 3 times, once each through a red, green and blue filter, so each line is made up of three 4K pixel scans, a total of just under 12k per line. Then the next line is scanned in the same manner all the way to the bottom of the frame. For a 35mm 1.33 aspect ratio film frame (4×3) that equates to roughly 4K x 3K. So the end result is that each 35mm film frame is sampled using 3 (RGB) x 4k x 3k, or 36 million samples. That is what 4k originally meant, a 4k x 3k x3 intermediate file.
Putting that into Red One perspective, it has a sensor with 8 Million pixels, so the highest possible sample size would be 8 million samples. Red Epic 13.8 million. But it doesn’t stop there because Red (like the F3) use a Bayer sensor where the pixels have to sample the 3 primary colours. As the human eye is most sensitive to resolution in the middle of the colour spectrum, twice as many of these pixel are used for green compared to red and blue. So you have an array made up of blocks of 4 pixels, BG above GR.
Now all video cameras (at least all correctly designed ones) include a low pass filter in the optical path, right in front of the sensor. This is there to prevent moire that would be created by the fixed pattern of the pixels or samples. To work correctly and completely eliminate moire and aliasing you have to reduce the resolution of the image falling on the sensor below that of the pixel sample rate. You don’t want fine details that the sensor cannot resolve falling on to the sensor, because the missing picture information will create strange patterns called moire and aliasing.
It is impossible to produce an Optical Low Pass Filter that has an instant cut off point and we don’t want any picture detail that cannot be resolved falling on the sensor, so the filter cut-off must start below the sensor resolution. Next we have to consider that a 4k bayer sensor is in effect a 2K horizontal pixel Green sensor combined with a 1K Red and 1K Blue sensor, so where do you put the low pass cut-off? As information from the four pixels in the bayer patter is interpolated, left/right/up/down there is some room to have the low pass cut off above the 2k pixel of the green channel but this can lead to problems when shooting objects that contain lots of primary colours.  If you set the low pass filter to satisfy the Green channel you will get strong aliasing in the R and B channels. If you put it so there would be no aliasing in the R and B channels the image would be very soft indeed. So camera manufacturers will put the low pass cut-off somewhere between the two leading to trade offs in resolution and aliasing. This is why with bayer cameras you often see those little coloured blue and red sparkles around edges in highly saturated parts of the image. It’s aliasing in the R and B channels. This problem is governed by the laws of physics and optics and there is very little that the camera manufacturers can do about it.
In the real world this means that a 4k bayer sensor cannot resolve more than about 1.5k to 1.8k TVL/ph without serious aliasing issues. Compare this with a 3 chip design with separate RGB sensors. With a three 1920×1080 pixel sensors, even with a sharp cut-off  low pass filter to eliminate any aliasing in all the channels you should still get at 1k TVL/ph. That’s one reason why bayer sensors despite being around since the 70s and being cheaper to manufacture than 3 chip designs (with their own issues created by big thick prisms) have struggled to make serious inroads into professional equipment. This is starting to change now as it becomes cheaper to make high quality, high pixel count sensors allowing you to add ever more pixels to get higher resolution, like the F35 with it’s (non bayer) 14.4 million pixels.
This is a simplified look at whats going on with these sensors, but it highlights the fact that 4k does not mean 4k, in fact it doesn’t even mean 2k TVL/ph, the laws of physics prevent that. In reality even the very best 4k pixels bayer sensor should NOT be resolving more than 1.8k TVL/ph. If it is it will have serious aliasing issues.
After all that, those that I have not lost yet are probably thinking: well hang on a minute, what about that film scan, why doesn’t that alias as there is no low pass filter there? Well two things are going on. One is that the dynamic structure of all those particles used to create a film image, which is different from frame to frame reduces the fixed pattern effects of the sampling, which causes the aliasing to be totally different from frame to frame so it is far less noticeable. The other is that those particles are of a finite size so the film itself acts as the low pass filter, because it’s resolution is typically lower than that of the 4k scanner.

Micro 4/3, Super 35, DSLR and the impact on traditional Pro Camcorders.

Micro 4/3, Super 35, DSLR and the impact on traditional Pro Camcorders.

I was asked by one of this blogs readers about my thoughts on this. It’s certainly a subject that I have spent a lot of time thinking about. Traditionally broadcast and television professionals have used large and bulky cameras that use 3 sensors arranged around a prism to capture the 3 primary colours. The 3 chip design gives excellent colour reproduction, full resolution images of the very highest quality. It’s not however without it’s problems. First its expensive matching 3 sensors and accurately placing them on a high precision prism made from very exotic glass. That prism also introduces image artefacts that have to be dealt with by careful electronic processing. The lenses that have to work with these thick prisms also require very careful design.
Single sensor colour cameras are not something new. I had a couple of old tube cameras that produced colour pictures from a single tube. Until recently, single chip designs were always regarded as inferior to multi-chip designs. However the rise of digital stills photography forced manufacturers to really improve the technologies used to generate a colour image from a single sensor. Sony’s F35 camera used to shoot movies and high end productions is a single chip design with a special RGB pixel matrix. The most common method used is by a single sensor is a bayer mask which places a colour filter array in front of the individual pixels on the sensor. Bayer sensors now rival 3 chip designs in most respects. There is still some leakage of colours between adjacent pixels and the colour separation is not as precise as with a prism, but in most applications these issues are extremely hard to spot and the stills pictures coming from DSLR’s speak for themselves.
A couple of years ago Canon really shook things up by adding video capabilities to some of their DSLR’s. Even now (at the time of writing at least) these are far from perfect as they are at the end of the day high resolution stills cameras so there are some serious compromises to the way video is done. But the Canons do show what can be done with a low cost single chip camera using interchangeable lenses. The shallow depth of field offered by the large near 35mm size sensors (video cams are normally 2/3?, 1/2? or smaller) can be very pleasing and the lack of a prism makes it easier to use a wide range of lenses. So far I have not seen a DSLR or other stills camera with video that I would swap for a current pro 3 chip camera, but I can see the appeal and the possible benefits. Indeed I have used a Canon DSLR on a couple of shoots as a B camera to get very shallow DoF footage.
Sony’s new NEX-VG10 consumer camcorder was launched a couple of weeks ago. It has the shape and ergonomics of a camcorder but with the sensor and lenses of a 4/3? stills camera. I liked it a lot, but there is no zoom rocker and for day to day pro use it’s not what I’m looking for. Panasonic and Sony both have professional large sensor cameras in the pipelines and it’s these that could really shake things up.
While shallow DoF is often desirable in narrative work, for TV news and fast action its not so desirable. When you are shooting the unexpected or something thats moving about a lot you need to have some leeway in focus. So for many applications a big sensor is not suitable. I dread to think what TV news would look like if it was all shot with DSLR’s!
Having said that a good video camera using a big sensor would be a nice piece of kit to have for those projects where controlling the DoF is beneficial.
What I am hoping is that someone will be clever enough to bring out a camera with a 35mm (or thereabouts) sized sensor that has enough resolution to allow it to be used with DSLR (or 4/3) stills camera lenses but also be windowed down and provided with an adapter to take 2/3? broadcast lenses without adding a focal length increase. This means that the sensor needs to be around 8 to 10 Mega Pixels so that when windowed down use just the center 2/3? and it still has around 3 million active pixels to give 1920×1080 resolution (you need more pixels than resolution with a bayer mask). This creates a problem though when you use the full sensor as the readout of the sensor will have to be very clever to avoid the aliasing issues that plague the current DSLR’s as you will have too much resolution when you use the full sensor. Maybe it will come with lens adapters that will have to incorporate optical low pass filters to give the correct response for each type of lens.
A camera like this would, if designed right dramatically change the industry. It would have a considerable impact on the sales of traditional pro video cameras as one camera could be used for everything from movie production to TV news. By using a single sensor (possibly a DSLR sensor) the cost of the camera should be lower than a 3 chip design. If it has a 10 MP sensor then it could also be made 3D capable through the use of a 3D lens like the 4/3? ones announced by Panasonic. These are exciting time we live in. I think the revolution is just around the corner. Having said all of this, I think it’s also fair to point out while you and I are clearly interested in the cutting edge (or bleeding edge) there are an awful lot of producers and production companies that are not, preferring traditional, tried and tested methods. It takes them years to change and adapt, just look at how long tape is hanging on! So the days of the full size 2/3? camera are not over yet, but those of us that like to ride the latest technology wave have great things to look forward to.

Why is Sensor Size Important: Part 2, Diffraction Limiting

Another thing that you must consider when looking at sensor size is something called “Diffraction Limiting”. For Standard Definition this is not as big a problem as it is for HD. With HD it is a big issue.

Basically the problem is that light doesn’t always travel in straight lines. When a beam of light passes over a sharp edge it gets bent, this is called diffraction. So when the light passes through the lens of a camera the light around the edge of the iris ring gets bent and this means that some of the light hitting the sensor is slightly de-focussed. The smaller you make the iris the greater the percentage of diffracted light with respect to non diffracted light. Eventually the amount of diffracted and thus de-focussed light will become large enough to start to soften the image.

With a very small sensor even a tiny amount of diffraction will bend the light enough to fall on the pixel adjacent to the one it’s supposed to be focussed on. With a bigger sensor and bigger pixels the amount of diffraction required to bend the light to the next pixel is greater. In addition the small lenses on cameras with small sensors means the iris will be smaller.

In practice, this means that an HD camera with 1/3? sensors will noticeably soften if it is more stopped down (closed) more than f5.6, 1/2? cameras more than f8 and 2/3? f11. This is one of the reasons why most pro level cameras have adjustable ND filters. The ND filter acts like a pair of sunglasses cutting down the amount of light entering the lens and as a result allowing you to use a wider iris setting. This softening happens with both HD and SD cameras, the difference is that with the low resolution of SD it was much less noticeable.

Why Is Sensor Size Important: Part 1.


Over the next few posts I’m going to look at why sensor size is important. In most situations larger camera sensors will out perform small sensors. Now that is an over simplified statement as there are many things that effect sensor performance, including continuing improvements in the technologies used, but if you take two current day sensors of similar resolution and one is larger than the other, the larger one will usually outperform the smaller one. Not only will the sensors themselves perform differently but other factors come in to play such as lens design and resolution, diffraction limiting and depth of field, I’ll look at those in subsequent posts, for today I’m just going to look at the actual sensor itself.

Pixel Size:

Pixel size is everything. If you have two sensors with 1920×1080 pixels and one is a 1/3? sensor and the other is a 1/2? sensor then the pixels themselves on the larger 1/2? sensor will be bigger. Bigger pixels will almost always perform better than smaller pixels. Why? Think of a pixel as a bucket that captures photons of light. If you relate that to a bucket that captures water, consider what happens if you put two buckets out in the rain. A large bucket with a large opening will capture more rain than a small bucket.

Small pixels capture less light each

Bigger pixels capture more light each.

It’s the same with the pixels on a CMOS or CCD sensor, the larger the pixel, the more light it will capture, so the more sensitive it will be. Taking that analogy a step further if the buckets are both of the same depth the large bucket will be able to hold more water before it overflows. It’s the same with pixels, a big pixel can store more charge of electrons before it overflows (photons of light get converted into electrical charge within the pixel). This increases the dynamic range of the sensor as a large pixel will be able to hold a bigger charge before overflowing than a small pixel.

Noise:

All the electronics within a sensor generate electrical noise. In a sensor with big pixels which is capturing more photons of light per pixel than a smaller sensor, the ratio of light captured to electrical noise is better, so the noise is less visible in the final image, in addition the heat generated in a sensor will increase the amount of unwanted noise. A big sensor will dissipate any heat better than a small sensor, so once again the big sensor will normally have a further noise advantage.

So as you can see, in most cases a large sensor has several electronic advantages over a smaller one. In the next post I will look at some of the optical advantages.

The Great CMOS Debate. More on CMOS vs CCD.

Until a couple of years ago CMOS sensors were definitely the underdog, they tended to be very noisy due to electrical noise generated the on chip by the readout circuits and A/D converters. In addition they lacked sensitivity due to the electronics on the face of the chip leaving less room for the light sensitive parts. Today, on chip noise reduction has made it possible to produce CMOS sensors with very low noise and micro lenses and better design has mitigated most of the sensitivity problems. In terms of a static image there is very little difference between a CMOS sensor and a CCD sensor. Dynamic range is remarkably similar (both types of sensor use essentially the same light gathering methods), in some respects CMOS has the edge as they are less prone to overload issues. CCD’s are very expensive to manufacture as the way they are read out requires near lossless transfer of minute charges through a thousand or more (for HD) memory cells. The first pixel to be read passes down through over 1000 memory cells, if it was to loose 5% of it’s charge in each cell, the signal would be seriously reduced by the time it left the chip. The last pixel to be read out only passes through one memory cell, so it would be less degraded, this variation could ruin an image making it uneven. Although there is more electronics on a CMOS sensor, as each pixel is read directly a small amount of loss in the transfer is acceptable as each pixel would have a similar amount of loss. So the chips are easier to make as although the design is more complex, it is less demanding and most semiconductor plants can make CMOS sensors while CCD needs much more specialised production methods. Yes, CMOS sensors are more prone to motion artifacts as the sensor is scanned from top to bottom, one pixel at a time (A CCD is read in it’s entirety just about instantaneously). This means that as you pan, at the start of the pan the top of the sensor is being read and as the pan progresses the scan moves down the chip. This can make things appear to lean over and it’s known as skew. The severity of the skew is dependent on the readout speed of the chip. Stills cameras and mobile phone cameras suffer from terrible skew as they typically have very slow readout speeds, the sensors used in an EX have a much higher readout speed and in most real world situations skew is not an issue. However there may be some circumstances where skew can cause problems but my experience is that these are few and far between. The other issue is Flash Banding. Again this is caused by the CMOS scan system. As a flash gun or strobe light is of very short duration compared to the CMOS scan it can appear that only part of the frame is illuminated by the flash of light. You can reduce the impact of Flash Banding by shooting at the slowest possible shutter speed (for example shooting 25P or 24P with no shutter) but it is impossible to completely eliminate. When I shoot lightning and thunderstorms I often use a 2 frame shutter, shooting this way I get very few partial bolts of lightning, maybe 1 in 50. If you shoot interlace then you can use the Flash Band removal tool in Sony’s Clip Browser software to eliminate flash gun problems. CMOS sensors are becoming much more common in high end cameras. Arri’s new Alexa film replacement camera uses a CMOS sensor rated at 800asa with 13 stops of latitude. Red uses CMOS as does SI2K. Slumdog Millionaire (SI2K) was the first electronically shot film to get an Oscar for cinematography, so certainly CMOS has come a long way in recent years. CMOS is here to stay, it will almost certainly make bigger and bigger inroads at higher levels. Read speeds will increase and skew etc will become less of an issue. IMHO skew is not an issue to loose sleep over with the EX’s anyway. I shoot all sorts from hurricanes and tornadoes to fast jets and race cars. I have yet to come across a shot spoilt by skew, generally motion blur tends to mask any skew long before it gets noticeable. If you shoot press conferences or red carpet events where flash guns will be going off, then you may prefer a CCD camera as this is harder to deal with, but the EXs are such good value for the money and bring many other advantages such as lower power and less weight that you have to look at the bigger picture and ask what you expect from your budget.