Category Archives: Workflow

What the hell is gamma anyway – video online now.

Last week I did an almost 2 hour long Webinar with Visual Impact on Gamma and gamma curves. The webinar was recorded and you can view the full session by following the link below.

Black and White LUT set.

This was asked for on one of the online groups I follow. It’s a simple Black and White LUT that gives a wide dynamic range black and white image. The LUT has been designed to give a pleasing and gentle highlight roll-off with a reasonably contrasty mid range. I decided not to go too contrasty via the LUT so that more contrast can be added in post if needed. The output is legal range (broadcast safe). There is a single camera LUT for use in camera as well as a set of exposure compensated LUT’s going from -2 to plus 2 stops in half stop steps.

As always (to date at least) I offer these as a free download available by clicking at the bottom of the page. It takes time to create them and money to host them. I feel that this LUT set is worth $5.00 and would really appreciate that being paid if you find the LUT’s useful. But I will let you pay what you feel is fair, all contributions are greatly appreciated and it really does help keep this website up and running. If you can’t afford to pay, then just download the LUT’s and enjoy using them. If in the future you should choose to use them on a paying project, please remember where you got them and come back and make a contribution. More contributions means more LUT offerings in the future.

Please feel free to share a link to this page if you wish to share these LUT’s with anyone else or anywhere else.

To make a contribution please use the drop down menu here, there are several contribution levels to choose from.


Your choice:



Click here to download AC Black and White LUT’s V1

SEE ALSO:

Film Emulation LUT’s set 1.

Venice Look LUT’s V3 (lower contrast and minus green)

Venice Look LUT’s V1 (high contrast)

HLG HDR LUT for FS7, F5 and F55.

WYSIWYG LUT’s (Designed to be baked in)

Adobe still can’t get XAVC levels right!

I’m often asked at the various workshops I run why I don’t grade in Adobe Premiere. Here’s why – they can’t even get basic import levels right.

Below are two screen grabs. The first is from Adobe Premiere CC 2019 and shows an ungraded, as shot, HLG clip. Shot with a Sony Z280 (love that little camera). Note how the clip appears over grossly exposed with a nuclear looking sky and clipped snow, it doesn’t look nice. Also note that the waveform suggest the clips peak levels exceed 110%. Now I know for a fact that if you shoot HLG with any Sony camera white will never exceed 100%.

Incorrect levels with an XAVC clip in Adobe Premier. Click on the image to view a larger version.

The second screen grab is from DaVinci Resolve and it shows the same clip. Note how in Resolve that although bright the clip certainly doesn’t look over exposed as it does in Premiere. Note also how the levels show by the waveform now no longer exceeds code value 869 (100% white is 940).  These are the correct and expected levels, this is how the clip is supposed to look. Not the utter nonsense that Adobe creates.

Same XAVC clip in Resolve and now the levels are correct. Click on the image to view a larger version.

Why can’t Adobe get this right. This problem has existed for ages and it really screws up your footage. If you are using S-Log and you try to add a LUT then things get even worse as the LUT expects the correct levels, not these totally incorrect levels.

Take the SDI or raw out from the camera and record a ProRes file on something like a Shogun while recording XAVC internally and the two files look totally different in Premiere but they look the same in Resolve. Come on Adobe – you should be doing better than this.

If they can’t even bring clips in at the correct levels, what hope is there of being able to get a decent grading output? I can make the XAVC clips look OK in Premiere but I have to bring the levels down to do this. I shouldn’t have to. I exposed it right when I shot it so I expect it to look right in my edit software.

ProRes Raw Over Exposure Magic Tricks – It’s all smoke and mirrors!

There are a lot of videos circulating on the web right now showing what appears to be some kind of magic trick where someone has shot over exposed, recorded the over exposed images using ProRes Raw and then as if by magic made some adjustments to the footage and it goes from being almost nothing but a white out of over exposure to a perfectly exposed image.

This isn’t magic, this isn’t raw suddenly giving you more over exposure range than you have with log, this is nothing more than a quirk of the way FCP-X handles ProRes Raw material.

Before going any further – this isn’t a put-down of raw or ProRes raw. It’s really great to be able to take raw sensor data and record that with only minimal processing. There are a lot of benefits to shooting with raw (see my earlier post showing all the extra data that 12 bit raw can give). But a magic ability to let you over expose by seemingly crazy amounts isn’t something raw does any better than log.

Currently to work with ProRes Raw you have to go through FCP-X. FCP-X applies a default sequence of transforms to the Raw footage to get it from raw data to a viewable image. These all expect the footage to be exposed exactly as per the camera manufacturers recommendations, with no leeway. Inside FCP-X it’s either exposed exactly right, or it isn’t.

The default decode settings include a heavy highlight roll-off. Apple call it “Tone Mapping”. Fancy words used to make it sound special but it’s really no different to a LUT or the transforms and processes that take place in other raw decoders. Like a LUT it maps very specific values in the raw data  to very specific output brightness values. So if you shoot just a bit bright – as you would often do with log to improve the signal to noise ratio – The ProRes raw appears to be heavily over exposed. This is because anything bright ends up crushed into nothing but flat white by the default highlight roll off that is applied by default.

In reality the material is probably only marginally over exposed, maybe just one to 2 stops which is something we have become used to doing with log. When you view brightly exposed log, the log itself doesn’t look over exposed, but if you apply a narrow high contrast 709 LUT to it, it then the footage looks over exposed until you grade it or add an exposure compensated LUT.  This is what is happening by default inside FCP-X, a transform is being applied that makes brightly exposed footage look very bright and possibly over exposed – because thats the way it was shot!

This is why in FCP-X  it is typical to change the color library to WCG (Wide Color Gamut) as this changes the way FCP-X processes the raw, changing the Tone Mapping and most importantly getting rid of the highlight roll off. With no roll-off, highlights and any even slight over exposure will still blow out as you can’t show 14 stops on a conventional 6 stop TV or monitor. Anything beyond the first 6 stops will be lost, the image will look over exposed until you grade or adjust the material to control the brighter parts of the image and bring them back into a viewable range. When you are in WCG mode in FCP-X the there is no longer a highlight roll off crushing the highlights and now because they are not crushed they can be recovered, but there isn’t any more highlight range than you would have if you shot with log on the same camera!

None of this is some kind of Raw over exposure magic trick as is often portrayed. It’s simply not really understanding how the workflow works and appreciating that if you shoot bright – well it’s going to look bright – until you normalise it in post. We do this all the time with log via LUT’s and grading too! It can be a little more straight forward to recover highlights from Linear Raw footage as comes form an FS5 or FS7 compared to log. That’s because of the way log maintains a constant data level in each highlight stop and often normal grading and colour correction tools don’t deal with this correctly. The highlight range is there, but it can be tricky to normalise the log without log grading tools such as the log controls in DaVinci Resolve.

Another problem is the common use of LUT’s on log footage. The vast majority of LUT’s add a highlight roll off, if you try to grade the highlights after adding a LUT with a highlight roll off it’s going to be next to impossible to recover the highlights. You must do the highlight recovery before the LUT is added or use a LUT that has compensation for any over exposure. All of these things can give the impression that log has less highlight range than the raw from the same camera. This is not normally the case, both will be the same as it’s the sensor that limits the range.

The difference in the highlight behaviour is in the workflows and very often both log and raw workflows are miss-understood. This can lead to owners and users of these cameras thinking that one process has more than the other, when in reality there is no difference, it’s appears to be different because the workflow works in a different way.

ProRes Raw Webinar – How to use ProRes Raw with the FS5 and FS7

Last week I presented a Webinar in conjunction with Visual Impact on how to shoot ProRes Raw with Sony’s FS5 and FS7. The Webinar was recorded, so if you missed it you can now watch it online. It’s almost 2 hours long and contains what I hope is a lot of useful information including what you need, exposure and how to get the footage in to FCP-X. I tried to structure the FCP-X part presentation in such a way that those that don’t normally use FCP-X (like me) will be able to get started quickly and understand what is going on under the hood.

Since the webinar it has been brought to my attention by Felipe Baez (thanks Felipe) that it is possible to add a LUT after the color panel and grading tools by adding the “custom LUT” effect to your clip. To do this you will set the raw input conversion to S-log3. Then add your color correction, then add the Custom LUT effect.

A big thank you to Visual Impact for making this possible, do check them out!

Here is the link to the video of the webinar:

 

ProRes Raw and Atomos Inferno and Sumo – BIG deal for the FS5 and FS7!!

Over the last few days there have been various rumours and posts coming from Apple about how they intend to get back to providing decent support for professional users of their computers. Apple have openly admitted that the Trash Can Mac Pro has thermal problems and as a result has become a dead end design, which is why there haven’t been any big updates to the flagship workstation from Apple. Apple have hinted that new workstations are on the way, although it would seem that we won’t see these until next year perhaps.
Another announcement came out today, a new version of FCP-X is to be released which includes support for a new ProRes codec called ProRes Raw. This is BIG!

PRORES RAW.

Raw recordings can be made from certain cameras that have bayer sensors such as the Sony FS5 and FS7. Recording the raw data from the sensor maximises your post production flexibility and normally offers the best possible image quality from the camera. Currently if you record 4K raw with these cameras using an Atomos Shogun or similar the bit rate will be close to 3Gb/s at 24p. These are huge files and the cDNG format used to record them is difficult and clunky to work with.  As a result most users take the raw output from the camera and transform it to S-Log2 or S-Log3 and record it as 10 bit ProRes on the external recorder. This is a bit of a shame as going from 12 bit linear raw to 10 bit S-log means you are not getting the full benefit of the raw output.

Enter ProRes Raw:  ProRes Raw will allow users to record the cameras raw output at a much reduced bit rate with no significant of quality. There are two versions, ProRes Raw and ProRes Raw HQ. The HQ bit rate is around 1Gb/s at 24fps. This is not significantly bigger than the ProRes HQ (880Mb/s) that most users are using now to record the raw, yet the full benefit of 12 bit linear will be retained. A 1TB SSD will hold around an hour of ProRes Raw, compare that to uncompressed raw where you only get around 20 mins and you can see that this is a big step forwards for users of the FS5 in particular.

ProRes Raw (the non HQ version) is even smaller! The files are smaller than typical ProRes HQ files. This is possible because recording raw is inherently more efficient than recording component video.

It is claimed by Apple that ProRes Raw will play back in real time on MacBook Pro’s and iMacs without any additional rendering or external graphics cards, so it obviously isn’t terribly processor intensive. This is excellent news! Within FCP-X the playback resolution can be decreased to bring improved playback performance in less powerful systems or mutistream playback.

It looks like you will be able to record from a 4K DCI  from an FS5 or FS7 at up to 60fps continuously. This breaks through the previous limits for the Shogun of 30fps. The FS7 will be able to record 2K raw at up to 240fps and the FS5 will be able to record 4K raw at 100 and 120fps for 4 seconds. Other raw cameras are also supported by the Atomos recorders at differing frame sizes and frame rates.

At the moment the only recorders listed as supporting ProRes Raw are the Atomos Shogun Inferno and the Sumo19 and it looks like it will be a free update. In addition the DJI Inspire 2 drone and Zenmuse X7 Super 35mm camera will also support ProRes Raw.

Whether you will be able to use ProRes Raw in other applications such as Resolve or Premiere is unclear at this time. I hope that you can (or at least will be able to in the near future).

SEE: Apple Press Release.

SEE: Apple ProRes Raw White Paper

SEE: ATOMOS ProRes Raw INFO PAGE.

 

 

DaVinci Resolve, ACES and the “Sony Raw” input transform.

A quick heads up for users of Resolve with Sony Raw and X-OCN. Don’t make the same mistake I have been making. For some time I have been unhappy with the way the Sony raw looked in DaVinci Resolve and ACES prior to grading. Apparently there used to be a small problem with the raw input transform that could lead to a red/pink hue getting added to the footage. This problem was fixed some time ago. You should now not use the the “Sony Raw” input transform, if you do, it will tint your Raw or X-OCN files slightly pink/red. Instead you should select “no transform”. With no transform selected my images look so much nicer and match Sony’s own Raw Viewer so much better. Thanks to Nick Shaw of Antler Post for helping me out on this and all on the CML list.

Banding in your footage. What Causes It, is it even there?

Once again it’s time to put pen to paper or fingers to keyboard as this is a subject that just keeps coming up again and again.

People really seem to have a lot of problems with banding in footage and I don’t really fully understand why as it’s something I only ever really encounter if I’m pushing a piece of material really, really hard in post production. General the vast majority of the content I shoot does not exhibit problematic banding, even the footage I shoot with 8 bit cameras.

First things first – Don’t blame it on the bits. Even an 8 bit recording  (from a good quality camera) shouldn’t exhibit noticeable banding. An 8 bit recording can contain up to 13 million tonal values. It’s extremely rare for us to shoot luma only, but even if you do it will still have 235 shades and these steps in standard dynamic range are too small for most people to discern so you shouldn’t ever be able to see them. I think that when most people see banding they are not seeing teeny, tiny almost invisible steps what most people see is something much more noticeable – so where is it coming from?

It’s worth considering at this stage that most TV’s, monitors and computer screens are only 8 bit, sometimes less! So if you are looking at one camera and it’s banding free and then you look at another and you see banding, in both cases you are probably looking at an 8 bit image, so it can’t just be the capture bit depth that causing the problem as you cant see 10 bit steps on an 8 bit monitor.

So what could it be?

A very common cause of banding is compression. DCT based codecs such as Jpeg, MJPEG, H264 etc break the image up into small blocks of pixels called macro blocks. Then all the pixels in each block is processed in a similar manner and as a result sometimes there may be a small step between each block or between groups of blocks across a gradient. This can show up as banding. Often we see this with 8 bit codecs because typically 8 bit codecs use older technology or are more highly compressed. It’s not because there are not enough code values. Decreasing the compression ratio will normally eliminate the stepping.

Scaling between bit depths or frame sizes is another very common cause of banding. It’s absolutely vital that you ensure that your monitoring system is up to scratch. It’s very common to see banding in video footage on a computer screen as the video data levels are different to computer data levels and in addition there may also be some small gamma differences so the image has to be scaled on the fly. In addition computer desktops runs at one bit range, the HDMI output another, so all kinds of conversions are taking place that can lead to all kinds of problems when you go from a video clip, to computer levels, to HDMI levels. See this article to fully understand how important it is to get your monitoring pipeline properly sorted. https://www.xdcam-user.com/2017/06/why-you-need-to-sort-out-your-post-production-monitoring/

Look Up Tables (LUT’s) can also introduce banding. LUT’s were never really intended to be used as a quick fix grade, the intention was to use them as an on-set reference or guide, not the final output. The 3D LUT’s that we typically use for grading break the full video range into bands and each band will apply a slightly different correction to the footage than the band above or below. These bands can show up as steps in the LUT’s output, especially with the most common 17x17x17 3D LUT’s. This problem gets even worse if you apply a LUT and then grade on top – a really bad practice.

Noise reduction – In camera or postproduction noise reduction will also often introduce banding. Very often pixel averaging is used to reduce noise. If you have a bunch of pixels that are jittering up and down taking an average value for all those pixels will reduce the noise, but then you can end up with steps across a gradient as you jump from one average value to the next. If you shoot log it’s really important that you turn off any noise reduction (if you can) when you are shooting because when you grade the footage these steps will get exaggerated. Raising the ISO (gain) in a camera also makes this much worse as the cameras built in NR will be working harder, increasing the averaging to compensate the increased noise.

Coming back to 8 bit codecs again – Of course a similar quality 10 bit codec will normally give you more picture information than an 8 bit one. But we have been using 8 bits for decades, largely without any problems. So if you can shoot 10 bit you might get a better end result. But also consider all the other factors I’ve mentioned above.

 

Can DaVinci Resolve steal the edit market from Adobe and Apple.

I have been editing with Adobe Premiere since around 1994. I took a rather long break from Premiere between 2001 and 2011 and switched over to Apple and  Final Cut Pro which in many ways used to be very similar to Premiere (I think some of the same software writers were used for FCP as Premiere). My FCP edit stations were always muti-core Mac Towers. The old G5’s first then later on the Intel Towers. Then along came FCP-X. I just didn’t get along with FCP-X when it first came out. I’m still not a huge fan of it now, but will happily concede that FCP-X is a very capable, professional edit platform.

So in 2011 I switch back to Adobe Premiere as my edit platform of choice. Along the way I have also used various versions of Avid’s software, which is another capable platform.

But right now I’m really not happy with Premiere. Over the last couple of years it has become less stable than it used to be. I run it on a MacBook Pro which is a well defined hardware platform, yet I still get stability issues. I’m also experiencing problems with gamma and level shifts that just shouldn’t be there. In addition Premiere is not very good with many long GOP codecs. FCP-X seems to make light work of XAVC-L compared to Premiere. Furthermore Adobe’s Media encoder which once used to be one of the first encoders to get new codecs or features is now lagging behind, Apples Compressor now has the ability to do at he full range of HDR files. Media Compressor can only do HDR10. If you don’t know, it is possible to buy Compressor on it’s own.

Meanwhile DaVinci Resolve has been my grading platform of choice for a few years now. I have always found it much easier to get the results and looks that I want from Resolve than from any edit software – this isn’t really a surprise as after all that’s what Resolve was originally designed for.

DaVinci Resolve a great grading software and it’s edit capabilities are getting better and better.

The last few versions of Resolve have become much faster thanks to some major processing changes under the hood and in addition there has been a huge amount of work on Resolves edit capabilities. It can now be used as a fully featured edit platform. I recently used Resolve to edit some simpler projects that were going to be graded as this way I could stay in the same software for both processes, and you know what it’s a pretty good editor. There are however a few things that I find a bit funky and frustrating in the edit section of Resolve at the moment. Some of that may simply be because I am less familiar with it for editing than I am Premiere.

Anyway, on to my point. Resolve is getting to be a pretty good edit platform and it’s only going to get better. We all know that it’s a really good and very powerful grading platform and with the recent inclusion of the Fairlight audio suite within Resolve it’s pretty good at handling audio too. Given that the free version of Resolve can do all of the edit, sound and grading functions that most people need, why continue to subscribe to Adobe or pay for FCP-X?

With the cost of the latest generations of Apple computers expanding the price gap between them and similar spec Windows machines – as well as the new Macbooks lacking built in ports like HDMI, USB3 that we all use every day (you now have to use adapters and dongles). The  Apple eco system is just not as attractive as it used to be. Resolve is cross platform, so an Mac user can stay with Apple if they wish, or move over to Windows or Linux whenever they want with Resolve. You can even switch platforms mid project if you want. I could start an edit on my MacBook and the do the grade on a PC workstation staying with Resolve through the complete process.

Even if you need the extra features of the full version like very good noise reduction, facial recognition, 4K DCI output or HDR scopes then it’s still good value as it currently only costs $299/£229 which is less than a years subscription to Premiere CC.

But what about the rest of the Adobe Creative suite? Well you don’t have to subscribe to the whole suite. You can just get Photoshop or After Effects. But there are also many alternatives. Again Blackmagic Design have Fusion 9 which is a very impressive VFX package used for many Hollywood movies and like Resolve there is also a free version with a very comprehensive tools set or again for just $299/£229 you get the full version with all it’s retiming tools etc.

Blackmagic Designs Fusion is a very impressive video effects package for Mac and PC.

For a Photoshop replacement you have GIMP which can do almost everything that Photoshop can do. You can even use Photoshop filters within GIMP. The best part is that GIMP is free and works on both Mac’s and PC’s.

So there you have it – It looks like Blackmagic Design are really serious about taking a big chunk of Adobe Premiere’s users. Resolve and Fusion are cross platform so, like Adobe’s products it doesn’t matter whether you want to use a Mac or a PC. But for me the big thing is you own the software. You are not going to be paying out rather a lot of money month on month for something that right now is in my opinion somewhat flakey.

I’m not quite ready to cut my Creative Cloud subscription yet, maybe on the next version of Resolve. But it won’t be long before I do.

Use the cameras media check to help ensure you don’t get file problems.

Any of the Sony cameras that use SxS or XQD cards include a media check and media restore function that is designed to detect any problems with your recording media or the files stored on that media.
However the media check is only normally performed when you insert a card into the camera, it is not done when you eject a card as the camera never knows when you are about to do that.
So my advice is: When you want to remove the card to offload your footage ensure you have a green light next to the card, this means it should be safe to remove. Pop the card out as you would do normally but then re-insert the card and wait for the light to go from red, back to green. Check the LCD/VF for any messages, if there are no messages, take the card out and do your offload as normal.
 
Why? Every time you put an XQD or SxS card into the camera the card and files stored on it are checked for any signs of any issues. If there is a problem the camera will give you a “Restore Media” warning. If you see this warning always select OK and allow the camera to repair whatever the problem is. If you don’t restore the media and you then make a copy from the card, any copy you make will also be corrupt and the files may be inaccessible.
Once the files have been copied from the card it is no longer possible to restore the media.  If there is a problem with the files on the card, the restore can only be done by the camera, before offload. So this simple check that takes just a few seconds can save a whole world of hurt. I wish there was a media check button you could press to force the check, but there isn’t. However this method works.
It’s also worth knowing that Catalyst Browse and the old Media Browser software performs a data integrity check if you directly attach an SxS card or XQD card to the computer and access the card from the software. If a problem is found you will get a message telling you to return the media to the camera and perform a media restore. But if this is some time after the shoot and you don’t have the camera to hand, this can be impossible. Which is why I like to check my media in the camera by re-inserting it back into the camera so that it gets checked for problems before the end of the shoot.