Facebook Live stream Thursday 26th March 4pm GMT/UTC on how to stream to Facebook and YouTube with the Sony PXW-FS5 (also applies to many other Sony cameras with similar streaming options).
I will show you how to connect the camera to a network via Wi-Fi, how to send the stream from the camera to a computer. I will show you how to set up VLC to receive the stream from the camera and then how to use OBS to convert the FS5’s stream (via VLC) and send it to YouTube.
The “How To” live stream will be on my Facebook page: https://www.facebook.com/alister.chapman.9. But if you also have YouTube you will be able to see the stream from the FS5 once it is connected and setup. Links will be shared during the presentation.
Do you have an FS5 and want to stream to Facebook or YouTube? It’s actually fairly straight forward and you don’t even need to buy anything extra! You can even connect a couple of FS5’s to a single computer and switch between them.
How do you do it?
First you will need to download and install two pieces of free software on your computer. The first is VLC. VLC is an open source video player but it also has the ability to act as a media server that can receive the UDP video streams that the FS5 sends and convert them into a live video clip on the computer. The computer and the camera will both need to be connected to the same wifi network and you will need to enter the IP address of the computer into the streaming server settings in the FS5. By connecting the FS5 to your computer via the network you can use VLC to decode the UDP stream . Go to “file” “open network” and click on “open RTP/UDP stream” and enter the computers IP address and the stream port, you should then save the FS5 stream as a playlist in VLC.
OBS is a clever open source streaming application that can convert any video feed connected to a computer into a web stream. From within OBS you can set the signal source to VLC and then the stream from the FS5 will become one of the “scenes” or inputs that OBS can stream to Facebook, YouTube etc.
For multi-camera use a different port for each of the UDP streams and then in VLC save each stream as a different playlist. Then each playlist can be attached to a different scene in OBS so that you can switch. cut and mix between them.
As noted in my previous post there can be some issues with the way ProRes is recorded on many external monitors as a legal range files rather than Data Range.
Another side effect of this is that LUT’s designed for post production as well as most camera LUT’s don’t work correctly in the monitor. So even when you apply the same LUT in the camera as in the monitor the images look different.
To address this I am providing here 2 sets of LUTs for S-Log3 and SGamut3.cine designed to match the built in s709 and 709(800) Luts included in many Sony cameras. These LUTs are specifically for external recorders and should not be used in camera. When you use these LUT’s the pictures on the monitor should now match the the images in the cameras viewfinder when the built in LUT has been applied.
You will find 3 LUTs of each type. One for the base exposure, one for footage exposed 1 stop brighter (minus1) and one for footage exposed 2 stops brighter than base (minus2).
As always (to date at least) I offer these as a free download available by clicking on the links below. Try them before you decide then pay what you feel is fair. All contributions are greatly appreciated and it really does help keep this website up and running. If you can’t afford to pay, then just download the LUT’s and enjoy using them, tell your friends and send them here. If in the future you should choose to use them on a paying project, please remember where you got them and come back and make a contribution. More contributions means more LUT offerings in the future.
I have written about this many times before, but I’ll try to be a bit more concise here.
So – You have recorded S-Log2 or S-Log3 on your Sony camera and at the same time recorded on an external ProRes Recorder such as an Atomos, Blackmagic or other ProRes recorder. But the pictures look different and they don’t grade in the same way. It’s a common problem. Often the external recording will look more contrasty and when you add a LUT the blacks and shadow areas come out very differently.
Video signals can be recorded using a several different data ranges. S-Log2 and S-Log3 signals are always Data Range. When you record in the camera the cameras adds information to the recording called metadata that tells your editing or grading software that the material is Data Range. This way the edit and grading software knows how to correctly handle the footage and how to apply any LUT’s.
However when you record to an external recorder the external recorder doesn’t have this extra metadata. So the recorder will record the Data Range signal that comes from the camera but it doesn’t add the metadata. The ProRes codec is normally used for Legal Range video and by default, unless there is metadata that says otherwise, edit and grading software will assume any ProRes recordings to be Legal Range.
So what happens is that your edit software takes the file, assumes it’s Legal Range and handles it as a Legal Range file when in fact the data in the file is Data Range. This results in the recording levels being transposed into incorrect levels for processing. So when you add a LUT it will look wrong, perhaps with very dark shadows or very bright over exposed looking highlights. It can also limit how much you can grade the footage.
What Can We Do About It?Premiere CC.
You don’t need to do anything in Premiere for the internal .mp4 or MXF recordings. They are handled correctly but Premiere isn’t handling the ProRes files correctly.
My approach for this has always been to use the legacy fast color corrector filter to transform the input range to the required output range. If you apply the fast color corrector filter to a clip you can use the input and output level sliders to set the input and output range. In this case we need to set the output black level to CV16 (as that is legal range black) and we need to set output white to CV235 to match legal range white. If you do this you will then see that the external recording appears to have almost exactly the same values as the internal recording. However there is some non-linearity in the transform, it’s not quite perfect.
Now when you apply a LUT the picture and the levels are more or less what you would expect and almost identical to the internal recordings. I say almost because there is a slight hue shift. I don’t know where the hue shift comes from. In Resolve the internal and external recordings look pretty much identical and there is no hue shift. In Premiere they are not quite the same. The hue is slightly different and I don’t know why. My recommendation – use Resolve, it’s so much better for anything that needs any form of grading or color correction.
It’s very easy to tell Resolve to treat the clips as Data Range recordings. In the media bin, right click on the clip and under “clip attributes” change the input range from “auto” to “full”. If you don’t do this DaVinci Resolve will assume the ProRes file to be legal range and it will scale the clip incorrectly in the same way as Premiere does. But if you tell Resolve the clip is full range then it is handled correctly.
This comes up again and again, hence why I am writing about it once again.
Raw should never be converted to log before recording if you want any benefit from the raw. You may as well just record the 10 bit log that most cameras are capable of internally. Or take log and output it via the cameras 10 bit output (if it has one) and record that directly on the ProRes recorder. It doesn’t matter how you do it but if you convert between different recording types you will always reduce the image quality and this is as bad a way to do it as you can get. This mainly relates to cameras like the PXW-FS7. The FS5 is different because it’s internal UHD recordings are only 8 bit, so even though the raw is still compromised by converting it to ProRes log, this can still be better than the internal 8 bit log.
S-Log like any other log is a compromise recording format. Log was developed to squash a big dynamic range into the same sized recording bucket as would normally be used for conventional low dynamic range gammas. It does this by discarding a lot of tonal and textural information from everything brighter than 1 stop above middle grey, instead of the amount of data doubling for each stop up you go in exposure, it’s held at a constant amount. Normally this is largely transparent as human vision is less acute in the highlight range, but it is still a compromise.
The idea behind Linear raw is that it should give nothing away, each stop SHOULD contain double the data as the one below. But if you only have 12 bit data that would only allow you to record 11 stops of dynamic range as you would quickly run out of code values. So Sony have to use floating point math or something very similar to reduce the size of each stop by diving down the number of code values each stop has. This has almost no impact on highlights where you start off with 100’s or 1000’s values but in the shadows where a stop may only have 8 or 16 values dividing by 4 means you now only have 2 or 4 tonal levels. So once again this is a compromise recording format. To record a big dynamic range using linear what you really need is 16 bit data.
In summary so far:
S-Log reduces the number of highlight tonal values to fit it a big DR in a normal sized bucket.
Sony’s FSRaw, 12 Bit Linear reduces the number of tonal Values across the entire range to fit it in a compact 12 bit recording bucket, but the assumption is that the recording will be at least 12 bit. The greatest impact of the reduction is in the shadows.
Convert 12 bit linear to 10 bit S-Log and now you are compromising both the highlight range and the shadow range. You have the worst of both, you have 10 bit S-Log but with much less shadow data than the S-log straight from the camera. It’s really not a good thing to do and the internally generated S-Log won’t have shadows compromised in the same way.
If you have even the tiniest bit of under exposure or you attempt to lift the shadows in any way this will accentuate the reduced shadow data and banding is highly likely as the values become stretched even further apart as you bring them up the output gamma range.
If you expose brightly and then reduce the shadows this has the effect of compressing the values closer together or pushing them further down the output curve, closing them together as they go down the output gamma range, this reduces banding. This is one of the reasons why exposing more brightly can often help both log and raw recordings. So a bit of over exposure might help, but any under exposure is really, really going to hurt. Again, you would probably be better off using the internally generated S-Log.
To make matters worse there is also often an issue with S-Log in a ProRes file.
If all that is not enough there is also a big problem in the way ProRes files record S-Log. S-Log should always be recorded as full range data. When you record an internal XAVC file the metadata in the clips tells the edit or grading software that the file is full range. Then when you apply a LUT or do your grading the correct transforms occur and all shadow textures are preserved. But ProRes files are by default treated as legal range files. So when you record full range S-Log inside a ProRes file there is a high likelihood that your edit or grading software will handle the data in the clip incorrectly and this too can lead to problems in the shadows including truncated data, clipping and banding, even though the actual recorded data may be OK. This is purely a metadata issue, grading software such as DaVinci resolve can be forced to treat the ProRes files as full range.
It’s a common problem. You are shooting a performance or event where LED lighting has been used to create dramatic coloured lighting effects. The intense blue from many types of LED stage lights can easily overload the sensor and instead of looking like a nice lighting effect the blue light becomes an ugly splodge of intense blue that spoils the footage.
Well there is a tool hidden away in the paint settings of many recent Sony cameras that can help. It’s called “adaptive matrix”.
When adaptive matrix is enabled, when the camera sees intense blue light such as the light from a blue LED light, the matrix adapts to this and reduces the saturation of the blue colour channel in the problem areas of the image. This can greatly improve the way such lights and lighting look. But be aware that if trying to shoot objects with very bright blue colours, perhaps even a bright blue sky, if you have the adaptive matrix turned on it may desaturate them. Because of this the adaptive matrix is normally turned off by default.
If you want to turn it on, it’s normally found in the cameras paint and matrix settings and it’s simply a case of setting adaptive matrix to on. I recommend that when you don’t actually need it you turn it back off again.
Most of Sony’s broadcast quality cameras produced in the last 5 years have the adaptive matrix function, that includes the FS7, FX9, Z280, Z450, Z750, F5/F55 and many others.
Sony have just released the latest version of their free viewing, copying and transcoding software Catalyst Browse and the more fully featured paid software Catalyst Prepare. These new versions includes support for the PXW-FX9’s metadata based image stabilisation. Hopefully the new Mac versions are also optimised for Catalina.
This is something a lot of people have been asking for. An extension or relocation cable that allows you to place devices that will be connected to a camera via the MI Shoe away from the shoe itself.
But in order to get the MI Shoe relocation cable you have to buy the whole XLR-K3M XLR adapter kit, you can’t get the cable on it’s own. This is a shame as I would like to use the cable with my UWP-D series radio mics. I’m not a fan of having the radio mic receiver right on top of the handle as it tends to stick out and get in the way when you put the camera into most camera bags. But, I don’t really need the XLR adapter.
Anyway, here’s a link to the XLR-K3M for those that really need that cable (or the new XLR adapter).
A completely useless bit of trivia for you is that the “E” in E-mount stands for eighteen. 18mm is the E-mount flange back distance. That’s the distance between the sensor and the face of the lens mount. The fact the e-mount is only 18mm while most other DSLR systems have a flange back distance of around 40mm means thare are 20mm or more in hand that can be used for adapters to go between the camera body and 3rd party lenses with different mounts.
Here’s a little table of some common flange back distances:
Sony FZ (F3/F5/F55)
Nikon F Mount
Sony A, Minolta
Camera setup, reviews, tutorials and information for pro camcorder users from Alister Chapman.