OCIO, Display Transforms and Misconceptions

Table Of Contents

Introduction

I often think about this great question asked by Doug Walker at one of our OpenColorIO (OCIO) meetings : what are you trying to solve ? So in our case, what is this post about ?

In the past year (2021), I have given several talks online about OCIO and the Academy Color Encoding System (ACES). I thought it would be useful to put those slides online for two reasons :

  • If there is any inaccurracy, anyone can correct me and I’ll be more than happy to update this page.
  • With these slides online, I hope I can reach more people and raise awareness about certain topics.

I do not consider myself an expert. I do not believe in this word anymore. We are only human beings, doing our best to understand things. We make mistakes, hopefully learn from them and keep going. So no, I’m not an expert on anything !

The only thing I can offer is an artist’s point of view on Color Management. Quite often, I have been lost on where to begin when it comes to OCIO and ACES. So by writing this post, I hope I can share with you some useful information.

There is a reason my book starts with two chapters about Color Management. It will give you the fundation to create beautiful images. So I thought it would be interesting to compare various OCIO configs, with a series of visual examples in order to study their strength and flaws.

Really the idea is to look back but NOT judge these historical OCIO configs. Not at all ! Huge respect to all the persons who more than ten years ago developed and shared these configs with the community.

A great community

Some people may find this post too obvious, boring or outdated. Probably ! But I was really interested to go back in time and write this historical analysis on OCIO and Display Transforms. This is post is primarly about my own misconceptions, but hopefully it will help others too !

Please note that I will only talk about OCIOv1 configs here.

First of all, I would like to thank the Academy of Motion Picture Arts and Sciences (AMPAS) because without them, I would have probably never discovered this crazy wonderful world of Color Management. ACES has been my entry point to Color Management and I will never be thankful enough for that.

Thanks to them, I have discovered this amazing community of colour nerds and colour scientists, and I would like to personnally thank the following persons for their generosity, sharing and patience with my questions :

It is really heart-warming to have this group of people passionate about colour and display transforms who are able to debate online at midnight over sRGB or scene-referred grading. It is a welcoming community, who does not agree on everything and who has still a lot to figure out… Which is really exciting !

Okay, enough with the name dropping, let’s start ! Yay !

The importance of terminology

I’m writing these lines with great humility because I have been dead wrong in the past and I’m probably still inacurrate about many things… But I’m learning and I think my only vantage is to explain things simply enough so hopefully fellow CG artists can understand me.

I’ll start here with a bit of OCIO vocabulary, so we’re all on the same page :

  • Display : monitor with its characteristics (primaries, white point, EOTF, nits…).
  • View : the way we see the images (ACES, Raw, Log, Un-tone-mapped…).
  • Look : creative preference or grading (LMT in the ACES terminology).

These three terms may sound really basic but they are out of the utmost importance for what comes next. As you probably know, I’m quite a practical person. I like to have visual and concrete examples to support my explanations. So we’ll start by having a look at these two OCIO Config repositories and study our different options :

  • nuke-default
  • spi-anim / spi-vfx
  • Filmic Blender
  • ACES

OCIO Configs examples

nuke-default

nuke-default description

Why should we start with an OCIO config that was shipped with Nuke 6.3 in 2011 ? First of all, because I believe it is still used in some schools and small shops. Second of all, because it contains an option that has misled me for ages.

displays:
  default:
    - !<View> {name: None, colorspace: raw}
    - !<View> {name: sRGB, colorspace: sRGB}
    - !<View> {name: rec709, colorspace: rec709}
    - !<View> {name: rec1886, colorspace: Gamma2.4}

active_displays: [default]
active_views: [sRGB, rec709, rec1886, None]

I won’t mention here that the “Display” is default (what ?) and that the “Views” are actually Transfer Functions…

Let’s focus on something else instead !

Revelation #1 : In this config it is important to notice that the Rec.709 Transfer Function is a camera encoding OETF ! It is NOT a display standard ! It took me a while to figure that one out ! So what everybody calls a “Rec.709 display” is actually a “BT.1886 EOTF with BT.709 primaries” display.

This camera encoding was designed to direct drive output on a CRT broadcast display, with an EOTF of approximately 2.4. It was designed to look appropriate on both what later became BT.1886 and sRGB 2.2.

From a fellow color nerd.

BT.709 and BT.1886

This was one of my first interrogations when I discovered ACES. Because the Rec.709 (ACES) Output Transform is “less contrasted” (for lack of a better words) than the sRGB (ACES) one, contrary to the nuke-default. Here are the approximative Gamma values :

  • BT.709 is a camera encoding OETF (~Gamma 1.95).
  • BT.1886 is an EOTF output display (~Gamma 2.4) and is used in the Rec.709 (ACES) Output Transform.

[…] using a tilde has become commonly used shorthand for differentiating between pure power functions and piecewise functions, when referring to display CCTFs, and is sometimes the only way to differentiate between the two, if the author isn’t explicitly using “power” or “piecewise”. For example, Baselight differentiates between “sRGB Display (2.2)” and “sRGB (~2.2)”. […] In the context of transfer function LUTs used for OCIO, BT.1886 is exactly Gamma 2.4.

Thanks Zach Lewis for the precision.

Old-school VS new-school

There is a couple of things to know about this :

  • Using “rec709” in your viewer (“old school” intention), you’ll never get 1:1 for all the values. Because the general power function difference is loosely a “surround/flare compensation” !
  • If you use “rec1886” on a pure BT.1886 display (“new school” intention), it is an idealized inverse for the EOTF. What we call a no-operation.

And it is the exact same issue with BT.709/sRGB piecewise encoding dumped through the proper 2.2 power function. There is a very interesting talk about this precise issue by Daniele Siragusano. It is a bit technical but definitely worth watching !

In 2020, flare compensation is part of the Display Rendering Transform and should not live in the distribution pipeline.

Daniele Siragusano.

We may add a couple of things : we should never ever use “Rec.709” nor “BT.709” without a further qualifier. Because “Rec.709” can either mean a transfer function or a set of primaries. For clarity, we need to separate the transfer function from the primaries.

I will also quickly mention that the use of the term “Gamma” is ambiguous : it’s a greek letter ! It has been used in so many different contexts. I now try to use the term “Power Function/Law” or “Transfer Functions” when appropriate.

nuke-default visual examples

We are now going to look at a sweep of sRGB/BT.709 primaries with this config. It will show an interesting phenomenon known as the “Notorious 6. The setup is super simple : in Nuke, I have set a range of sRGB/BT.709 primaries vertically and I increase by 1 stop for each column (horizontally). Let’s have a look !

140_misconceptions_0010_srgb_sweep_FHD
140_misconceptions_0020_srgb_sweep_FHD
140_misconceptions_0030_srgb_sweep_FHD
140_misconceptions_0040_srgb_sweep_FHD
140_misconceptions_0050_light_sabers_FHD
Exit full screenEnter Full screen
previous arrow
next arrow
 

Revelation #2 : With this config, no matter how many values you use, you will ALWAYS end up on these six values at display : Red, Green, Blue, Cyan, Magenta and Yellow. Pay attention to this behaviour since we will observe it with several Color Management Workflows.

This config should be avoided at all cost for CGI work. It may be useful in some edge cases such as albedo textures. But that’s about it !

A friendly reminder…

spi-anim / spi-vfx

spi-anim description

There is a very complete description of the spi-vfx config in the OCIO docs (which is like a “sister” config of spi-anim). For the record, both configs are from Sony Picture Imageworks (spi).

Since we focus on Display Transforms in this post, we will have a look at the “Film (sRGB)” View from the spi-anim OCIO Config. Here is a quick quote :

The vd spaces are mappings of linear image data into display space. The main part of the transformation is defined as a single curve that is conceptually two parts. The first is a ln to lg (linear to logarithmic) conversion. The second is lg to sRGB conversion. This is based off the neutral channel response of the sRGB film emulation lut used in the profile.

The spi-anim OCIO config uses a 1D LUT for its “Film” View. For sRGB Displays, the LUT file used is called “vd16.spi1d“. Here is an example from the config which I believe uses the correct OCIO terminology :

displays:
  DCIP3:
    - !<View> {name: Film, colorspace: p3dci8}
    - !<View> {name: Log, colorspace: lm10}
    - !<View> {name: Raw, colorspace: nc10}
  sRGB:
    - !<View> {name: Film, colorspace: vd16}
    - !<View> {name: Log, colorspace: lm10}
    - !<View> {name: Raw, colorspace: nc10}

We do have access to two displays : “sRGB” and “DCIP3“. And three different ways of “viewing” our renders : “Film“, “Log” and “Raw“. This is a simple and correct OCIO setup, very clear and easy to understand.

spi-anim visual examples

Let’s now have a look at some renders !

140_misconceptions_0060_srgb_sweep_FHD
140_misconceptions_0070_srgb_sweep_FHD
140_misconceptions_0080_srgb_sweep_FHD
140_misconceptions_0090_srgb_sweep_FHD
140_misconceptions_0100_light_sabers_FHD
Exit full screenEnter Full screen
previous arrow
next arrow
 

spi-vfx description

For completeness, I have also used the same renders through the “srgb8” View from the spi-vfx OCIO config. Here its description from the OCIO docs :

srgb8 bakes in the film3d emulation lut. This table can be used for either QuickTime generation or output to the sRGB display. The transformation is a 3d film emulation table with gray balance compensation, so a value of 445,445,445 in lg10 space is modified to become equal RGB values in sRGB. Additionally the lut is scaled so that at least one of the color channels on maximum white uses the display max. The transformation takes place in three parts. First the linear data is converted to the show log space. Then a film emulation table is applied. Then the grey balance and white scaling compensation are applied. This table is designed to be evaluated in a dimly lit office environment on a sRGB display.

spi-vfx visual examples

And here are our usual renders for comparison :

140_misconceptions_0110_srgb_sweep_FHD
140_misconceptions_0120_srgb_sweep_FHD
140_misconceptions_0130_srgb_sweep_FHD
140_misconceptions_0140_srgb_sweep_FHD
140_misconceptions_0150_light_sabers_FHD
Exit full screenEnter Full screen
previous arrow
next arrow
 

For your information, I haven’t used any compositing trick for my render. The “glow” around the light sabers is actually a volumetrics shader, affected by the Mesh Light. Straight from Guerilla Render.

I got really really really surprised when I looked at these images… I did not expect a Display Transform called “Film” to behave like that ! My two main surprises were :

  • sRGB/BT.709 primaries do not go to white !
  • We end up on the “Notorious 6” values, just like a simple sRGB EOTF !

Which made me arrive at this earth-shaking conclusion… Revelation #3 : an s-curve has nothing to do with the “Filmic” behaviour. Certainly, it will add a nice “contrast” to the image and perform some kind of luminance lookup. But does this really qualify as “Filmic” ? I’m not sure…

To me, “filmic tone-mapping” means that there is a visual “path-to-white” (some people calls it “desaturated highlights“). The fact that sRGB/BT.709 primaries light sabers do not go to white with a “film3d emulation lut” looks wrong to me. It tells me that maybe something else is at play here ! So, yeah, s-curves have nothing to do with the “path-to-white“.

I did look for a “filmic” reference to illustrate my thoughts. And “Star Wars” was the best example I could come with. This is the “Look” I would expect for my light sabers : a white “core” with a colored glow.

140_misconceptions_0160_light_sabers_FHD
Exit full screenEnter Full screen
 

Agreed, these OCIO configs are a bit old but they’re still used by some small studios. And I love doing this OCIO archeology because we learn so much on the way.

Edge cases ?

One may argue that my examples are edge cases : “hey Chris, stupid you, just use 0.001 instead of 0 and you’re good to go !” And this is true. It is a common and known practice to limit the range of input colors to get some kind of “path-to-white“. But doesn’t it reveal something ?

140_misconceptions_0170_srgb_sweep_FHD
140_misconceptions_0180_srgb_sweep_FHD
140_misconceptions_0190_srgb_sweep_FHD
140_misconceptions_0200_srgb_sweep_FHD
140_misconceptions_0210_light_sabers_FHD
Exit full screenEnter Full screen
previous arrow
next arrow
 

To be clear, I am not even using any wide gamut rendering in these examples. But rather the lowest common denominator : sRGB/BT.709 primaries through several Display Transforms. Since lasers are BT.2020 primaries, I believe that using sRGB/BT.709 primaries is not an edge case. Not at all !

I reckon no one should have to overcome the stimulus encoding to avoid errors.

A fellow color nerd.

But, of course, we need to put things back in their context here. These two configs from Sony did a great job back in 2011, color managing amazing movies such as “Cloudy with a chance of meatballs“. Again, I’m trying to dig a bit in History to understand where my own misconceptions come from. Nothing more.

The VFX config was film-scan Cineon based… whereas the Anim was using per-channel 1D LUTs for the Display Rendering Transform (DRT). spi-vfx is well beyond its shelf-life… spi-anim can still be informative to simple animation productions, but the naive RGB tonecurve will make their lives unpleasant for multi-deliverable scenarios.

Great summary by Sean Cooper.

Filmic Blender

Filmic Blender description

Not the easiest OCIO config to start with, but it is actually super simple. And a pretty damn good option ! I highly recommend to read the README from GitHub as it describes exactly what “Filmic” does.

It compresses the scene referred linear radiometric energy values down to the display / output referred range. This aspect is known as a transfer function or tone mapping. The shape of the Filmic Base Log with a contrast aesthetic roughly emulates a photographic film curve.

It compresses the gamut for high intensity values. As colour ratios increase in intensity, highly saturated ratios tend to be resistant to transfer function compression, which results in peculiar feeling imagery with some regions feeling appropriately over-exposed and others “lingering” behind. Filmic considers all colour values fair game, and attempts to blend colours into a consistent output that matches our learned expectations from film emulsion-like media.

Great description by Troy Sobotka.

Filmic Blender Displays and Views

It should be needless to say that the “Filmic” OCIO Config will work in any DCC software that supports OCIO. Let’s have a quick look at the Displays and Views available in the config :

displays:
  sRGB:
    - !<View> {name: sRGB OETF, colorspace: sRGB OETF}
    - !<View> {name: Non-Colour Data, colorspace: Non-Colour Data}
    - !<View> {name: Linear Raw, colorspace: Linear}
    - !<View> {name: Filmic Log Encoding Base, colorspace: Filmic Log Encoding}
  BT.1886:
    - !<View> {name: BT.1886 EOTF, colorspace: BT.1886 EOTF}
    - !<View> {name: Non-Colour Data, colorspace: Non-Colour Data}
    - !<View> {name: Linear Raw, colorspace: Linear}
    - !<View> {name: Filmic Log Encoding Base, colorspace: BT.1886 Filmic Log Encoding}
  Apple Display P3:
    - !<View> {name: sRGB OETF, colorspace: AppleP3 sRGB OETF}
    - !<View> {name: Non-Colour Data, colorspace: Non-Colour Data}
    - !<View> {name: Linear Raw, colorspace: Linear}
    - !<View> {name: Filmic Log Encoding Base, colorspace: AppleP3 Filmic Log Encoding}
# VRay users should uncomment the Filmic views below as VRay doesn't permit Looks
#    - !<View> {name: Filmic Very High Contrast, colorspace: Filmic Log Encoding, look: +Very High Contrast}
#    - !<View> {name: Filmic High Contrast, colorspace: Filmic Log Encoding, look: +High Contrast}
#    - !<View> {name: Filmic Medium High Contrast, colorspace: Filmic Log Encoding, look: +Medium High Contrast}
#    - !<View> {name: Filmic Very Low Contrast, colorspace: Filmic Log Encoding, look: +Very Low Contrast}
#    - !<View> {name: Filmic Medium Low Contrast, colorspace: Filmic Log Encoding, look: +Medium Low Contrast}
#    - !<View> {name: Filmic Low Contrast, colorspace: Filmic Log Encoding, look: +Low Contrast}
#    - !<View> {name: Filmic Base Contrast, colorspace: Filmic Log Encoding, look: +Base Contrast}
#    - !<View> {name: Filmic False Colour, colorspace: Filmic Log Encoding, look: +False Colour}
#    - !<View> {name: Debug, colorspace: Debug}

Just like the spi-anim/vfx OCIO configs, the Displays and Views are setup correctly, according to the OCIO terminology. But there is one more element available : Looks. And it can be unsettling if you’re not used to them. Let’s check the documentation again :

This basic view (Filmic Log Encoding Base) is designed to be coupled with one of the contrast looks.

But there is a tiny issue here. Only a couple of DCC softwares permit interactive “Look Overrides” : Blender and Renderman 24. So most users modify the OCIO config to combine the “Look” with the “View” as a workaround.

OCIO Viewing experience and Roles

When this article was published, the following revelation has caused a bit of waves. So I have updated it to be as accurate as I could.

Nowadays, OCIO itself does behave the same way in all DCCs that I’m aware of, and that’s been the case since at least 2017 — which is to say, viewing is a consistent experience across DCCs. That’s fundamentally the most important thing about OCIO for most folks — that once you set up an OCIO config for a show or a facility, you can rely on DCC implementations of OCIO to provide a consistent viewing experience throughout the pipeline, without any additional technical intervention.

Proper explanation by Zach Lewis.

Revelation #4 : OCIO is NOT implemented the exact same way in all softwares. And yes, that’s a big one ! With OCIOv2, there is an effort trying to normalize it though. But if you look at the “color_picking” role for instance, it does not have the same behaviour between Maya and any other DCC softwares !

OCIO implementation and UX

I will repeat for clarity : yes, OCIO provides a consistent viewing experience accross DCCs. But it is also worth noting that :

There are some slight differences in how the GPU and CPU paths process color in OCIOv1, so if one DCC uses the GPU path and another uses only the CPU path, that would provide a slight non-visual discrepancy across DCCs; but I think it would be disingenuous to suggest that DCCs aren’t capable of providing a consistent viewing experience across DCCs.

Rv is a clear example of this behaviour. Thanks Zach !

So, I reckon I entangled a couple of things in the term “implementation” :

  • Most DCC softwares do not permit interactive “Look Overrides“, like a dropdown menu that would let you easily choose them from a list. This more an UX thing.
  • DCCs may use different OCIO roles, or the same roles for different purposes (“reference” and “color_picking” are ambiguous). I’d personnally say that this is an implementation thing.
  • With OCIOv1, there are some visual differences between the CPU and the GPU processor that may cause you a little bit of headache. This is now fixed with OCIOv2.

Editing an OCIO Config

So if you want to use “Looks” in your OCIO Config, there is an easy way to do that. Most users have this workaround, which looks like this :

displays:
  sRGB:
    - !<View> {name: sRGB OETF, colorspace: sRGB OETF}
    - !<View> {name: Non-Colour Data, colorspace: Non-Colour Data}
    - !<View> {name: Linear Raw, colorspace: Linear}
    - !<View> {name: Filmic High Contrast, colorspace: Filmic Log Encoding, look: High Contrast}

active_displays: [sRGB]
active_views: [Filmic High Contrast, Non-Colour Data]

As you can see, it is pretty simple to modify an OCIO config based on your needs. Just open it in notepad and you can directly edit it ! And that’s our Revelation #5 : it is recommended to edit the standard OCIO configs. So don’t be afraid to dive in !

This is very well explained in this thread :

It’s very common practice to make a custom config by editing the standard ACES config in a text editor. […] Not only that but recommended ! One of the goals of the ACES config has always been to be a starting point for people to tailor it to their needs.

Nick Shaw and Thomas Mansencal for the win !

Filmic Blender visual examples

Let’s have a look at our images with Filmic !

140_misconceptions_0220_srgb_sweep_FHD
140_misconceptions_0230_srgb_sweep_FHD
140_misconceptions_0240_srgb_sweep_FHD
140_misconceptions_0250_srgb_sweep_FHD
140_misconceptions_0260_light_sabers_FHD
Exit full screenEnter Full screen
previous arrow
next arrow
 

A few observations :

  • Light sabers are white ! Yay !
  • We still end up on the “Notorious 6” ! Noooo ! Why ?

If you ever wondered what animation feature film has been color managed by Troy’s Filmic, here is the trailer of this great movie (I loved it !) :

Let’s keep on going with our investigations and see if we can find more answers to our questions !

ACES

ACES Displays and Views

I have already written a complete chapter about ACES… So I’ll try to come up with some new information here.

First of all, if we have a look at the ACES 1.2 OCIO Config, you may notice something weird. Revelation #6 : Displays and Views have been inverted/swapped in the ACES OCIOv1 configs. This has caused me a bit of headache when I tried to learn about OCIO. Here is a sample from the current config :

displays:
  ACES:
    - !<View> {name: sRGB, colorspace: Output - sRGB}
    - !<View> {name: DCDM, colorspace: Output - DCDM}
    - !<View> {name: DCDM P3D60 Limited, colorspace: Output - DCDM (P3D60 Limited)}
    - !<View> {name: DCDM P3D65 Limited, colorspace: Output - DCDM (P3D65 Limited)}
    - !<View> {name: P3-D60, colorspace: Output - P3-D60}
    - !<View> {name: P3-D65 ST2084 1000 nits, colorspace: Output - P3-D65 ST2084 (1000 nits)}
    - !<View> {name: P3-D65 ST2084 2000 nits, colorspace: Output - P3-D65 ST2084 (2000 nits)}
    - !<View> {name: P3-D65 ST2084 4000 nits, colorspace: Output - P3-D65 ST2084 (4000 nits)}
    - !<View> {name: P3-DCI D60 simulation, colorspace: Output - P3-DCI (D60 simulation)}
    - !<View> {name: P3-DCI D65 simulation, colorspace: Output - P3-DCI (D65 simulation)}
    - !<View> {name: P3D65, colorspace: Output - P3D65}
    - !<View> {name: P3D65 D60 simulation, colorspace: Output - P3D65 (D60 simulation)}
    - !<View> {name: P3D65 Rec.709 Limited, colorspace: Output - P3D65 (Rec.709 Limited)}
    - !<View> {name: P3D65 ST2084 108 nits, colorspace: Output - P3D65 ST2084 (108 nits)}
    - !<View> {name: Rec.2020, colorspace: Output - Rec.2020}
    - !<View> {name: Rec.2020 P3D65 Limited, colorspace: Output - Rec.2020 (P3D65 Limited)}
    - !<View> {name: Rec.2020 Rec.709 Limited, colorspace: Output - Rec.2020 (Rec.709 Limited)}
    - !<View> {name: Rec.2020 HLG 1000 nits, colorspace: Output - Rec.2020 HLG (1000 nits)}
    - !<View> {name: Rec.2020 ST2084 1000 nits, colorspace: Output - Rec.2020 ST2084 (1000 nits)}
    - !<View> {name: Rec.2020 ST2084 2000 nits, colorspace: Output - Rec.2020 ST2084 (2000 nits)}
    - !<View> {name: Rec.2020 ST2084 4000 nits, colorspace: Output - Rec.2020 ST2084 (4000 nits)}
    - !<View> {name: Rec.709, colorspace: Output - Rec.709}
    - !<View> {name: Rec.709 D60 sim., colorspace: Output - Rec.709 (D60 sim.)}
    - !<View> {name: sRGB D60 sim., colorspace: Output - sRGB (D60 sim.)}
    - !<View> {name: Raw, colorspace: Utility - Raw}
    - !<View> {name: Log, colorspace: Input - ADX - ADX10}

In the above example, the Display is “ACES” which is incorrect since we do not buy ACES monitors… And the Views are “sRGB“, “Rec.709” or “P3D65” which are basically “monitors”. It should be the other way around !

True, the configs are being fixed with the OCIOv2 Configs, which is great ! They will be soon available here.

Thanks Thomas !

Good news is that it is super easy to fix it ! And if you don’t want very long “Display” menus in Nuke or Maya, you may simply modify the OCIO config file. Let’s have a look ! For instance, these are the “active_displays” and “active_views” in the default ACES 1.2 OCIO Config :

active_displays: [ACES]
active_views: [sRGB, DCDM, DCDM P3D60 Limited, DCDM P3D65 Limited, P3-D60, P3-D65 ST2084 1000 nits, P3-D65 ST2084 2000 nits, P3-D65 ST2084 4000 nits, P3-DCI D60 simulation, P3-DCI D65 simulation, P3D65, P3D65 D60 simulation, P3D65 Rec.709 Limited, P3D65 ST2084 108 nits, Rec.2020, Rec.2020 P3D65 Limited, Rec.2020 Rec.709 Limited, Rec.2020 HLG 1000 nits, Rec.2020 ST2084 1000 nits, Rec.2020 ST2084 2000 nits, Rec.2020 ST2084 4000 nits, Rec.709, Rec.709 D60 sim., sRGB D60 sim., Raw, Log]

So, if we want to shorten the long list of options and fix the “Display/View” menu, you may simply do this :

displays:
  sRGB:
    - !<View> {name: ACES, colorspace: out_srgb}
    - !<View> {name: Raw, colorspace: raw}
    - !<View> {name: Log, colorspace: acescct}

active_displays: [sRGB]
active_views: [ACES, Raw, Log]

And in case you didn’t know, there is an ACES 1.1 OCIOv1 CG config (which is very light, like less than 48 MB) available at the CAVE ACADEMY website. It can be used for full CG stuff for example.

ACES visual examples

Let’s have a look at our images now !

140_misconceptions_0270_srgb_sweep_FHD
140_misconceptions_0280_srgb_sweep_FHD
140_misconceptions_0290_srgb_sweep_FHD
140_misconceptions_0300_srgb_sweep_FHD
140_misconceptions_0310_light_sabers_FHD
Exit full screenEnter Full screen
previous arrow
next arrow
 

Hum, what’s going on ? Two interesting things to observe :

  • Light sabers are white ! But the red light saber goes orange, the green one goes yellow and the blue neons go magenta… These are called Hue Skews/Shifts. It’s basically a “Look“.
  • We still observe hue paths going towards the “Notorious 6” ! Noooo ! Why ?

We will detail the “ACES Look” a bit later. But it is important to notice that this look is embedded in the Output Transform. You cannot escape from it.

For completeness, I will run the same renders using ACEScg primaries this time. Let’s have a look !

140_misconceptions_0320_acescg_sweep_FHD
140_misconceptions_0330_acescg_sweep_FHD
140_misconceptions_0340_acescg_sweep_FHD
140_misconceptions_0350_acescg_sweep_FHD
140_misconceptions_0360_light_sabers_FHD
Exit full screenEnter Full screen
previous arrow
next arrow
 

Same observations can be made… By using ACEScg primaries, we can see :

  • Gamut Clipping.
  • We still end up on the “Notorious 6” values, just like an sRGB EOTF ! Why does this keep happening ?

Revelation #7 : Contrary to what I may have written in the past, the ACES Output Transforms do not solve anything. We observe with ACES a similar behaviour than the others OCIO configs.

It’s a bit tricky to know exactly which projects have used the ACES Output Transform. So I’ll just add a link to this video game since it has been confirmed on ACESCentral :

It took me a while to understand this but “going-to-white” (like on the light sabers for instance) is not an end in itself. I will repeat in bold and with emphasis because it is CRITICALRevelation #8 : having these sweeps “going-to-white” does not mean necessarily that the Display Transform is working properly. The path itself is important ! And this is where it gets tricky and awesome !

Comparison and Summary

We have compared four OCIO configs with some visual examples. Here is a little summary with pros and cons for each of them.

ProsConsDate
nuke-defaultSuper light, less than 3 MB. Should be avoided at all costs for CGI work. Useful in some edge cases for albedo textures, for example.2011
spi-anim/vfxSuper light, less than 30 MB.
Used on several Sony movies.
Free.
No path-to-whiteon sRGB/BT.709 primaries.
“Notorious 6”.
No Wide Gamut rendering.
No HDR Display Transforms.
2011
Filmic BlenderSuper light, less than 20 MB.
Path-to-white.
Contrast Looks.
Free.
“Notorious 6”.
No Wide Gamut rendering.
No HDR Display Transforms.
2017
ACES 1.2Cross platform “standard”.
Wide Gamut rendering.
Free.
Super heavy, more than 1.89 GB.
“Notorious 6”.
Hue Shifts.
Gamut clipping.
Color appearance mismatch.
2014 – 2021

We could stop this post right there, right ? We have compared several OCIO configs, pointed out their strength and flaws. So the image maker can make an informed choice when it comes to pick a Color Management Workflow. We could even finish on this beautiful quote :

Each approach reflects the ZeitGeist of an era.

This is a great quote by Daniele Siragusano. Zeitgeist is the defining spirit or mood of a particular period of history as shown by the ideas and beliefs of the time.

But no… We cannot stop here, we’re only half-way. We need to dig deeper to understand where the “Notorious 6” come from and what Tone Mapping really is about.

So, what we have learned so far ?

  • We have compared several Color Management OCIO Configs.
  • All of them have shown the same “Notorious 6” behaviour.
  • Maybe because they all use the same fundamental mechanics !
  • So where these mechanics would come from ? Let’s see…

Lost in misconception

In the context of the ACES Output Transforms VWG, we have been challenged to define some terms such as “Tone” or “Filmic“. Since I was not able to define them precisely, I wanted to go back to the first time I ever heard them. Friendly reminder : yes, my goal is still the same ! Understanding how my own misconceptions have lead me on the wrong path for the past 15 years.

Here are the few questions I have been torturing myself in 2021 :

  • When was the first time I ever heard about tone-mapping ?
  • What does tone-mapping really mean ? What does it do ?
  • What does an s-curve do ?
  • What does “filmic” really mean ?

All these concepts may seem overloaded but I was curious to question my whole knowledge about them. And funny thing, I was wrong about most of them ! So let’s go back in time for another beautiful journey through color.

A bit of History

First time I ever heard the words “tone-mapping” was at Axis Studios back in 2012… They were mentioned in this fx guide article about Mass Effect 3 and Digic Pictures. It confused me a lot back then !

Then at Illumination I heard those terms again ! They came from the John Hable’s talk at GDC about Uncharted 2 (Video and slides), which was probably a game changer in the industry ! In his presentation John Hable describes a 1D LUT called “Filmic” that originally came from Haarm-Pieter Duiker (here are two of his presentations : one at Electronic Arts and one at SIGGRAPH 2010).

Revelation #9 : this 1D s-curve LUT is not applied on open domain, aka on infinite values or scene-linear values. It is applied on logarithmic values !

140_misconceptions_0370_log_color_FHD
140_misconceptions_0380_filmic_curve_FHD
140_misconceptions_0390_filmic_tonemap_FHD
Exit full screenEnter Full screen
previous arrow
next arrow
 

So back in 2010 during his second talk, Haarm-Pieter Duiker was already ackowledging the limitations of a 1d LUT and mentioning Gamut Mapping !

This Gamut Mapping/Compress thing seems like an important thing…

The per-channel approach

So here is the fundamental mechanics at play :

  • The “filmic s-curve” is a 1d LUT…
  • Applied on the channels R, G, B separately…
  • This is called per-channel/RGB lookup…
  • And it will ALWAYS results in the “Notorious 6” !

Revelation #10 : Any curve that asymptotes/converges/saturates at 1, like this “filmic s-curve“, will end up on the “Notorious 6“.

Generic per channel means taking the light data emissions and rendering them on a per-channel basis, one by one, through some sort of curve. Aka the existing flawed approach used all over.

A fellow colour nerd.

And if that was not clear enough :

Per-channel (or RGB) lookup is a skewed and posterized solution, that does not respect the intention of the ratios. It changes the “hue”, changes the “saturation” and changes the “luminance” of the intended light mixture.

A fellow colour nerd.

There’s nothing better than a CIE 1976 diagram to show what’s happening. But before that I would strongly recommend to read Question#15 and Question#16 by Troy Sobotka. It explains very clearly what the CIE 1976 chromaticity diagram is. And it has this very cool diagram where it shows the connection between “spectral light energy” (aka the wavelengths) to “numbers” (aka chromaticities). A must-read !

Plots and Chromaticity Diagrams

Let’s have a look at some plots :

140_misconceptions_0400_chromaticity_paths_FHD
140_misconceptions_0410_chromaticity_paths_FHD
140_misconceptions_0420_chromaticity_paths_FHD
140_misconceptions_0430_chromaticity_paths_FHD
140_misconceptions_0440_chromaticity_paths_FHD
140_misconceptions_0450_chromaticity_paths_FHD
140_misconceptions_0460_chromaticity_paths_FHD
140_misconceptions_0470_chromaticity_paths_FHD
140_misconceptions_0480_chromaticity_paths_FHD
140_misconceptions_0490_chromaticity_paths_FHD
140_misconceptions_0500_chromaticity_paths_FHD
140_misconceptions_0510_chromaticity_paths_FHD
Exit full screenEnter Full screen
previous arrow
next arrow
 

Revelation #11 : All the scenes I have lit, rendered and displayed for the past 15 years have the “Notorious 6” look. Yes, that’s a biggie ! It took me a while to accept it, like a leap of faith almost : I have NEVER seen the colours as encoded. They have always been distorted at display by a “filmic s-curve 1d LUT“. OMG.

This is how I have been working for the past 15 years :

Most CG artists do not have a good understanding of what happens when the image data is rendered for a display device and assume it is “the correct colors”. Then they fight with “weird stuff” they don’t understand until they end up with something they are happy with.

Jed Smith.

And at the risk of repeating myself, I’ll add a similar explanation here :

But at the most basic level, yes you’ve probably never looked at stimulus that matched what the texture artists poured many hours into. They just looked at their work “under a LUT” and made something they thought was acceptable.

A fellow colour nerd.

Values for the plots

To be very clear about which values have been used for each plot, here is a recap of the colors used in the “Red to Yellow Notorious 6 sweep” :

Color ManagementSweep PrimariesRMore GMore GMore GMore GY
Filmic HablesRGB1, 0.00001, 0.000011, 0.0001, 0.000011, 0.0003, 0.000061, 0.0003, 0.00011, 0.0003, 0.00021, 1, 0.00001
spi-animsRGB1, 0.00001, 0.000011, 0.0001, 0.000011, 0.0003, 0.000061, 0.0003, 0.00011, 0.0003, 0.00021, 1, 0.00001
spi-vfxsRGB1, 0.00001, 0.000011, 0.0001, 0.000011, 0.0003, 0.000061, 0.0003, 0.00011, 0.0003, 0.00021, 1, 0.00001
Filmic High ContrastACEScg1, 0, 01, 0.003, 0 1, 0.01, 0 1, 0.05, 0 1, 0.2, 01, 1, 0
sRGB (ACES)ACEScg1, 0, 01, 0.05, 01, 0.2, 0 1, 0.5, 01, 0.8, 01, 1, 0
sRGB (ACES) ACEScg1, 0, 01, 0.005, 01, 0.015, 01, 0.03, 01, 0.1, 01, 1, 0

Please notice what kind of crazy tiny values I had to use in order to get these “nice readable” plots… Insane ! We’ll do more plots later with a regular step, so you can see the differences !

More about per-channel

So are we doomed ? Since all these open-source approaches use the same mechanics. Well, let’s have a look at other rendering solutions ?

Wait, what ? Rendering solutions ? What do you mean ? Like Arnold, Renderman or Guerilla Render ? No, I’m talking about a complete different rendering operation here. So before we have a look at these other possibilities, we need to talk about something super important and just mind-blowing… Buckle up !

Rendering exr files

When I was starting in the industry, my supervisor Barbara Meyers gave me this great definition of “rendering” :

Rendering is the transition from a 3d scene (of your favorite DCC application) to a 2d image.

It has taken me 15 years to realize that this definition could be split in two because there are actually two different types of rendering in the CG world ! Yes ! This is HUGE ! Let’s take a deep breath and explain it properly.

When you hit “render” in a 3d software and wait for sampling to happen, what you are generating ? Most likely, you’ll be saving exr files from a batch render. Can this exr file be considered an image ? Well, not exactly… It becomes an image once it is “Viewed” on a “Display“.

Revelation#12 : a “beauty” exr file is NOT an image, it is a file containing light data. It only becomes an image when displayed on your monitor. Hence the expression “Image Formation Chain” : we’re trying to form/create an image from “light data” that can be “Viewed” on a “Display”.

EXRs are data encodings. They could hold normals, depth, random spectral data, you name it.

A fellow colour nerd.

Two types of rendering

Therefore we can distinguish two types of rendering in CG. Just like this :

3D scene -> Data -> Image

Revelation#13 : There are two rendering operations in CG. The first one is from scene to data and the second one from data to image.

Here are some quick definitions in case of it is not clear enough :

  • Scene rendering : this is what I have been doing for the past fifteen years… Hit “render” and wait for sampling to happen and generate an exr file. This can take hours.
  • Display rendering : then load this exr and display it in your Render View, maybe using OCIO… And yes ! This is a color rendering operation, that most likely will be in real-time.

If you’re not convinced, just think at we have been doing since the beginning of this post. The same “beauty” exr can generate many different images. They’re just light data.

And here is the best part ! These two rendering operations are both dependent on the primaries ! Yes, this is crazy ! Revelation#14 : applying an s-curve on different primaries will yield a different result. Boom ! Mind blown !

Custom Display Rendering Primaries experiment

So when we talk about “Display Transforms“, or “Display Rendering Transforms” as Baselight calls them… We must acknowledge that these words have a meaning :

  • Display : the monitor/screen.
  • Rendering : from data to image.
  • Transform : we need to change the values from scene-referred to display-referred (change of primaries, white point, EOTF…)

Jed Smith did an amazing per-channel experiment to show the importance of primaries for “Display Rendering“. It is a super simple setup based on ARRI K1S1 model, described here by Jed :

[ARRI K1S1 is] Essentially a tonemap in Alexa Wide Gamut, and a display encoding, which includes a custom AWG → Rec.709 matrix which desaturates the red and green primaries a little. So I decided to make my own, inspired by this simple though highly effective approach to a display transform. I built a little setup that lets you position the coordinate of the three primaries positions using a 2d position knob, which then automatically calculates a matrix to convert your incoming image into this “custom rendering gamut”.

Jed Smith and another amazing experiment !

Here are a few examples, comparing a tonemap applied on RED Wide Gamut primaries and ACEScg ones :

140_misconceptions_0520_srgb_sweep_FHD
140_misconceptions_0530_srgb_sweep_FHD
140_misconceptions_0540_srgb_sweep_FHD
140_misconceptions_0550_srgb_sweep_FHD
140_misconceptions_0560_srgb_sweep_FHD
140_misconceptions_0570_srgb_sweep_FHD
140_misconceptions_0580_srgb_sweep_FHD
140_misconceptions_0590_srgb_sweep_FHD
140_misconceptions_0600_srgb_sweep_FHD
140_misconceptions_0610_srgb_sweep_FHD
Exit full screenEnter Full screen
previous arrow
next arrow
 

Crazy, right ? It is the exact same setup using different “Display Rendering” primaries to apply the tonescale. Here is the list of operations :

  • First, we choose the source gamut of the input. In this case, it is ACEScg.
  • Then we convert to some “Display Rendering” primaries. Using a simple Matrix 3×3.
  • Then we apply the tone-mapping. In this example, it is the “Siragusano” one.
  • And then we convert back to the”Display Gamut”, using again a simple Matrix 3×3.

And for clarity, this beautiful quote :

The working space (primaries) and the Display Rendering Transform need to be considered together.

Daniele Siragusano.

ARRI ALF-2

So, back to our main thread. Let’s have a look at an another rendering solution. I would have loved to talk about ARRI K1S1 since it is probably the most famous colour rendering solution out there. But I was not able to generate an OCIO config based on it. So I will use its newer version as an example : ARRI Look File 2 (ALF-2), which is quite similar.

The ALF-2 transforms don’t diverge too far aesthetically from the K1S1, and include HDR versions. The K1S1 itself was only ever SDR AFAIK. The K1S1 is very simple. Harald explains pretty much exactly what’s in it here.

From Nick Shaw.

Jed Smith was kind enough to summarize the whole K1S1 process here :

Scene-Linear AWG -> Tonemap to display-linear AWG -> 3x3 Matrix to BT.709 -> 3x3 Matrix 10% BT.709 weighted desaturation -> Custom 2.4 Power Inverse EOTF with Linear Extension

Here are the two K1S1 matrices if you’re curious :

ALEXA Wide Gamut to ITU-R BT.709:

 [[ 1.61752447 -0.53728621 -0.08023815]
 [-0.07057271  1.33461286 -0.26404021]
 [-0.02110205 -0.22695361  1.24805592]] 

ALEXA Wide Gamut to ITU-R BT.709, 90% saturation:

 [[ 1.48496742 -0.40117349 -0.08379383]
 [-0.03432004  1.28353567 -0.24921569]
 [ 0.01020356 -0.12187415  1.11167083]]

Okay, let’s have a look at our images :

140_misconceptions_0620_srgb_sweep_FHD
140_misconceptions_0630_srgb_sweep_FHD
140_misconceptions_0640_srgb_sweep_FHD
140_misconceptions_0650_srgb_sweep_FHD
140_misconceptions_0660_light_sabers_FHD
140_misconceptions_0670_chromaticity_paths_FHD
140_misconceptions_0680_chromaticity_paths_FHD
Exit full screenEnter Full screen
previous arrow
next arrow
 

This an interesting colour rendering solution, that is used really often. It is still using the per-channel mechanics that will skew towards the “Notorious 6“. But the change of primaries makes it more “acceptable“. To be very clear about the ALF-2 plot, here is a recap of the values used in the “Red to Yellow Notorious 6 sweep” :

Color ManagementSweep PrimariesRMore GMore GMore GMore GY
Filmic HablesRGB1, 0, 01, 0.03, 01, 0.1, 01, 0.25, 01, 0.6, 01, 1, 0

The fact that I did not have to use crazy small values to get a nice readable plot is also a good sign.

We’re on the right path !

I am also mentioning ARRI K1S1 because it is somehow related to the ACES 2.0 talks and some inacurracies I did in a talk last year (summer 2020). Let’s dive in !

Mea Culpa

ACES History

First of all, a few informations I have recently discovered about ACES worth sharing :

  • The making of ACES 1.0 was “complex” because of “contradictory design requirements”.
  • No CG images were available during the conception of ACES 1.0.
  • Originally ACES was just the color space and encoding (IIF – Image Interchange Framework).
  • ACES 0.1.1 is still available and used by some colorists out there since it is quite popular.
  • ACES 0.1.1 was arguably an interesting color rendering solution as the images below show.
140_misconceptions_0690_aces_example_FHD
140_misconceptions_0700_aces_example_FHD
140_misconceptions_0710_aces_example_FHD
140_misconceptions_0720_aces_example_FHD
140_misconceptions_0730_aces_example_FHD
140_misconceptions_0740_aces_example_FHD
140_misconceptions_0750_aces_example_FHD
140_misconceptions_0760_aces_example_FHD
140_misconceptions_0770_aces_example_FHD
140_misconceptions_0780_aces_example_FHD
Exit full screenEnter Full screen
previous arrow
next arrow
 

All these examples can be found at this Dropbox Paper link. And here’s an article mentioning ACES issues on Spider-Man : Homecoming. The purple fringing issue can be seen on the cars’ headlights examples.

Lego movies and ACES

During the talk I mentioned this video about ACES. A few things worth noticing though :

  • The Lego movies were rendered in Linear – P3-D60, not ACEScg.
  • Lego 1 used ACES 0.1.1.
  • Animal Logic has an in-house colorist using BaseLight…
  • Who used the Gamut compress tool from Baselight 5.0 to create the Rec.709 delivery for Lego Batman.

Hum, another mention of Gamut Compression… This may look like an important ingredient of an Output Transform.

List of ACES movies and Output Transforms

Another mistake I did in the talk was to mention a list of movies done with ACES. Here is the quote :

“ACES has been used in more than 300 movies and is a standard for Netflix. It is currently used by many VFX studios such as ILM, Animal Logic…”

This is not exactly true. Kevin Wheatley mentioned it in the TAC meeting of February 2021 (1:02:28). If I understood correctly, ACES is mostly used for exchanging files and archival (in ACES 2065-1). Not necessarily for colour reproduction at display. I guess it will be impossible anyway to know exactly which movies are using the ACES Output Transforms.

VFX workflows and studio requirements

Here is the quote :

I would come from the to the pragmatic “what actually happens in visual effects world”, and all of the ACES projects that we see do do not use the RRT as is out of the box. Or anything that we’ve come across. Maybe one or two projects that actually supply an LMT and a RRT as intended. But, it’s very rare, mostly, they just throw it all out and replace it with something else. Including ones that are on the ACES website promoting how they were ACES.

Kevin Wheatley.

OMG. I did not expect that ! And it goes on like this :

From a visual effects standpoint, we get given a bunch of lookup tables, that are not at all like the RRT and that probably weren’t derived from it. And so, therefore, I believe they have just wholeheartedly replaced them. And then I believe the DI houses are then stuck trying to reverse engineer how they can shoe-horn this into an ACES project so they can tick the box for a studio requirement.And so, they struggled with that.

Kevin Wheatley.
140_misconceptions_0790_aces_workflow_FHD
140_misconceptions_0800_aces_workflow_FHD
Exit full screenEnter Full screen
previous arrow
next arrow
 

I reckon Joshua Pines said a similar thing at meeting#14. And this exact topic was discussed extensively during the TAC meeting of September 2021. It is worth a watch !

ACES looking better

Finally I have provided two examples in this talk, which are still on my website, saying that ACES was making them look better. Terrible choice of words ! Because, and I hope this post will make up for it, if you pay attention to these renders :

  • Overexposed green goes yellow…
  • Overexposed red goes orange…
  • We don’t see it but overexposed blue goes magenta !

It is basically a “look”, which doesn’t give you a choice. There is a look embedded in the Ouput Transform that you cannot compensate for. And you have every right to like the “ACES look” but I think it also important to be aware of it.

140_misconceptions_0810_aces_example_FHD
140_misconceptions_0820_aces_example_FHD
140_misconceptions_0830_aces_example_FHD
140_misconceptions_0840_aces_example_FHD
Exit full screenEnter Full screen
previous arrow
next arrow
 

Yes, in these examples I have used a Display Transform which is called “DisplayTransform_JzAzBz“. But you don’t need to worry about that yet. The point here is to compare an Output Transform that will distort the colors in an accidental way and one that doesn’t.

It’s just another great experiment from Jed.

Compensating the behaviour ?

There is a couple of animated GIFS (done by Nick Shaw) that shows the issue very well. Could we tweak the values to compensate the ACES Output Transform’s behaviour ? Like in the example above with the blue SRGB sphere : “Hey Chris, why don’t you add just a bit of green to compensate ?” Sure, let’s have a look :

140_misconceptions_0850_hue_skew_FHD
Exit full screenEnter Full screen
 

So, we may think that adding a bit of green would fix the issue ? But it only makes the colors take a different path. In this case, towards cyan instead of magenta. That’s all. You cannot escape the “Notorious 6“.

140_misconceptions_0860_hue_skew_FHD
Exit full screenEnter Full screen
 

Revelation#15 : There is no compensation possible since the Display Transform is behaving like a kitchen sink, or a funnel if you will.

These animated GIFs were originally posted on ACESCentral here. Interesting thread !

Rendering in ACEScg

My final inacurracy was about ACEScg being the “ultimate rendering space” or thinking that ACEScg would make my renders look better. First of all, “looking better” does not mean anything. It is just a “bait” word.

Secondly, I’m not convinced that all animated feature films should be rendered in ACEScg. The rendering primaries (or working space) could be a choice based on the Art Direction, just like the Display Transform.

And that’s our Revelation #16 : RGB rendering is kind of broken from first principles. The different rendering spaces just try to make it a bit less broken.

RGB rendering spectrally speaking is so far away from a real scene that everyone would agree that RGB based rendering is not really what we would like to do. But constraints force us to use RGB based engines […].

Daniele Siragusano.

The only thing I can say is that rendering in ACEScg brings us closer to Spectral Rendering. Not that it looks better. Here a few tests done with some colored spheres. I’ll let you guess which ones have been rendered in ACEScg and which ones in “Linear_sRGB/Rec.709” :

140_misconceptions_0870_acescg_rendering_FHD
140_misconceptions_0880_linsrgb_rendering_FHD
140_misconceptions_0890_acescg_rendering_FHD
140_misconceptions_0900_linsrgb_rendering_FHD
140_misconceptions_0910_acescg_rendering_FHD
140_misconceptions_0920_linsrgb_rendering_FHD
140_misconceptions_0930_acescg_rendering_FHD
140_misconceptions_0940_linsrgb_rendering_FHD
140_misconceptions_0950_acescg_rendering_FHD
140_misconceptions_0960_linsrgb_rendering_FHD
140_misconceptions_0970_acescg_rendering_FHD
140_misconceptions_0980_linsrgb_rendering_FHD
140_misconceptions_0990_acescg_rendering_FHD
140_misconceptions_1000_linsrgb_rendering_FHD
140_misconceptions_1010_acescg_rendering_FHD
140_misconceptions_1020_linsrgb_rendering_FHD
Exit full screenEnter Full screen
previous arrow
next arrow
 

ACEScg is the jack of all trades.

Thomas Mansencal, about RGB Scene Rendering.

So, yeah, normally the “typical” answer about these colored spheres is : “Easy ! Test B has been rendered in ACEScg ! There is more bounce and Global Illumination (GI).” But nope ! That’s the other way around. All of these “Test B” images have been rendering in “Linear_sRGB/Rec.709” and all the “Test A” in ACEScg. Crazy, right ?

But I will not repeat my past mistake here : I cannot honestly say one colorspace is better than the other. They give different results and ACEScg is closer to Spectral Rendering. That’s all I can say.

Ground Truth

So, we have seen different Display Transforms, all based on Per-Channel/RGB lookup. Some of them work better than others. But where does this leaves us ? Is there some kind of neutral ground truth somewhere ? How can we possibly know if what we’re viewing is “correct” ?

I’ll just quickly add that even though a per-channel approach may give “acceptable” results (such as ARRI ALF-2), you’ll probably face its limitations if you are doing multiple deliveries on a show (such as SDR/HDR for instance).

See Daniele’s post about that.

Chromaticity linear approach

I can only think of two options so far… Remember the plots we have done for each Output Transform ? We could consider straight lines in the chromaticity domain as “neutral”, aka no deformation of the chromaticities through the Display Transforms.

These plots with Open DRT (using v0.0.83b2) were the only ones where I did not have to put crazy tiny values to show properly the chromaticities’ paths towards the achromatic axis. Just look how the stepping is regular and neat. No distortion at all !

140_misconceptions_1030_chromaticity_paths_FHD
140_misconceptions_1040_chromaticity_paths_FHD
140_misconceptions_1050_chromaticity_paths_FHD
140_misconceptions_1060_chromaticity_paths_FHD
140_misconceptions_1070_chromaticity_paths_FHD
140_misconceptions_1080_chromaticity_paths_FHD
140_misconceptions_1090_chromaticity_paths_FHD
140_misconceptions_1100_chromaticity_paths_FHD
Exit full screenEnter Full screen
previous arrow
next arrow
 

Here are the values used for these plots. A more regular step to show exactly what’s going on :

Color ManagementSweep PrimariesR+ 0.2 G+ 0.4 G + 0.6 G+ 0.8 G Y
Open DRTsRGB1, 0, 01, 0.2, 01, 0.4, 01, 0.6, 01, 0.8, 01, 1, 0
Filmic HablesRGB1, 0.00001, 0.000011, 0.2, 0.000011, 0.4, 0.000011, 0.6, 0.000011, 0.8, 0.000011, 1, 0.00001
spi-animsRGB1, 0.00001, 0.000011, 0.2, 0.000011, 0.4, 0.000011, 0.6, 0.000011, 0.8, 0.000011, 1, 0.00001
spi-vfxsRGB1, 0.00001, 0.000011, 0.2, 0.000011, 0.4, 0.000011, 0.6, 0.000011, 0.8, 0.000011, 1, 0.00001
Filmic High ContrastACEScg1, 0, 0 1, 0.2, 01, 0.4, 01, 0.6, 01, 0.8, 01, 1, 0
sRGB (ACES)ACEScg1, 0, 0 1, 0.2, 01, 0.4, 01, 0.6, 01, 0.8, 01, 1, 0
sRGB (ACES) ACEScg1, 0, 01, 0.2, 01, 0.4, 01, 0.6, 01, 0.8, 01, 1, 0

I had to use 0.00001 instead of 0 for spi-anim, spi-vfx and Hable’s Filmic in order to get the “path-to-white” working and a proper plot. Hopefully the comparisons still stand !

Chromaticity linear definition

It is probably worth taking a bit of time to define what those words mean. There has been some very interesting debate about them on ACESCentral. Several expressions have been used to describe what we’re trying to achieve : “hue-linear“, “hue-preserving“, “chromaticity preserving” or “dominant wavelength preserving“… I won’t go in which expression seems more appropriate since it has been debated by greater minds here.

So I’ll just quote Thomas because his explanation is simple and efficient :

We are saying that one of its quality [aka a chromaticity linear path to white] is that it preserves the dominant wavelength of any chromaticity coordinates it affects.

Great definition by Thomas Mansencal.

Achromatic images

The second option I have been playing with is super simple and almost “stupid“. We can desaturate at 100% our images and the tonality is back ! It sounds ridiculous but it is an interesting exercise to do. It was inspired to me by the excellent video by Peter Postma (tree example at 04:55).

There are probably more sophisticated ways to calculate the Luminance/Brightness of an image. I have used the simplest one because… I don’t know better !

Let’s have a look !

140_misconceptions_1110_acescg_sweep_FHD
140_misconceptions_1120_acescg_sweep_FHD
140_misconceptions_1130_acescg_sweep_FHD
140_misconceptions_1140_acescg_sweep_FHD
140_misconceptions_1150_acescg_sweep_FHD
140_misconceptions_1160_acescg_sweep_FHD
140_misconceptions_1170_acescg_sweep_FHD
140_misconceptions_1180_acescg_sweep_FHD
140_misconceptions_1190_light_sabers_FHD
140_misconceptions_1200_light_sabers_FHD
Exit full screenEnter Full screen
previous arrow
next arrow
 

You need to use the proper “luminance math.” in order to desaturate correctly based on the primaries. You can use Jed’s Saturation tool for ACEScg.

That’s what I did in this post.

I think it is an interesting exercise because it can be very difficult to define exactly what looks “wrong” with the “colour versions“. Let’s try :

“It looks weird”, ‘it looks flat”, “it doesn’t look natural”, “it looks clipped”… are few expressions I may have used to describe these images.

What is tone ?

And we’re coming to a key element here ! All these previous examples use a s-curve for tone-mapping. But what exactly does this mean ? What do these two words mean ?

  • Tone” as in “Tonality“, which is a range (or a system) of tones used in a picture.
  • A range of tones ? Like luminance for instance ? Or brightness ?

So what would Tone-Mapping mean then ? This is best definition I have found online :

The handling of gray scale information in a patterned, predictable way.

Rory Gordon nailing it.

I also like very much this definition :

We seek to map luminance (or brightness) to the display.

This is what Tone-Mapping should do. Definition by a fellow color nerd.

Revelation #17 : According to this definition, we have never tone-mapped. Never. Yeah, thats’s a big one. Hard to believe, right ? But if you followed my steps closely, you should see and realize that none of the OCIO Configs described in this post maps properly the brightness to the display. None of them.

And if you have difficulties to trust me, I don’t blame you ! But just look at these two images and tell me which one is using a simple sRGB EOTF and which one is “color managed” ?

140_misconceptions_1210_cornell_box_FHD
140_misconceptions_1220_cornell_box_FHD
Exit full screenEnter Full screen
previous arrow
next arrow
 

And even better, there is an amazing post on ACESCentral about tone-mapping that was a game changer for me ! Let’s have a look !

An historic post

So let’s dive in this great post by Sean Cooper :

I think the term “Tone Mapping” and “Tone Mapping Operator” largely comes from the computer graphics side of the coin whereas in historical terminology the topic as a whole is referred to as “Tone Reproduction” and the “Tone Reproduction Curve”. Where the former is largely used in the context of Image to Image “mapping”: HDR EXR to SDR JPEG for example. Whereas the latter definition is largely used in reference to Scene to Image “reproduction”, where both the Scene and resulting Image have “Tone” or “Tone Scale”, but the objective quality metric is the “reproduction” of “Tone” from one domain to another.

About tone

Sean goes on like this :

The term “Tone” its self is not scientific. There is a reason this does not show up in any color appearance models, nor the CIE Vocabulary, it is exclusive to the media/imaging community. I would place the origin of this concept in our community/domain first to Hurter and Driffield and their pioneering work on sensitometry, and later to L.A. Jones for his investigation of Tone Reproduction across the full imaging chain. I’ll leave it to the reader to explore the works of these individuals.

About Jones Diagram

This is really an historic post :

Specifically with respect to Jones and his 1920 “On the theory of tone reproduction, with a graphic method for the solution of problems”. He defines the problem space as “…the extent to which it is possible by the photographic process to produce a pictorial representation of an object which will, when viewed, excite in the mind of the observer the same subjective impression as that produced by the image formed on the the retina when the object its self is observed…”. He then goes on to say “The proper reproduction of brightness and brightness differences…is of preeminent importance…”.

About Brightness

It is really an amazing explanation :

Brightness does have a definition in the color science community, to quote Fairchild “…visual sensation according to which an area appears to emit more or less light”. Jones utilises an explicit simplifying assumption, which I believe is largely implicit in most discussions of “Tone Curves” or “Tone Mapping Operators”. That is, that all objects in the scene are non-selective, and the imaging forming mechanism is also non-selective. Non-selective meaning it absorbs/reflects all incident electromagnetic energy equally. He uses this simplification because “under such conditions values of visual brightness are directly proportional to photographic brightness.”

There’s more to it. Let’s keep going !

There are of course deviations in perceived brightness of the scene and reproduction with respect to many factors, reflective selectivity being one of them, that breaks the simplifying assumption. However, the simplifying assumption is used both for convenience, and to focus analysis on the primary principal component of Image Reproduction, or more specifically Tone Reproduction. With the added assumption that the dominant illuminant in the scene and the image viewing environment is the adopted illuminant, you could call this problem space more specially “Neutral Scale Tone Reproduction”.

About Reproduction

All that to say, “Tone Reproduction” is largely focused on the reproduction of Scene brightness (and relative brightness) of non-selective objects onto a non-selective Image medium.

About “Film”

I would argue that the implicit assumption (of neutrality/ non-selectivity) is generally held by most that discuss the topic, and that rather the reference to “Film” is significantly more vague and less defined than “Tone”. “Film” being an incredibly varied and complex quantum mechanical piece of technology with over a century’s worth of innumerable manifestations. “Film” has done many things, from image my teeth at the dentist, help prove Einstein’s general theory of relativity, to being scratched with a knife by an animator and projected.

This is the part where I struggle the most : define “Film”.

We would do well to define “Film” in the same rigour as “Tone”, because I don’t believe the objective here is to emulate a single batch of a single brand of photo-chemical film processed on a certain day.

S-curves and Path-to-white

Finally this leads us directly to our next point… We have explained earlier that “s-curves” have nothing to do with this desired “path-to-white“. Nothing ! So what is at play here ?

Revelation #18 : the s-curve is only one module out of several for a Display Transform. This took me a while as well ! And without this enlightening post from Jed, I would have probably never figured it out myself !

If you have read carefully this post, you may have noticed that I have hinted at a very special ingredient. So special that it is only used in “Filmic Blender” and “OpenDRT“. Yes, I’m talking about Gamut Compress/Mapping. Check this demo out :

140_misconceptions_1230_gamut_visualization_FHD
140_misconceptions_1240_gamut_visualization_FHD
140_misconceptions_1250_gamut_visualization_FHD
140_misconceptions_1260_gamut_visualization_FHD
140_misconceptions_1270_gamut_visualization_FHD
Exit full screenEnter Full screen
previous arrow
next arrow
 

Revelation #19 : In digital RGB land, the path-to-white must be engineered. Values do NOT go magically to white on a digital monitor. They can “only reach” 100% emission on R, G and B channels.

But wait ! If all these OCIO configs use the same mechanics (per-channel or RGB lookup), why some of them “go-to-white” and some don’t ? Here are a few hints :

  • A change of primaries in the Output Transform, like ACES.
  • Gamut compression, like Filmic Blender.
  • Some secret sauce as well may interfere like the ACES “sweeteners” or the “ARRI Desaturation Matrix“.

I won’t go more in detail here but yeah, that’s a big one. The path-to-white must be engineered properly for digital displays. I’m still properly amazed by this finding !

Chromaticity linear looks weird

Okay, let’s take a deep breath. Where were we ? Yes, we were looking at Open DRT and its “chromaticity-linear” approach on the path-to-white. So, problem solved ? Colors don’t get distorted… Are we happy with that ? Well not exactly ! And this is where the most fascinating part kicks in.

Revelation #20 : Even if a “chromaticity linear” approach is desirable, it does not create necessarily pleasing images. It is just a layer of “ground truth” that we may build upon.

As suprising as it sounds, we do NOT want necessarily straight lines in the chromaticity domain to display our images. Because if you have a look at the BT.709 blue primary going to white on a straight line in the 1976 CIE diagram, you may observe what we call the Abney Effect.

140_misconceptions_1280_srgb_sweep_FHD
140_misconceptions_1290_srgb_sweep_FHD
140_misconceptions_1300_chromaticity_paths_FHD
140_misconceptions_1310_chromaticity_paths_FHD
Exit full screenEnter Full screen
previous arrow
next arrow
 

Yup, we have just opened a big can of worms : psychophysical effects happening in our brain (or to be more accurate in the “Human Visual System“). And you can only imagine the challenge it is to model those phenomenons with mathematics ! Because it is not only the Abney Effect… There’s a whole bunch of them !

It was Abney’s experiments which showed that a straight line to adopted white in chromaticity domain does not produce a hue isoline.

Another great post by Sean Cooper.

Color Appearance Models

We have now entered the fascinating world of Color Appearance Models, which divide in four types :

  • Contrast appearance (Stevens Effect, Simultaneous contrast…)
  • Colorfulness appearance (Hunt effect…)
  • Hue appearance (Abney effect, Bezold-Brücke effect…)
  • Brightness appearance (Helmholtz-Kohlrausch effect, lateral-brightness adaptation…)

And Jed is using (in v0.0.83b2) the ICtCp colorspace to address this perceptual facet. Fascinating ! This is another argument in favor of a “chromaticity linear” approach. How can you build a perceptual system (based on sensation) without nailing first the stimulus part (the chromaticities) ?

I would strongly encourage anyone to read Sean’s answer where he describes hue isolines in the Dolby ICtCp colorspace.

What an eye opener !

This is what per-channel may never give you : a strong fundation to build your house on top. It is a happy engineered accident. There is this quote about Color Appearance Models that I found very interesting. It is from Ed Giorgianni, Color Management for Digital Cinema. I have slightly modified it to avoid any ambiguity :

Revelation#21 : For color appearance, the colorimetry of a displayed image always must be altered (rendered) in an engineered and controlled way from that of an original live scene.

Because per-channel is altering the colorimetry in an accidental way…

If you want to know more about the ACES Output Transform and Color Appearance models, I would suggest to read this thread from… 2017 !

TCAMv2

So now we have understood what Display Transforms are about (hopefully !), you may ask yourself : is there an OCIO config out there that addresses the different topics discussed in this post ? And the answer is yes ! There is one !

TCAMv2 description

I would like to state clearly that TCAMv2 is NOT open-source. It is the work of Filmlight’s engineers and I am only mentioning it in this article because I have been given permission. Let’s have a look at the config itself :

displays:
  TCAMv2 100 nits office:
    - !<View> {name: "sRGB Display: 2.2 Gamma - Rec.709", colorspace: "sRGB Display: 2.2 Gamma - Rec.709"}
  TCAMv2 100 nits video dim:
    - !<View> {name: "Rec1886: 2.4 Gamma - Rec.709", colorspace: "Rec1886: 2.4 Gamma - Rec.709"}
  TCAMv2 48 nits cinema dark:
    - !<View> {name: "DCI: 2.6 Gamma - P3D65", colorspace: "DCI: 2.6 Gamma - P3D65"}
  TCAMv2 48 nits cinema dark - MWP D65:
    - !<View> {name: "DCI: 2.6 Gamma - XYZ", colorspace: "DCI: 2.6 Gamma - XYZ"}
  TCAMv2 100 nits videoWide dim:
    - !<View> {name: "Rec1886: 2.4 Gamma - Rec.2020", colorspace: "Rec1886: 2.4 Gamma - Rec.2020"}
  TCAMv2 100 nits officeWide:
    - !<View> {name: "AdobeRGB: 2.2 Gamma - AdobeRGB", colorspace: "AdobeRGB: 2.2 Gamma - AdobeRGB"}
  TCAMv2 1000 nits videoWide dim - OOTF 1.2:
    - !<View> {name: "Rec2100: HLG - 2020 - 1000 nits", colorspace: "Rec2100: HLG - 2020 - 1000 nits - OOTF 1.2"}
  TCAMv2 1000 nits videoWide dim:
    - !<View> {name: "Rec2100: PQ - 2020 - 1000 nits", colorspace: "Rec2100: PQ - 2020 - 1000 nits"}
  TCAMv2 4000 nits videoWide dim:
    - !<View> {name: "Rec2100: PQ - 2020 - 4000 nits", colorspace: "Rec2100: PQ - 2020 - 4000 nits"}
  USE ONLY FOR CHECKING COMPS:
    - !<View> {name: LowContrast, colorspace: "FilmLight: T-Log - E-Gamut"}
  USE ONLY FOR COMPARING RENDERS:
    - !<View> {name: ReferenceLINEAR, colorspace: "FilmLight: Linear - E-Gamut"}

Displays” and “Views” are bit mixed up. But it’s not a big deal since we know how to fix this. This is what I have done in my custom TCAMv2 OCIO config :

displays:
  sRGB 100 nits office:
    - !<View> {name: TCAMv2, colorspace: TCAMv2 - sRGB - 2.2 Gamma}
    - !<View> {name: Raw, colorspace: acescg}
    - !<View> {name: Log, colorspace: acescct}
  Rec.1886 100 nits video dim:
    - !<View> {name: TCAMv2, colorspace: TCAMv2 - Rec.709 - 2.4 Gamma}
    - !<View> {name: Raw, colorspace: acescg}
    - !<View> {name: Log, colorspace: acescct}
  P3D65 48 nits cinema dark:
    - !<View> {name: TCAMv2, colorspace: TCAMv2 - P3D65 - 2.6 Gamma}
    - !<View> {name: Raw, colorspace: acescg}
    - !<View> {name: Log, colorspace: acescct}

active_displays: [sRGB 100 nits office, Rec.1886 100 nits video dim, 
P3D65 48 nits cinema dark]
active_views: [TCAMv2, Raw, Log]

One possible improvement in terms of terminology in the TCAM OCIO Config (and that’s what I have done in mine) is that the “Views” name are ambiguous… If the View is called “sRGB Display: 2.2 Gamma – Rec.709“, there is absolutely no way for me to know if this is using a “simple” 2.2 power law or the actual TCAMv2 DRT.

TCAMv2 settings

I have been asked a couple of times about this config’s terminology. Yes, it is different from the ACES’ one and it can be unsettling… But don’t be afraid by a couple of terms that are here to help you ! Let’s see if we can classify the information from the config :

NameScreen’s brightnessTypeViewing EnvironmentPrimariesWhite PointTransfer FunctionACES “equivalent”
sRGB Display100 nitsofficeXRec.709 X2.2 GammaOutput – sRGB
Rec.1886100 nitsvideodimRec.709X2.4 GammaOutput – Rec.709
DCI48 nitscinemadarkP3D652.6 GammaOutput – P3D65

sRGB and Rec.709 colorspaces have a fixed White Point of D65. I guess this is why it is not specified in the OCIO configs. And I’m not sure why the name “DCI” was used for the “P3-D65” Display Transform. […] Normally “DCI” would imply a DCI whitepoint. […] DCI has a “technical whitepoint” of the greenish DCI white (~D63). However you can encode any whitepoint you want within that container. Commonly this is D60 white in a DCI container. Could be D65 or D55.

Zach and Jed to the rescue.

By having these details directly in the OCIO config, you don’t have to guess which primaries or EOTF is being used. The TCAM OCIO Config is pretty handy since you have most of the important information in one place.

With the ACES Output Transform, if you have any doubt, you can check the CTL reference code. It is actually super easy to read and has lots of useful information in it. For instance, it clearly states for the sRGB ODT :

// Display EOTF :
//  The reference electro-optical transfer function specified in 
//  IEC 61966-2-1:1999.
//  Note: This EOTF is *NOT* gamma 2.2

So please note because this is dead important, that the ACES Output – sRGB uses the sRGB piecewise function when the TCAMv2 sRGB Display uses the Gamma 2.2. That’s a big difference.

TCAMv2 terminology

The TCAMv2 OCIO Config is super light (26 MB) and only contains two families : “SceneReferred” and “DisplayReferred“. Simple and neat ! As you can see in the comparison below, the colorspaces and “general concepts” are the same. So it is fairly simple to switch from one to another (especially if you use “roles” properly) :

ACES OCIO terminologyTCAM OCIO terminology
ACES – ACEScg“ACEScg: Linear – AP1”
Input – ARRI – Linear – ALEXA Wide Gamut“ARRI: Linear – WG”
Output – sRGB “sRGB Display: 2.2 Gamma – Rec.709”
Utility – sRGB – TextureX
Utility – Linear – sRGB “CGI: Linear – Rec.709”

True, “Utility – sRGB – Texture” is missing from the TCAM config. You can either add it yourself or switch to a full “scene-linear” texture workflow.

TCAMv2 visual examples

Cool ! Let’s have a look at some images now to see how TCAMv2 performs.

140_misconceptions_1320_srgb_sweep_FHD
140_misconceptions_1330_srgb_sweep_FHD
140_misconceptions_1340_srgb_sweep_FHD
140_misconceptions_1350_srgb_sweep_FHD
140_misconceptions_1360_light_sabers_FHD
Exit full screenEnter Full screen
previous arrow
next arrow
 

A couple of comments :

  • These are pretty good sweeps ! Path-to-white on sRGB primaries and no “Notorious6” !
  • This looks like a neutral Display Rendering Transform. Finally !

Here is a great explanation about the TCAMv2’s philosophy :

TCAM is different from the other Colour Management workflows because it does not impart a creative “Look” to the display rendering transform. So the creative look belongs to the grading stack and not to the display transform.

Daniele Siragusano.

And this is a great vantage of this Color Management Workflow. For the first time in my career, I’m going to be able to lookdev an asset in a neutral way, without any arbitrary “happy accident” distortion. And then, when we move to the Lighting/Rendering stage, we can choose a “Look” based on our preferences. Awesome ! If you want to know more about TCAMv2, here is a great video of explanation :

But, wait… Should we check our ACEScg images through TCAMv2 to see how they look ? Sure, let’s do that ! But before a quick recap on the different Color Management Workflows and their colorspaces.

Different colorspaces

I have been told that in order to do a fair comparison between different Color Management Workflows, I should do sweeps in their “native primaries“. For this article I rather decided to do sRGB sweeps for two main reasons :

  • sRGB/BT.709 primaries are our lowest common denominator (as explained above).
  • In the studios where I have worked, textures were still done in this colorspace. Probably for the following reasons :
  • Some softwares (such as Substance Painter) have no OCIO implementation.
  • By painting in ACEScg, you can reach extra-saturated values (such as lasers)… Not ideal for albedos !
  • Most resources (such as Megascans) are still in “Linear -sRGB” as far as I know.
  • Legacy. When you have a library of several terabytes in “Linear – sRGB” in several formats and bit depths…

For sure, some places (such as Rising Sun Pictures) have been texturing in ACEScg since 2017. I am mostly talking about my “limited” experience in a couple of animation studios here.

But I thought it was interesting to observe (as I was building these OCIO configs) that each Color Management Workflow comes with a “Linear space” and a “Log space“. I will just list them here for your information :

Linear” spaceLog” space
ACESACES 2065-1 / ACEScgACEScc / ACEScct
ARRILinear – ALEXA Wide GamutLogC – ALEXA Wide Gamut
REDLinear – REDWideGamutRGBLog3G10 – REDWideGamutRGB
SonyLinear – S-Gamut3S-Log3 – S-Gamut3
TCAMv2Linear – E-GamutT-Log – E-Gamut

“Native” primaries sweeps

And yes, for completeness I have done these sweeps. I’m not sure if they tell us anything relevant, especially in a Full CG context. But here they are :

140_misconceptions_1370_acescg_sweep_FHD
140_misconceptions_1380_alexa_sweep_FHD
140_misconceptions_1390_red_sweep_FHD
140_misconceptions_1400_egamut_sweep_FHD
Exit full screenEnter Full screen
previous arrow
next arrow
 

And finally here are our ACEScg images viewed with TCAMv2 :

140_misconceptions_1410_acescg_sweep_FHD
140_misconceptions_1420_acescg_sweep_FHD
140_misconceptions_1430_acescg_sweep_FHD
140_misconceptions_1440_acescg_sweep_FHD
140_misconceptions_1450_light_sabers_FHD
Exit full screenEnter Full screen
previous arrow
next arrow
 

One main observation :

  • We do see some Gamut Clipping and Hue Skews by using ACEScg primaries. This is far from ideal.

Why is this happening ? Well, there is actually a reason for that. Let’s dive in !

TCAMv2 philosophy

Revelation#22 : just like the Filmic Blender, the Display Rendering Transforms from TCAMv2 should not be used without a “Look“. Ideally, you should not use one without the other. As Daniele explains :

TCAM does not include Preferred Colour Reproduction by design but PCR is vital for a successful colour management workflow.

Daniele Siragusano.

So a choice has been made for TCAMv2 to have a “loose” Display Transform, that does the strict minimum to give colourists as much freedom as they need. Otherwise, if the Display Transform is too “restrictive”, it can become a pain to achieve a certain “Look“. There is this great video that explains it properly :

We are going to describe now a very important component of the TCAMv2 success : “Looks“. For clarity, I would like to state that TCAMv2 Looks are proprietary of FilmLight and should not be used without buying a license.

TCAMv2 looks

What are “Looks” exactly ? What do we do with them ? They’re just a creative grade baked into a LUT, right ? How hard can it be to generate them ? That’s what I have thought for a very long time until Daniele said during one of the VWG meetings :

Building a good LMT is not trivial.

Daniele Siragusano, during meeting #14.

Hum… I have probably misunderstood something then… This comment got me thinking really really hard ! Because a “Look” (or “LMT”) is generally a color correction baked into a LUT and that’s pretty much it. So I asked and I got this amazing answer :

Where an LMT is “just a grade” (done in ACES) it’s trivial. Where you want to match a reference achieved by a completely different method, and often also define what happens with values which could not exist in your reference (e.g. match a film print look without limiting the dynamic range to that of a film print) it is absolutely not trivial.

Nick Shaw to the rescue.

Okay, I did not expect that. Fascinating, really ! Revelation#23 : a look can contain much more than a grading node baked in. You can put way more complex and engineered stuff in a 3d LUT than a “simple” grading node.

Looks and Studios

I completely acknowledge here that my understanding of “Looks” is pretty limited. I’d say that “Looks” are not very common in most animation studios I have worked at. I’ll quote Kevin Campbell here :

I’d make a stringer distinction between normal working practice in feature animation shows vs. normal working practice on vfx shows. On animation shows (with a normal budget) we only really start to think about lmt choices close to the end when a DI house is chosen. On vfx projects, I’d expect the lmt to be chosen before they shoot on set, and a version of that lut would be used for on-set dailies. […] I’m used to having the look/lmt come from the client at the start of a vfx project, and being different on every project.

Thanks Kevin for the precision. There are indeed differences between VFX and Animation workflows.

So let’s study one “Look” from TCAMv2 called “C-105 Vision” ! I have especially chosen this look on purpose because it is one of the most impressive from TCAMv2 and has a making-of available online about its genesis.

C-105 Vision Look

Here is its description :

C-105 Vision – This look is a spectral simulation of a modern North-American camera negative and print film process. Main difference to other core looks is the correct rendering of saturated dark green tones.

There is a making-of about this look (at 00:26:25) where Daniele explains his method to spectrally simulate a film-negative. Yep, that’s crazy impressive ! This is probably one of the best Color Rendering solution out there. Do not hesitate to watch :

So, let’s have a look at our images with the C-105 Vision Look :

140_misconceptions_1460_acescg_sweep_FHD
140_misconceptions_1470_acescg_sweep_FHD
140_misconceptions_1480_acescg_sweep_FHD
140_misconceptions_1490_acescg_sweep_FHD
140_misconceptions_1500_light_sabers_FHD
Exit full screenEnter Full screen
previous arrow
next arrow
 

You have every right to like or dislike these images : the look is a creative and subjective choice. But here is the key : you have a choice ! So you could imagine for each feature film, to sit with the Director and the Art Director looking at sweeps, images, references and to choose your “Preferred Color Reproduction“.

This “Look” could be based on the story, the characters or the emotion you want to convey. It could be set at a movie, sequence or even shot level. Really the possiblities are endless since you’re starting with a neutral Display Transform. Power to the image makers.

This is exactly what Robert Eggers and Jarin Blaschke have done for their movie “The Lighthouse” (trailer, article and video about its look).

Revelation #24 : in a full CG feature film, chosing an OCIO Config (and even the rendering primaries) is as important as chosing the camera, lenses and film stock for a live-action Director of Photography !

True; ARRI, RED and Sony also provide some looks for their Color Management Workflows.

Thanks Scott for the links !

Quality and Idendity

As a studio or freelance artist, you may be looking for two things :

  • Quality.
  • Visual Identity.

And if all studios and 3D shools end up using the same per-channel solution, we may never have both. So here is a little experiment I did to show you better the differences, especially the reange of values between different Display Transforms. Most of it was already described here on ACESCentral.

This test has been inspired by this thread on AC. I had been looking for a while to render something “iconic” for the Virtual Working Group (VWG). And by reading the notes on Gamut Compression, I thought : what is more iconic than a can of Coke ?

So I did a bit of research to get the coca-cola red values, convert them to “linear-sRGB” and ended up with an exr texture of (0.9047, 0, 0). I then lit with an HDRI (the “treasure island” one) and a distant light. Here are the results :

140_misconceptions_1510_coke_can_FHD
140_misconceptions_1520_coke_can_FHD
Exit full screenEnter Full screen
previous arrow
next arrow
 

To avoid any confusion, here are the exact steps of this test :

  1. A single render in the BT.709 footprint.
  2. Which was then taken to the wider gamut working space of ACEScg (in Nuke).
  3. Each was rendered from ACEScg through their respective transforms.

Conclusion

Over the past years, colour has become a real obsession for me. I’m pretty sure I have lost some brain cells trying to understand this stuff. But hopefully, it was worth it ! Per-channel has given us for “free” a behaviour that seemed “filmic” (reduced saturation in highlights for instance)… But hopefully we know now the limits of this approach.

There is a desperate need for an open-source Color Management Workflow and ACES has filled this spot. Especially for the CG community. But we must not forget that ACES is just one solution among many. And hopefully, in the near future, more solutions will emerge because the research is still going on.

Revelation #25 : Digital Colour Rendering is not achieved. It is a work-in-progress and there are still many things to understand.

This is my original sin and I have learned from it the hard way : there is no perfect Color Management Workflow. Hopefully, by sharing with you these new informations, I have made up for my past errors. Because when someone’s wrong, you can either hide your head in the sand or learn from your mistakes and fix them.

The journey

On a personal level, I have learned so many things about myself and how truly studying things. Almost like a philosophical journey. Here are a few of the lessons learned :

  • We need brains, not role models.
  • Use your eyes, trust your guts.
  • No appeal to authority, test for yourself !
  • The goal of this post is to think. Please examine what is said !

As the “most dangerous man in America” once said :

Think for yourself. Question authority.

Timothy Leary.

Obviously, I would like to thank Jed Smith for providing all these wonderful Nuke scripts and his support during the writing of this post. I would not have been able to generate any of the plots without him ! And I would like to finish with my favorite quote about Display Transforms, ever :

I cannot understand why anybody would like to limit all of their productions to use the same output transform. It would be the same as limiting the Production to use a single camera. Documentaries, features, animation, hand drawn, all of them have their unique challenges. Do you think film would have flourished in the last 100 years if the Academy would have standardised the chemical receipt ? Instead the academy standardised the transport mechanism. The 35mm perf. And this was exactly the right thing to do. People could innovate and interchange.

Daniele Siragusano.

Finally, once you have seen images through a neutral display transform, there is no going back. It is like seeing new and true colors for the first time. It is a leap of faith. Are you willing to take it ?

140_misconceptions_1530_morpheus_pills_FHD
Exit full screenEnter Full screen
 

Summary

Through a series of 25 revelations, hopefully we have understood some stuff about Display Transforms :

  1. In the “nuke-default” OCIO config, the Rec.709 Transfer Function is a camera encoding OETF !
  2. With the “nuke-default” OCIO config, you will ALWAYS end up on the “Notorious 6“.
  3. An s-curve has nothing to do with the “Filmic” behaviour.
  4. OCIO is NOT implemented the exact same way in all softwares.
  5. It is rather recommended to edit the standard OCIO configs.
  6. Displays and Views have been swapped in the ACES OCIOv1 configs.
  7. The ACES Output Transforms do not solve anything. They’re essentially a Matrix 3×3 and a tonescale.
  8. Having a path-to-white does not mean that the Display Transform is “correct” nor “neutral”.
  9. A 1D s-curve LUT is applied on logarithmic (or semi-log) values.
  10. Any curve that converges at 1 will end up on the “Notorious 6“.
  11. All the scenes I have lit, rendered and displayed for the past 15 years have the “Notorious 6” look.
  12. A “beauty” exr file is NOT an image, it is a file containing light data.
  13. There are two rendering operations in CG : one from scene to data and one from data to image.
  14. Applying an s-curve on different primaries will yield a different result.
  15. There is no “compensation” possible with the ACES Output Transforms.
  16. RGB rendering is kind of broken from first principles (compared to real/spectral light transport).
  17. I have never used a proper tone-mapping solution. Does it exist ?
  18. The s-curve is only one module out of several for a “modern” Display Transform.
  19. In digital RGB land, the path-to-white should be engineered.
  20. Even if a “chromaticity linear” approach is desirable, it does not necessarily create pleasing images.
  21. The colorimetry of the displayed image must always be altered from that of the scene.
  22. The Display Rendering Transforms from TCAMv2 should not be used without a “Look“.
  23. A look can contain much more than a grading node baked in.
  24. Chosing an OCIO Config is as important as chosing the camera, lenses and film stock !
  25. Digital Colour Rendering has NOT been achieved yet.

Sources

A series of threads on ACESCentral discussing the ACES 2.0 Output Transforms :

Some links to OCIO archeology :