Digital Cinema Image Representation Signal Flow

the butterfly in the studio set. The viewer will ... of the visual stimuli being reflected off the illuminated .... distinguish between the finest color differences of neg-.
3MB taille 13 téléchargements 375 vues
Digital Cinema Image Representation Signal Flow By John Silva

The purpose of this paper is to initiate the reader to digital cinema as a technology, to discuss image signal flow down the system pipeline in feature post-production, as well as through the distribution network and associated theater system; to provide short tutorials on related technologies, and to enumerate important technical considerations and related aspects that need to be understood in making the system perform as intended.

SMPTE Motion Imaging Journal, April 2006 • www.smpte.org

pproximately 200 professional volunteers have laboriously worked over the years, each in various ways, to study related technologies and engineering practices needed to form this new and unique digital cinema moving image technology. Along the way, special technical ad-hoc committees were established to develop a practical digital cinema system having appropriate technical standards, recommended practices, and engineering guidelines. The latter will serve to guide and support a successful business model that will benefit and include major film studios, feature post-production facilities, satellite and high-speed network carriers, digital cinema theater owners and crews, and last but not least, satisfied theater audiences. It should be noted that even though the digital cinema system must be well defined, there is an extremely fine-line between related standards (requirements) and the many implementations involved to make the system capable of delivering overall cinema picture quality that will exceed that for 35mm film answer prints as viewed by audiences in film-based theaters. In particular, this includes existing, and yet to be developed theater digital projection systems, at affordable prices to be extended to theater owners. Therefore, standards written must define system requirements, but must not exclude manufacturers from developing improved and competitively priced supporting equipment, including digital projection systems. This was and is being made possible through the formulation of related Recommended Practices and Engineering Guidelines as supporting documents to the standards. These implementations will be discussed in this paper and will serve as examples to

A

137

DIGITAL CINEMA IMAGE REPRESENTATION SIGNAL FLOW allow a full description of successful and proven methods that can be used within the system. Beyond this, it will be up to manufacturers to develop digital projectors for digital cinema theaters and other related equipment that will deliver improved picture and sound quality with reasonable and affordable pricing.

History Until a few years ago, the two entities of TV broadcast and motion picture film production chose to ignore each other; each acknowledging that the other was to be tolerated, but not respected. Television-oriented people felt comfortable with their agenda of breaking news, sporting events, talk shows, syndicated shows recorded on videotape. Film-oriented people, on the other hand, concluded that the mediocre picture quality produced by television, compared with that of film, would never serve the needs of higher quality theater markets all over the world. Then digital image technology was born. This was followed in later years with the introduction of high-definition television (HDTV), which ultimately provided TV-generated picture quality far superior than that produced by NTSC, the first, and still existing, Monochrome/Color-compatible Television System in the U.S., named in honor of the National Television Service Committee that developed it in 1941. In the years following, digital image technology became the backbone of post-production workflow, first for HDTV production, and then for the creation of special effects used within feature films. Later, its usage was extended to signal correction and processing, as provided in feature film post-production, with eventual transformation back to film for theater exhibition. This was termed the Digital Intermediate (DI) process. Time proved that digitally-performed special effects and signal processing for film were not only considerably less expensive, but could be accomplished in far less time and with noticeable improved picture quality. As the years progressed, digital cinema, the technology behind electronic movies, became a viable concept, offering a promising future business model for major film studios and producers. Now digital projectors are beginning to replace existing film projectors in movie theaters all over the world. 138

Feature program content is originated from both film and digital cameras. Also, digitally captured image data signals for both types of origination sources are being processed through digital cinema post-production pipelines before reaching intended theaters via high-speed networks, including satellite communications. Standards for this new technology that will support the new digital cinema business model are now being written. In the meantime, the digital post-production workflow, through significant technology advancements, has made giant steps forward in providing improvements in digital data signal processing, color-correction/enhancement, and working data storage network devices. Together, this has resulted in the elimination of multitudes of nonrealtime signal processing bottlenecks, providing vast improvements in resultant perceived picture quality as viewed by audiences in theaters. As a result, motion picture feature film producers have required that their directors and cinematographers follow the new digital workflow all the way through from beginning to end. This will ensure that the storytelling can be extended and/or enhanced by the use of digital color processing to produce image enhancements in colors, shades of gray, and textures, which will serve to produce desired emotional feelings to audiences. Now, meaningful dialogs between film and television camps are taking place on a daily basis.

Digital Cinema Moving Image Technology By definition, digital cinema is a modern, electronic moving image technology that was conceived and designed to provide a completely new business model for producing digitized feature movies to be shown on screens in digital cinema theaters throughout the world, without the necessity of film prints and filmbased projectors. Digital cinema signal flow process throughout the system is divided into three phases: mastering, distribution, and exhibition. • Mastering includes post-production development of the Digital Cinema Distribution Master (DCDM) from Digital Source Master (DSM) playout. • Distribution includes the transmission of the completed and uncompressed DCDM down the digital cinema network, followed by essence signal compression, encryption, packaging, transport via high-speed SMPTE Motion Imaging Journal, April 2006 • www.smpte.org

DIGITAL CINEMA IMAGE REPRESENTATION SIGNAL FLOW networks to intended theaters. Distribution concludes with all layers of DCDM being stored on disk memory. • Exhibition includes playout from disk memory, DCDM layer separation, implementation of intense security measures, use of auxiliary data to control lights, curtains, etc., at the theater, essence signal decryption and decompression where required, signal transformation, and final projection for audience viewing. The digital cinema system flow path, for description purposes in this paper, will be extended to the system front-end to include both film and digital camera scene acquisition, production, and feature post-production. Digital cinema is datacentric in form, meaning selected origination signals from film or digital cameras are processed, edited, distributed, and projected in digital form. It has two basic acquisition modes, each relating to scene image capture (origination). These are either motion picture film or digital camera acquisition. (1) Motion picture film acquisition Scene images are captured with motion picture film cameras. This is followed by chemically developing the exposed film into what is termed the original negative. In most cases an interpositive, second-generation copy, is then made from the original negative for film scanning use. This content will consist of film clips of feature segments, and camera shots in bits and pieces, which will then be assigned to designated working film reels in accordance with the feature script. Using this origination content, an edit decision list (EDL) will be developed in an offline session to determine the actual frame sequences that will be filmscanned in feature post-production. Once filmscanned, the resultant digitized image representation signals will be transformed into image representation coded data files, and then transferred to a working digital archive in feature post-production, to become working digital negatives called “digital reels.” In this state, the selected raw program content is immediately ready for post-production feature assembly. Stored data in the digital archive can be acquired almost instantly, to be processed and edited as desired at specifically designated workstations without the necessity of first making copies for usage. Once processed by a workstation, the digitized content is

returned to the working digital archive to reside as the “feature-in-process” workmaster, and is then available to be subsequently processed or modified by other workstations along the pipeline. Within the combined array of workstation operations in feature post-production, the following types of signal correction and processing can be made to the delivered digital negative after film origination:

SMPTE Motion Imaging Journal, April 2006 • www.smpte.org

139

(1) Color correction/grading (2) Gamma (3) Cropping (4) Lift (5) Painting/special effects (6) Dust busting (7) Grain matching (8) Noise reduction (9) Compositing (10) Final editing (2) Digital camera origination In this mode, scene images will be captured with digital cameras that will produce digitized image representation signals in coded-data form, on playout. As in the case for film acquisition, the digitized playout signals will be transferred to the working digital archive in feature post-production. From there they will become working digital negatives, immediately available for image processing and editing, with the exception that dust-busting and grain-matching will not be needed. Due to recent significant advancements in digital technology, such as high-speed, high-bandwidth, and uncompressed digital data flow, as well as the evolution of the Storage Area Network (SAN) with a common file system; equipment supporting the Gigabyte System Network (GSN), High-Speed Data Link (HSDL), and other related technologies, are now becoming available. Therefore, when implemented in feature post-production, this equipment will not only allow immediate acquisition by workstations of working digital archival data content, but will further provide realtime and semi-realtime processing, which has not been available until now. In his paper, “A Datacentric Approach to Cinema Mastering,”1 Thomas True clearly explains what has and is happening in mastering methodology, which is currently available to digital cinema, and represents good news for its implementation for the present and

DIGITAL CINEMA IMAGE REPRESENTATION SIGNAL FLOW capture and digitize feature film footage. For SDTV and 1920 x 1080 HDTV content, telecine film chains will work. For 4096 horizontal pixel counts, such as will be employed for digital cinema, the upgrade to high-quality motion picture film scanners will be required. Briefly stated, motion picture film scanners, which are somewhat similar to telecine film chains, but like digital cameras, do not define a direct color space and associated primary set. This will be defined as the image representation Figure 1. Scene object as perceived from reflected visual stimuli. signals progress along the system where they will be used to feed a digital cinema reference display device, such as, feature post-production workstation control or screening room projector or monitor.

How Light Translates to Dye Densities on Negative Film

Figure 2. Motion picture film raw stock light-sensitive layers.

the future. Back to system flow realities, the colorless butterfly is the object in the scene that the film camera lens is focused on, as shown in Fig. 1. It is represented here in achromatic (monochromatic) form because it has no outward physical properties of color. The butterflyobject instead, reflects specific visible wavelengths of illuminating visual stimuli (electromagnetic energy), which the camera lens then focuses on to successive sprocket-driven frames of raw motion picture film stock. In the diagram, the observer is shown viewing the butterfly in the studio set. The viewer will perceive its various colors by virtue of the visual stimuli being reflected off the illuminated insect and entering his eyes and passing through his internal visual system. Therefore, the observer’s perceived butterfly-object appearance is represented as a colored object. The above applies as well to all illuminated scene objects, cameras, and scene observers in the Digital Camera Acquisition mode.

Regarding Film Scanners Telecine film chains or film scanners are used to 140

Figure 2 represents a cross-section of 35mm motion picture raw film stock before exposure. It consists of four separate layers, three of which are individual coatings of silver-halide crystal grains, which provide super-imposed mosaics of blue, green, and red-lightsensitive surfaces, all sequentially coated onto the top surface of a transparent support structure underneath. The top layer is blue-light-sensitive. The third layer is green-light-sensitive, and the fourth layer is redlight-sensitive. Each light-sensitive layer is chemically treated during the manufacturing process to provide its desired individual spectral sensitivity. The second (yellow-colored) layer sequentially coated between the blue- and green-light-sensitive layers acts as a blue filter protecting the green- and red-light-sensitive layers, which have a discernable sensitivity to blue light. This is due to certain wavelengths of blue spectral sensitivity overlapping with wavelengths of those for green and red. This yellow filter layer will become colorless once the film is chemically processed (developed). When film stock is exposed to scene visual stimuli via a film camera and lens, each layer of silver-halide crystals change in chemical character. This occurs in accordance with scene light exposures incrementally reaching each of the overlaid light-sensitive surfaces of each film frame in the form of latent negative images. When the film is chemically processed: • Blue layer mosaics of exposures are replaced with SMPTE Motion Imaging Journal, April 2006 • www.smpte.org

DIGITAL CINEMA IMAGE REPRESENTATION SIGNAL FLOW equivalent respective mosaics of proportional densities of complementary-colored yellow dye. • Green layer exposures are replaced with equivalent respective mosaics of proportional densities of complementary-colored magenta dye. • Red layer exposures are replaced with equivalent respective mosaics of proportional densities of complementary-colored cyan dye. The collective summation of these three layers of complementary-colored dye densities, frame-byframe, make up the original negative film, which in this negative form contains valid image representations of visual stimuli from the original scene.

Also, the film scanner will not change the color representation of the film material. This means that if the color space and primary-set of the visual display includes all possible film colors, the above top-of-theline film scanner will be compatible with this color space as well.

The Generation and Progression of Image Representation Signals in Digital Cameras

Most transmission scanners are essentially tricolor densitometers. The process begins by transmitting a white-light source through sequential sprocket-driven frames of chemically processed motion picture film consisting of overlaid mosaics of yellow, magenta, and cyan dye densities that serve as light-modulated filters. This results in three combined sources of yellow, magenta, and cyan visual stimuli, which represent individual residual amounts of the respective dye-filtered white source light. These combined sources, which have retained their individual identities to respective sources of visual stimuli from the original scene, are next sorted by color separation optics and then directed over separate paths into individual associated photodetector sensors. In combination, these produce triplets of complementary color-formed, negative analog image representation signals. From here, the triplet signals are each quantized and coded in digital form for subsequent processing down the pipeline, but have yet to define an associated color space and primary-set. This will not be done until the image representation signals reach a point in system flow where they will feed a feature post-production color-control or screening projector or monitor. To make this happen, a matrix will be applied that will translate from film dye code values (valid image representation signals) to the primaries of the display device. As a top-of-the-line film scanner will be designed to distinguish between the finest color differences of negative (and positive) film material, the digital data output will be related to the full colorimetric content of the film.

Digital cameras do not have direct primaries. Instead, they have “taking characteristics,” which, for practical reasons in manufacturing, are altered versions of the calculated ideal color-matching function curves for digital cameras. The plot-points for these curves are calculated, starting with the primaries of the control monitor or projector used, to adjust the digital camera controls to ensure or produce acceptable pictures. These ideal color-matching functions are spectral responsivity curves related to perceptual color vision of the average, color normal, human observer (i.e., CIE 1931 2° standard observer2). Originally, in 1931,2 the plot-points for these ideal curves were determined by the use of a colorimeter, which allowed a qualified observer to provide perceptual color matches between two adjacent semicircular areas, called fields. The first field was formed by a projection of successive predetermined single, monochromatic wavelengths of visual stimuli (the reference field). The second matching field was formed by the resultant visual stimuli produced by superimposed projections of individual and adjustable intensities of a particular set of three independent red, green, and blue primary light sources. The process in 1931 was done by changing the reference stimuli in incremental steps, wavelength by wavelength, throughout the visible spectrum, and providing color matches by the observer, adjusting the individual intensities of the three RGB tristimuli. As can be seen, this was a somewhat tedious process. However, in practice, it is not necessary to repeat the experiment for different sets of primaries. Instead, the color-matching functions corresponding to any given set of primaries, such as those of a particular image display device (e.g., CRT monitor or digital projector) can be readily computed.

SMPTE Motion Imaging Journal, April 2006 • www.smpte.org

141

Basic Film-Scanner Action

DIGITAL CINEMA IMAGE REPRESENTATION SIGNAL FLOW

Figure 3. Ideal RGB color-matching function curves (1931).

Figure 4. Ideal RGB color-matching function curves (1931) (modified in camera design to produce all positive spectral sensitivities).

In the image acquisition process for digital cameras, preliminary image representation signals are formed from scene visual stimuli through the sequential action of a taking lens, separation optics, RGB optical trim filters, followed by individual R, G, and B pickup devices. The combined optical action of these elements provide a similar but altered filtering of the primary sources of visual stimuli, in a manner somewhat, but not directly related to the calculated ideal colormatching function curves as mentioned. However, in actual practice, using a direct relationship with visual display color matching functions will not produce the correct or desired results. Calculated color-matching functions corresponding to any set of physically realizable visual display primaries will have some negative lobes, such as is shown in Fig. 3. To compound the issue, in actual practice, the color-matching function portions of negativity are even greater than shown in Fig. 3. The original CIE experiments were done with monochromatic matching primaries (each having a 1 nm bandwidth), which produced curves with less negativity than those in real-world situations where primary stimuli having much greater bandwidths exist. Since the calculated color-matching functions that define the theoretically desired spectral sensitivities of the digital camera have significant portions that represent negative values, those respective sensitivities cannot be

physically realized as such. Therefore, real cameras are built with optical components and sensors that produce all-positive spectral sensitivities that will be somewhat similar, but not identical, to the CIE XYZ curves as shown in Fig. 4. As a result, signal values produced by a sensor having such spectral sensitivities are always positive. However, interpolating forward, those sensitivities, and all other sets of all-positive sensitivities, will correspond to display primaries that are not physically realizable. Therefore, real cameras designed under these criteria would correspond to imaginary displays, and real displays would correspond to imaginary cameras. To resolve this dilemma, a matrix is applied to these image signals to transform the signal values to those that would have been formed if the camera had been able to implement the theoretical sensitivities corresponding to the color-matching functions of the actual display primaries. It is at this point in the signal path that some negative signal values are created in the process. How and when these signals are processed as they proceed along the digital cinema pipeline, must be determined by system/equipment designed to meet post-production Reference Projector and theater projector viewing requirements. For example, they might simply be clipped, or remapped (gamma mapped) to produce a more pleasing, or intended, appearance. Nothing can be done to increase the color gamut defined by the chromaticity boundaries of the actual display devices used.

142

SMPTE Motion Imaging Journal, April 2006 • www.smpte.org

Digital Camera Image Representation Signal Creation and Processing

DIGITAL CINEMA IMAGE REPRESENTATION SIGNAL FLOW For this reason, the display matrix should be applied as late as possible in the pipeline, or some means will need to be employed to retain the negative values for use with other types of displays having larger color gamuts. In addition, the actual spectral sensitivities of a film scanner or digital camera will not correspond directly to any set of color-matching functions. For many practical reasons, including minimizing noise buildup, deliberate departures from color-matching functions are made. Despite the fact that the resultant pictures are obtained using manufacture-skewed taking-characteristics, the image representation signals serve to be satisfactorily representative of the color-producing performance of the feature post-production workstation image display devices, as well as of the mastering Reference Projector and the digital cinema theater digital projectors that follow. This does not mean that control room monitor or projector development in the future will not produce image control display devices having primaries and displayable picture quality equal or comparable to those required of the Reference Projector and digital cinema theater projection systems of that time. When this does occur, the new representative display primaries will be used to determine associated and calculated ideal color-matching function curves as a starting point in the process, as was mentioned above.3

Content providers have been given the flexibility of producing DSM-coded image data files having color spaces, color primary sets, and white points of their choice. There are also no restrictions regarding pixel matrix resolutions, frame rates, bit depths used, as well as other related metrics. However, they are expected to be given yet-to-be-defined requirements in processing the DSM image data files to produce the individual master distribution formats and for Image DCDM file development.

Digital Cinema Image Color Data Flow

The Digital Source Master (DSM) is a master recording that is developed from digital camera or filmacquired origination content. It is subsequently color corrected, color processed, and edited in feature postproduction. All necessary distribution formats, such as, NTSC, PAL, DTV, DVD, HDTV, and DCDM, etc., are derived totally, or in part, from DSM program playout content, as well as from archival storage. The DSM, when originated from digital cameras, will be processed by individual workstations in feature post-production for working data file archive, signal correction and processing, timing and color correction, editing and conforming, and final color grading. For film-acquisition, the operations are the same as above, but the dust-busting and grain-matching processes are added.

Figure 5 is the digital cinema image flow diagram. It starts with assembling the digital source master in feature post-production, shown in the large blue box in the upper left corner. All signal processing action in feature post-production will be recorded in memory and distributed to theater projectors via metadata track files. These will synchronously pass down the digital cinema network with and via the associated Image DCDM layered files. Also, note the two smaller boxes within the Feature Post-Production box. The blue box labeled DSM represents the finished master designated for DCDM development. The yellow box labeled DSM1 represents a copy of the DSM that provides the playout that serves as a feed for development of the DCDM. The purpose of the DSM copy is to isolate the DCDM development signal media from those intended for additional format sources, such as, DVD, HDTV, and so on. As shown in the flow diagram, the next step in the process is to transform the DSM1 output image reference signals having a particular primary set, color space, whitepoint, and quantization bit rate, to linearized XYZ primary signals with a CIE linearized Image DCDM-specified primary-set, color space, and white point. This operation is shown in the yellow box labeled “DSM TO XYZ Transform.” (Fig. 5.) Again, implementation of these actions will be left to the manufacturers of equipment containing this circuitry. At this system point, the image signals are ready to be encoded into finalized uncompressed Image DCDM signals. This is shown next to the yellow box to the right in the flow diagram. The operation involves applying an inverse 2.6 transfer characteristic of the trio of image representa-

SMPTE Motion Imaging Journal, April 2006 • www.smpte.org

143

Digital Source Master

DIGITAL CINEMA IMAGE REPRESENTATION SIGNAL FLOW

Figure 5. Digital cinema image data flow diagram.

least-significant bit components, to accommodate the full 16-bit TIFF file requirement. In this form, the constrained TIFF frame files will then be fed to the Reference Projector in post-production and subsequently into the compression engine, which performs the first major step in distribution to associated digital cinema theaters. On arriving at these first two destinations, the 12 most-significant bits of the image data will first be selected to reestablish the original 12-bit DCDM encoded form (minus the “zeros”), and then will be processed through the Reference Projector and compression engine as was intended.

tion signals and transcoding them from the DSM output signal bit rate to those having a 12-bit quantization. The reason for processing the data signals to form this combination is to prevent contouring artifacts Presently Recommended Image DCDM on scene objects, that will be viewed by audiences. Specifications Up until recently, layered DPX frame files had been The CIE XYZ primary-set, and its respective chosen to serve as carriers of DCDM image represeninverse-matrixed RGB primary-set, both encompass tation data to feed the Reference Projector in post-prothe complete spectrum locus. In fact, both have very duction and the compression engine down the netsimilar, but not identical, respective spectrum loci work. However, DPX files were found to support 10-bit chromaticity coordinates (Figs. 6 and 7). data, but could not work with 12-bit data, which is needed. As a result, the use of DPX files as DCDM carriers has been abandoned, and constrained TIFF frame files, which are designed to work with 16-bit data, have been considered instead. It has been proposed that the 12-bit quantized Image DCDM signals be transposed into the 12 most significant bit components of the constrained TIFF frame files, and to place Figure 6. CIE XYZ primary set. Figure 7. CIE RGB primary set. “zeros” in each of the 4 unused 144

SMPTE Motion Imaging Journal, April 2006 • www.smpte.org

DIGITAL CINEMA IMAGE REPRESENTATION SIGNAL FLOW displays, there are several aspect ratios used for features, as shown on visual display screens. To accommodate this, active pixels are reduced horizontally and vertically, accordingly, by introducing black pixels on respective picture raster edges (Fig. 8). Note that on a given screen raster, the maximum possible number of active pixels equal number of black pixels plus number of active pixels, both horizontally and vertically. However, to keep the bit rate as low as possiFigure 8. Showing active and black pixel arrangement to accommodate ble, only data representing active pixels will particular projected aspect ratios. be sent down the network via the Image The major specifications are: DCDM container. (1) Full bandwidth XYZ (not RGB) image represen(6) Encoding primary chromaticities presently rectation signals will be coded for compression, without ommended are shown in Table 1. color subsampling. (7) The recommended white point will be EE (equal (2) Digital cinema frame rates will be 24 and 48 energy white point), for which the chromaticities are frames/sec. shown in Table 2. (3) Pixel formats supported are those that can The basic encoding formulas that are applied to each achieve horizontal and vertical image resolutions respective X, Y, and Z tristimulus component in the producing resultant picture detail better than that triplet to accomplish the desired nonlinearity, as reprerealized in 35mm film production. sented within the dotted enclosure of the flow diagram (4) The Image DCDM (image distribution master) mentioned above are: serves as an image representation signal container for all elements that make up pictorial content. 1(a) Among these, pixels are the smallest visible picture elements for all displayed images on the screen. 1(b) The maximum numbers of active horizontal and vertical pixels that make up projected image content in a screen raster are designated horizontal and 1(c) vertical resolutions, respectively. (5) Presently, digital cinema has two classes of Table 1 active pixel resolutions related to the number of active pixels that make up image content across a Primaries x y u’ v’ digital display raster horizontally, and the number of Red 1.0000 0.0000 4.0000 0.0000 active pixels that make up image content vertically Green 0.0000 1.0000 0.0000 0.6000 down the display raster. The first class is 2048 x Blue 0.0000 0.0000 0.0000 0.0000 1080; the second (producing higher resolution) is 4096 x 2160. The DCDM, as a carrier, delivers the image signals that represent streams of pixel conTable 2 tent as related to associated visual stimuli, to the Reference Projector screen in mastering and to White point intended digital cinema theater projector screens. (Illuminant) x y u’ v’ As a point of clarification, both classes of pixel resoEE 0.33330 0.3333 0.2105 0.4737 lutions represent the maximum possible number of horizontal and vertical pixels in each case. For visual SMPTE Motion Imaging Journal, April 2006 • www.smpte.org

145

DIGITAL CINEMA IMAGE REPRESENTATION SIGNAL FLOW Where CV is the calculated code value for the specified encoded tristimulus component. CVmax = 2b  1, where b is the signal quantization bit depth of 12 bits. X, Y, and Z are the linear image representation triplet components, prior to being processed for  = 1/ 2.6 nonlinearity in the encoding process. These will ultimately produce reflected triplets of respective R, G, and B linear visual stimuli, as measured off the intended projector screen. When X', Y', and Z' encoded triplet components are all equal to CVmax, the resultant pixel triplet of visual stimuli as measured off the front of the projector screen will produce a luminance value equal to the presently recommended 14.0 fL (48 cd/m2). Presently, K is the normalizing (scaling) constant, recommended to be 52.37, as compared to the previously recommended 48.0 value. Both of these values will produce a maximum luminance value of 48 cd/m2. The rationale used in choosing the 52.37 value is related to the present selection of the EE chromaticity for the DCDM white point. This value will expand the encoded color gamut to include D65 as a potential alternate for the DCDM white point, if desired in the future.  (gamma) is the power coefficient, recommended to be 2.6. ( is the desired nonlinearity determinant). It has been calculated that by conforming to all of the above, a contrast ratio of approximately 10,000:1 can be accommodated, even though a universal use of a 2000:1 Reference Projector and exhibition projector contrast ratio is desired for digital cinema at the present time.

where each is subtracted from the luminance signal (e.g., Y-R and Y-B). In constant luminance encoding, the color difference signals carry luminance content, making the eventual decoded R, G, and B signals subject to inherited luminance noise. When the above noise-effect does occur, the ultimate picture quality, as viewed on a theater screen, will be reduced accordingly.

The Reference Projector

All luminance information is carried by the green primary signal, which also carries its own color components. The red and blue primary signals carry only their respective color component information. This allows separate-luminance encoding to be implemented, using full bandwidth RGB component image signals. It also prevents degradation in subsequently perceived picture quality, but additionally avoids a constant luminance encoding problem where the R and B primary signals form separate color-difference signals

The Reference Projector, located in the mastering post-production screening room, is the most important working visual display in the entire digital cinema system, from scene-to-screen. It is used by cinematographers, directors, and other key post-production personnel to make creative and critical color judgments and decisions on all feature program content. It is highly recommended that the Reference Projector: • Be a real, color-calibrated, working projector in post-production; • Have performance specifications that meet the requirements defined in the SMPTE Recommended Practice for Reference Projector and Environment; • Operate within a controlled viewing environment as specified in the Reference Projector Recommended Practice; and • Serve as the visual display reference for all digital cinema theater projectors having the same target performance specifications, so that consistent and repeatable color quality for both mastering and exhibition can be achieved. Ideally, all post-production workstation displays and their environments where creative decisions are made, should also match the Reference Projector in regard to image and color appearance and performance parameters, within the specified tolerances. This can be better accomplished if the workstation visual displays are digital projectors. This will allow meaningful creative decisions to be made at workstations before and during the final color grading step on the Reference Projector. This is particularly important for occasions where cinematographers and other key feature production personnel sit beside the colorist at a workstation to mutually make creative in-process decisions for features. This will increase the likelihood that such content will be accepted by cinematogra-

146

SMPTE Motion Imaging Journal, April 2006 • www.smpte.org

An Important Advantage in Using the CIE Color Space and Primary Set for the Image DCDM

DIGITAL CINEMA IMAGE REPRESENTATION SIGNAL FLOW

Figure 9. Xenon primary-set compared to that for Rec. 709.

phers and directors when combined to make up the finished feature, and viewed on the Reference Projector screen containing unnecessary post-production costs.

DCDM Image Signal Flow to the Reference Projector As mentioned above, the uncompressed Image DCDM X'Y'Z' image data, which is output-referenced, starts its distribution journey in one or both of two distinct paths. The first leads to the Reference Projector in feature post-production for director and cinematographer viewing. Along this path the now uncompressed and completed DCDM X'Y'Z' data must first be transformed to output-referenced linear RGB input data in the projectors native RGB color space, primary-set, and designated white point. Output-referenced means referenced to projected visual stimuli representing program content, as reflected and viewed, or measured off the front of the projector screen. Note in the flow diagram (Fig. 5), that this operation is indicated within the dotted rectangle enclosing the “X'Y'Z' to XYZ Transform” box followed by the “XYZ to Projector RGB Transform” box. This shows an implementation consisting of two separate operations that together perform the X'Y'Z' to projector RGB transformation. This is an implementation method that has SMPTE Motion Imaging Journal, April 2006 • www.smpte.org

been tested and proven to work. However, if manufacturers can devise a more efficient method, they are encouraged to incorporate it into their equipment. In this implementation, nonlinear X'Y'Z' image data is first transposed to linear XYZ data, which can be done by applying a 2.6 gamma transfer characteristic to the input data via a lookup table. This is followed by transforming the required linear XYZ data through a 3 x 3 matrix to the Reference Projector light modulator as linear RGB image representation input signals. The reason for this needed linearity is explained in the following paragraph. It is very important to understand that transform matricies, as used in digital cinema, are linear signal processing entities. As such, they require linear signal inputs, and in turn, produce linear output signals. As linear entities, the associated arithmetic is reasonable in complexity and in cost, as opposed to the alternative of transformation of nonlinear signals by other matricies. The matrix is shown below in Equation 2.

(2)

The Reference Projector’s light modulator then processes these image representation data triplets to ultimately produce, project, and focus respective R, G, and B linear pixels of light (visual stimuli) derived from the projector’s Xenon light source onto the screen for producer, director, and cinematographer viewing. At this present state of the art, the Reference Projector color space, primary-set, and white point are defined by commercially available projectors having Xenon light sources. These are shown in Fig. 9, and Tables 3 and 4. 147

DIGITAL CINEMA IMAGE REPRESENTATION SIGNAL FLOW Table 3 White point (Illuminant)

x

Xenon

0.3140 0.3510 0.1908 0.4798

y

u’

v’

not needed to be compressed or encrypted. Open/close curtains, lights up/lights down, etc., are some of the functions in theaters that are controlled by the auxiliary data.

JPEG2000 Compression

The DCDM was originally conceived to be a layered set of DPX-formatted files, to separately contain Image DCDM binary-coded data files, subtitle DCDM binary-coded data files, captions DCDM multiple tracks, audio DCDM multiple tracks, and presentation supporting auxiliary multiple tracks. As discussed above, the DPX-formatted files are now being replaced with 16-bit constrained TIFF files for the reasons mentioned. Image DCDM files must be prepared for compression, which follows, and then be encrypted for security purposes. Subtitle DCDM files must also be prepared for compression and then optionally be encrypted, as decided by the content provider. Audio DCDM, which includes multiple tracks, may be optionally compressed and may be optionally encrypted. Captions DCDM, which includes multiple tracks, is not needed to be compressed, but may be optionally encrypted. Auxiliary DCDM, which includes multiple tracks, is

The compression technology chosen for digital cinema is JPEG2000. This was selected by Digital Cinema Initiatives, referred to as “DCI.” DCI is a Limited Liability Company formed by the seven major film studios: Disney, Fox, Metro Goldwyn Mayer, Paramount Pictures, Sony Pictures Entertainment, Universal Studios, and Warner Brothers, to ensure that digital cinema standards were written to adequately protect and enhance their combined business model for this technology. Their working members as a team have become an integral part of the digital cinema standardforming process, as well as for the Recommended Practices and Engineering Guideline supporting documents. The technical description for this technology and its adaptation to digital cinema has been adequately and deeply covered in a book and white paper written by David S. Taubman and Michael W. Marcellin.4 It should be noted that JPEG2000 was selected because of its tremendous flexibility, as well as its ability to deliver excellent picture quality. One important feature is its ability to compress both 2k (2048 x 1080) and 4k (4096 x 2160) pixel resolutions with one pass of 4096 x 2160 down the network. Either 2k and/or 4k resolutions can be programmed to be sent to selected digital cinema theaters. Theaters where only 4k is sent will have the choice of using it as such or to extract the 2k from the 4k, depending on the projector capability, without any loss of picture quality. A second feature is that the JPEG2000 compression engine is primary-set independent. Another advantage is that separate related signals simultaneously passing through can be selectively compressed or ignored and seamlessly passed along together. For example, accompanying metadata in MXF track files are not compressed, but are sent along in the output codestreams, in sync with the compressed image representations. Among other features, JPEG2000 also has the ability to extract subframe objects within full frames without any loss of quality.

148

SMPTE Motion Imaging Journal, April 2006 • www.smpte.org

Table 4 Xenon Primaries

x

Red Green Blue

0.6800 0.3200 0.4964 0.5255 0.2650 0.6900 0.0986 0.5777 0.1400 0.0600 0.1628 0.1570

y

u’

v’

Exhibition Projector Input Color Data Flow The second path for the uncompressed DCDM data leads to the distribution network, where in sequential order, the following operations take place: compression, encryption, packaging, transport to the intended theater, and all related program content is stored on disk drives at the digital cinema theater. These will be discussed below.

DCDM Data File Layers

DIGITAL CINEMA IMAGE REPRESENTATION SIGNAL FLOW

Final Steps in the Delivery of Features to Intended Theaters Once the applicable DCDM files have met their individual compression and/or encryption requirements, all of the above layers of files will be combined together into a media package for transport to intended theaters by (1) high-speed terrestrial network, (2) lowspeed data service, satellite, or (3) courier. An important advantage of digital cinema is that content providers are able to simultaneously transport digital features to intended theaters worldwide. It also provides a much shorter distribution time compared to that required for film, and millions of dollars are saved by not having to generate and physically deliver multigeneration release prints to theaters. In addition, if for any reason multigeneration copies are required during the distribution phase, digital features do not lose picture quality. This is not the case for film release prints. Once the packaged DCDM content is received at a theater facility and recorded on storage disks, the digital cinema distribution phase is complete.

coded, nonlinear X'Y'Z' triplet form. From this point on, as shown in the flow diagram (Fig. 5), the recovered DCDM image representation signals are processed in the same manner that the original uncompressed DCDM signals were when directed to the Reference Projector in mastering. To enumerate, again as an example implementation, the nonlinear DCDM X'Y'Z' triplets are first transposed to linear XYZ data. This is done by applying a 2.6 gamma transfer characteristic to the input data via a lookup table. This also applies to the signal flow to the Reference Projector as described above, as do the decoding formulas that follow. This provides a resultant transfer characteristic of 1.0 (1/2.6 x 2.6 = 1) between this system point at the exhibition projector input and the first instance where linear XYZ image signals were created by transformation from a copy of the DSM (the DSM1) in feature post-production. The decoding formulas that accomplish this are: 3(a)

3(b)

The Exhibition Phase The exhibition phase begins when the stored DCDM 3(c) combined content is played-out for audience viewing, by either an associated data-push server, or a datapull playout device, depending on the theater equipWhere: ment installed. • X, Y, and Z are the linear triplet component code On playout from disk memory, the individual layered values that, when all are equal to CVmax, will ultimately content, of which only some are compressed and produce a recommended 14.0 fL (48 cd/m2) of lumiencrypted, enter what has been named the media nance, as measured off the front of the projector block. There, the combined layers of production conscreen. tent are first individually extracted from the group and • K is a normalizing (scaling) constant recommended then separated into individual data codestreams or data components. From here, for example, the Image DCDM codestream is first decrypted and then decompressed. Then, the image representation signals once again become authentic DCDM image representation codestreams in 12-bit, binaryFigure. 10. Screen object as perceived from reflected visual stimuli off of screen SMPTE Motion Imaging Journal, April 2006 • www.smpte.org

149

DIGITAL CINEMA IMAGE REPRESENTATION SIGNAL FLOW at this time to be 52.37. • CV is the code value for the specified tristimulus component. • CVmax = 2b  1, where b is the bit depth (12-bits). •  (gamma) is the power coefficient (recommended to be 2.6). ( is the desired nonlinearity determinant). This is followed by transforming the required linear XYZ triplets through a final 3 x 3 matrix, producing linear triplets of R, G, and B image representation signals having the target projector’s primary-set, color space, and designated Figure. 11. Effect of differing projector contrast ratios on displayable low-luminance picture white point. content that can be displayed. For the system end-point (measured in candelas per square meter (cd/m2)) that in the theater, Fig. 10 illustrates the final image of the can be revealed to an audience in a given digital cineoriginal camera-captured butterfly that was signal-corma theater. The measurement of screen luminance rected and modified as desired in feature post-produclevel to determine theater black will be made with the tion, and finally projected on the digital cinema theater projector input data triplet code values set to 0, 0, and screen. Notice that the butterfly, as in Fig. 1, appears 0. graphically as a colorless object. This again is done to The maximum theater projector-generated, screenillustrate that projected images are not physically colreflected luminance is presently recommended to be ored objects. Instead, they are arrays of pixilated visu14 fL, or 48 cd/m 2. The ratio of maximum screenal stimuli, produced by the projector in accordance reflected luminance to theater black, without light-spill with respective RGB image representation input signal on the screen, is the projector contrast ratio. triplets. These were modified versions of the image After subjective testing, it has been decided that the representations of visual stimuli originally reflected Reference Projector in post-production must be able from the butterfly at the scene. to deliver a contrast ratio of 2000:1, as measured and An observer viewing the theater screen is also calculated from reflected luminance off the viewing shown. His view on the screen is made up of electroscreen and with no spill-light contamination. In this magnetic energy in the visible-spectrum. However, as case, the darkest luminance revealed off the screen Fig. 10 shows, this array of pixilated electromagnetic (theater black), will be measured at 0.024 cd/m2. This energy allows the observer in the theater to perceive luminance level is subjectively considered equivalent (in color) an acceptable cinematographer-desired to pure black, because the average observer cannot replica of the image of the original butterfly captured at recognize luminance differences below that luminance the scene. level (Fig. 11). Theater Black Observe the luminance (black straight-line) curve for Theater Black is the term used for the darkest proprojectors having a contrast ratio of 2000:1. As illusjector screen luminance level of reflected visual stimuli trated, screen-reflected luminance levels, without light 150

SMPTE Motion Imaging Journal, April 2006 • www.smpte.org

DIGITAL CINEMA IMAGE REPRESENTATION SIGNAL FLOW spill, will change continuously (no horizontal curve displacement) throughout the projector’s lower luminance range of projected visual stimuli. Thus, dark objects occurring in darkened regions of the picture content will be viewed on the screen by the audience, as was intended by the cinematographer, director, and producer. This curve verifies that the lowest luminance that can be produced is 0.024 cd/m2. This is calculated by dividing the maximum luminance (48 cd/m 2) value by the projector contrast ratio (48 cd/m2 / 2000 = 0.024 cd/m2). Unfortunately, two sets of detrimental conditions that cannot be ignored exist in digital cinema theaters, causing low-luminance picture objects not to be viewed by audiences, as was intended by the content providers. The first detrimental condition involves the projector contrast ratio in a given theater. If it is 2000:1, as mentioned, dark objects in dark regions of the picture down to 0.024 cd/m2 luminance levels will be observed on the theater screen, as was intended by the content provider; barring any light-spill on the screen. At the present time, content providers are satisfied that a projector with a 2000:1 maximum contrast ratio will allow them to view as many low luminance objects as needed in post-production, to analyze and make creative decisions regarding image appearance of this program content. The problem, however, is that most digital cinema theater projectors have contrast ratios less than 2000:1, because of the cost. For example, if a theater projector has a contrast ratio of 1000:1, as shown in the red curve in Fig. 1, the minimum luminance level at which it can project on a screen will be 48/1000 = 0.048 cd/m2. As a result, black objects with screen luminance values below 0.048 cd/m2 will not have the same appearances as was desired by the content provider. The curves show that intended black picture content, as input to projectors having contrast ratios lower than 2000:1, will be seen on the screen as being offset upward into the gray luminance regions of the picture. This means that intended black objects in the lower luminance regions will be viewed on the screen as gray objects within gray surrounds. This will most likely not be considered acceptable to the content provider. The solution, of course, is to improve the maximum contrast ratio capability of all digital cinema

theater projectors so that they will have at least 2000:1 contrast ratios—at affordable cost to theater owners. The second detrimental condition that occurs in many digital cinema theaters is spill-light on the projector screen, caused by aisle safety lighting and theater exit lights required by building and safety codes. Unfortunately, this light contamination eliminates any chance of displaying dark objects in darkened regions of the picture as desired. It is extremely important that research be done to find ways to eliminate light-spill on theater screens. Whether this means finding ways to deflect this extraneous light from the screen, or by other means, it needs to be done to maximize the potential image quality of digital cinema features shown in theaters around the world. As a comparative note, the Reference Projector screen in post-production viewing rooms, for all subjective purposes, will adequately reveal desired black program content when projector image representation triplet codes of 0, 0, and 0 are input. The reason being, the Reference Projector used in each will have contrast ratios of 2000:1. Further, the post-production viewing room environment will be effectively void of extraneous spill-light onto the screen, as the facility will not be encumbered with theater safety lighting requirements. In consideration of all of the above, when all digital cinema mastering and theater projectors universally have 2000:1 contrast ratios, and when their individual environments are void of screen spill light, the darkest screen-reflected black level, as projected and measured on the front of the screen—consistent with projector input triplet code values of 0, 0, and 0—will be: 48/2000 = 0.024 cd/m2. When this condition universally exists, from feature post-production to associated theaters, cinematographers, directors, and producers will be confident that all their picture enhancements and overall image appearance subjectively made and approved in postproduction will be viewed as intended, with expected appreciation by audiences at all associated digital cinema theaters. Assuming that the complete digital cinema system works as specified, both observers at the scene and in the theater, as shown in Figs. 1 and 10, will have acceptably equivalent color perceptions of the butter-

SMPTE Motion Imaging Journal, April 2006 • www.smpte.org

151

DIGITAL CINEMA IMAGE REPRESENTATION SIGNAL FLOW fly, as they will for all other picture-objects passing through the system. The observer at the theater, of course, will have the added advantage of viewing enhanced versions of original program content, as was desired by the producer, director, and cinematographer in feature post-production.

Conclusion A lot of information has been included and discussed in this paper. The goal was to initiate readers to digital cinema as a technology, or to serve as a solid review to professionals directly involved with its standards-making process. The discussion started with an overview of the digital cinema standards process, followed by a brief history of the film/television relationship over the years, in which most parties are now participating together. This was followed with (1) an elemental discussion of digital cinema as a moving image technology; (2) film and film scanners and digital camera design considerations and scene content acquisition; (3) a minitutorial on how light translates to dye densities on negative film; (4) basic film scanner action; (5) generation and progression of image representation signals in digital cameras; (6) digital cameras and their relationship to visual displays; (7) the digital source master and its role in DCDM development; (8) the digital cinema image color data flow diagram, followed by (9) signal processing in feature post-production in creating the DSM; (10) constrained TIFF frame files as preliminary DCDM carriers; (11) presently recommended Image DCDM specifications; (12) digital cinema pictures consist of screen rasters of active horizontal and vertical pixels; (13) the DCDM encoding equations with their formula element descriptions; (14) the Reference Projector and its special role in digital cinema, (15) DCDM signal flow to the Reference Projector; (16) exhibition projector input color data flow; (16) DCDM data file layers; (17) JPEG2000 compression; (18) the final steps in the delivery of features to digital cinema theaters; (19) the exhibition phase in digital cinema theaters; (20) the understanding of theater black and the two detrimental theater conditions

152

that must be eliminated before digital cinema can be declared supremely successful.

References 1. Thomas J. True, “A Datacentric Approach to Cinema Mastering,” SMPTE Mot. Imag. J. , 112:347, Oct./Nov. 2003. 2. CIE, 1931, www.cie.org. 3. Edward J. Giorgianni and Thomas E. Madden, Digital Color Management Encoding Solutions, Addison Westley: Reading, MA, 1997. 4. David S. Taubman, and Michael W. Marcellin, JPEG2000 Image Compression Fundamentals, Standards and Practice, Kluwer Academic Publishers: Boston, Dordrecht, London, 2002.

Bibliography Berns, Roy S., Principals of Color Technology, Third Edition, Wiley Inter-Science: New York, 2000. Poynton, Charles, A Technical Introduction to Digital Video, John Wiley & Sons, Inc.: NY, 1969. Poynton, Charles, Digital Video and HDTV Algorithms and Interfaces, Morgan Kaufmann Publishers: NY, 2002. Rast, R. M, “SMPTE Technology Committee on Digital Cinema—DC28: A Status Report, ”SMPTE J., 110:78, Feb. 2001. Wyszecki and Stiles, Color Science, Second Edition, Wiley Inter-Science: New York, 2000.

A contribution received December 2005. Copyright © 2006 by SMPTE.

THE AUTHOR John Silva is known as the “father of airborne news gathering.” In 1958, as chief engineer of television station KTLA in Los Angeles, he conceived, designed, and developed the world’s first airborne news helicopter which was named the “KTLA Telecopter.” In 1970 he received an Emmy for “outstanding achievement in newsgathering.” In 1974 he received a second Emmy for “concept, design, and expertise of the KTLA Telecopter. Silva also designed and developed the world’s first frame-by-frame videotape editor called the TVola in 1961. In 1977 he received the “NAB Engineer of the Year” award. Silva is an active participant and contributor on all four SMPTE Digital Cinema Standards Committees.

SMPTE Motion Imaging Journal, April 2006 • www.smpte.org