For list of authors, see Credits (Chapter 13).
This document is an informal draft of the PNG development group.
It is a proposal, and the format is subject to change.
Comments on this document can be sent to the PNG specification maintainers at one of the following addresses:
Distribution of this memo is unlimited.
At present, the latest version of this document is available on the World Wide Web from
ftp://swrinde.nde.swri.edu/pub/mng/documents/.
In the case of any discrepancy between this extract and the full MNG specification, the full MNG specification shall take precedence.
This document presents MNG-LC (Multiple-image Network Graphics, Low Complexity) and MNG-VLC (Multiple-image Network Graphics, Very Low Complexity) which are proper subsets of the MNG (Multiple-image Network Graphics) format.
MNG is a multiple-image member of the PNG (Portable Network Graphics) format family that can contain animations (slide shows) or complex still frames, comprised of multiple PNG single-image datastreams.
The MNG-LC format uses the same chunk structure that is defined in the PNG specification and shares other features of the PNG format. Any valid PNG datastream is also a valid MNG-LC or MNG-VLC datastream.
A MNG-LC frame normally contains a two-dimensional image or a two-dimensional layout of smaller images. It could also contain three-dimensional "voxel" data arranged as a series of two-dimensional planes (or tomographic slices), each plane being represented by a PNG datastream.
This document includes examples that demonstrate various capabilities of MNG-LC including simple movies and composite frames.
If "231
" looks like
the number "231
"
instead of 2
raised to the power
31
, your viewer is not
recognizing the HTML 4.0 <SUP> tag; you need
to look at the HTML 2.0, ASCII text, or PostScript
version of this document instead.
This document presents low-complexity and very-low-complexity versions (MNG-LC and MNG-VLC, proper subsets) of the MNG (Multiple-image Network Graphics) format.
Note: This specification depends on the PNG (Portable Network Graphics) [PNG] and, for MNG-LC and MNG-VLC applications that are enhanced with JNG support, the JNG (JPEG Network Graphics) specifications. The PNG specification is available at the PNG web site,
http://www.cdrom.com/pub/png/and the JNG (JPEG Network Graphics) specification and the full MNG specification are available at the MNG web site,
http://www.cdrom.com/pub/mng/
A MNG datastream describes a sequence of single frames, each of which can be composed of zero or more embedded images or directives to show previously defined images.
The embedded images can be PNG, JNG, or Delta-PNG datastreams. MNG-LC and MNG-VLC datastreams do not contain JNG (JPEG Network Graphics) datastreams, which are allowed in full MNG datastreams, but MNG-LC and MNG-VLC applications can be enhanced to recognize and process those as well.
A typical MNG-LC or MNG-VLC datastream consists of:
MNG is pronounced "Ming."
When a MNG datastream is stored in a file, it is recommended that ".mng" be used as the file suffix. In network applications, the Media Type "video/x-mng" can be used. Registration of the media type "video/mng" might be pursued at some future date.
The first eight bytes of a MNG datastream are
138 77 78 71 13 10 26 10
(decimal) which is similar to the PNG signature with "\212 M N G" instead of "\211 P N G" in bytes 0-3.
Chunk structure (length, name, data, CRC) and the chunk-naming system are identical to those defined in the PNG specification. As in PNG, all integers that require more than one byte must be in network byte order.
The chunk copying rules for MNG employ the same mechanism as PNG, but with rules that are explained more fully in the full MNG specification.
Note that decoders are not required to follow any decoding models described in this specification nor to follow the instructions in this specification, as long as they produce results identical to those that could be produced by a decoder that did use this model and did follow the instructions.
Each chunk of the MNG datastream or of any embedded object is an independent entity, i.e., no chunk is ever enclosed in the data segment of another chunk.
An independent PNG datastream, with a PNG signature, is also a valid MNG-LC datastream that must be recognized and decoded by MNG-LC and MNG-VLC decoders. An independent JNG datastream must also be recognized and decoded by any MNG-LC or MNG-VLC decoder that has been enhanced to include JNG support. This kind of MNG-LC or enhanced MNG-LC datastream will contain only a single embedded image.
Because the embedded objects making up a MNG are normally in PNG format, MNG shares the good features of PNG:
In addition:
See also the glossary in the PNG and the "terminology" section of the full MNG specification.
`object_id=N'
".
In MNG-LC, only image 0 is permitted.
An embedded visible PNG or JNG datastream generates a single layer, even though it might be interlaced or progressive.
object_id
is
an unsigned sixteen-bit number that serves as the identifier of a set of
object attributes.
In MNG-LC, only object 0 is permitted.
do_not_show
flag to zero.
This chapter describes chunks that can appear at the top level of a MNG datastream.
This section describes critical MNG control chunks. MNG-compliant decoders must recognize and process them ("processing" a chunk sometimes can consist of simply recognizing it and ignoring it; some chunks have been declared to be critical only to prevent them from being relocated by MNG editors).
The MHDR chunk is always first in all MNG datastreams except for those that consist of a PNG datastream with a PNG or JNG signature.
The MHDR chunk contains exactly 28 bytes:
Frame width: 4 bytes (unsigned integer). Frame height: 4 bytes (unsigned integer). Ticks per second: 4 bytes (unsigned integer). Nominal layer count: 4 bytes (unsigned integer). Nominal frame count: 4 bytes (unsigned integer). Nominal play time: 4 bytes (unsigned integer). Simplicity profile: 4 bytes:(unsigned integer). bit 0: 1: Presence or absence of certain features is specified by the remaining bits of the simplicity profile. (must be 1 in MNG-LC datastreams) bit 1: 0: Simple MNG features are absent. 1: Simple MNG features are present. (must be 0 in MNG-VLC datastreams) bit 2: 0: Complex MNG features are absent. (must be 0 in MNG-LC datastreams) bit 3: 0: transparency is absent or can be ignored. 1: transparency is present and is essential (critical) for proper display of the images. bit 4: 0: JNG is absent. 1: JNG is present. bit 5: 0: Delta-PNG is absent. bits 6-15: Reserved for public expansion. Must be zero in this version. bits 16-30: Available for private or experimental expansion. Undefined in this version and can be ignored. bit 31: Must be zero.
Decoders can ignore the "informative" frame-count, layer-count, play-time, and simplicity-profile fields.
The frame_width
and frame_height
fields
give the intended display size (measured in
pixels) and provide default clipping boundaries
(see below).
It is strongly recommended that these be set to zero if
the MNG datastream contains no visible images.
The ticks_per_second
field gives the
unit used by the FRAM chunk to specify frame duration and sync timeout.
It must be nonzero if the datastream contains an animation. When the
datastream contains exactly one frame,
this field should be set to zero. When this field is
zero, the length of a tick is infinite, and decoders will ignore any
attempt to define interframe delay, timeout, or any other variable that
depends on the length of a tick. If the frames are intended to be
displayed one at a time under user control, such as a slide show or
a multi-page FAX, the tick length can be set to any positive number
and a FRAM chunk can be used to set an infinite sync_timeout.
Unless the user intervenes, viewers will only display the first frame
in the datastream.
When ticks_per_second
is nonzero, and there is no other
information available about frame duration, viewers should display
animations at the rate of one frame per tick.
If the frame-count field contains a zero, the frame
count is unspecified. If it is nonzero, it contains the number
of frames that would be displayed, ignoring the
TERM chunk. If the frame count is greater
than 231-1
,
encoders should write 231-1
, representing an infinite frame count.
If the layer-count field contains a zero, the layer
count is unspecified. If it is nonzero, it contains the number of
layers in the datastream, ignoring the
TERM chunk.
If the layer count is greater than 231-1
, encoders
should
write 231-1
, representing an infinite layer count.
If the nominal-play-time field contains a zero, the
nominal play time is unspecified. Otherwise, it gives the play time,
in ticks, when the file is displayed ignoring the
TERM chunk.
Authors who write this field should choose a
value of "ticks_per_second" that will allow the nominal play time
to be expressed in a four-bit integer. If the nominal play time is greater
than 231-1
ticks, encoders should write 231-1
,
representing an infinite nominal play time.
When bit 0 of the simplicity profile is 0, the simplicity (or complexity) of the MNG datastream is unspecified, and all bits of the simplicity profile must be 0. The simplicity profile must be nonzero in MNG-LC and MNG-VLC datastreams.
If the simplicity profile is nonzero, it can be regarded as a 32-bit profile, with bit 0 (the least significant bit) being a "profile-validity" flag, bit 1 being a "simple MNG" flag, bit 2 being a "complex MNG" flag, bit 3 being a "transparency" flag, bit 4 being a "JNG" flag, and bit 5 being a "Delta-PNG" flag. The upper 15 bits (except for the most significant bit, which must be zero) are available for private test or experimental versions, and the remaining bits are reserved for future MNG versions, and must be zero in this version. If a bit is zero, the corresponding feature is guaranteed to be absent, and if a bit is one, the corresponding feature may be present in the MNG datastream.
When bit 1 is 0 ("simple" MNG features are absent), the datastream does not contain the DEFI, FRAM, or global PLTE and tRNS chunks, and if the BACK chunk is present it only defines an advisory background color and can therefore be ignored.
"Transparency is absent or can be ignored" means that the MNG or PNG tRNS chunk is not present and no PNG or JNG image has an alpha channel, or that if they are present they are not essential (or critical) for displaying the images.
A MNG-LC (i.e., a "low-complexity MNG") datastream must have a simplicity profile with bit 0 equal to 1 and all other bits except possibly for bits 1 and 3 ("simple MNG" MNG features and transparency) equal to 0. If bit 4 (JNG) is 1, the datastream is a "MNG-LC with JNG" datastream. MNG-LC decoders are allowed to reject such datastreams unless they have been enhanced with JNG capability.
A MNG-VLC (i.e., a "very low-complexity MNG") datastream must have a simplicity profile with bit 0 equal to 1 and all other bits except possibly for bit 3 (transparency) equal to 0. If bit 4 (JNG) is 1, the datastream is a "MNG-VLC with JNG" datastream. MNG-VLC decoders are allowed to reject such datastreams unless they have been enhanced with JNG capability.
Encoders that write a nonzero simplicity profile should endeavor to be accurate, so that decoders that process it will not unnecessarily reject datastreams. For example, the simplicity profile 31 indicates that JNG, critical transparency, and at least one "complex" MNG feature are all present, but Delta-PNG is not. If the simplicity profile promises that certain features are absent, but they are actually present in the MNG datastream, the datastream is invalid.
The MEND chunk's data length is zero. It signifies the end of a MNG datastream.
The LOOP chunk provides a "shorthand" notation that can be used to avoid having to repeat identical chunks in a MNG datastream. If the LOOP chunk is present, it can be ignored by MNG-LC decoders, along with the ENDL chunk.
The DEFI chunk sets the default object attribute
set (object_id
, potential_visibility
,
concrete_flag
, location, and clipping boundaries) for
any subsequent images that are defined with IHDR-IEND,
or JHDR-IEND datastreams.
Bit 1 of the MHDR simplicity profile can be used to promise that the DEFI chunk is not present.
The DEFI chunk contains 2, 3, 4, 12, or 28 bytes:
Object id: 2 bytes (unsigned integer) identifier to be given to the objects that follow the DEFI chunk. This field must be zero in MNG-LC files. Do_not_show flag: 1 byte (unsigned integer) 0: Make the objects potentially visible. 1: Do not make the objects potentially visible. This field can be omitted if the concrete_flag, location, and clipping boundary fields are also omitted. When it is omitted, the image is potentially visible (do_not_show=0). Concrete flag: 1 byte (unsigned integer) 0: Make the objects "abstract" (image can not be the source for a Delta-PNG) 1: Make the objects "concrete" (object can be the source for a Delta-PNG). This field can be omitted if the location and clipping boundary fields are also omitted. When it is omitted, the object is made "abstract" (concrete_flag=0). MNG-LC decoders can ignore this flag. X_location: 4 bytes (signed integer). The X_location and Y_location fields can be omitted if the clipping boundaries are also omitted. If so, decoders must assume default values {X_location=0, Y_location=0}. Y_location: 4 bytes (signed integer). Left_cb: 4 bytes (signed integer). Left clipping boundary. The left_cb, right_cb, top_cb, and bottom_cb fields can be omitted as a group. If so, decoders must assume default values {0, frame_width, 0, frame_height}. Right_cb: 4 bytes (signed integer). Top_cb: 4 bytes (signed integer). Bottom_cb: 4 bytes (signed integer).
Negative values are permitted for the X and Y location and clipping boundaries. The positive directions are downward and rightward from the frame origin.
The DEFI chunk values are discarded after the object's IEND chunk is processed. If no DEFI chunk precedes an object, the decoder must use the following default values:
Object_id = 0 Do_not_show = 0 Concrete_flag = 0 X_location = 0 Y_location = 0 Left_cb = 0 Right_cb = frame_width Top_cb = 0 Bottom_cb = frame_height
The PLTE chunk has the same format as a PNG PLTE chunk. It provides a global palette that is inherited by PNG datastreams that contain an empty PLTE chunk.
The tRNS chunk has the same format as a PNG tRNS chunk. It provides a global transparency array that is inherited along with the global palette by PNG datastreams that contain an empty PLTE chunk.
If a PNG datastream is present that does not contain an empty PLTE chunk, neither the global PLTE nor the global tRNS data is inherited by that datastream. If it is an indexed-color PNG, it must supply its own PLTE (and tRNS, if it has transparency) chunks.
A PNG (Portable Network Graphics) datastream.
See the PNG specification [PNG] and the PNG Special Purpose Chunks document [PNG-EXT] for the format of the PNG chunks.
Any chunks between IHDR and IEND are written and decoded according to the PNG specification.
If a global PLTE chunk appears in the top-level MNG datastream, the PNG datastream can have an empty PLTE chunk, to direct that the global PLTE and tRNS data be used. If an empty PLTE chunk is not present, the data is not inherited. MNG applications that recreate PNG files must write the global PLTE chunk rather than the empty one in the output PNG file, along with the global tRNS data if it is present. The global tRNS data can be subsequently overridden by a tRNS chunk in the PNG datastream. It is an error for the PNG datastream to contain an empty PLTE chunk when the global PLTE chunk is not present or has been nullified.
The PNG oFFs and pHYs chunks and any future chunks that attempt to set the pixel dimensions or the drawing location must be ignored by MNG viewers and simply copied (according to the copying rules) by MNG editors.
The PNG gIFg, gIFt, and gIFx chunks must be ignored by viewers and must be copied according to the copying rules by MNG editors.
If do_not_show=0
for the image when the IHDR
chunk is encountered, a viewer can choose to display the image while
it is being decoded, perhaps taking advantage of the PNG interlacing
method, or to display it after decoding is complete.
A JNG (JPEG Network Graphics) datastream.
See the JNG specification for the format of the JNG datastream.
Any chunks between JHDR and IEND are written and decoded according to the JNG specification.
The remaining discussion in the previous paragraph about PNG datastreams also applies to JNG datastreams.
The BACK chunk suggests a background color against which transparent, clipped, or less-than-full-frame images can be displayed.
Red_background: 2 bytes (unsigned integer). Green_background: 2 bytes (unsigned integer). Blue_background: 2 bytes (unsigned integer). Mandatory background: 1 byte (unsigned integer). 0: Background color is advisory. Applications can use it if they choose to. 1: Background color is mandatory. Applications must use it. This byte can be omitted. If so, the background color is advisory. In "MNG-VLC" datastreams, this byte must be 0 or omitted.
Viewers are expected to composite every subframe in the MNG datastream against a fresh copy of the background, if the framing mode given in the FRAM chunk is 3 or 4. The images and the background are both clipped to the frame boundaries given in the FRAM chunk.
The three BACK components are always interpreted in the current color space as defined by any top-level gAMA, cHRM, iCCP, sRGB chunks that have appeared prior to the BACK chunk in the MNG datastream. If no such chunks appear, the color space is unknown.
When the BACK chunk appears between FRAM chunks, it applies to the upcoming frame, not to the current one. When the framing_mode is 3, it takes effect immediately prior to the next IHDR or JHDR chunk in the datastream.
Multiple instances of the BACK chunk are permitted in a MNG datastream.
The BACK chunk can be omitted. If a background is required and the BACK chunk is omitted, then the viewer must supply its own background.
In practice, most applications that use MNG as part of a
larger composition should ignore the BACK data if
mandatory_background=0
and the application already has
its own background definition. This will frequently be the case in
World Wide Web pages, to achieve nonrectangular transparent animations
displayed against the background of the page.
The FRAM chunk provides information that a decoder needs for generating layers, subframes, and frames. The FRAM parameters govern how the decoder is to behave when it encounters a FRAM chunk, or an embedded image. The FRAM chunk also delimits subframes.
Bit 1 of the MHDR simplicity profile can be used to promise that the FRAM chunk is not present.
An empty FRAM chunk is just a subframe delimiter. A nonempty one is a subframe delimiter, and it also changes FRAM parameters, either for the upcoming subframe or until reset. When the FRAM chunk is not empty, it contains a framing-mode byte, an optional name string, a zero-byte separator, plus four 1-byte fields plus a variable number of optional fields.
Framing mode: 1 byte. 0: Don't change framing mode. 1: Each embedded image generates a subframe consisting of a single image layer. There is no background layer. Use this mode to avoid unnecessary clearing of the display when the first image covers the entire frame area, and subsequent frames can be displayed properly by simply overlaying them on the prior frame. This is the default framing mode. 2: The group of embedded images appearing prior to the next FRAM chunk form a composite subframe consisting of zero or more image layers. The level, or stacking order, of each layer is given by its order of appearance in the datastream. No interframe delay occurs between the layers. 3: This is the same as framing_mode=1, except that a background layer is generated. Each embedded image generates a subframe consisting of a background layer followed by an image layer. No interframe delay occurs between the background layer and the image layer. 4: This is the same as framing_mode=2, except that a background layer is generated prior to the image layers. Subframe name: 0 or more bytes (Latin-1 Text). Can be omitted; if so, the subframe is nameless. Separator: 1 byte: (null). Must be omitted if all remaining fields are also omitted. Change interframe delay: 1 byte. 0: No. 1: Yes, for the next subframe only. 2: Yes, also reset default. This field and the next three must be omitted as a group if no frame parameters other than the framing mode are changed. Change sync timeout and termination: 1 byte 0: No. 1: Deterministic, for the next subframe only. 2: Deterministic, also reset default. 3: Decoder-discretion, for the next subframe only. 4: Decoder-discretion, also reset default. 5: User-discretion, for the next subframe only. 6: User-discretion, also reset default. 7: External-signal, for the next subframe only. 8: External-signal, also reset default. Change subframe clipping boundaries: 1 byte. 0: No. 1: Yes, for the next subframe only. 2: Yes, also reset default. Change sync id list: 1 byte. 0: No. 1: Yes, for the next subframe only. 2: Yes, also reset default list. Interframe delay: 4 bytes (unsigned integer). Must be omitted if change_interframe_delay=0. The range is [0..2^31-1] ticks. Sync timeout: 4 bytes (unsigned integer). Omit if change_sync_timeout=0. The range is [0..2^31-1]. The value 2^31-1 (0x7fffffff) ticks represents an infinite timeout period. Subframe boundary delta type: 1 byte (unsigned integer). 0: Subframe clipping boundary values are given directly. 1: Subframe clipping boundaries are determined by adding the FRAM data to their previous values. This and the following four fields must be omitted if change_frame_clipping_boundaries=0. Left_fb or delta left_fb: 4 bytes (signed integer). Right_fb or delta right_fb: 4 bytes (signed integer). Top_fb or delta top_fb: 4 bytes (signed integer). Bottom_fb or delta bottom_fb: 4 bytes (signed integer). Sync id: 4 bytes (unsigned integer). Omit if change_sync_id_list=0 or if the new list is empty; repeat until all sync_id's have been listed. The range is [0..2^31-1].
When the FRAM parameters are changed, the new parameters affect the subframe that is about to be defined, not the one that is being terminated by the FRAM chunk.
Framing modes:
framing_mode=1
, each image generates a separate
subframe consisting of a single image layer. In the usual case, the
interframe delay is nonzero, so each subframe is a frame. FRAM
chunks need not appear to separate them.
The following events generate a subframe:
For example (assuming a nonzero interframe delay time), the sequence
FRAM 1 IHDR ... IDAT ... IEND IHDR ... IDAT ... IEND IHDR ... IDAT ... IEND
which will generate three frames, each containing a single subframe consisting of an image layer.
If the BACK chunk is present, encoders must insert a background layer, with a zero delay, ahead of the first image layer in the datastream, even when the framing_mode is 1. This layer must be included in the layer count but not in the frame count.
framing_mode=2
, viewers are expected
to display all of the images at once, if possible, or as fast as can be
managed, without clearing the display or restoring the background. The
next FRAM chunk delimits the subframe. A subframe boundary
also occurs when a SEEK chunk or the MEND chunk
appears.
For example, the sequence
FRAM 2 DEFI 0 0 0 x y IHDR ... IDAT ... IEND DEFI 0 0 0 x y IHDR ... IDAT ... IEND DEFI 0 0 0 x y IHDR ... IDAT ... IEND FRAM
will result in a single subframe containing three layers, each consisting of one image displayed according to its location and clipping boundaries. If the images do not cover the entire frame, whatever was already on the display shows through.
When images in a subframe overlap, viewers are expected to composite the later images against the partially completed subframe that includes all earlier images.
This framing_mode
is fundamentally declarative; it
describes the elements that go into an individual subframe. It is up
to the decoder to work out an efficient way of making the screen match
the desired composition. Simple decoders can handle it as if it were
procedural, compositing the images into the frame buffer in the order
that they appear, but efficient decoders might do something different,
as long as the final appearance of the frame is the same.
If the BACK chunk is present, encoders should insert a background layer, with a zero delay, ahead of the first image layer in the datastream, even when the framing_mode is 2. This layer must be included in the layer count but not in the frame count.
framing_mode=3
, a subframe consisting of the
background layer and an image layer is generated for each image.
A subframe boundary occurs after each image appears. Otherwise,
framing_mode=3
is identical to framing_mode=1
.
Subframes are triggered by the same events that trigger a subframe
when framing_mode=1
.
You can use this mode to show a frame containing only the background, with its own time delay, as in
FRAM 3 (shows background color with interframe delay) FRAM 1 DEFI 0 0 0 x y IHDR ... IDAT ... IEND DEFI 0 0 0 x y IHDR ... IDAT ... IEND DEFI 0 0 0 x y IHDR ... IDAT ... IENDIn this example, the background and the three images will be displayed one at a time against the background, like cards being dealt.
framing_mode=4
, a subframe boundary and interframe
delay occurs when each FRAM chunk appears. A
background layer, consisting of the background image composited against
the background color, is generated immediately after the FRAM chunk.
Otherwise, framing_mode=4
is
identical to framing_mode=2
. A subframe boundary also occurs
when a SEEK chunk or the MEND chunk appears, but neither
of these generates a background layer.
You can make a composite frame consisting of four layers (a background layer and 3 images) with
FRAM 4 (show background, no delay) DEFI 0 0 0 x y IHDR ... IDAT ... IEND DEFI 0 0 0 x y IHDR ... IDAT ... IEND DEFI 0 0 0 x y IHDR ... IDAT ... IEND FRAM (interframe delay, then start next frame)
+--------------+--------------------+-------------------+ | Framing mode | Restore background | Interframe delay | +--------------+--------------------+-------------------+ | 1 | Before first image*| Before each image | | | in the datastream | after the first | | | | in the datastream | +--------------+--------------------+-------------------+ | 2 | Before first image | Before each FRAM | | | after first FRAM | after the first | | | in the datastream | in the datastream | +--------------+--------------------+-------------------+ | 3 | Before each image | Before each image | | | | after the first | | | | in the datastream | +--------------+--------------------+-------------------+ | 4 | Before first image | Before each FRAM | | | following each | after the first | | | FRAM chunk | in the datastream | +--------------+--------------------+-------------------+ | *"image" means an image layer that is generated in | | response to decoding an embedded object, even if no | | pixels are actually drawn due to the image being | | outside the clipping boundaries. | +-------------------------------------------------------+
The subframe name must conform to the same formatting rules as
those for a PNG tEXt keyword: It must consist only of printable
Latin-1 characters and must not have leading or trailing blanks, but
can have single embedded blanks. There must be at least one (unless
the subframe name is omitted) and no more than 79 characters in the
keyword. Keywords are case-sensitive. There is no null byte within
the keyword.
Applications can use this
field for such purposes as constructing an external list of subframes
in the datastream. The subframe name only applies to the upcoming
subframe; subsequent subframes are unnamed unless they also have their
own frame_name
field. It is recommended that the same name
not appear in any other FRAM
chunk. Subframe names should not begin with the
case-insensitive
strings "clock(", "frame(", or "frames(",
which are reserved for use in URI queries and
fragments, as explained in the full MNG specification.
The interframe delay value is the desired minimum time to elapse from the beginning of displaying one frame until the beginning of displaying the next frame. When the interframe delay is nonzero, which will probably be the usual case, subframes are frames. When it is zero, a frame consists of any number of consecutive subframes, until a nonzero delay subframe is encountered and completed. Decoders are not obligated to display such subframes individually; they can composite them offscreen and only display the complete frame.
The sync timeout field can be a number or <infinity>. Infinity can be represented by 0x7fffffff.
The termination condition given in the
change_sync_timeout_and_termination
field specifies how much
longer, after the normal interframe delay has elapsed, the frame will
endure. It can take the following values:
sync_id
, but
must wait no longer than the timeout.
The sync_id
list can be omitted if the termination
condition is not "external-signal".
When the sync_id
list is changed, the number of
sync_id
entries is determined by the remaining length of the
chunk data, divided by four. This number can be zero, which either
inactivates the existing sync_id
list for one frame or
deletes it.
The initial values of the FRAM parameters are:
Framing mode = 1 Subframe name = <empty string> Interframe delay = 0 Left subframe boundary = 0 Right subframe boundary = frame width Top subframe boundary = 0 Bottom subframe boundary = frame height termination = deterministic Sync timeout = 0x7fffffff (infinite) Sync id = <empty list>
The DEFI chunk can be used to specify the placement of each image within the layer. The DEFI chunk can be used to specify clipping boundaries for each image. The subframe boundaries are only used for clipping, not for placement. Even when the left and top frame boundaries are nonzero, the image locations are measured with respect to the {0,0} position in the display area. If the layers are transparent or do not cover the entire area defined by the subframe clipping boundaries, they are composited against the background defined by the BACK chunk, or against an application-defined background, if the BACK chunk is not present or does not define a mandatory background. The background, as well as the images, is clipped to the subframe clipping boundaries. Any pixels outside the subframe clipping boundaries remain unchanged.
The frame_duration
field gives the duration of
display, which is the minimum time that must elapse from the
beginning of displaying one frame until the beginning of displaying
the next (or between images, when framing_mode=1
). It
is measured in "ticks" using the tick length determined from
ticks_per_second
defined in the MHDR chunk.
A viewer does not actually have to follow the procedure of erasing the screen, redisplaying the background, and recompositing the images against it, but what is displayed when the frame is complete must be the same as if it had. It is sufficient to redraw the parts of the display that change from one frame to the next.
The sync_id
list provides a point at which the processor
must wait for all pending processes to reach the synchronization
point having the same sync_id
before resuming, perhaps
because of a need to synchronize a sound datastream (not defined
in this specification) with the display, to synchronize stereo
images, and the like. When the period defined by the sum of the
frame_duration
and the sync_timeout
fields
elapses, processing can resume even though the processor has not
received an indication that other processes have reached the
synchronization point.
Note that the synchronization point does not occur immediately, but
at the end of the subframe that follows the FRAM chunk. If it
is necessary to establish a synchronization point immediately, this can
be done by using two consecutive FRAM chunks, the first setting
a temporary frame_duration=0
, sync_timeout
, and
sync_id
, and the second establishing the synchronization
point:
FRAM 2 0 1 1 0 1 0000 sync_timeout sync_id FRAM 0 name
The identifier sync_id=0
is reserved to represent
synchronization with a user input from a keyboard or pointing device.
The sync_id
values 1-255 are reserved to represent the
corresponding ASCII letter, received from the keyboard (or a simulated
keyboard), and values 256-1023 are reserved for future definition
by this specification. If multiple channels (not defined in this
specification) are not present, viewers can ignore other values
appearing in the sync_id
list.
The TERM chunk suggests how the end of the MNG datastream should be handled, when a MEND chunk is found. It contains either a single byte or ten bytes:
Termination action: 1 byte (unsigned integer) 0: Show the last frame indefinitely. 1: Cease displaying anything. 2: Show the first frame after the TERM chunk. 3: Repeat the animation starting immediately after the TERM chunk. Action after iterations: 1 byte 0: Show the last frame indefinitely after iteration_max iterations have been done. 1: Cease displaying anything. 2: Show the first frame after the TERM chunk. This and the remaining fields must be present if termination_action is 3, and must be omitted otherwise. Delay: 4 bytes (unsigned integer). Delay, in ticks, before repeating the animation. Iteration max: 4 bytes (unsigned integer). Maximum number of times to repeat the animation.
The TERM chunk, if present, must appear either immediately after the MHDR chunk or immediately prior to a SEEK chunk. The TERM chunk is not considered to be a part of any segment for the purpose of determining the copy-safe status of any chunk. Only one TERM chunk is permitted in a MNG datastream.
Simple viewers and single-frame viewers can ignore the TERM chunk. It has been made critical only so MNG editors will not inadvertently relocate it.
Simple decoders that only read MNG datastreams sequentially can safely ignore the SAVE and SEEK chunks.
This section describes ancillary MNG chunks. MNG-compliant decoders are not required to recognize and process them.
The MNG pHYs chunk is identical in syntax to the PNG pHYs chunk. It applies to complete MNG layers and not to the individual images within them.
The MNG top-level pHYs chunk can be nullified by a subsequent empty pHYs chunk appearing in the MNG top level.
The namespace for MNG chunk names is separate from that of PNG. Only those PNG chunks named in this paragraph are also defined at the MNG top level. They have exactly the same syntax and semantics as when they appear in a PNG datastream:
A MNG editor that writes PNG datastreams should not include the top-level iTXt, tEXt, tIME, and zTXt chunks in the generated PNG datastreams.
The following PNG chunks are also defined at the MNG top level. They provide default values to be used in case they are not provided in subsequent PNG datastreams. Any of these chunks can be nullified by the appearance of a subsequent empty chunk with the same chunk name. Such empty chunks are not legal PNG or JNG chunks and must only appear in the MNG top level.
A MNG editor that writes PNG or JNG datastreams is expected to include the top-level cHRM, gAMA, iCCP, and sRGB chunks in the generated PNG or JNG datastreams, if the embedded image does not contain its own chunks that define the color space. When it writes the sRGB chunk, it should write the gAMA chunk (and perhaps the cHRM chunk), in accordance with the PNG specification, even though no gAMA or cHRM chunk is present in the MNG datastream.
JNG (JPEG Network Graphics) is the lossy sub-format for MNG objects. It is described in the full MNG specification and is also available as a separate extract from the full MNG specification. Both documents are available at the MNG home page,
http://www.cdrom.com/pub/mng/
MNG-LC applications can choose to support JNG or not. Those that do not can check bit 4 (JNG is present/absent) of the MHDR simplicity profile to decide whether it can process the datastream.
The chunk copying rules for MNG are similar to those in PNG. Authors of MNG editing applications should consult the full MNG specification for details.
This section specifies the minimum level of support that is expected of MNG-compliant decoders, and provides recomendations for viewers that will support slightly more than the minimum requirements. All critical chunks must be recognized, but some of them can be ignored after they have been read and recognized. Ancillary chunks can be ignored, and do not even have to be recognized.
Anything less than this level of support requires subsetting. Applications that provide less than minimal support should check the MHDR "simplicity profile" for the presence of features that they are unable to support. A specific subset, in which "complex MNG features" and JNG are absent, is called "MNG-LC". In MNG-LC datastreams, bit 0 of the simplicity profile must be 1 and bit 2 must be 0. In MNG-VLC datastreams, "simple MNG features" are also absent, and bit 1 must also be 0.
Some subsets of MNG-LC support are listed in the following table, more or less in decreasing order of complexity.
Level of support MHDR Profile bits Profile 31 30-6 5 4 3 2 1 0 value MNG-LC with JNG 0 0 0 1 1 0 1 1 19 MNG-LC 0 0 0 0 1 0 1 1 11 MNG-VLC with JNG 0 0 0 1 0 0 0 1 17 MNG-VLC 0 0 0 0 1 0 0 1 9 MNG-VLC without transparency 0 0 0 0 0 0 0 1 1
One reasonable path for an application developer to follow might be to develop and test the application at each of the following levels of support in turn:
We are allowing conformant decoders to skip twelve-bit JNGs because those are likely to be rarely encountered and used only for special purposes. There is no profile flag to indicate the presence or absence of 12-bit JNGs.
color_type
, bit_depth
,
compression_method
, filter_method
and
interlace_method
must be supported (interlacing, as in PNG,
need not necessarily be displayed on-the-fly; the image can be displayed
after it is fully decoded). The alpha-channel must be supported, at
least to the degree that fully opaque pixels are opaque and fully
transparent ones are transparent. It is recommended that alpha be fully
supported. Alpha can be ignored if bit 3 of the simplicity profile is
0.
Bit 4 of the simplicity profile can be used to promise that JNG chunks are not present. Viewers that choose not to support JNG can check this bit before deciding to proceed. MNG-LC and MNG-VLC viewers are not required to support JNG.
color_type
, bit_depth
,
compression_method
, filter_method
and
interlace_method
must be supported (interlacing, as in PNG,
need not necessarily be displayed on-the-fly; the image can be displayed
after it is fully decoded). The alpha-channel must be supported, at
least to the degree that fully opaque pixels are opaque and fully
transparent ones are transparent. It is recommended that alpha be fully
supported.
sample_depth=8
must be supported. The JSEP
chunk must be recognized and must be used by minimal decoders to select
the eight-bit version of the image, when both eight-bit and twelve-bit versions
are present, as indicated by JDAT_sample_depth=20
in the
JHDR chunk. When JDAT_sample_depth=12
, minimal
decoders are not obligated to display anything. Such decoders can
choose to display nothing or an empty rectangle of the width and height
specified in the JHDR chunk.
ticks_per_second
must be supported by animation viewers.
The simplicity profile, frame count, layer count, and nominal play time
can be ignored. Decoders that provide less than minimal support can use
the simplicity profile to identify datastreams that they are incapable
of processing.
framing_mode
and clipping parameters must be
supported. The interframe_delay
must be supported
except by single-frame viewers. The sync_id
and
sync_timeout
data can be ignored.
The following recommendations do not form a part of the specification.
It is a good idea to use a single color space for all of the layers in an animation, where speed and fluidity are more important than exact color rendition. This is best accomplished by defining a single color space at the top level of MNG, using gAMA and cHRM chunks, and either an sRGB or iCCP chunk, and removing any color space chunks from the individual images after converting them to the common color space.
When the encoder converts all images to a single color space before putting them in the MNG datastream, this will allow decoders to improve the speed and consistency of the display.
For single-frame MNG datastreams, however, where decoding speed is less important and exact color rendition might be more important, it is best to leave the images in their original color space, as recommended in the PNG specification, to avoid any loss of data due to conversion, and to retain the individual color space chunks if the images have different color spaces.
Always use framing mode 1 or 2 when all of the images are opaque. This avoids unnecessary screen clearing, which can cause flickering.
When a JNG datastream contains an alpha channel, and the file is intended for transmission over a network, it is useful to interleave the IDAT and JDAT chunks. In the case of sequential JPEG, the interleaving should be arranged so that the alpha data arrives more or less in sync with the color data for the scanlines. In the case of progressive JPEG, the alpha data should be interleaved with the first JPEG pass, so that all of the alpha data has arrived before beginning the second JPEG pass.
The PNG specification gives a good explanation of how to composite a partially transparent image over an opaque image, but things get more complicated when both images are partially transparent.
Pixels in PNG and JNG images are represented using gamma-encoded RGB (or gray) samples along with a linear alpha value. Alpha processing can only be performed on linear samples. This chapter assumes that R, G, B, and A values have all been converted to real numbers in the range [0..1], and that any gamma encoding has been undone.
For a top pixel {Rt,Gt,Bt,At} and a bottom pixel {Rb,Gb,Bb,Ab}, the composite pixel {Rc,Gc,Bc,Ac} is given by:
Ac = 1 - (1 - At)(1 - Ab) if (Ac != 0) then s = At / Ac t = (1 - At) Ab / Ac else s = 0.0 t = 1.0 endif Rc = s Rt + t Rb Gc = s Gt + t Gb Bc = s Bt + t Bb
When the bottom pixel is fully opaque (Ab = 1.0), the function reduces to:
Ac = 1 Rc = At Rt + (1 - At) Rb Gc = At Gt + (1 - At) Gb Bc = At Bt + (1 - At) Bb
When the bottom pixel is not fully opaque, the function is much simpler if premultiplied alpha is used. A pixel that uses non-premultiplied alpha can be converted to premultiplied alpha by multiplying R, G, and B by A.
For a premultiplied top pixel {Rt,Gt,Bt,At} and a premultiplied bottom pixel {Rb,Gb,Bb,Ab}, the premultiplied composite pixel {Rc,Gc,Bc,Ac} is given by:
Ac = 1 - (1 - At)(1 - Ab) Rc = Rt + (1 - At) Rb Gc = Gt + (1 - At) Gb Bc = Bt + (1 - At) Bb
As mentioned in the PNG specification, the equations become much simpler when no pixel has an alpha value other than 0.0 or 1.0, and the RGB samples need not be linear in that case.
When a fatal error is encountered, such as a bad CRC or an unknown critical MNG chunk, minimal viewers should simply abandon the MNG datastream.
Decoders are required to be able to interpret datastreams that contain interlaced PNG images, but are only required to display the completed frames; they are not required to display the images as they evolve. Viewers that are decoding datastreams coming in over a slow communication link might want to do that, but MNG authors should not assume that the frames will be displayed in other than their final form.
When a PLTE chunk is received, it only affects the display of the PNG datastream that includes or inherits it. Decoders must take care that it does not retroactively affect anything that has already been decoded.
If a frame contains two or more images, the PLTE chunk in one image does not affect the display of the other.
A composite frame consisting only of indexed-color images should not be assumed to contain 256 or fewer colors, since the individual palettes do not necessarily contain the same set of colors.
Viewers that can only display a single frame must display the first frame that they encounter.
MNG-LC provides three types of clipping, in addition to any clipping that might be required due to the physical limitations of the display device.
frame_width
and frame_height
are defined in
the MHDR chunk and cannot be changed.
This is the only type of clipping available in MNG-VLC datastreams.
Decoders can use these parameters to establish the size of a window
in which to display the MNG frames. When the frame_width
or frame_height
exceeds the physical dimensions of the
display hardware, the contents of the area outside those dimensions is
undefined. If a viewer chooses, it can create "scroll bars" or the
like, to enable persons to pan and scroll to the offscreen portion
of the frame. If this is done, then the viewer is responsible for
maintaining and updating the offscreen portion of the frame.
In the case of a MNG datastream that consists of a PNG or JNG
datastream, with the PNG or JNG signature, the frame_width
and frame_height
are defined by the width
and
height
fields of the IHDR (or JHDR) chunk.
frame_width
and frame_height
)
will remain on the display from frame to frame without being explicitly
redisplayed.
On systems where file names customarily include an extension
signifying file type, the extension .mng
is recommended for
MNG (including MNG-LC and MNG-VLC) files. Lowercase .mng
is
preferred if file names are case-sensitive. The extension .jng
is
recommended for JNG files.
See also the PNG-1.1 draft:
Randers-Pehrson, G., et. al., "PNG (Portable Network Graphics
Format) Version 1.1", which is available at
ftp://swrinde.nde.swri.edu/pub/png/documents/.
Security considerations are addressed in the basic PNG specification.
Some people may experience epileptic seizures when they are exposed to certain kinds of flashing lights or patterns that are common in everyday life. This can happen even if the person has never had any epileptic seizures. All graphics software and file formats that support animation and/or color cycling make it possible to encode effects that may induce an epileptic seizure in these individuals. It is the responsibility of authors and software publishers to issue appropriate warnings to the public in general and to animation creators in particular.
No known additional security concerns are raised by this format.
We use the "#" character to denote commentary in these examples; such comments are not present in actual MNG datastreams.
The simplest MNG datastream is a single-image PNG datastream. The simplest way to create a MNG from a PNG is:
copy file.png file.mng
The resulting MNG file looks like:
\211 P N G \r \n ^z \n # PNG signature. IHDR 720 468 8 0 0 0 0 # Width and Height, etc. IDAT ... IEND
This example demonstrates a very simple movie, such as might result from directly converting an animated GIF that contains a simple series of full-frame images:
\212 M N G \r \n ^z \n # MNG signature. MHDR 256 300 # Width and height. 30 # 30 ticks per second. 7 6 180 # Layers, frames, play time 1 # MNG-VLC simplicity TERM 3 0 120 10 # When done, repeat animation 10 times. FRAM 1 0 2 0 0 0 30 # Set framing_mode=1 (because the # images are opaque) and frame_duration to 1 sec. IHDR ... IDAT ... IEND # Six PNG datastreams IHDR ... IDAT ... IEND # are read and displayed. IHDR ... IDAT ... IEND IHDR ... IDAT ... IEND IHDR ... IDAT ... IEND IHDR ... IDAT ... IEND MEND
\212 M N G \r \n ^z \n # MNG signature. MHDR 720 468 # Width and height. 1 6 5 0 # 1 tick per second. 6 5 5 # Layers, frames, play time. 1 # Simplicity profile (VLC) FRAM 1 0 2 2 0 2 1 600 0 # Set frame_duration to 1, # sync_timeout to 600 sec, and sync_id list to {0}. SAVE SEEK "Briefing to the Workforce" IHDR ... IDAT ... IEND # DEFI 0, visible, abstract SEEK "Outline" # is implied. IHDR ... IDAT ... IEND SEEK "Our Vision" IHDR ... IDAT ... IEND SEEK "Our Mission" IHDR ... IDAT ... IEND SEEK "Downsizing Plans" IHDR ... IDAT ... IEND MEND
These examples in the full MNG specification use features that are not available in MNG-LC.
Outline of a program to convert simple GIF animations that do not use the "restore-to-previous" disposal method to "simple" MNG (or "MNG-LC") format:
begin write "MHDR" and "mandatory BACK" chunks Frame_duration := 0; Previous_mode := 1; Framing_mode := 1; if(loops>1) "write TERM 3 0 0 loops" for subimage in gif89a file do if(Frame_duration != gif_duration) then Frame_duration := gif_duration write "FRAM 0 0 2 2 0 2 0 Frame_duration 0" endif if(X_loc != 0 OR Y_loc != 0) then write "DEFI 0 0 0 X_loc Y_loc" chunk endif write "<image>" if (gif_disposal_method < 1) then /* (none or keep) */ Framing_mode := 1; else if (gif_disposal_method == 2) then /* (restore background) */ Framing_mode := 3; else if (gif_disposal_method == 3) then /* (restore previous) */ error ("can't do gif_disposal method = previous.") endif if(Framing_mode != Previous_mode) then write "FRAM Framing_mode" chunk Previous_mode := Framing_mode; endif end write "MEND" chunk end
Where "<image>" represents a PNG datastream containing a GIF frame that has been converted to PNG format.
Caution: if you write such a program, you might have to pay royalties in order to convey it to anyone else.
Contributors' names are presented in alphabetical order:
This document was built from the file mng-master-19990316 on 16 March 1999.
Copyright © 1998, 1999 by: Glenn Randers-Pehrson
This specification is being provided by the copyright holder under the following license. By obtaining, using and/or copying this specification, you agree that you have read, understood, and will comply with the following terms and conditions:
Permission to use, copy, and distribute this specification for any purpose and without fee or royalty is hereby granted, provided that the full text of this NOTICE appears on ALL copies of the specification or portions thereof, including modifications, that you make.
THIS SPECIFICATION IS PROVIDED "AS IS," AND COPYRIGHT HOLDER MAKES NO REPRESENTATIONS OR WARRANTIES, EXPRESS OR IMPLIED. BY WAY OF EXAMPLE, BUT NOT LIMITATION, COPYRIGHT HOLDERS MAKE NO REPRESENTATIONS OR WARRANTIES OF MERCHANTABILITY OR FITNESS FOR ANY PARTICULAR PURPOSE OR THAT THE USE OF THE SPECIFICATION WILL NOT INFRINGE ANY THIRD PARTY PATENTS, COPYRIGHTS, TRADEMARKS OR OTHER RIGHTS. COPYRIGHT HOLDER WILL BEAR NO LIABILITY FOR ANY USE OF THIS SPECIFICATION.
The name and trademarks of copyright holder may NOT be used in advertising or publicity pertaining to the specification without specific, written prior permission. Title to copyright in this specification and any associated documentation will at all times remain with copyright holder.