summaryrefslogtreecommitdiff
path: root/vendor/exr/src/meta/mod.rs
diff options
context:
space:
mode:
authorValentin Popov <valentin@popov.link>2024-01-08 00:21:28 +0300
committerValentin Popov <valentin@popov.link>2024-01-08 00:21:28 +0300
commit1b6a04ca5504955c571d1c97504fb45ea0befee4 (patch)
tree7579f518b23313e8a9748a88ab6173d5e030b227 /vendor/exr/src/meta/mod.rs
parent5ecd8cf2cba827454317368b68571df0d13d7842 (diff)
downloadfparkan-1b6a04ca5504955c571d1c97504fb45ea0befee4.tar.xz
fparkan-1b6a04ca5504955c571d1c97504fb45ea0befee4.zip
Initial vendor packages
Signed-off-by: Valentin Popov <valentin@popov.link>
Diffstat (limited to 'vendor/exr/src/meta/mod.rs')
-rw-r--r--vendor/exr/src/meta/mod.rs821
1 files changed, 821 insertions, 0 deletions
diff --git a/vendor/exr/src/meta/mod.rs b/vendor/exr/src/meta/mod.rs
new file mode 100644
index 0000000..0c36af8
--- /dev/null
+++ b/vendor/exr/src/meta/mod.rs
@@ -0,0 +1,821 @@
+
+//! Describes all meta data possible in an exr file.
+//! Contains functionality to read and write meta data from bytes.
+//! Browse the `exr::image` module to get started with the high-level interface.
+
+pub mod attribute;
+pub mod header;
+
+
+use crate::io::*;
+use ::smallvec::SmallVec;
+use self::attribute::*;
+use crate::block::chunk::{TileCoordinates, CompressedBlock};
+use crate::error::*;
+use std::fs::File;
+use std::io::{BufReader};
+use crate::math::*;
+use std::collections::{HashSet};
+use std::convert::TryFrom;
+use crate::meta::header::{Header};
+use crate::block::{BlockIndex, UncompressedBlock};
+
+
+// TODO rename MetaData to ImageInfo?
+
+/// Contains the complete meta data of an exr image.
+/// Defines how the image is split up in the file,
+/// the number and type of images and channels,
+/// and various other attributes.
+/// The usage of custom attributes is encouraged.
+#[derive(Debug, Clone, PartialEq)]
+pub struct MetaData {
+
+ /// Some flags summarizing the features that must be supported to decode the file.
+ pub requirements: Requirements,
+
+ /// One header to describe each layer in this file.
+ // TODO rename to layer descriptions?
+ pub headers: Headers,
+}
+
+
+/// List of `Header`s.
+pub type Headers = SmallVec<[Header; 3]>;
+
+/// List of `OffsetTable`s.
+pub type OffsetTables = SmallVec<[OffsetTable; 3]>;
+
+
+/// The offset table is an ordered list of indices referencing pixel data in the exr file.
+/// For each pixel tile in the image, an index exists, which points to the byte-location
+/// of the corresponding pixel data in the file. That index can be used to load specific
+/// portions of an image without processing all bytes in a file. For each header,
+/// an offset table exists with its indices ordered by `LineOrder::Increasing`.
+// If the multipart bit is unset and the chunkCount attribute is not present,
+// the number of entries in the chunk table is computed using the
+// dataWindow, tileDesc, and compression attribute.
+//
+// If the multipart bit is set, the header must contain a
+// chunkCount attribute, that contains the length of the offset table.
+pub type OffsetTable = Vec<u64>;
+
+
+/// A summary of requirements that must be met to read this exr file.
+/// Used to determine whether this file can be read by a given reader.
+/// It includes the OpenEXR version number. This library aims to support version `2.0`.
+#[derive(Clone, Copy, Eq, PartialEq, Debug, Hash)]
+pub struct Requirements {
+
+ /// This library supports reading version 1 and 2, and writing version 2.
+ // TODO write version 1 for simple images
+ pub file_format_version: u8,
+
+ /// If true, this image has tiled blocks and contains only a single layer.
+ /// If false and not deep and not multilayer, this image is a single layer image with scan line blocks.
+ pub is_single_layer_and_tiled: bool,
+
+ // in c or bad c++ this might have been relevant (omg is he allowed to say that)
+ /// Whether this file has strings with a length greater than 31.
+ /// Strings can never be longer than 255.
+ pub has_long_names: bool,
+
+ /// This image contains at least one layer with deep data.
+ pub has_deep_data: bool,
+
+ /// Whether this file contains multiple layers.
+ pub has_multiple_layers: bool,
+}
+
+
+/// Locates a rectangular section of pixels in an image.
+#[derive(Copy, Clone, Debug, Hash, Eq, PartialEq)]
+pub struct TileIndices {
+
+ /// Index of the tile.
+ pub location: TileCoordinates,
+
+ /// Pixel size of the tile.
+ pub size: Vec2<usize>,
+}
+
+/// How the image pixels are split up into separate blocks.
+#[derive(Copy, Clone, Debug, PartialEq, Eq, Hash)]
+pub enum BlockDescription {
+
+ /// The image is divided into scan line blocks.
+ /// The number of scan lines in a block depends on the compression method.
+ ScanLines,
+
+ /// The image is divided into tile blocks.
+ /// Also specifies the size of each tile in the image
+ /// and whether this image contains multiple resolution levels.
+ Tiles(TileDescription)
+}
+
+
+/*impl TileIndices {
+ pub fn cmp(&self, other: &Self) -> Ordering {
+ match self.location.level_index.1.cmp(&other.location.level_index.1) {
+ Ordering::Equal => {
+ match self.location.level_index.0.cmp(&other.location.level_index.0) {
+ Ordering::Equal => {
+ match self.location.tile_index.1.cmp(&other.location.tile_index.1) {
+ Ordering::Equal => {
+ self.location.tile_index.0.cmp(&other.location.tile_index.0)
+ },
+
+ other => other,
+ }
+ },
+
+ other => other
+ }
+ },
+
+ other => other
+ }
+ }
+}*/
+
+impl BlockDescription {
+
+ /// Whether this image is tiled. If false, this image is divided into scan line blocks.
+ pub fn has_tiles(&self) -> bool {
+ match self {
+ BlockDescription::Tiles { .. } => true,
+ _ => false
+ }
+ }
+}
+
+
+
+
+
+/// The first four bytes of each exr file.
+/// Used to abort reading non-exr files.
+pub mod magic_number {
+ use super::*;
+
+ /// The first four bytes of each exr file.
+ pub const BYTES: [u8; 4] = [0x76, 0x2f, 0x31, 0x01];
+
+ /// Without validation, write this instance to the byte stream.
+ pub fn write(write: &mut impl Write) -> Result<()> {
+ u8::write_slice(write, &self::BYTES)
+ }
+
+ /// Consumes four bytes from the reader and returns whether the file may be an exr file.
+ // TODO check if exr before allocating BufRead
+ pub fn is_exr(read: &mut impl Read) -> Result<bool> {
+ let mut magic_num = [0; 4];
+ u8::read_slice(read, &mut magic_num)?;
+ Ok(magic_num == self::BYTES)
+ }
+
+ /// Validate this image. If it is an exr file, return `Ok(())`.
+ pub fn validate_exr(read: &mut impl Read) -> UnitResult {
+ if self::is_exr(read)? {
+ Ok(())
+
+ } else {
+ Err(Error::invalid("file identifier missing"))
+ }
+ }
+}
+
+/// A `0_u8` at the end of a sequence.
+pub mod sequence_end {
+ use super::*;
+
+ /// Number of bytes this would consume in an exr file.
+ pub fn byte_size() -> usize {
+ 1
+ }
+
+ /// Without validation, write this instance to the byte stream.
+ pub fn write<W: Write>(write: &mut W) -> UnitResult {
+ 0_u8.write(write)
+ }
+
+ /// Peeks the next byte. If it is zero, consumes the byte and returns true.
+ pub fn has_come(read: &mut PeekRead<impl Read>) -> Result<bool> {
+ Ok(read.skip_if_eq(0)?)
+ }
+}
+
+fn missing_attribute(name: &str) -> Error {
+ Error::invalid(format!("missing or invalid {} attribute", name))
+}
+
+
+/// Compute the number of tiles required to contain all values.
+pub fn compute_block_count(full_res: usize, tile_size: usize) -> usize {
+ // round up, because if the image is not evenly divisible by the tiles,
+ // we add another tile at the end (which is only partially used)
+ RoundingMode::Up.divide(full_res, tile_size)
+}
+
+/// Compute the start position and size of a block inside a dimension.
+#[inline]
+pub fn calculate_block_position_and_size(total_size: usize, block_size: usize, block_index: usize) -> Result<(usize, usize)> {
+ let block_position = block_size * block_index;
+
+ Ok((
+ block_position,
+ calculate_block_size(total_size, block_size, block_position)?
+ ))
+}
+
+/// Calculate the size of a single block. If this is the last block,
+/// this only returns the required size, which is always smaller than the default block size.
+// TODO use this method everywhere instead of convoluted formulas
+#[inline]
+pub fn calculate_block_size(total_size: usize, block_size: usize, block_position: usize) -> Result<usize> {
+ if block_position >= total_size {
+ return Err(Error::invalid("block index"))
+ }
+
+ if block_position + block_size <= total_size {
+ Ok(block_size)
+ }
+ else {
+ Ok(total_size - block_position)
+ }
+}
+
+
+/// Calculate number of mip levels in a given resolution.
+// TODO this should be cached? log2 may be very expensive
+pub fn compute_level_count(round: RoundingMode, full_res: usize) -> usize {
+ usize::try_from(round.log2(u32::try_from(full_res).unwrap())).unwrap() + 1
+}
+
+/// Calculate the size of a single mip level by index.
+// TODO this should be cached? log2 may be very expensive
+pub fn compute_level_size(round: RoundingMode, full_res: usize, level_index: usize) -> usize {
+ assert!(level_index < std::mem::size_of::<usize>() * 8, "largest level size exceeds maximum integer value");
+ round.divide(full_res, 1 << level_index).max(1)
+}
+
+/// Iterates over all rip map level resolutions of a given size, including the indices of each level.
+/// The order of iteration conforms to `LineOrder::Increasing`.
+// TODO cache these?
+// TODO compute these directly instead of summing up an iterator?
+pub fn rip_map_levels(round: RoundingMode, max_resolution: Vec2<usize>) -> impl Iterator<Item=(Vec2<usize>, Vec2<usize>)> {
+ rip_map_indices(round, max_resolution).map(move |level_indices|{
+ // TODO progressively divide instead??
+ let width = compute_level_size(round, max_resolution.width(), level_indices.x());
+ let height = compute_level_size(round, max_resolution.height(), level_indices.y());
+ (level_indices, Vec2(width, height))
+ })
+}
+
+/// Iterates over all mip map level resolutions of a given size, including the indices of each level.
+/// The order of iteration conforms to `LineOrder::Increasing`.
+// TODO cache all these level values when computing table offset size??
+// TODO compute these directly instead of summing up an iterator?
+pub fn mip_map_levels(round: RoundingMode, max_resolution: Vec2<usize>) -> impl Iterator<Item=(usize, Vec2<usize>)> {
+ mip_map_indices(round, max_resolution)
+ .map(move |level_index|{
+ // TODO progressively divide instead??
+ let width = compute_level_size(round, max_resolution.width(), level_index);
+ let height = compute_level_size(round, max_resolution.height(), level_index);
+ (level_index, Vec2(width, height))
+ })
+}
+
+/// Iterates over all rip map level indices of a given size.
+/// The order of iteration conforms to `LineOrder::Increasing`.
+pub fn rip_map_indices(round: RoundingMode, max_resolution: Vec2<usize>) -> impl Iterator<Item=Vec2<usize>> {
+ let (width, height) = (
+ compute_level_count(round, max_resolution.width()),
+ compute_level_count(round, max_resolution.height())
+ );
+
+ (0..height).flat_map(move |y_level|{
+ (0..width).map(move |x_level|{
+ Vec2(x_level, y_level)
+ })
+ })
+}
+
+/// Iterates over all mip map level indices of a given size.
+/// The order of iteration conforms to `LineOrder::Increasing`.
+pub fn mip_map_indices(round: RoundingMode, max_resolution: Vec2<usize>) -> impl Iterator<Item=usize> {
+ 0..compute_level_count(round, max_resolution.width().max(max_resolution.height()))
+}
+
+/// Compute the number of chunks that an image is divided into. May be an expensive operation.
+// If not multilayer and chunkCount not present,
+// the number of entries in the chunk table is computed
+// using the dataWindow and tileDesc attributes and the compression format
+pub fn compute_chunk_count(compression: Compression, data_size: Vec2<usize>, blocks: BlockDescription) -> usize {
+
+ if let BlockDescription::Tiles(tiles) = blocks {
+ let round = tiles.rounding_mode;
+ let Vec2(tile_width, tile_height) = tiles.tile_size;
+
+ // TODO cache all these level values??
+ use crate::meta::attribute::LevelMode::*;
+ match tiles.level_mode {
+ Singular => {
+ let tiles_x = compute_block_count(data_size.width(), tile_width);
+ let tiles_y = compute_block_count(data_size.height(), tile_height);
+ tiles_x * tiles_y
+ }
+
+ MipMap => {
+ mip_map_levels(round, data_size).map(|(_, Vec2(level_width, level_height))| {
+ compute_block_count(level_width, tile_width) * compute_block_count(level_height, tile_height)
+ }).sum()
+ },
+
+ RipMap => {
+ rip_map_levels(round, data_size).map(|(_, Vec2(level_width, level_height))| {
+ compute_block_count(level_width, tile_width) * compute_block_count(level_height, tile_height)
+ }).sum()
+ }
+ }
+ }
+
+ // scan line blocks never have mip maps
+ else {
+ compute_block_count(data_size.height(), compression.scan_lines_per_block())
+ }
+}
+
+
+
+impl MetaData {
+
+ /// Read the exr meta data from a file.
+ /// Use `read_from_unbuffered` instead if you do not have a file.
+ /// Does not validate the meta data.
+ #[must_use]
+ pub fn read_from_file(path: impl AsRef<::std::path::Path>, pedantic: bool) -> Result<Self> {
+ Self::read_from_unbuffered(File::open(path)?, pedantic)
+ }
+
+ /// Buffer the reader and then read the exr meta data from it.
+ /// Use `read_from_buffered` if your reader is an in-memory reader.
+ /// Use `read_from_file` if you have a file path.
+ /// Does not validate the meta data.
+ #[must_use]
+ pub fn read_from_unbuffered(unbuffered: impl Read, pedantic: bool) -> Result<Self> {
+ Self::read_from_buffered(BufReader::new(unbuffered), pedantic)
+ }
+
+ /// Read the exr meta data from a reader.
+ /// Use `read_from_file` if you have a file path.
+ /// Use `read_from_unbuffered` if this is not an in-memory reader.
+ /// Does not validate the meta data.
+ #[must_use]
+ pub fn read_from_buffered(buffered: impl Read, pedantic: bool) -> Result<Self> {
+ let mut read = PeekRead::new(buffered);
+ MetaData::read_unvalidated_from_buffered_peekable(&mut read, pedantic)
+ }
+
+ /// Does __not validate__ the meta data completely.
+ #[must_use]
+ pub(crate) fn read_unvalidated_from_buffered_peekable(read: &mut PeekRead<impl Read>, pedantic: bool) -> Result<Self> {
+ magic_number::validate_exr(read)?;
+
+ let requirements = Requirements::read(read)?;
+
+ // do this check now in order to fast-fail for newer versions and features than version 2
+ requirements.validate()?;
+
+ let headers = Header::read_all(read, &requirements, pedantic)?;
+
+ // TODO check if supporting requirements 2 always implies supporting requirements 1
+ Ok(MetaData { requirements, headers })
+ }
+
+ /// Validates the meta data.
+ #[must_use]
+ pub(crate) fn read_validated_from_buffered_peekable(
+ read: &mut PeekRead<impl Read>, pedantic: bool
+ ) -> Result<Self> {
+ let meta_data = Self::read_unvalidated_from_buffered_peekable(read, !pedantic)?;
+ MetaData::validate(meta_data.headers.as_slice(), pedantic)?;
+ Ok(meta_data)
+ }
+
+ /// Validates the meta data and writes it to the stream.
+ /// If pedantic, throws errors for files that may produce errors in other exr readers.
+ /// Returns the automatically detected minimum requirement flags.
+ pub(crate) fn write_validating_to_buffered(write: &mut impl Write, headers: &[Header], pedantic: bool) -> Result<Requirements> {
+ // pedantic validation to not allow slightly invalid files
+ // that still could be read correctly in theory
+ let minimal_requirements = Self::validate(headers, pedantic)?;
+
+ magic_number::write(write)?;
+ minimal_requirements.write(write)?;
+ Header::write_all(headers, write, minimal_requirements.has_multiple_layers)?;
+ Ok(minimal_requirements)
+ }
+
+ /// Read one offset table from the reader for each header.
+ pub fn read_offset_tables(read: &mut PeekRead<impl Read>, headers: &Headers) -> Result<OffsetTables> {
+ headers.iter()
+ .map(|header| u64::read_vec(read, header.chunk_count, u16::MAX as usize, None, "offset table size"))
+ .collect()
+ }
+
+ /// Skip the offset tables by advancing the reader by the required byte count.
+ // TODO use seek for large (probably all) tables!
+ pub fn skip_offset_tables(read: &mut PeekRead<impl Read>, headers: &Headers) -> Result<usize> {
+ let chunk_count: usize = headers.iter().map(|header| header.chunk_count).sum();
+ crate::io::skip_bytes(read, chunk_count * u64::BYTE_SIZE)?; // TODO this should seek for large tables
+ Ok(chunk_count)
+ }
+
+ /// This iterator tells you the block indices of all blocks that must be in the image.
+ /// The order of the blocks depends on the `LineOrder` attribute
+ /// (unspecified line order is treated the same as increasing line order).
+ /// The blocks written to the file must be exactly in this order,
+ /// except for when the `LineOrder` is unspecified.
+ /// The index represents the block index, in increasing line order, within the header.
+ pub fn enumerate_ordered_header_block_indices(&self) -> impl '_ + Iterator<Item=(usize, BlockIndex)> {
+ crate::block::enumerate_ordered_header_block_indices(&self.headers)
+ }
+
+ /// Go through all the block indices in the correct order and call the specified closure for each of these blocks.
+ /// That way, the blocks indices are filled with real block data and returned as an iterator.
+ /// The closure returns the an `UncompressedBlock` for each block index.
+ pub fn collect_ordered_blocks<'s>(&'s self, mut get_block: impl 's + FnMut(BlockIndex) -> UncompressedBlock)
+ -> impl 's + Iterator<Item=(usize, UncompressedBlock)>
+ {
+ self.enumerate_ordered_header_block_indices().map(move |(index_in_header, block_index)|{
+ (index_in_header, get_block(block_index))
+ })
+ }
+
+ /// Go through all the block indices in the correct order and call the specified closure for each of these blocks.
+ /// That way, the blocks indices are filled with real block data and returned as an iterator.
+ /// The closure returns the byte data for each block index.
+ pub fn collect_ordered_block_data<'s>(&'s self, mut get_block_data: impl 's + FnMut(BlockIndex) -> Vec<u8>)
+ -> impl 's + Iterator<Item=(usize, UncompressedBlock)>
+ {
+ self.collect_ordered_blocks(move |block_index|
+ UncompressedBlock { index: block_index, data: get_block_data(block_index) }
+ )
+ }
+
+ /// Validates this meta data. Returns the minimal possible requirements.
+ pub fn validate(headers: &[Header], pedantic: bool) -> Result<Requirements> {
+ if headers.len() == 0 {
+ return Err(Error::invalid("at least one layer is required"));
+ }
+
+ let deep = false; // TODO deep data
+ let is_multilayer = headers.len() > 1;
+ let first_header_has_tiles = headers.iter().next()
+ .map_or(false, |header| header.blocks.has_tiles());
+
+ let mut minimal_requirements = Requirements {
+ // according to the spec, version 2 should only be necessary if `is_multilayer || deep`.
+ // but the current open exr library does not support images with version 1, so always use version 2.
+ file_format_version: 2,
+
+ // start as low as possible, later increasing if required
+ has_long_names: false,
+
+ is_single_layer_and_tiled: !is_multilayer && first_header_has_tiles,
+ has_multiple_layers: is_multilayer,
+ has_deep_data: deep,
+ };
+
+ for header in headers {
+ if header.deep { // TODO deep data (and then remove this check)
+ return Err(Error::unsupported("deep data not supported yet"));
+ }
+
+ header.validate(is_multilayer, &mut minimal_requirements.has_long_names, pedantic)?;
+ }
+
+ // TODO validation fn!
+ /*if let Some(max) = max_pixel_bytes {
+ let byte_size: usize = headers.iter()
+ .map(|header| header.total_pixel_bytes())
+ .sum();
+
+ if byte_size > max {
+ return Err(Error::invalid("image larger than specified maximum"));
+ }
+ }*/
+
+ if pedantic { // check for duplicate header names
+ let mut header_names = HashSet::with_capacity(headers.len());
+ for header in headers {
+ if !header_names.insert(&header.own_attributes.layer_name) {
+ return Err(Error::invalid(format!(
+ "duplicate layer name: `{}`",
+ header.own_attributes.layer_name.as_ref().expect("header validation bug")
+ )));
+ }
+ }
+ }
+
+ if pedantic {
+ let must_share = headers.iter().flat_map(|header| header.own_attributes.other.iter())
+ .any(|(_, value)| value.to_chromaticities().is_ok() || value.to_time_code().is_ok());
+
+ if must_share {
+ return Err(Error::invalid("chromaticities and time code attributes must must not exist in own attributes but shared instead"));
+ }
+ }
+
+ if pedantic && headers.len() > 1 { // check for attributes that should not differ in between headers
+ let first_header = headers.first().expect("header count validation bug");
+ let first_header_attributes = &first_header.shared_attributes;
+
+ for header in &headers[1..] {
+ if &header.shared_attributes != first_header_attributes {
+ return Err(Error::invalid("display window, pixel aspect, chromaticities, and time code attributes must be equal for all headers"))
+ }
+ }
+ }
+
+ debug_assert!(minimal_requirements.validate().is_ok(), "inferred requirements are invalid");
+ Ok(minimal_requirements)
+ }
+}
+
+
+
+
+impl Requirements {
+
+ // this is actually used for control flow, as the number of headers may be 1 in a multilayer file
+ /// Is this file declared to contain multiple layers?
+ pub fn is_multilayer(&self) -> bool {
+ self.has_multiple_layers
+ }
+
+ /// Read the value without validating.
+ pub fn read<R: Read>(read: &mut R) -> Result<Self> {
+ use ::bit_field::BitField;
+
+ let version_and_flags = u32::read(read)?;
+
+ // take the 8 least significant bits, they contain the file format version number
+ let version = (version_and_flags & 0x000F) as u8;
+
+ // the 24 most significant bits are treated as a set of boolean flags
+ let is_single_tile = version_and_flags.get_bit(9);
+ let has_long_names = version_and_flags.get_bit(10);
+ let has_deep_data = version_and_flags.get_bit(11);
+ let has_multiple_layers = version_and_flags.get_bit(12);
+
+ // all remaining bits except 9, 10, 11 and 12 are reserved and should be 0
+ // if a file has any of these bits set to 1, it means this file contains
+ // a feature that we don't support
+ let unknown_flags = version_and_flags >> 13; // all flags excluding the 12 bits we already parsed
+
+ if unknown_flags != 0 { // TODO test if this correctly detects unsupported files
+ return Err(Error::unsupported("too new file feature flags"));
+ }
+
+ let version = Requirements {
+ file_format_version: version,
+ is_single_layer_and_tiled: is_single_tile, has_long_names,
+ has_deep_data, has_multiple_layers,
+ };
+
+ Ok(version)
+ }
+
+ /// Without validation, write this instance to the byte stream.
+ pub fn write<W: Write>(self, write: &mut W) -> UnitResult {
+ use ::bit_field::BitField;
+
+ // the 8 least significant bits contain the file format version number
+ // and the flags are set to 0
+ let mut version_and_flags = self.file_format_version as u32;
+
+ // the 24 most significant bits are treated as a set of boolean flags
+ version_and_flags.set_bit(9, self.is_single_layer_and_tiled);
+ version_and_flags.set_bit(10, self.has_long_names);
+ version_and_flags.set_bit(11, self.has_deep_data);
+ version_and_flags.set_bit(12, self.has_multiple_layers);
+ // all remaining bits except 9, 10, 11 and 12 are reserved and should be 0
+
+ version_and_flags.write(write)?;
+ Ok(())
+ }
+
+ /// Validate this instance.
+ pub fn validate(&self) -> UnitResult {
+ if self.file_format_version == 2 {
+
+ match (
+ self.is_single_layer_and_tiled, self.has_deep_data, self.has_multiple_layers,
+ self.file_format_version
+ ) {
+ // Single-part scan line. One normal scan line image.
+ (false, false, false, 1..=2) => Ok(()),
+
+ // Single-part tile. One normal tiled image.
+ (true, false, false, 1..=2) => Ok(()),
+
+ // Multi-part (new in 2.0).
+ // Multiple normal images (scan line and/or tiled).
+ (false, false, true, 2) => Ok(()),
+
+ // Single-part deep data (new in 2.0).
+ // One deep tile or deep scan line part
+ (false, true, false, 2) => Ok(()),
+
+ // Multi-part deep data (new in 2.0).
+ // Multiple parts (any combination of:
+ // tiles, scan lines, deep tiles and/or deep scan lines).
+ (false, true, true, 2) => Ok(()),
+
+ _ => Err(Error::invalid("file feature flags"))
+ }
+ }
+ else {
+ Err(Error::unsupported("file versions other than 2.0 are not supported"))
+ }
+ }
+}
+
+
+#[cfg(test)]
+mod test {
+ use super::*;
+ use crate::meta::header::{ImageAttributes, LayerAttributes};
+
+ #[test]
+ fn round_trip_requirements() {
+ let requirements = Requirements {
+ file_format_version: 2,
+ is_single_layer_and_tiled: true,
+ has_long_names: false,
+ has_deep_data: true,
+ has_multiple_layers: false
+ };
+
+ let mut data: Vec<u8> = Vec::new();
+ requirements.write(&mut data).unwrap();
+ let read = Requirements::read(&mut data.as_slice()).unwrap();
+ assert_eq!(requirements, read);
+ }
+
+ #[test]
+ fn round_trip(){
+ let header = Header {
+ channels: ChannelList::new(smallvec![
+ ChannelDescription {
+ name: Text::from("main"),
+ sample_type: SampleType::U32,
+ quantize_linearly: false,
+ sampling: Vec2(1, 1)
+ }
+ ],
+ ),
+ compression: Compression::Uncompressed,
+ line_order: LineOrder::Increasing,
+ deep_data_version: Some(1),
+ chunk_count: compute_chunk_count(Compression::Uncompressed, Vec2(2000, 333), BlockDescription::ScanLines),
+ max_samples_per_pixel: Some(4),
+ shared_attributes: ImageAttributes {
+ pixel_aspect: 3.0,
+ .. ImageAttributes::new(IntegerBounds {
+ position: Vec2(2,1),
+ size: Vec2(11, 9)
+ })
+ },
+
+ blocks: BlockDescription::ScanLines,
+ deep: false,
+ layer_size: Vec2(2000, 333),
+ own_attributes: LayerAttributes {
+ layer_name: Some(Text::from("test name lol")),
+ layer_position: Vec2(3, -5),
+ screen_window_center: Vec2(0.3, 99.0),
+ screen_window_width: 0.19,
+ .. Default::default()
+ }
+ };
+
+ let meta = MetaData {
+ requirements: Requirements {
+ file_format_version: 2,
+ is_single_layer_and_tiled: false,
+ has_long_names: false,
+ has_deep_data: false,
+ has_multiple_layers: false
+ },
+ headers: smallvec![ header ],
+ };
+
+
+ let mut data: Vec<u8> = Vec::new();
+ MetaData::write_validating_to_buffered(&mut data, meta.headers.as_slice(), true).unwrap();
+ let meta2 = MetaData::read_from_buffered(data.as_slice(), false).unwrap();
+ MetaData::validate(meta2.headers.as_slice(), true).unwrap();
+ assert_eq!(meta, meta2);
+ }
+
+ #[test]
+ fn infer_low_requirements() {
+ let header_version_1_short_names = Header {
+ channels: ChannelList::new(smallvec![
+ ChannelDescription {
+ name: Text::from("main"),
+ sample_type: SampleType::U32,
+ quantize_linearly: false,
+ sampling: Vec2(1, 1)
+ }
+ ],
+ ),
+ compression: Compression::Uncompressed,
+ line_order: LineOrder::Increasing,
+ deep_data_version: Some(1),
+ chunk_count: compute_chunk_count(Compression::Uncompressed, Vec2(2000, 333), BlockDescription::ScanLines),
+ max_samples_per_pixel: Some(4),
+ shared_attributes: ImageAttributes {
+ pixel_aspect: 3.0,
+ .. ImageAttributes::new(IntegerBounds {
+ position: Vec2(2,1),
+ size: Vec2(11, 9)
+ })
+ },
+ blocks: BlockDescription::ScanLines,
+ deep: false,
+ layer_size: Vec2(2000, 333),
+ own_attributes: LayerAttributes {
+ other: vec![
+ (Text::try_from("x").unwrap(), AttributeValue::F32(3.0)),
+ (Text::try_from("y").unwrap(), AttributeValue::F32(-1.0)),
+ ].into_iter().collect(),
+ .. Default::default()
+ }
+ };
+
+ let low_requirements = MetaData::validate(
+ &[header_version_1_short_names], true
+ ).unwrap();
+
+ assert_eq!(low_requirements.has_long_names, false);
+ assert_eq!(low_requirements.file_format_version, 2); // always have version 2
+ assert_eq!(low_requirements.has_deep_data, false);
+ assert_eq!(low_requirements.has_multiple_layers, false);
+ }
+
+ #[test]
+ fn infer_high_requirements() {
+ let header_version_2_long_names = Header {
+ channels: ChannelList::new(
+ smallvec![
+ ChannelDescription {
+ name: Text::new_or_panic("main"),
+ sample_type: SampleType::U32,
+ quantize_linearly: false,
+ sampling: Vec2(1, 1)
+ }
+ ],
+ ),
+ compression: Compression::Uncompressed,
+ line_order: LineOrder::Increasing,
+ deep_data_version: Some(1),
+ chunk_count: compute_chunk_count(Compression::Uncompressed, Vec2(2000, 333), BlockDescription::ScanLines),
+ max_samples_per_pixel: Some(4),
+ shared_attributes: ImageAttributes {
+ pixel_aspect: 3.0,
+ .. ImageAttributes::new(IntegerBounds {
+ position: Vec2(2,1),
+ size: Vec2(11, 9)
+ })
+ },
+ blocks: BlockDescription::ScanLines,
+ deep: false,
+ layer_size: Vec2(2000, 333),
+ own_attributes: LayerAttributes {
+ layer_name: Some(Text::new_or_panic("oasdasoidfj")),
+ other: vec![
+ (Text::new_or_panic("xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx"), AttributeValue::F32(3.0)),
+ (Text::new_or_panic("y"), AttributeValue::F32(-1.0)),
+ ].into_iter().collect(),
+ .. Default::default()
+ }
+ };
+
+ let mut layer_2 = header_version_2_long_names.clone();
+ layer_2.own_attributes.layer_name = Some(Text::new_or_panic("anythingelse"));
+
+ let low_requirements = MetaData::validate(
+ &[header_version_2_long_names, layer_2], true
+ ).unwrap();
+
+ assert_eq!(low_requirements.has_long_names, true);
+ assert_eq!(low_requirements.file_format_version, 2);
+ assert_eq!(low_requirements.has_deep_data, false);
+ assert_eq!(low_requirements.has_multiple_layers, true);
+ }
+}
+