Tutorial¶
Units of measurement¶
Many of the metadata interfaces provide methods to get or set values with an associated unit of measurement. The reason for this is to ensure that values are always associated with an appropriate unit, and to enforce compile-time or run-time sanity checks to ensure that the correct unit type is used, and that any unit conversions performed are legal. The following terminology is used:
- dimension
- A measured property, for example length, pressure, temperature or time.
- unit system
- A system of units for a given dimension, for example the SI units for the length dimension are metre and its derived units. For the pressure, temperature and time dimensions, pascal, celsius and second are used respectively, along with any derived units. Multiple systems may be provided for a given dimension, such as Imperial length units, bar or Torr for pressure or Fahrenheit for temperature. Different unit systems for the same dimension will typically be inter-convertible, but this is not a requirement.
- unit
- A unit of measure within a given unit system, for example cm, µm and nm are all scaled units derived from m.
- base unit
- The primary unit for a given unit system; all other units are scaled relative to this unit. Automatic conversion between unit systems is defined in conversion of base units. For example, m is the base unit for the SI length unit system.
- quantity
- A measured value with an associated unit. For example, 3.5 mm.
Model units¶
The metadata interfaces make use of these unit types, which are based upon
- Unit type enumerations
- The Quantity class
The Quantity
class is the user-visible part of the units
support. It trades compile-time correctness for run-time flexibility.
It is templated, specialized for a given unit type enumeration, for
example Quantity<UnitLength>
for length quantities. It may
represent any valid unit from the enumeration.
A Quantity
is constructed using a numerical value and a
unit enumeration value. Basic arithmetic operations such as
assignment, multiplication, subtraction, division and multiplication
are supported. Note however that complex arithmetic with different
unit types is not supported; if this is required, please use the basic
units directly (see below). Also note that multiplication and
division are permitted only with scalars, and addition and subtraction
require operands to be of the same type; the unit of the left-hand
operand will be preserved, and the value of the right-hand operand
will be implicitly converted to match.
Unit conversion is performed by using the free convert()
function, which requires a quantity and destination unit. If
conversion is not supported, an exception will be thrown. The
conversion operations are all performed in terms of the basic
quantities, described below.
The following example demonstrates these concepts:
typedef Quantity<UnitsLength, double> Length;
// Micrometre units.
Length a(50.0, UnitsLength::MICROMETER);
Length b(25.3, UnitsLength::MICROMETER);
std::cout << "a=" << a << '\n';
std::cout << "b=" << b << '\n';
// Arithmetic operations.
Length c(a + b);
Length d(a * 8.0);
Length e(a / 4.0);
std::cout << "c=" << c << '\n';
std::cout << "d=" << d << '\n';
std::cout << "e=" << e << '\n';
// Unit conversion to SI and Imperial units.
Length f(convert(c, UnitsLength::NANOMETER));
Length g(convert(c, UnitsLength::THOU));
Length h(convert(c, UnitsLength::INCH));
std::cout << "f=" << f << '\n';
std::cout << "g=" << g << '\n';
std::cout << "h=" << h << '\n';
// Unit sytems which do not permit interconversion.
Length i(34.8, UnitsLength::PIXEL);
Length j(2.922, UnitsLength::REFERENCEFRAME);
std::cout << "i=" << i << '\n';
std::cout << "j=" << j << '\n';
// Will throw since conversion is impossible.
try
{
Length k(convert(i, UnitsLength::MICROMETER));
}
catch (const std::logic_error&)
{
std::cout << "i not convertible to µm\n";
}
try
{
Length l(convert(j, UnitsLength::MICROMETER));
}
catch (const std::logic_error&)
{
std::cout << "j not convertible to µm\n";
}
Model units are also used in the following examples such as the Metadata store.
The model units are implemented using a lower-level units interface, which we term “basic units”.
Basic units¶
Unlike the model units, which provide run-time checking, the basic units enforce correctness during compilation. Basic units are provided by a static compile-time type-safe unit system based upon Boost.Units. All the data types provided are simply typedefs or specializations of the Boost.Units library unit types. See the Boost.Units documentation for further information.
As a user of the library, quantity types are the primary type which you will use. The other types are implementation details which will not be needed unless you have a need to add additional custom units to the existing set of registered units. However, a comprehensive set of SI, Imperial and other customary units are provided, so most of the conceivable usage should be fully catered for. A full list of quantity types is listed in the units namespace. Should you wish to add your own quantities, units, unit systems and even entirely new dimensions, this is all possible using Boost.Units.
To create a quantity, use the quantity::from_value()
method,
which will return a quantity
of the desired type. Since
the unit is encoded in the type name, the unit type is immutable and
correctness is enforced at compile time. This differs from the model
units where the unit type is passed to the constructor. Once
constructed, the quantity may be used in a similar manner to the model
quantities: a quantity is assignable to any quantity of the same type,
and arithmetic operations with scalar quantities or other units are
permitted. Note that a quantity is not assignable to other quantities
in the same unit system which are not of exactly the same type;
implicit unit conversions are not allowed, even between scaled units
in the same unit system. Also note that multiplication and division
with quantities may change the unit type, such as squaring when
multiplying the same unit type, and that adding and subtracting scalar
values is not permitted since there is no associated unit.
Explicit conversion is as simple as instantiating a quantity type with a different quantity as the construction parameter. If the conversion is not permitted, this will result in a compilation failure, enforcing correctness.
The following example demonstrates these concepts:
// Micrometre units.
micrometre_quantity a(micrometre_quantity::from_value(50.0));
micrometre_quantity b(micrometre_quantity::from_value(25.3));
std::cout << "a=" << a << '\n';
std::cout << "b=" << b << '\n';
// Arithmetic operations.
micrometre_quantity c(a + b);
micrometre_quantity d(a * 8.0);
micrometre_quantity e(a / 4.0);
std::cout << "c=" << c << '\n';
std::cout << "d=" << d << '\n';
std::cout << "e=" << e << '\n';
// Unit conversion to SI and Imperial units.
nanometre_quantity f(c);
thou_quantity g(c);
inch_quantity h(c);
std::cout << "f=" << f << '\n';
std::cout << "g=" << g << '\n';
std::cout << "h=" << h << '\n';
// Unit sytems which do not permit interconversion.
pixel_quantity i(pixel_quantity::from_value(34.8));
reference_frame_quantity j(reference_frame_quantity::from_value(2.922));
std::cout << "i=" << i << '\n';
std::cout << "j=" << j << '\n';
// Compilation will fail if uncommented since conversion is
// impossible.
// micrometre_quantity k(i);
// micrometre_quantity l(j);
As a general rule, the user-facing API will not use basic units, but they are available should you wish to make use of strict unit type checking and unit type conversion in your own code. They are also potentially more performant when performing conversions since there is much greater scope for optimization.
Metadata¶
OME-Files supports several different classes of metadata, from very basic information about the image dimensions and pixel type to detailed information about the acquisition hardware and experimental parameters. From simplest to most complex, these are:
- Core metadata
- Basic information describing an individual 5D image (series), including dimension sizes, dimension order and pixel type
- Original metadata
- Key-value pairs describing metadata from the original file format for the image. Two forms exist: global metadata for an entire dataset (image collection) and series metadata for an individual 5D image
- Metadata store
- A container for all image metadata providing interfaces to get and set individual metadata values. This is a superset of the core and original metadata content (it can represent all values contained within the core and original metadata). It is an alternative representation of the OME-XML data model objects, and is used by the OME-Files reader and writer interfaces.
- OME-XML data model objects
- The abstract OME-XML data model is realized as a collection of model objects. Classes are generated from the elements of the OME-XML data model schema, and a tree of the model objects acts as a representation of the OME data model which may be modified and manipulated. The model objects may be created from an OME-XML text document, and vice versa.
For the simplest cases of reading and writing image data, the core metadata interface will likely be sufficient. If specific individual parameters from the original file format are needed, then original metadata may also be useful. For more advanced processing and rendering, the metadata store should be the next source of information, for example to get information about the image scale, stage position, instrument setup including light sources, light paths, detectors etc., and access to plate/well information, regions of interest etc. Direct access to the OME-XML data model objects is an alternative to the metadata store, but is more difficult to use; certain modifications to the data model may only be made via direct access to the model objects, otherwise the higher-level metadata store interface should be preferred.
The header file ome/files/MetadataTools.h provides several convenience functions to work with and manipulate the various forms of metadata, including conversion of Core metadata to and from a metadata store.
Core metadata¶
Core metadata is accessible through the getter methods in the
FormatReader
interface. These operate on the current
series, set using the setSeries()
method. The
CoreMetadata
objects are also accessible directly using
the getCoreMetadataList
method. The
FormatReader
interface should be preferred; the objects
themselves are more of an implementation detail at present.
void
readMetadata(const FormatReader& reader,
std::ostream& stream)
{
// Get total number of images (series)
dimension_size_type ic = reader.getSeriesCount();
stream << "Image count: " << ic << '\n';
// Loop over images
for (dimension_size_type i = 0 ; i < ic; ++i)
{
// Change the current series to this index
reader.setSeries(i);
// Print image dimensions (for this image index)
stream << "Dimensions for Image " << i << ':'
<< "\n\tX = " << reader.getSizeX()
<< "\n\tY = " << reader.getSizeY()
<< "\n\tZ = " << reader.getSizeZ()
<< "\n\tT = " << reader.getSizeT()
<< "\n\tC = " << reader.getSizeC()
<< "\n\tEffectiveC = " << reader.getEffectiveSizeC();
for (dimension_size_type channel = 0;
channel < reader.getEffectiveSizeC();
++channel)
{
stream << "\n\tChannel " << channel << ':'
<< "\n\t\tRGB = " << (reader.isRGB(channel) ? "true" : "false")
<< "\n\t\tRGBC = " << reader.getRGBChannelCount(channel);
}
stream << '\n';
// Get total number of planes (for this image index)
dimension_size_type pc = reader.getImageCount();
stream << "\tPlane count: " << pc << '\n';
// Loop over planes (for this image index)
for (dimension_size_type p = 0 ; p < pc; ++p)
{
// Print plane position (for this image index and plane
// index)
std::array<dimension_size_type, 3> coords =
reader.getZCTCoords(p);
stream << "\tPosition of Plane " << p << ':'
<< "\n\t\tTheZ = " << coords[0]
<< "\n\t\tTheT = " << coords[2]
<< "\n\t\tTheC = " << coords[1]
<< '\n';
}
}
}
If implementing a reader, it is fairly typical to set the basic image
metadata in CoreMetadata
objects, and then use the
fillMetadata()
function in
ome/files/MetadataTools.h to fill the reader’s metadata store
with this information, before filling the metadata store with
additional (non-core) metadata as required. When writing an image, a
metadata store is required in order to provide the writer with all the
metadata needed to write an image. If the metadata store was not
already obtained from a reader, fillMetadata()
may also be
used in this situation to create a suitable metadata store:
// OME-XML metadata store.
auto meta = make_shared<OMEXMLMetadata>();
// Create simple CoreMetadata and use this to set up the OME-XML
// metadata. This is purely for convenience in this example; a
// real writer would typically set up the OME-XML metadata from an
// existing MetadataRetrieve instance or by hand.
std::vector<shared_ptr<CoreMetadata>> seriesList;
shared_ptr<CoreMetadata> core(make_shared<CoreMetadata>());
core->sizeX = 512U;
core->sizeY = 512U;
core->sizeC.clear(); // defaults to 1 channel with 1 subchannel; clear this
core->sizeC.push_back(3U); // replace with single RGB channel
core->pixelType = ome::xml::model::enums::PixelType::UINT16;
core->interleaved = false;
core->bitsPerPixel = 12U;
core->dimensionOrder = DimensionOrder::XYZTC;
seriesList.push_back(core);
seriesList.push_back(core); // add two identical series
fillMetadata(*meta, seriesList);
Full example source: metadata-formatreader.cpp
,
metadata-formatreader.cpp
See also
Original metadata¶
Original metadata is stored in two forms: in a
MetadataMap
which is accessible through the
FormatReader
interface, which offers access to individual
keys and the whole map for both global and series metadata. It is
also accessible using the metadata store; original metadata is stored
as an XMLAnnotation
. The following example demonstrates
access to the global and series metadata using the
FormatReader
interface to get access to the maps:
void
readOriginalMetadata(const FormatReader& reader,
std::ostream& stream)
{
// Get total number of images (series)
dimension_size_type ic = reader.getSeriesCount();
stream << "Image count: " << ic << '\n';
// Get global metadata
const MetadataMap& global = reader.getGlobalMetadata();
// Print global metadata
stream << "Global metadata:\n" << global << '\n';
// Loop over images
for (dimension_size_type i = 0 ; i < ic; ++i)
{
// Change the current series to this index
reader.setSeries(i);
// Print series metadata
const MetadataMap& series = reader.getSeriesMetadata();
// Print image dimensions (for this image index)
stream << "Metadata for Image " << i << ":\n"
<< series
<< '\n';
}
}
It would also be possible to use getMetadataValue()
and
getSeriesMetadataValue()
to obtain values for individual
keys. Note that the MetadataMap
values can be scalar
values or lists of scalar values; call the flatten()
method
to split the lists into separate key-value pairs with a numbered
suffix.
Full example source: metadata-formatreader.cpp
Metadata store¶
Access to metadata is provided via the MetadataStore
and
MetadataRetrieve
interfaces. These provide setters and
getters, respectively, to store and retrieve metadata to and from an
underlying abstract metadata store. The primary store is the
OMEXMLMetadata
which stores the metadata in OME-XML data
model objects (see below), and implements both interfaces. However,
other storage classes are available, and may be used to filter the
stored metadata, combine different stores, or do nothing at all.
Additional storage backends could also be implemented, for example to
allow metadata retrieval from a relational database, or JSON/YAML.
When using OMEXMLMetadata
the convenience function
createOMEXMLMetadata()
is the recommended method for
creating a new instance and then filling it with the content from an
OME-XML document. This is overloaded to allow the OME-XML to be
obtained from various sources. For example, from a file:
// Create metadata directly from file
shared_ptr<meta::OMEXMLMetadata> filemeta(createOMEXMLMetadata(filename));
Alternatively from a DOM tree:
// XML platform (required by Xerces)
xml::Platform xmlplat;
// XML DOM tree containing parsed file content
xml::dom::Document inputdoc(ome::xml::createDocument(filename));
// Create metadata from DOM document
shared_ptr<meta::OMEXMLMetadata> dommeta(createOMEXMLMetadata(inputdoc));
The convenience function getOMEXML()
may be used to reverse
the process, i.e. obtain an OME-XML document from the store. Note the
use of convert()
. Only the OMEXMLMetadata
class can dump an OME-XML document, therefore if the source of the
data is another class implementing the MetadataRetrieve
interface, the stored data will need to be copied into an
OMEXMLMetadata
instance first.
meta::OMEXMLMetadata *omexmlmeta = dynamic_cast<meta::OMEXMLMetadata *>(&meta);
shared_ptr<meta::OMEXMLMetadata> convertmeta;
if (!omexmlmeta)
{
convertmeta = make_shared<meta::OMEXMLMetadata>();
meta::convert(meta, *convertmeta);
omexmlmeta = &*convertmeta;
}
// Get OME-XML text from metadata store (and validate it)
std::string omexml(getOMEXML(*omexmlmeta, true));
Conceptually, the metadata store contains lists of objects, accessed
by index (insertion order). In the example below,
getImageCount()
method is used to find the number of images.
This is then used to safely loop through each of the available images.
Each of the getPixelsSizeA()
methods takes the image index
as its only argument. Internally, this is used to find the
Image
model object for the specified index, and then call
the getSizeA()
method on that object and return the result.
Since objects can contain other objects, some accessor methods require
the use of more than one index. For example, an Image
object can contain multiple Plane
objects. Similar to
the above example, there is a getPlaneCount()
method,
however since it is contained by an Image
it has an
additional image index argument to get the plane count for the
specified image. Likewise its accessors such as
getPlaneTheZ()
take two arguments, the image index and
the plane index. Internally, these indices will be used to find the
Image
, then the Plane
, and then call
getTheZ()
. When using the MetadataRetrieve
interface with an OMEXMLMetadata
store, the methods are
simply a shorthand for navigating through the tree of model objects.
void
queryMetadata(const meta::MetadataRetrieve& meta,
const std::string& state,
std::ostream& stream)
{
// Get total number of images (series)
index_type ic = meta.getImageCount();
stream << "Image count: " << ic << '\n';
// Loop over images
for (index_type i = 0 ; i < ic; ++i)
{
// Print image dimensions (for this image index)
stream << "Dimensions for Image " << i << ' ' << state << ':'
<< "\n\tX = " << meta.getPixelsSizeX(i)
<< "\n\tY = " << meta.getPixelsSizeY(i)
<< "\n\tZ = " << meta.getPixelsSizeZ(i)
<< "\n\tT = " << meta.getPixelsSizeT(i)
<< "\n\tC = " << meta.getPixelsSizeC(i);
// These are optional, so handle failure gracefully.
try
{
stream << "\n\tPhysicalX = " << meta.getPixelsPhysicalSizeX(i);
}
catch (const meta::MetadataException&)
{
}
try
{
stream << "\n\tPhysicalY = " << meta.getPixelsPhysicalSizeY(i);
}
catch (const meta::MetadataException&)
{
}
try
{
stream << "\n\tPhysicalZ = " << meta.getPixelsPhysicalSizeZ(i);
}
catch (const meta::MetadataException&)
{
}
stream<< '\n';
// Get total number of planes (for this image index)
index_type pc = meta.getPlaneCount(i);
stream << "\tPlane count: " << pc << '\n';
// Loop over planes (for this image index)
for (index_type p = 0 ; p < pc; ++p)
{
// Print plane position (for this image index and plane
// index)
stream << "\tPosition of Plane " << p << ':'
<< "\n\t\tTheZ = " << meta.getPlaneTheZ(i, p)
<< "\n\t\tTheT = " << meta.getPlaneTheT(i, p)
<< "\n\t\tTheC = " << meta.getPlaneTheC(i, p)
<< '\n';
}
}
}
The methods for storing data using the MetadataStore
interface are similar. The set methods use the same indices as the
get methods, with the value to set as an additional initial argument.
The following example demonstrates how to update dimension sizes for
images in the store:
void
updateMetadata(meta::Metadata& meta)
{
// Get total number of images (series)
index_type ic = meta.getImageCount();
// Loop over images
for (index_type i = 0 ; i < ic; ++i)
{
// Change image dimensions (for this image index)
meta.setPixelsSizeX(12, i);
meta.setPixelsSizeY(24, i);
meta.setPixelsSizeZ(6, i);
meta.setPixelsSizeT(30, i);
meta.setPixelsSizeC(4, i);
meta.setPixelsPhysicalSizeX
(PositiveLength(118.2, model::enums::UnitsLength::MICROMETER), i);
meta.setPixelsPhysicalSizeY
(PositiveLength(118.2, model::enums::UnitsLength::MICROMETER), i);
meta.setPixelsPhysicalSizeZ
(PositiveLength(26.5, model::enums::UnitsLength::MICROMETER), i);
}
}
When adding new objects to the store, as opposed to updating existing
ones, some additional considerations apply. An new object is added to
the store if the object corresponding to an index does not exist and
the index is the current object count (i.e. one past the end of the
last valid index). Note that for data model objects with a
setID()
method, this method alone will trigger insertion and
must be called first, before any other methods which modify the
object. The following example demonstrates the addition of a new
Image
to the store, plus contained Plane
objects.
void
addMetadata(meta::Metadata& meta)
{
// Get total number of images (series)
index_type i = meta.getImageCount();
// Size of Z, T and C dimensions
index_type nz = 3;
index_type nt = 1;
index_type nc = 4;
// Create new image; the image index is the same as the image
// count, i.e. one past the end of the current limit; createID
// creates a unique identifier for the image
meta.setImageID(createID("Image", i), i);
// Set Pixels identifier using createID and the same image index
meta.setPixelsID(createID("Pixels", i), i);
// Now set the dimension order, pixel type and dimension sizes for
// this image, using the same image index
meta.setPixelsDimensionOrder(model::enums::DimensionOrder::XYZTC, i);
meta.setPixelsType(model::enums::PixelType::UINT8, i);
meta.setPixelsSizeX(256, i);
meta.setPixelsSizeY(256, i);
meta.setPixelsSizeZ(nz, i);
meta.setPixelsSizeT(nt, i);
meta.setPixelsSizeC(nc, i);
// Plane count
index_type pc = nz * nc * nt;
// Loop over planes
for(index_type p = 0; p < pc; ++p)
{
// Get the Z, T and C coordinate for this plane index
array<dimension_size_type, 3> coord =
getZCTCoords("XYZTC", nz, nc, nt, pc, p);
// Set the plane position using the image index and plane
// index to reference the correct plane
meta.setPlaneTheZ(coord[0], i, p);
meta.setPlaneTheT(coord[2], i, p);
meta.setPlaneTheC(coord[1], i, p);
}
// Add MetadataOnly to Pixels since this is an example without
// TiffData or BinData
meta::OMEXMLMetadata *omexmlmeta = dynamic_cast<meta::OMEXMLMetadata *>(&meta);
if (omexmlmeta)
addMetadataOnly(*omexmlmeta, i);
}
In addition to this basic metadata, it is possible to create and modify extended metadata elements. In the following example, we describe the setup of the microscope during acquisition, including its objective and detector parameters. Only a few parameters are set here; it is possible to completely describe the instrument configuration, including the settings on a per-image and per-channel basis if they vary during the course of acquisition.
// There is one image with one channel in this image.
MetadataStore::index_type image_idx = 0;
MetadataStore::index_type channel_idx = 0;
MetadataStore::index_type annotation_idx = 0;
// Create an Instrument.
MetadataStore::index_type instrument_idx = 0;
std::string instrument_id = createID("Instrument", instrument_idx);
store->setInstrumentID(instrument_id, instrument_idx);
// Create an Objective for this Instrument.
MetadataStore::index_type objective_idx = 0;
std::string objective_id = createID("Objective",
instrument_idx, objective_idx);
store->setObjectiveID(objective_id, instrument_idx, objective_idx);
store->setObjectiveManufacturer("InterFocal", instrument_idx, objective_idx);
store->setObjectiveNominalMagnification(40, instrument_idx, objective_idx);
store->setObjectiveLensNA(0.4, instrument_idx, objective_idx);
store->setObjectiveImmersion(Immersion::OIL, instrument_idx, objective_idx);
store->setObjectiveWorkingDistance({0.34, UnitsLength::MILLIMETER},
instrument_idx, objective_idx);
// Create a Detector for this Instrument.
MetadataStore::index_type detector_idx = 0;
std::string detector_id = createID("Detector", instrument_idx, detector_idx);
store->setDetectorID(detector_id, instrument_idx, detector_idx);
store->setObjectiveManufacturer("MegaCapture", instrument_idx, detector_idx);
store->setDetectorType(DetectorType::CCD, instrument_idx, detector_idx);
// Create Settings for this Detector for the Channel on the Image.
store->setDetectorSettingsID(detector_id, image_idx, channel_idx);
store->setDetectorSettingsBinning(Binning::TWOBYTWO, image_idx, channel_idx);
store->setDetectorSettingsGain(56.89, image_idx, channel_idx);
If the existing data model elements and attributes are insufficient
for describing the complexity of your hardware or experimental setup,
it is possible to extend it with custom annotations. These
annotations exist globally, but may be referenced by a model element
where needed, and may be referenced by multiple model elements if
required. In the following example, we create and attach an
annotation to the Detector
element, and then create and attach two
annotations to the first Image
element.
// Create a MapAnnotation.
MetadataStore::index_type map_annotation_idx = 0;
std::string annotation_id = createID("Annotation", annotation_idx);
store->setMapAnnotationID(annotation_id, map_annotation_idx);
store->setMapAnnotationNamespace
("https://microscopy.example.com/colour-balance", map_annotation_idx);
ome::xml::model::primitives::OrderedMultimap map;
map.push_back({"white-balance", "5,15,8"});
map.push_back({"black-balance", "112,140,126"});
store->setMapAnnotationValue(map, map_annotation_idx);
// Link MapAnnotation to Detector.
MetadataStore::index_type detector_ref_idx = 0;
store->setDetectorAnnotationRef(annotation_id, instrument_idx, detector_idx,
detector_ref_idx);
// Create a LongAnnotation.
++annotation_idx;
MetadataStore::index_type long_annotation_idx = 0;
annotation_id = createID("Annotation", annotation_idx);
store->setLongAnnotationID(annotation_id, long_annotation_idx);
store->setLongAnnotationValue(239423, long_annotation_idx);
store->setLongAnnotationNamespace
("https://microscopy.example.com/trigger-delay", long_annotation_idx);
// Link LongAnnotation to Image.
MetadataStore::index_type image_ref_idx = 0;
store->setImageAnnotationRef(annotation_id, image_idx, image_ref_idx);
// Create a second LongAnnotation.
++annotation_idx;
++long_annotation_idx;
annotation_id = createID("Annotation", annotation_idx);
store->setLongAnnotationID(annotation_id, long_annotation_idx);
store->setLongAnnotationValue(934223, long_annotation_idx);
store->setLongAnnotationNamespace
("https://microscopy.example.com/sample-number", long_annotation_idx);
// Link second LongAnnotation to Image.
++image_ref_idx = 0;
store->setImageAnnotationRef(annotation_id, image_idx, image_ref_idx);
// Update all the annotation cross-references.
store->resolveReferences();
Full example source: metadata-io.cpp
and metadata-formatwriter.cpp
OME-XML data model objects¶
The data model objects are not typically used directly, but are
created, modified and queried using the Metadata
interfaces (above), so in practice these examples should not be
needed.
To create a tree of OME-XML data model objects from OME-XML text:
// XML DOM tree containing parsed file content
xml::dom::Document inputdoc(ome::xml::createDocument(filename));
// OME Model (needed only during parsing to track model object references)
model::detail::OMEModel model;
// OME Model root object
shared_ptr<model::OME> modelroot(make_shared<model::OME>());
// Fill OME model object tree from XML DOM tree
modelroot->update(inputdoc.getDocumentElement(), model);
In this example, the OME-XML text is read from a file into a DOM tree.
This could have been read directly from a string or stream if the
source was not a file. The DOM tree is then processed using the
OME
root object’s update()
method, which uses
the data from the DOM tree elements to create a tree of corresponding
model objects contained by the root object.
To reverse the process, taking a tree of OME-XML model objects and converting them back of OME-XML text:
// Schema version to use
const std::string schema("http://www.openmicroscopy.org/Schemas/OME/2016-06");
// XML DOM tree (initially containing an empty OME root element)
xml::dom::Document outputdoc(xml::dom::createEmptyDocument(schema, "OME"));
// Fill output DOM document from OME-XML model
modelroot->asXMLElement(outputdoc);
// Dump DOM tree as text to stream
xml::dom::writeDocument(outputdoc, stream);
Here, the OME
root object’s asXMLElement()
method is used to copy the data from the OME root object and its
children into an XML DOM tree. The DOM tree is then converted to text
for output.
As shown previously for the MetadataStore
API, it is also
possible to create and modify extended metadata elements using the
model objects directly. The following example demonstrates the setup
of the microscope during acquisition, including its objective and
detector parameters, to achieve the same effect as in the example
above.
// Get root OME object
std::shared_ptr<ome::xml::meta::OMEXMLMetadataRoot>
root(std::dynamic_pointer_cast<ome::xml::meta::OMEXMLMetadataRoot>
(store->getRoot()));
if (!root)
return;
MetadataStore::index_type annotation_idx = 0;
// Create an Instrument.
auto instrument = make_shared<ome::xml::model::Instrument>();
instrument->setID(createID("Instrument", 0));
root->addInstrument(instrument);
// Create an Objective for this Instrument.
auto objective = make_shared<ome::xml::model::Objective>();
objective->setID(createID("Objective", 0));
auto objective_manufacturer = std::make_shared<std::string>("InterFocal");
objective->setManufacturer(objective_manufacturer);
auto objective_nominal_mag = std::make_shared<double>(40.0);
objective->setNominalMagnification(objective_nominal_mag);
auto objective_na = std::make_shared<double>(0.4);
objective->setLensNA(objective_na);
auto objective_immersion = std::make_shared<Immersion>(Immersion::OIL);
objective->setImmersion(objective_immersion);
auto objective_wd = std::make_shared<Quantity<UnitsLength>>
(0.34, UnitsLength::MILLIMETER);
objective->setWorkingDistance(objective_wd);
instrument->addObjective(objective);
// Create a Detector for this Instrument.
auto detector = make_shared<ome::xml::model::Detector>();
std::string detector_id = createID("Detector", 0);
detector->setID(detector_id);
auto detector_manufacturer = std::make_shared<std::string>("MegaCapture");
detector->setManufacturer(detector_manufacturer);
auto detector_type = std::make_shared<DetectorType>(DetectorType::CCD);
detector->setType(detector_type);
instrument->addDetector(detector);
// Get Image and Channel for future use. Note for your own code,
// you should check that the elements exist before accessing them;
// here we know they are valid because we created them above.
auto image = root->getImage(0);
auto pixels = image->getPixels();
auto channel0 = pixels->getChannel(0);
auto channel1 = pixels->getChannel(1);
auto channel2 = pixels->getChannel(2);
// Create Settings for this Detector for each Channel on the Image.
auto detector_settings0 = make_shared<ome::xml::model::DetectorSettings>();
{
detector_settings0->setID(detector_id);
auto detector_binning = std::make_shared<Binning>(Binning::TWOBYTWO);
detector_settings0->setBinning(detector_binning);
auto detector_gain = std::make_shared<double>(83.81);
detector_settings0->setGain(detector_gain);
channel0->setDetectorSettings(detector_settings0);
}
auto detector_settings1 = make_shared<ome::xml::model::DetectorSettings>();
{
detector_settings1->setID(detector_id);
auto detector_binning = std::make_shared<Binning>(Binning::TWOBYTWO);
detector_settings1->setBinning(detector_binning);
auto detector_gain = std::make_shared<double>(56.89);
detector_settings1->setGain(detector_gain);
channel1->setDetectorSettings(detector_settings1);
}
auto detector_settings2 = make_shared<ome::xml::model::DetectorSettings>();
{
detector_settings2->setID(detector_id);
auto detector_binning = std::make_shared<Binning>(Binning::FOURBYFOUR);
detector_settings2->setBinning(detector_binning);
auto detector_gain = std::make_shared<double>(12.93);
detector_settings2->setGain(detector_gain);
channel2->setDetectorSettings(detector_settings2);
}
Creating annotations and linking them to model objects is also possible using model objects directly:
// Add Structured Annotations.
auto sa = std::make_shared<ome::xml::model::StructuredAnnotations>();
root->setStructuredAnnotations(sa);
// Create a MapAnnotation.
auto map_ann0 = std::make_shared<ome::xml::model::MapAnnotation>();
std::string annotation_id = createID("Annotation", annotation_idx);
map_ann0->setID(annotation_id);
auto map_ann0_ns = std::make_shared<std::string>
("https://microscopy.example.com/colour-balance");
map_ann0->setNamespace(map_ann0_ns);
ome::xml::model::primitives::OrderedMultimap map;
map.push_back({"white-balance", "5,15,8"});
map.push_back({"black-balance", "112,140,126"});
map_ann0->setValue(map);
sa->addMapAnnotation(map_ann0);
// Link MapAnnotation to Detector.
detector->linkAnnotation(map_ann0);
// Create a LongAnnotation.
auto long_ann0 = std::make_shared<ome::xml::model::LongAnnotation>();
++annotation_idx;
annotation_id = createID("Annotation", annotation_idx);
long_ann0->setID(annotation_id);
auto long_ann0_ns = std::make_shared<std::string>
("https://microscopy.example.com/trigger-delay");
long_ann0->setNamespace(long_ann0_ns);
long_ann0->setValue(239423);
sa->addLongAnnotation(long_ann0);
// Link LongAnnotation to Image.
image->linkAnnotation(long_ann0);
// Create a second LongAnnotation.
auto long_ann1 = std::make_shared<ome::xml::model::LongAnnotation>();
++annotation_idx;
annotation_id = createID("Annotation", annotation_idx);
long_ann1->setID(annotation_id);
auto long_ann1_ns = std::make_shared<std::string>
("https://microscopy.example.com/sample-number");
long_ann1->setNamespace(long_ann1_ns);
long_ann1->setValue(934223);
sa->addLongAnnotation(long_ann1);
// Link second LongAnnotation to Image.
image->linkAnnotation(long_ann1);
Full example source: model-io.cpp
and metadata-formatwriter2.cpp
See also
Pixel data¶
The Bio-Formats Java implementation stores and passes pixel values in
a raw byte
array. Due to limitations with C++ array
passing, this was not possible for the OME-Files C++ implementation.
While a vector or other container could have been used, several
problems remain. The type and endianness of the data in the raw bytes
is not known, and the dimension ordering and dimension extents are
also unknown, which imposes a significant burden on the programmer to
correctly process the data. The C++ implementation provides two types
to solve these problems.
The PixelBuffer
class is a container of pixel data. It
is a template class, templated on the pixel type in use. The class
contains the order of the dimensions, and the size of each dimension,
making it possible to process pixel data without need for
externally-provided metadata to describe its structure. This class
may be used to contain and process pixel data of a specific pixel
type. Internally, the pixel data is contained within a
boost::multi_array
as a 9D hyper-volume, though its usage
in this release of OME-Files is limited to 5D. The class can either
contain its own memory allocation for pixel data, or it can reference
memory allocated or mapped externally, allowing use with memory-mapped
data, for example.
In many situations, it is desirable to work with arbitrary pixel
types, or at least the set of pixel types defined in the OME data
model in its PixelType
enumeration. The
VariantPixelBuffer
fulfills this need, using
boost::variant
to allow it to contain a
PixelBuffer
specialized for any of the pixel types in the
OME data model. This is used to allow transfer and processing of any
supported pixel type, for example by the FormatReader
class’ getLookupTable()
and openBytes()
methods,
and the corresponding FormatWriter
class’
setLookupTable()
and saveBytes()
methods.
An additional problem with supporting many different pixel types is
that each operation upon the pixel data, for example for display or
analysis, may require implementing separately for each pixel type.
This imposes a significant testing and maintenance burden.
VariantPixelBuffer
solves this problem through use of
boost::apply_visitor()
and
boost::static_visitor
, which allow algorithms to be
defined in a template and compiled for each pixel type. They also
allow algorithms to be specialized for different classes of pixel
type, for example signed vs. unsigned, integer vs. floating point, or
simple vs. complex, or special-cased per type e.g. for bitmasks. When
boost::apply_visitor()
is called with a specified algorithm
and VariantPixelBuffer
object, it will select the
matching algorithm for the pixel type contained within the buffer, and
then invoke it on the buffer. This permits the programmer to support
arbitrary pixel types without creating a maintenance nightmare, and
without unnecessary code duplication.
The 9D pixel buffer makes a distinction between the logical dimension order (used by the API) and the storage order (the layout of the pixel data in memory). The logical order is defined by the values in the Dimensions enum. The storage order is specified by the programmer when creating a pixel buffer.
The following example shows creation of a pixel buffer with a defined size, and default storage order:
// Language type for FLOAT pixel data
typedef PixelProperties<PixelType::FLOAT>::std_type float_pixel_type;
// Create PixelBuffer for floating point data
// X=512 Y=512 Z=16 T=1 C=3 S/z/t/c=1
PixelBuffer<float_pixel_type> buffer
(boost::extents[512][512][16][1][3][1][1][1][1], PixelType::FLOAT);
The storage order may be set explicitly. The order may be created by hand, or with a helper function. While the helper function is limited to supporting the ordering defined by the data model, specifying the order by hand allows additional flexibility. Manual ordering may be used to allow the indexing for individual dimensions to run backward rather than forward, which is useful if the Y-axis requires inverting, for example. The following example shows creation of two pixel buffers with defined storage order using the helper function:
// Language type for UINT16 pixel data
typedef PixelProperties<PixelType::UINT16>::std_type uint16_pixel_type;
// Storage order is XYSCTZctz; subchannels are not interleaved
// ("planar") after XY; lowercase letters are unused Modulo
// dimensions
PixelBufferBase::storage_order_type order1
(PixelBufferBase::make_storage_order(DimensionOrder::XYCTZ, false));
// Create PixelBuffer for unsigned 16-bit data with specified
// storage order
// X=512 Y=512 Z=16 T=1 C=3 S/z/t/c=1
PixelBuffer<uint16_pixel_type> buffer1
(boost::extents[512][512][16][1][3][1][1][1][1],
PixelType::UINT16,
ome::files::ENDIAN_NATIVE,
order1);
// Language type for INT8 pixel data
typedef PixelProperties<PixelType::INT8>::std_type int8_pixel_type;
// Storage order is SXYZCTzct; subchannels are interleaved
// ("chunky") before XY; lowercase letters are unused Modulo
// dimensions
PixelBufferBase::storage_order_type order2
(PixelBufferBase::make_storage_order(DimensionOrder::XYZCT, true));
// Create PixelBuffer for signed 8-bit RGB data with specified storage
// order
// X=1024 Y=1024 Z=1 T=1 C=1 S=3 z/t/c=1
PixelBuffer<int8_pixel_type> buffer2
(boost::extents[1024][1024][1][1][1][3][1][1][1],
PixelType::INT8,
ome::files::ENDIAN_NATIVE,
order2);
Note that the logical order of the dimension extents is unchanged.
Sometimes it may be necessary to change the storage order of data, for example to give it the appropriate structure to pass to another library with specific ordering requirements. This can be done by a simple assignment between two buffers having a different storage order; the dimension extents must be of the same size for the buffers to be compatible. The following example demonstrates conversion of planar data to contiguous:
// Language type for FLOAT pixel data
typedef PixelProperties<PixelType::FLOAT>::std_type float_pixel_type;
// Storage order is XYSZCTzct; subchannels are not interleaved
// ("planar") after XY; lowercase letters are unused Modulo
// dimensions
PixelBufferBase::storage_order_type planar_order
(PixelBufferBase::make_storage_order(DimensionOrder::XYZCT, false));
// Storage order is SXYZCTzct; subchannels are interleaved
// ("chunky" or "contiguous") before XY; lowercase letters are
// unused Modulo dimensions
PixelBufferBase::storage_order_type contiguous_order
(PixelBufferBase::make_storage_order(DimensionOrder::XYZCT, true));
// Create PixelBuffer for float data with planar order
// X=512 Y=512 Z=16 T=1 C=3 S/z/t/c=1
PixelBuffer<float_pixel_type> planar_buffer
(boost::extents[512][512][16][1][3][1][1][1][1],
PixelType::FLOAT,
ome::files::ENDIAN_NATIVE,
planar_order);
// Create PixelBuffer for float data with contiguous order
// X=512 Y=512 Z=16 T=1 C=3 S/z/t/c=1
PixelBuffer<float_pixel_type> contiguous_buffer
(boost::extents[512][512][16][1][3][1][1][1][1],
PixelType::FLOAT,
ome::files::ENDIAN_NATIVE,
contiguous_order);
// Transfer the pixel data from the planar buffer to the
// contiguous buffer; the pixel data will be reordered
// appropriately during the transfer.
contiguous_buffer = planar_buffer;
In-place conversion is not yet supported.
In practice, it is unlikely that you will need to create any
PixelBuffer
objects directly. The
FormatReader
and FormatWriter
interfaces use
VariantPixelBuffer
objects, and in the case of the reader
interface the getLookupTable()
and openBytes()
methods can be passed a default-constructed
VariantPixelBuffer
and it will be set up automatically,
changing the image dimensions, dimension order and pixel type to match
the data being fetched, if the size, order and type do not match. For
example, to read all pixel data in an image using
openBytes()
:
void
readPixelData(const FormatReader& reader,
std::ostream& stream)
{
// Get total number of images (series)
dimension_size_type ic = reader.getSeriesCount();
stream << "Image count: " << ic << '\n';
// Loop over images
for (dimension_size_type i = 0 ; i < ic; ++i)
{
// Change the current series to this index
reader.setSeries(i);
// Get total number of planes (for this image index)
dimension_size_type pc = reader.getImageCount();
stream << "\tPlane count: " << pc << '\n';
// Pixel buffer
VariantPixelBuffer buf;
// Loop over planes (for this image index)
for (dimension_size_type p = 0 ; p < pc; ++p)
{
// Read the entire plane into the pixel buffer.
reader.openBytes(p, buf);
// If this wasn't an example, we would do something
// exciting with the pixel data here.
stream << "Pixel data for Image " << i
<< " Plane " << p << " contains "
<< buf.num_elements() << " pixels\n";
}
}
}
To perform the reverse process, writing pixel data with
saveBytes()
:
// Total number of images (series)
dimension_size_type ic = writer.getMetadataRetrieve()->getImageCount();
stream << "Image count: " << ic << '\n';
// Loop over images
for (dimension_size_type i = 0 ; i < ic; ++i)
{
// Change the current series to this index
writer.setSeries(i);
// Total number of planes.
dimension_size_type pc = 1U;
pc *= writer.getMetadataRetrieve()->getPixelsSizeZ(i);
pc *= writer.getMetadataRetrieve()->getPixelsSizeT(i);
pc *= writer.getMetadataRetrieve()->getChannelCount(i);
stream << "\tPlane count: " << pc << '\n';
// Loop over planes (for this image index)
for (dimension_size_type p = 0 ; p < pc; ++p)
{
// Pixel buffer; size 512 × 512 with 3 subchannels of type
// uint16_t. It uses the native endianness and has a
// storage order of XYZTC without interleaving
// (subchannels are planar).
shared_ptr<PixelBuffer<PixelProperties<PixelType::UINT16>::std_type>>
buffer(make_shared<PixelBuffer<PixelProperties<PixelType::UINT16>::std_type>>
(boost::extents[512][512][1][1][1][3][1][1][1],
PixelType::UINT16, ome::files::ENDIAN_NATIVE,
PixelBufferBase::make_storage_order(DimensionOrder::XYZTC, false)));
// Fill each subchannel with a different intensity ramp in
// the 12-bit range. In a real program, the pixel data
// would typically be obtained from data acquisition or
// another image.
for (dimension_size_type x = 0; x < 512; ++x)
for (dimension_size_type y = 0; y < 512; ++y)
{
PixelBufferBase::indices_type idx;
std::fill(idx.begin(), idx.end(), 0);
idx[DIM_SPATIAL_X] = x;
idx[DIM_SPATIAL_Y] = y;
idx[DIM_SUBCHANNEL] = 0;
buffer->at(idx) = (static_cast<float>(x) / 512.0f) * 4096.0f;
idx[DIM_SUBCHANNEL] = 1;
buffer->at(idx) = (static_cast<float>(y) / 512.0f) * 4096.0f;
idx[DIM_SUBCHANNEL] = 2;
buffer->at(idx) = (static_cast<float>(x+y) / 1024.0f) * 4096.0f;
}
VariantPixelBuffer vbuffer(buffer);
stream << "PixelBuffer PixelType is " << buffer->pixelType() << '\n';
stream << "VariantPixelBuffer PixelType is " << vbuffer.pixelType() << '\n';
stream << std::flush;
// Write the the entire pixel buffer to the plane.
writer.saveBytes(p, vbuffer);
stream << "Wrote " << buffer->num_elements() << ' '
<< buffer->pixelType() << " pixels\n";
}
}
Both buffer classes provide access to the pixel data so that it may be
accessed, manipulated and passed elsewhere. The
PixelBuffer
class provides an at
method.
This allows access to individual pixel values using a 9D coordinate:
// Set all pixel values for Z=2 and C=1 to 0.5
// 9D index, default values to zero if unused
PixelBuffer<float_pixel_type>::indices_type idx;
// Set Z and C indices
idx[ome::files::DIM_SPATIAL_Z] = 2;
idx[ome::files::DIM_CHANNEL] = 1;
idx[ome::files::DIM_TEMPORAL_T] =
idx[ome::files::DIM_SUBCHANNEL] =
idx[ome::files::DIM_MODULO_Z] =
idx[ome::files::DIM_MODULO_T] =
idx[ome::files::DIM_MODULO_C] = 0;
for (uint16_t x = 0; x < 512; ++x)
{
idx[ome::files::DIM_SPATIAL_X] = x;
for (uint16_t y = 0; y < 512; ++y)
{
idx[ome::files::DIM_SPATIAL_Y] = y;
buffer.at(idx) = 0.5f;
}
}
Conceptually, this is the same as using an index for a normal 1D
array, but extended to use an array of nine indices for each of the
nine dimensions, in the logical storage order. The
VariantPixelBuffer
does not provide an at
method for efficiency reasons. Instead, visitors should be used for
the processing of bulk pixel data. For example, this is one way the
minimum and maximum pixel values could be obtained:
// Visitor to compute min and max pixel value for pixel buffer of
// any pixel type
// The static_visitor specialization is the required return type of
// the operator() methods and boost::apply_visitor()
struct MinMaxVisitor : public boost::static_visitor<std::pair<double, double>>
{
// The min and max values will be returned in a pair. double is
// used since it can contain the value for any pixel type
typedef std::pair<double, double> result_type;
// Get min and max for any non-complex pixel type
template<typename T>
result_type
operator() (const T& v)
{
typedef typename T::element_type::value_type value_type;
value_type *min = std::min_element(v->data(),
v->data() + v->num_elements());
value_type *max = std::max_element(v->data(),
v->data() + v->num_elements());
return result_type(static_cast<double>(*min),
static_cast<double>(*max));
}
// Less than comparison for real part of complex numbers
template <typename T>
static bool
complex_real_less(const T& lhs, const T& rhs)
{
return std::real(lhs) < std::real(rhs);
}
// Greater than comparison for real part of complex numbers
template <typename T>
static bool
complex_real_greater(const T& lhs, const T& rhs)
{
return std::real(lhs) > std::real(rhs);
}
// Get min and max for complex pixel types (COMPLEXFLOAT and
// COMPLEXDOUBLE)
// This is the same as for simple pixel types, except for the
// addition of custom comparison functions and conversion of the
// result to the real part.
template <typename T>
typename boost::enable_if_c<
boost::is_complex<T>::value, result_type
>::type
operator() (const std::shared_ptr<PixelBuffer<T>>& v)
{
typedef T value_type;
value_type *min = std::min_element(v->data(),
v->data() + v->num_elements(),
complex_real_less<T>);
value_type *max = std::max_element(v->data(),
v->data() + v->num_elements(),
complex_real_greater<T>);
return result_type(static_cast<double>(std::real(*min)),
static_cast<double>(std::real(*max)));
}
};
void
applyVariant()
{
// Make variant buffer (int32, 16×16 single plane)
VariantPixelBuffer variant(boost::extents[16][16][1][1][1][1][1][1][1],
PixelType::INT32);
// Get buffer size
VariantPixelBuffer::size_type size = variant.num_elements();
// Create sample random-ish data
std::vector<int32_t> vec;
for (VariantPixelBuffer::size_type i = 0; i < size; ++i)
{
int32_t val = static_cast<int32_t>(i + 42);
vec.push_back(val);
}
std::random_shuffle(vec.begin(), vec.end());
// Assign sample data to buffer.
variant.assign(vec.begin(), vec.end());
// Create and apply visitor
MinMaxVisitor visitor;
MinMaxVisitor::result_type result = boost::apply_visitor(visitor, variant.vbuffer());
std::cout << "Min is " << result.first
<< ", max is " << result.second << '\n';
}
This example demonstrates several features:
- The visitor operators can return values to the caller (for more complex algorithms, the visitor class could use member variables and additional methods)
- The operator is expanded once for each pixel type
- The operators can be special-cased for individual pixel types; here we use the SFINAE rule to implement a specialization for an entire category of pixel types (complex numbers), but standard function overloading and templates will also work for more common cases
- Pixel data can be assigned to the buffer with a single
assign()
call.
The OME-Files source uses pixel buffer visitors for several purposes, for example to load pixel data into OpenGL textures, which automatically handles pixel format conversion and repacking of pixel data as needed.
While the pixel buffers may appear complex, they do permit the OME Files library to support all pixel types with relative ease, and it will allow your applications to also handle multiple pixel types by writing your own visitors. Assignment of one buffer to another will also repack the pixel data if they use different storage ordering (i.e. the logical ordering is used for the copy), which can be useful if you need the pixel data in a defined ordering.
If all you want is access to the raw data, as in the Java API, you are
not required to use the above features. Simply use the
data()
method on the buffer to get a pointer to the raw
data. Note that you will need to multiply the buffer size obtained
with num_elements()
by the size of the pixel type (use
bytesPerPixel()
or sizeof
on the buffer
value_type
).
Alternatively, it is also possible to access the underlying
boost::multi_array
using the array()
method, if
you need access to functionality not wrapped by
PixelBuffer
.
Full example source: pixeldata.cpp
Reading images¶
Image reading is performed using the FormatReader
interface. This is an abstract reader interface implemented by
file-format-specific reader classes. Examples of readers include
TIFFReader
, which implements reading of Baseline TIFF
(optionally with additional ImageJ metadata), and
OMETIFFReader
which implements reading of OME-TIFF (TIFF
with OME-XML metadata).
Using a reader involves these steps:
- Create a reader instance.
- Set options to control reader behavior.
- Call
setId()
to specify the image file to read. - Retrieve desired metadata and pixel data.
- Close the reader.
These steps are illustrated in this example:
// Create TIFF reader
shared_ptr<FormatReader> reader(make_shared<TIFFReader>());
// Set reader options before opening a file
reader->setMetadataFiltered(false);
reader->setGroupFiles(true);
// Open the file
reader->setId(filename);
// Display series core metadata
readMetadata(*reader, std::cout);
// Display global and series original metadata
readOriginalMetadata(*reader, std::cout);
// Read pixel data
readPixelData(*reader, std::cout);
// Explicitly close reader
reader->close();
Here we create a reader to read TIFF files, set two options (metadata
filtering and file grouping), and then call setId()
. At
this point the reader has been set up and initialized, and we can then
read metadata and pixel data, which we covered in the preceding
sections. You might like to combine this example with the
MinMaxVisitor
example to make it display the minimum and
maximum values for each plane in an image; if you try running the
example with TIFF images of different pixel types, it will
transparently adapt to any supported pixel type.
Note
Reader option-setting methods may only be called before
setId()
. Reader state changing and querying methods such
as setSeries()
and getSeries()
, metadata
retrieval and pixel data retrieval methods may only be called
after setId()
. If these constraints are violated, a
FormatException
will be thrown.
Full example source: metadata-formatreader.cpp
See also
Writing images¶
Image writing is performed using the FormatWriter
interface. This is an abstract writer interface implemented by
file-format-specific writer classes. Examples of writers include
MinimalTIFFWriter
, which implements writing of Baseline
TIFF and OMETIFFWriter
which implements writing of
OME-TIFF (TIFF with OME-XML metadata).
Using a writer involves these steps:
- Create a writer instance.
- Set metadata store to use.
- Set options to control writer behavior.
- Call
setId()
to specify the image file to write. - Store pixel data for each plane of each image in the specified dimension order.
- Close the writer.
These steps are illustrated in this example:
// Create minimal metadata for the file to be written.
auto meta = createMetadata();
// Add extended metadata.
addExtendedMetadata(meta);
// Create TIFF writer
auto writer = make_shared<OMETIFFWriter>();
// Set writer options before opening a file
auto retrieve = static_pointer_cast<MetadataRetrieve>(meta);
writer->setMetadataRetrieve(retrieve);
writer->setInterleaved(false);
writer->setTileSizeX(256);
writer->setTileSizeY(256);
// Open the file
writer->setId(filename);
// Write pixel data
writePixelData(*writer, std::cout);
// Explicitly close writer
writer->close();
Here we create a writer to write OME-TIFF files, set the metadata
store using metadata we create, then set a writer option (sample
interleaving), and then call setId()
. At this point the
writer has been set up and initialized, and we can then write the
pixel data, which we covered in the preceding sections. Finally we
call close()
to flush all data.
Note
Metadata store setting and writer option-setting methods may only be
called before setId()
. Writer state changing and
querying methods such as setSeries()
and
getSeries()
, and pixel data storage methods may only be
called after setId()
. If these constraints are
violated, a FormatException
will be thrown.
Note
close()
should be called explicitly to catch any errors.
While this will be called by the destructor, the destructor can’t
throw exceptions and any errors will be silently ignored.
Full example source: metadata-formatwriter.cpp
See also