Background & Context
- ImageType gives a label representing the number type used internally to represent each pixel of an image. The Wolfram Language supports images encoded by 1-bit integers (i.e. two-level binary), 8-bit unsigned integers, 16-bit unsigned integers, and 32- and 64-bit real numbers. These different bit depths allow a variety of image encodings, with smaller bit depths used for compact (i.e. lower memory) representation of images and larger bit depths providing higher resolution.
- Many image processing functions automatically convert integer-type images to a real type for higher accuracy prior to computation.
- The function Image can be used to convert between image types, and ImageAdjust can be used to bring all values of a real-valued image into the range 0 to 1. Another important property of an image is the number of channels the image contains, which may be obtained using the function ImageChannels.
Introduced in 2008Updated in 2019