An output image produced by a Core ML image analysis request.


class VNPixelBufferObservation : VNObservation


This type of observation results from performing a VNCoreMLRequest image analysis with a Core ML model whose role is image-to-image processing. For example, this observation would result from a model that analyzes the style of one image and then transfers that style to a different image.

Vision infers that an MLModel object is an image-to-image model if that model includes an image. Its modelDescription object includes an image-typed feature description in its outputDescriptionsByName dictionary.


Parsing Observation Content

var pixelBuffer: CVPixelBuffer

The image that results from a request with image output.

var featureName: String?

The name used in the model description of the CoreML model that produced this observation.


Inherits From

Conforms To

See Also

Machine-Learning Image Analysis

Classifying Images with Vision and Core ML

Preprocess photos using the Vision framework and classify them with a Core ML model.

Training a Create ML Model to Classify Flowers

Train a flower classifier using Create ML in Swift Playgrounds, and apply the resulting model to real-time image classification using Vision.

class VNCoreMLRequest

An image analysis request that uses a Core ML model to process images.

class VNClassificationObservation

Classification information produced by an image analysis request.

class VNCoreMLFeatureValueObservation

A collection of key-value information produced by a Core ML image analysis request.