A collection of key-value information produced by a Core ML image analysis request.


@interface VNCoreMLFeatureValueObservation : VNObservation


This type of observation results from performing a VNCoreMLRequest image analysis with a Core ML model whose role is prediction rather than classification or image-to-image processing.

Vision infers that an MLModel object is a predictor model if that model predicts multiple features. You can tell that a model predicts multiple features when its modelDescription object has a nil value for its predictedFeatureName property, or when it inserts its output in an outputDescriptionsByName dictionary.


Obtaining Feature Values


The feature result of a VNCoreMLRequest that outputs neither a classification nor an image.


The name used in the model description of the CoreML model that produced this observation.


Inherits From

See Also

Machine-Learning Image Analysis

Classifying Images with Vision and Core ML

Preprocess photos using the Vision framework and classify them with a Core ML model.

Training a Create ML Model to Classify Flowers

Train a flower classifier using Create ML in Swift Playgrounds, and apply the resulting model to real-time image classification using Vision.


An image analysis request that uses a Core ML model to process images.


Classification information produced by an image analysis request.


An output image produced by a Core ML image analysis request.