Sessions

  • Detect animal poses in Vision

    8:00 a.m.

    Go beyond detecting cats and dogs in images. We’ll show you how to use Vision to detect the individual joints and poses of these animals as well — all in real time — and share how you can enable exciting features like animal tracking for a camera app, creative embellishment on an animal photo, and more. We’ll also explore other important enhancements to Vision and share best practices. To learn even more about what’s new in the Vision framework, watch "Explore 3D body pose and person segmentation in Vision” and "Lift subjects from images in your app.” And to learn more about building live camera-tracking experiences, check out "Integrate with motorized iPhone stands using DockKit"

    Watch

  • Discover machine learning enhancements in Create ML

    8:00 a.m.

    Find out how Create ML can help you do even more with machine learning models. Learn about the latest updates to image understanding and text-based tasks with multilingual BERT embeddings. Discover how easy it is to train models that can understand the content of images using multi-label classification. We’ll also share information about interactive model evaluation and the latest APIs for custom training data augmentations. To learn more about the latest updates to machine learning, watch “Explore Natural Language multilingual models” and “Improve Core ML integration with async prediction" from WWDC23.

    Watch

  • Lift subjects from images in your app

    8:00 a.m.

    Discover how you can easily pull the subject of an image from its background in your apps. Learn how to lift the primary subject or to access the subject at a given point with VisionKit. We’ll also share how you can lift subjects using Vision and combine that with lower-level frameworks like Core Image to create fun image effects and more complex compositing pipelines. For more information about the latest updates to VisionKit, check out “What’s new in VisionKit." And for more information about person segmentation in images, watch "Explore 3D body pose and person segmentation in Vision" from WWDC23.

    Watch

  • What’s new in VisionKit

    8:00 a.m.

    Discover how VisionKit can help people quickly lift subjects from images in your app and learn more about the content of an image with Visual Look Up. We’ll also take a tour of the latest updates to VisionKit for Live Text interaction, data scanning, and expanded support for macOS apps. For more on VisionKit, check out “Lift subjects from images in your app" from WWDC23.

    Watch

Labs

  • Machine Learning & Computer Vision lab

    Tuesday @ 1:00 - 3:00 p.m.

    Request an appointment with an Apple engineer for guidance and conversation about machine learning in your app including Core ML, Create ML, Vision, Natural Language, Speech, Sound Analysis, as well as Core ML Tools for model conversion and compression.

    Request

Activities

  • Q&A: Create ML

    Tuesday @ 9:00 - 10:00 a.m.

    Ask Apple engineers about creating models using Create ML during this 1 hour text-based Q&A. Stop in to request guidance, ask for clarifications, or learn from others.

    Sign up

  • Q&A: Machine learning open forum

    Tuesday @ 10:00 - 11:00 a.m.

    Join us for an open forum to discuss anything related to machine learning in your app during this 1 hour text-based Q&A. Topics include Core ML, Create ML, Vision, Natural Language, Speech, Sound Analysis, TensorFlow and PyTorch model conversion, model compression, and more. Stop in to request guidance, ask for clarifications, or learn from others.

    Sign up

Sessions

  • Explore 3D body pose and person segmentation in Vision

    8:00 a.m.

    Discover how to build person-centric features with Vision. Learn how to detect human body poses and measure individual joint locations in 3D space. We’ll also show you how to take advantage of person segmentation APIs to distinguish and segment up to four individuals in an image. To learn more about the latest features in Vision, check out “Detect animal poses in Vision” from WWDC23.

    Watch

  • Explore Natural Language multilingual models

    8:00 a.m.

    Learn how to create custom Natural Language models for text classification and word tagging using multilingual, transformer-based embeddings. We’ll show you how to train with less data and support up to 27 different languages across three scripts. Find out how to use these embeddings to fine-tune complex models trained in PyTorch and TensorFlow. For more on Natural Language, check out "Make apps smarter with Natural Language” from WWDC20.

    Watch

Labs

  • Machine Learning & Computer Vision lab

    Wednesday @ 9:00 - 11:00 a.m.

    Request an appointment with an Apple engineer for guidance and conversation about machine learning in your app including Core ML, Create ML, Vision, Natural Language, Speech, Sound Analysis, as well as Core ML Tools for model conversion and compression.

    Request

  • Live text, subject lifting, and data scanning lab

    Wednesday @ 4:00 - 6:00 p.m.

    Request an appointment with an Apple engineer for guidance and conversation about supporting Live Text, subject lifting, and data scanning in your app through VisionKit.

    Request

Activities

  • Q&A: Subject lifting

    Wednesday @ 10:00 - 11:00 a.m.

    Ask Apple engineers about lifting subjects from images using both VisionKit and the Vision framework in this 1 hour text-based Q&A. Stop in to request code-level guidance, ask for clarifications, or learn from others.

    Sign up

  • Q&A: Image understanding and the Vision framework

    Wednesday @ 2:00 - 3:00 p.m.

    Ask Apple engineers about the many ways you can bring image understanding to your app using the Vision framework during this 1 hour text-based Q&A. Stop in to request guidance, ask for clarifications, or learn from others.

    Sign up

  • Q&A: Machine learning open forum

    Wednesday @ 3:00 - 4:00 p.m.

    Join us for an open forum to discuss anything related to machine learning in your app during this 1 hour text-based Q&A. Topics include Core ML, Create ML, Vision, Natural Language, Speech, Sound Analysis, TensorFlow and PyTorch model conversion, model compression, and more. Stop in to request guidance, ask for clarifications, or learn from others.

    Sign up

Sessions

  • Improve Core ML integration with async prediction

    8:00 a.m.

    Learn how to speed up machine learning features in your app with the latest Core ML execution engine improvements and find out how aggressive asset caching can help with inference and faster model loads. We’ll show you some of the latest options for async prediction and discuss considerations for balancing performance with overall memory usage to help you create a highly responsive app. Discover APIs to help you understand and maximize hardware utilization for your models. For more on optimizing Core ML model usage, check out "Use Core ML Tools for machine learning model compression" from WWDC23.

    Watch

  • Integrate with motorized iPhone stands using DockKit

    8:00 a.m.

    Discover how you can create incredible photo and video experiences in your camera app when integrating with DockKit-compatible motorized stands. We’ll show how your app can automatically track subjects in live video across a 360-degree field of view, take direct control of the stand to customize framing, directly control the motors, and provide your own inference model for tracking other objects. Finally, we’ll demonstrate how to create a sense of emotion through dynamic device animations. To learn more techniques for image tracking, check out “Detect animal poses in Vision” from WWDC23 and "Classify hand poses and actions with Create ML” from WWDC21.

    Watch

  • Optimize machine learning for Metal apps

    8:00 a.m.

    Discover the latest enhancements to accelerated ML training in Metal. Find out about updates to PyTorch and TensorFlow, and learn about Metal acceleration for JAX. We’ll show you how MPS Graph can support faster ML inference when you use both the GPU and Apple Neural Engine, and share how the same API can rapidly integrate your Core ML and ONNX models. For more information on using Metal for machine learning, check out “Accelerate machine learning with Metal” from WWDC22.

    Watch

  • Use Core ML Tools for machine learning model compression

    8:00 a.m.

    Discover how to reduce the footprint of machine learning models in your app with Core ML Tools. Learn how to use techniques like palettization, pruning, and quantization to dramatically reduce model size while still achieving great accuracy. Explore comparisons between compression during the training stages and on fully trained models, and learn how compressed models can run even faster when your app takes full advantage of the Apple Neural Engine. For more on optimizing Core ML, check out “Improve Core ML integration with async prediction" from WWDC23.

    Watch

Labs

  • Live text, subject lifting, and data scanning lab

    Thursday @ 9:00 - 11:00 a.m.

    Request an appointment with an Apple engineer for guidance and conversation about supporting Live Text, subject lifting, and data scanning in your app through VisionKit.

    Request

  • Machine Learning & Computer Vision lab

    Thursday @ 1:00 - 3:00 p.m.

    Request an appointment with an Apple engineer for guidance and conversation about machine learning in your app including Core ML, Create ML, Vision, Natural Language, Speech, Sound Analysis, as well as Core ML Tools for model conversion and compression.

    Request

Activities

  • Q&A: Core ML

    Thursday @ 9:00 - 10:00 a.m.

    Ask Apple engineers about bringing machine learning to your app with Core ML. Learn about techniques for model compression, optimizations for handling model loading, and getting improving app responsiveness with async prediction during this 1 hour text-based Q&A. Stop in to request guidance, ask for clarifications, or learn from others.

    Sign up

  • Q&A: Machine learning open forum

    Thursday @ 10:00 - 11:00 a.m.

    Join us for an open forum to discuss anything related to machine learning in your app during this 1 hour text-based Q&A. Topics include Core ML, Create ML, Vision, Natural Language, Speech, Sound Analysis, TensorFlow and PyTorch model conversion, model compression, and more. Stop in to request guidance, ask for clarifications, or learn from others.

    Sign up

  • Q&A: Live Text and data scanning

    Thursday @ 2:00 - 3:00 p.m.

    Ask Apple engineers about bringing Live Text and data scanning support to your app through VisionKit in this 1 hour text-based Q&A. Stop in to request code-level guidance, ask for clarifications, or learn from others.

    Sign up

Sessions

  • Customize on-device speech recognition

    8:00 a.m.

    Find out how you can improve on-device speech recognition in your app by customizing the underlying model with additional vocabulary. We’ll share how speech recognition works on device and show you how to boost specific words and phrases for more predictable transcription. Learn how you can provide specific pronunciations for words and use template support to quickly generate a full set of custom phrases — all at runtime. For more on the Speech framework, check out “Advances in Speech Recognition” from WWDC19.

    Watch

Labs

  • Machine Learning & Computer Vision lab

    Friday @ 9:00 - 11:00 a.m.

    Request an appointment with an Apple engineer for guidance and conversation about machine learning in your app including Core ML, Create ML, Vision, Natural Language, Speech, Sound Analysis, as well as Core ML Tools for model conversion and compression.

    Request

Activities

  • Q&A: Natural Language and Speech

    Friday @ 1:00 - 2:00 p.m.

    Ask Apple engineers about the Natural Language framework during this 1 hour text-based Q&A. Stop in to request guidance, ask for clarifications, or learn from others.

    Sign up

  • Q&A: Machine learning open forum

    Friday @ 2:00 - 3:00 p.m.

    Join us for an open forum to discuss anything related to machine learning in your app during this 1 hour text-based Q&A. Topics include Core ML, Create ML, Vision, Natural Language, Speech, Sound Analysis, TensorFlow and PyTorch model conversion, model compression, and more. Stop in to request guidance, ask for clarifications, or learn from others.

    Sign up