Apple Intelligence

RSS for tag

Apple Intelligence is the personal intelligence system that puts powerful generative models right at the core of your iPhone, iPad, and Mac and powers incredible new features to help users communicate, work, and express themselves.

Posts under Apple Intelligence subtopic

Post

Replies

Boosts

Views

Activity

"failed to processImage" in videoProcessor
Hello, I’m working on a program that analyzes video files frame by frame to detect human poses in each frame. However, during the process of reading observations from the stream, the analysis frequently stops with the following error: [LOG_ERROR] /Library/Caches/com.apple.xbs/Sources/MediaAnalysis/VideoProcessing/VCPHumanPoseImageRequest.mm[85]: code -18 [LOG_ERROR] /Library/Caches/com.apple.xbs/Sources/MediaAnalysis/VideoProcessing/VCPHumanPoseImageRequest.mm[178]: code -18 The error was caught and printed using a do-catch block, and here is the output: Error Domain=NSOSStatusErrorDomain Code=-18 "Error: failed to processImage" UserInfo={NSLocalizedDescription=Error: failed to processImage} While the do-catch block helps prevent the app from crashing, the frames following the error cannot be analyzed. I’m hoping to understand the cause of this error, or find a way to skip the problematic frames and continue analyzing the subsequent ones. My development environment is Xcode Version 16.0 (16A242d) and iOS 18.0. Thank you for your help. (Attaching my code below.) let videoProcessor = VideoProcessor(videoURL) let bodyPoseRequest = DetectHumanBodyPoseRequest() let asset = AVURLAsset(url: videoURL) let videoTrack = try await asset.loadTracks(withMediaType: .video).first let bodyPoseStream = try await videoProcessor.addRequest(bodyPoseRequest) videoProcessor.startAnalysis() do { for try await observations in bodyPoseStream { guard let observation = observations.first else { continue } if let timeRange = observation.timeRange { /// do something... } } } catch { print("\(error.localizedDescription)") }
0
1
385
Oct ’24
Seeking Feedback on an Idea: Real-Time Siri Running Coach for iOS
Hello everyone, I hope you’re all doing well. I’m not a developer, but I have an idea for an iOS app that I’d love to get your thoughts on. I wanted to share it here to gather feedback from this knowledgeable community and to learn from your expertise. Idea Overview: Real-Time AI Running Coach for iOS The concept is an iOS application that provides personalized, real-time running coaching by leveraging on-device data sources and Apple’s latest technologies. The app aims to offer an adaptive and motivating running experience while ensuring user privacy through on-device processing. Key Features: • Personalized Coaching: • Utilize real-time biometric data and personal insights to deliver AI-driven coaching tailored to the user’s mental and physical state. • Analyze health metrics, activity data, mood check-ins, and more to provide context-based motivational feedback. • Privacy First: • All data processing occurs on-device using Apple’s frameworks like Core ML, ensuring no personal data leaves the device. • Adaptive Motivation: • Implement Natural Language Processing to analyze user inputs like journal entries or mood check-ins. • Generate personalized coaching cues based on historical performance and mood trends. • Performance Enhancement: • Offer dynamic adjustments to pace, route, and strategy in real time to help improve running performance. • Seamless integration with Apple Watch for real-time data collection and haptic feedback. Technologies and Frameworks Involved: • HealthKit: Access health metrics such as heart rate, distance run, VO₂ max, sleep patterns, etc. • Core ML: On-device machine learning for real-time data analysis without latency. • Natural Language Processing: Analyze personal inputs for better coaching personalization. • Core Motion & Core Location: Track motion data and location services for runs. • AVFoundation & Speech: Provide real-time voice feedback and coaching cues. • SiriKit Integration: Allow users to initiate workouts and receive updates via Siri. Target Audience: • Runners of all levels seeking personalized coaching that adapts to their mental and physical states. • Users who prioritize privacy and want AI-driven insights without their data leaving the device. • Tech-savvy fitness enthusiasts who use iOS devices and Apple wearables. Questions for the Community: 1. Feasibility: Is this idea technically achievable using current iOS frameworks and technologies? 2. Data Access: Are there limitations in accessing and processing the necessary data on-device, especially regarding privacy and permissions? 3. Potential Challenges: What hurdles might developers face in creating such an app, and how could they be addressed? 4. Advice: As someone without a technical background, what steps would you recommend I take to move this idea forward? I truly appreciate any feedback or insights you can provide. I’m excited about the potential of this idea but also aware there may be complexities I’m not considering. Thank you for taking the time to read this! Best regards, Paul
0
0
465
Oct ’24
Tapping Take Photo reloads Image Playground sheet on iOS
I integrated the image playground sheet in my app, however when I select Take Photo on the iOS version of my app it just reloads the sheet. After several attempts I get the below error message. This issue doesn’t occur on the macOS version of my app, where it first requests camera permission before allowing me to take the photo. I’m not sure if this is happening because I don’t request the camera permission anywhere in my app. My app doesn’t use the camera at all apart from the Take Photo feature which is part of the image playground sheet. Feedback ID: FB15591786
0
0
487
Oct ’24
From Apple Dev.
Early access to Image Playground, Genmoji, and Image Wand  Apple Oct 25, 2024 at 5:58 PM With the iOS & iPadOS 18.2 and macOS Sequoia 15.2 betas, you can join the waitlist for early access to Image Playground, Genmoji, and Image Wand in order to test and help improve these features. You can request access within any one of these experiences: Image Playground app Image Playground integration in Messages or Freeform Genmoji integration in the emoji keyboard, or Image Wand within the Apple Pencil tool palette in Notes We will roll out access to Image Playground, Genmoji, and Image Wand over the coming weeks. When the features are ready for you to test, you will be notified. After you receive access, you can tap the thumbs up or thumbs down that appear with each result in Image Playground, Genmoji, and Image Wand in order to provide feedback.
0
0
392
Oct ’24
Developing apps with local LLM Foundation Models
I just watched the October 30 MacBook Pro Announcement where they talked about on-device local LLMs for the M4s. What developer training resources are available, where we can learn how to use custom llm models and build our Swift apps to use both Apple Intelligence and other llm models on device? Is the guidance to follow MLX github repos, or were those experimental and now there is an approved workflow and tooling? https://www.youtube.com/watch?v=G0cmfY7qdmY
0
2
1.2k
Oct ’24
Apple intelligence
So recently i updated from 18.1 to 18.2 beta on 15 pro max. I lost my access to apple intelligence. I was excited to see the use the image playground in 18.2 update but it need apple intelligence to use that app but it went back to older siri version. what are tye solutions now because i do not have any access to apple intelligence even though i have ios 18.2?
0
0
336
Nov ’24
Apple intelligence
I recently updated my iPhone to iOS 14.2 and the playground app came and then I couldn't open it because I hadn't released it yet. Today it released it and I rediled the notification but I went to look for it to test it and I couldn't find it anymore. How do I make the PLAYGROUND app come back from Apple smart. Como faço para baixar novamente o app?
0
0
341
Nov ’24
Trouble getting the waitlist to work
I recently upgraded my Ipad(M4) pro to ios18.1 and the apple intelligence waitlist option popped up. I‘ve been trying to click join waitlist but nothing happens and the apple intellignece pop up menu just dissapears each time I clicked join waitlist. I made sure my region is in the States.
0
0
272
Nov ’24
Apple Intelligence & iMac M1
Trying to get on the waitlist for the above and the computer is saying: “Apple Intelligence is not available when Mac is set to English (Singapore)”. When just a few more bullet points below my Language selection shows “English (United States)”. That’s the only thing I can see, of course you guys are the experts. I would like to be part of this AI experiment/experience. Thanks for any help you can give to this 35+ year Mac user. Lee W
0
0
369
Nov ’24
Apple Intelligence stuck on "Pending" since 1st availability of MacOS 15.2 Beta 1
Once available, I immediately installed the MacOS 15.2 beta and configured MacOS and Siri enable Apple Intelligence. I joined the waiting list, and soon after the downloading process started. I have since then been stuck in Pending. I have recently (yesterday) installed 15.2 Beta 2 (Public) - and there is no difference. I have restarted multiple times; I have left my mac on - connected to WiFi - over night multiple times; I have changed language - no only of Siri but also on my Mac (requiring a restart) - multiple time. I am frustrated that I cannot see what is causing the Pending status- pending on what? I am frustrated that I cannot just start Apple Intelligence enrolment from scratch - no restart button. Any help and advice would be greatly welcome.
0
0
425
Nov ’24