So experimenting with the new SpeechTranscriber, if I do: let transcriber = SpeechTranscriber( locale: locale, transcriptionOptions: [], reportingOptions: [.volatileResults], attributeOptions: [.audioTimeRange] ) only the final result has audio time ranges, not the volatile results. Is this a performance consideration? If there is no performance problem, it would be nice to have the option to also get speech time ranges for volatile responses. I'm not presenting the volatile text at all in the UI, I was just trying to keep statistics about the non-speech and the speech noise level, this way I can determine when the noise level falls under the noisefloor for a while. The goal here was to finalize the recording automatically, when the noise level indicate that the user has finished speaking.
Search results for
Popping Sound
19,349 results found
Selecting any option will automatically load the page
Post
Replies
Boosts
Views
Activity
On this page https://developer.apple.com/library/ios/documentation/IDEs/Conceptual/AppDistributionGuide/ConfiguringYourApp/ConfiguringYourApp.html#//apple_ref/doc/uid/TP40012582-CH28-SW3, titled Configuring Your Xcode Project for Distribution, in the section titled, Before You Begin Configuring Your Project, it says:To open the project navigator, choose View > Navigators > Show Project Navigator. Choose the target from the Project/Targets pop-up menu or in the Targets section of the second sidebar if it appears. Click General to view settings discussed in this chapter.I cannot get the second sidebar to appear. How do I get to the distribution settings?
Let me start by returning to what I said here: Indeed, the fact this works in iOS 26 is an accidental oversight, not an intentional choice. That is not an exaggeration. The ability to report outgoing calls from the background was a compatibility workaround we preserved in iOS 13 to support PTT apps. The ONLY reason it continues to work in iOS 26 is because I didn't think about it when we were disabling the PTT entitlement. It will not continue to work and continuing to rely on it is a mistake. User has now ended call in CallKit UI, To resume the ongoing audio call - client requires to report to callKi, which is on end call action and no early call process at this moment I'm sorry, but this isn't something CallKit will continue to support. You need to stop doing this, as it WILL break in the future. __ Kevin Elliott DTS Engineer, CoreOS/Hardware
Topic:
App & System Services
SubTopic:
General
Tags:
I started playing which transcription of audio files on macOS today, latest beta of Xcode and latest beta of Tahoe. Transcription itself works really well, but for some reason the majority of the results contain no audioTimeRange. I got 22 single-word results with time ranges, spread out all over total file of 53 minutes. Is there something I can do to improve this? To my understanding, I have followed sample code and instructions very closely, but the SwiftTranscriptionSampleApp and other examples I've seen lead me to believe I should be getting a lot more time ranges than I actually do.
Ah, nice, let's see, first baseline without prepareToAnalyze: The KPI I'm interested is the time between the last audio above the noise-ground level and the final transcript (e.g. between the user stopping to speak and the transcription being ready to trigger actions): n: 11, avg: 2.2s, Var: 0.75 Then, with calling prepareToAnalyze: n: 11, avg: 1.45s, Var: 1.305 (the delay varied greatly between 0.05s and 3s) So yeah, based on this small sample, preparing did seem to decrease the delay.
Topic:
Media Technologies
SubTopic:
Audio
Tags:
So, I've been wondering how fast a an offline STT -> ML Prompt -> TTS roundtrip would be. Interestingly, for many tests, the SpeechTranscriber (STT) takes the bulk of the time, compared to generating a FoundationModel response and creating the Audio using TTS. E.g. InteractionStatistics: - listeningStarted: 21:24:23 4480 2423 - timeTillFirstAboveNoiseFloor: 01.794 - timeTillLastNoiseAboveFloor: 02.383 - timeTillFirstSpeechDetected: 02.399 - timeTillTranscriptFinalized: 04.510 - timeTillFirstMLModelResponse: 04.938 - timeTillMLModelResponse: 05.379 - timeTillTTSStarted: 04.962 - timeTillTTSFinished: 11.016 - speechLength: 06.054 - timeToResponse: 02.578 - transcript: This is a test. - mlModelResponse: Sure! I'm ready to help with your test. What do you need help with? Here, between my audio input ending and the Text-2-Speech starting top play (using AVSpeechUtterance) the total response time was 2.5s. Of that time, it took the SpeechAnalyzer 2.1s to get the transcript finalized, Foundat
[quote='852438022, DTS Engineer, /thread/793663?answerId=852438022#852438022'] That statement was not a vague warning. We are ACTIVELY shutting down the PTT workarounds we created in iOS 13. That includes starting outgoing calls from the background. Indeed, the fact this works in iOS 26 is an accidental oversight, not an intentional choice. [/quote] Client is already migrated to PTC framework. we have a use case which actually needs to report Audio and video calls to CallKit (in this case BLE permission is enabled.) As mentioned earlier: Reported a incoming Video call using CallKit User has now initiated a Audio call (Full duplex call) which is using the active callKit session. User has now ended call in CallKit UI, To resume the ongoing audio call - client requires to report to callKi, which is on end call action and no early call process at this moment
Topic:
App & System Services
SubTopic:
General
Tags:
I've already searched extensively on Apple Developer Forums and Stack Overflow, and didn't really find what I need or I missed it. I'm developing a macOS music player app that uses cuesheet files paired with audio files. The core functionality is working, opening and playing files works without issues, but I'm struggling with implementing proper file handling features due to my limited experience with UI frameworks. The current state of my app is intentionally simple: Single window interface representing a music player with track list Opening cuesheet files changes the “disc” and updates the window Built with SwiftUI (not AppKit) Not created as a Document-Based app since the user doesn't need to edit, save, or work with multiple documents simultaneously What I Need to Implement: Open Recent menu that actually works Recent files accessible from Dock menu Opening cuesheet files from Finder Drag-and-drop cuesheet files onto app window (lower priority) Problems I've Encountered: I've tried multiple appro
Hello @brother_z , thank you for your question! If you are seeing a virtual object drift from its original position, that sounds like unexpected behavior and I would recommend submitting a bug report via Feedback Assistant. However there are many things that could be preventing your device from tracking its position correctly, such as obstructed cameras or high velocity motion (like wearing Apple Vision Pro on a train), so it's hard to diagnose without more details about what you're trying to do. Access to the main camera requires an entitlement. The extrinsic value of the camera will be a 4x4 matrix representing its pose relative to the device. The math you've shared here looks correct, although I'm not sure what your tag object is? Are the tag objects you are referring to Entities you've created? If you do file a feedback request, I recommend sharing as much of your project in the request as you are able to, and then share the number here so we can track it on our end. Thank you!
Topic:
Spatial Computing
SubTopic:
ARKit
Tags:
Has iOS 18 introduced new permission requirements or entitlements for VoIP push notifications? No. The last significant change here was the iOS 13 CallKit requirements. There hasn't been any change to them since then. Do I need to explicitly request a new type of user permission for VoIP notifications? No. Are there additional background modes, Info.plist keys, or PushKit changes required for VoIP to work in background and terminated states on iOS 18? No. Looking over your list, I did notice this: . Background modes for Voice over IP and Background Processing are enabled. Did you also include audio/Audio, AirPlay, and Picture in Picture? Historically, the architecture of VoIP apps has always relied on two different background categories in order to function: voip” -> Allows apps to use PushKit for call notifications and CallKit for call management. audio” -> Keeps the app awake in the background while the app is actually on a call (the same way it would keep any long-playing audio
Topic:
App & System Services
SubTopic:
Notifications
Tags:
Thank you for the detailed reply. I've submitted a bug report as requested: FB19421676 – Perfect, thank you. It turns out that isVoiceProcessingInputMuted was set to true when starting a transmission, and only reverted to false once audio output stopped. This was the source of the delay between initiating transmission and receiving valid microphone input. Good find! I'm still relatively new to Swift and iOS audio development, and I was wondering if there are any sample projects or best practices that demonstrate integrating audio with the Push-to-Talk framework. No, there isn't any direct sample for it. Practically speaking, the PushToTalk framework was actually created to support an existing set of developers who'd previously built PTT apps using the voip background category and CallKit, so that they could migrate away from the unrestricted PTT entitlement. That's why we didn't create a sample— most of the framework's adopters were integrating the sample into an existing large-scal
Topic:
Media Technologies
SubTopic:
Audio
Tags:
Thanks. So it sounds like the NSImage is getting over-released or being incorrectly deallocated, although it's unclear to me what exactly is trying to form a weak reference, since my code (as shown above) has a strong reference. Unfortunately, the higher up function calls are made by AppKit, which is not open source, so I cannot look it up. Seeing that code wouldn't actually help. The thing that makes over-release issues (and most other memory crashes ) hard to debug is that the crash log you’re looking at is NOT why your app crashed. At some earlier point in your app, something happened in your app that caused an extra release that would not normally occur. In other words, what the crash log shows is a victim of an underlying issue, not its direct cause. As far as I understand, the link you posted helps investigating memory issues in Xcode, but since the crash reports are downloaded by Xcode from other users and I cannot reproduce it myself... I should have posted more direct links, but most of the
Topic:
UI Frameworks
SubTopic:
AppKit
Tags:
Thanks. So it sounds like the NSImage is getting over-released or being incorrectly deallocated, although it's unclear to me what exactly is trying to form a weak reference, since my code (as shown above) has a strong reference. Unfortunately the higher up function calls are made by AppKit, which is not open source, so I cannot look it up. As far as I understand, the link you posted helps investigating memory issues in Xcode, but since the crash reports are downloaded by Xcode from other users and I cannot reproduce it myself... The Xcode statistics seem to show that it only happens with macOS 15.3 or newer. I don't know if it's because there's not enough space to show older releases, or if it's really a clue that it's a change introduced with macOS 15.3 that causes this issue. From my perspective it would make sense that it's a new issue with macOS 15.3, because I haven't changed the code that generates or assigns that image in a very long time, and this issue didn't happen for a previous version of
Topic:
UI Frameworks
SubTopic:
AppKit
Tags:
I'm struggling to convert Swift 5 to Swift 6. As advised in doc, I first turned strict concurrency ON. I got no error. Then, selected swift6… and problems pop up. I have a UIViewController with IBOutlets: eg a TextField. computed var eg duree func using UNNotification: func userNotificationCenter I get the following error in the declaration line of the func userNotificationCenter: Main actor-isolated instance method 'userNotificationCenter(_:didReceive:withCompletionHandler:)' cannot be used to satisfy nonisolated requirement from protocol 'UNUserNotificationCenterDelegate' So, I declared the func as non isolated. This func calls another func func2, which I had also to declare non isolated. Then I get error on the computed var used in func2 Main actor-isolated property 'duree' can not be referenced from a nonisolated context So I declared duree as nonsilated(unsafe). Now comes the tricky part. The computed var references the IBOutlet dureeField if dureeField.text == X leading to the error Main actor-
I’m currently developing a spam number blocking app using CallKit. I’ve confirmed that up to iOS 26 beta 5, there is a bug where number blocking doesn’t work. In my current tests, the ringtone doesn’t sound and the blocking works fine, but the call still appears in the missed calls list, which is bothersome. If the bug is fixed in future versions (as it was in previous versions), is there a way to block the number so that it also does not appear in missed calls?