Search results for

“Popping Sound”

20,024 results found

Post

Replies

Boosts

Views

Activity

Reply to Building macOS apps with Xcode 26 on macOS 26 VM
@RickMaddy I think that might be a little out of scope for this thread. But in any case, I was able to update my macOS 26.1b2 VM to 26.1b3 via my logged-in Apple Account, so I don't think it's universally broken. Although my host is still on Sequoia 15.7.1. The closest thing I've experienced to what you're describing is an issue that I've seen for a while where logging in with an Apple Account hasn't been working for me if I have physical security keys enabled on my account for 2FA (FB19614234). Unfortunately if you're in this situation I don't know of any workaround other than removing all the security keys (and even if you re-enable them after logging in with the VM, eventually the VM will start complaining with pop-ups asking you to login again, although beta updates will still work despite this). This isn't new to 26.1, though. If this isn't the issue you're experiencing you might find better luck creating a new thread since I don't think many people with similar issues are going to find this rep
Oct ’25
Reply to Matter Media Playback Cluster
Kevin and ssmith_c for your assistance. I fortunately managed to file in the bug: FB20677188 Thank you! Looking over the bug, there was one point I wanted to comment on: For the Matter Media Device Basic Video Player Type Apple Home does support the commissioning (adding the device to the list of accessories displayed in Home App), but no controls are displayed. Not even the mandatory required controls: ... This type of Matter Device, which is part of the standard, should at least expose the mandatory controls and commands. The idea that Apple (and general ecosystem vendors more broadly) should try and support the full range of the Matter specification is, IMHO, one of those ideas which sounds reasonable on the surface but is, in fact, a very bad idea. First off, quoting myself: The Matter specification itself is sufficiently broad and complex that, in practice, I'm not sure it will ever really be possible to create a high-quality controller app* that actually supports EVERY possible accessory config
Oct ’25
iOS26 UINavigationController swipe from middle on maps causing issue
I have a UINavigation controller which on the second viewController has a Google Map view. In iOS 18 you would navigate normally around the map, and swipe your finger from the left side of the screen to move the map, this worked correctly. However, in iOS 26 swiping from left to right from anywhere but the far right side of the screen will cause the UI to try to pop the viewController. This does not happen when I add Apple Maps or other map frameworks to the VC. Is there a way to disable the swipe from middle in iOS 26?
Topic: UI Frameworks SubTopic: UIKit Tags:
1
0
178
Oct ’25
Reply to XCode 18.6 not selectable
I've recently updated Xcode and it says I have to download ios 26 platform It sounds like you have Xcode 26 — to do development for iOS, you do need to download those additional components, as they are used as part of the compilation of your app, even if you're focused on debugging your app on an iOS 18 device. — Ed Ford,  DTS Engineer
Topic: Community SubTopic: Apple Developers Tags:
Oct ’25
Reply to iPAD: achieve UserClient's privileges in background to have robust IOServiceAddMatchingNotification(...) as if ~[Background modes][Enable external communications].
First off, I need to repeat what I asked here: Why? What are you trying to do? More specifically, why does your app need to be awake in the background? The system has multiple APIs for managing background work and it's entirely possible there IS an API that would meet your needs. Notably, we just introduced BGContinuedProcessingTask which will allow your app to stay awake in the background for an extended period of time (several minutes). If you're trying to complete work on the user's behalf, that API is the solution. In my tests 2 weeks ago, I (might be by my mistake) saw a correspondence between external-accessoryand the facts of callbacks from IOServiceAddMatchingNotification. Were you testing using Xcode? The issue here is that in order to debug your app, Xcode disables normal app suspension, which ends up keeping your app awake in the background forever. That then makes all sorts of things work with normally wouldn't. Having said that, the word correlation did raise a possibility I hadn't really conside
Topic: App & System Services SubTopic: Drivers Tags:
Oct ’25
Reply to When debugging with VisionPro in Xcode, the program will block and it will take a long time to execute
Hi @iOS-LI, Please file a feedback at https://feedbackassitant.apple.com and post the number back here. You should expect the first time you connect to any device to take a bit of extra time to load symbols but that should only happen once. From your description it sounds like it happens quite often. A feedback will help us track down and prioritize your findings. Thank you
Oct ’25
AlarmKit alarm UI strings are lost after updating OS
I have a working AlarmKit app, but I've noticed that after any iOS update (e.g. the 26.0.1 update from a few days ago), my scheduled alarms seem to lose their UI strings, so instead of the Stop button saying Stop, it says alarm_ui_stop_button (which is my localization key for the button text). If I delete the alarm and re-add it, then it works again... until the next software update. It seems like OS updates are interfering with the link between scheduled alarms and app localization strings, which I believe are dynamically looked up at alarm time (not at configuration time). I am settings the strings in the standard way like this: AlarmManager.AlarmConfiguration( schedule: .fixed(date), attributes: AlarmAttributes( presentation: AlarmPresentation( alert: AlarmPresentation.Alert( title: alarm_ui_title, stopButton: .init(text: alarm_ui_stop_button, textColor: .yellow, systemImageName: xmark) ) ), metadata: SDAlarmMetadata(title: title, subtitle: subtitle), tintColor: .yellow ), sound: .default ) Has an
3
0
194
Oct ’25
How can third-party iOS apps obtain real-time waveform / spectrogram data for Apple Music tracks (similar to djay & other DJ apps)?
Hi everyone, I’m working on an iOS MusicKit app that overlays a metronome on top of Apple Music playback. To line the clicks up perfectly I’d like access to low-level audio analysis data—ideally a waveform / spectrogram or beat grid—while the track is playing. I’ve noticed that several approved DJ apps (e.g. djay, Serato, rekordbox) can already: • Display detailed scrolling waveforms of Apple Music songs • Scratch, loop or time-stretch those tracks in real time That implies they receive decoded PCM frames or at least high-resolution analysis data from Apple Music under a special entitlement. My questions: 1. Does MusicKit (or any public framework) expose real-time audio buffers, FFT bins, or beat markers for streaming Apple Music content? 2. If not, is there an Apple program or entitlement that developers can apply for—similar to the “DJ with Apple Music” initiative—to gain that deeper access? 3. Where can I find official documentation or a point of contact for this kind of request? I’ve sea
2
0
469
Oct ’25
How can third-party iOS apps obtain real-time waveform / spectrogram data for Apple Music tracks (similar to djay & other DJ apps)?
Hi everyone, I’m working on an iOS MusicKit app that overlays a metronome on top of Apple Music playback. To line the clicks up perfectly I’d like access to low-level audio analysis data—ideally a waveform / spectrogram or beat grid—while the track is playing. I’ve noticed that several approved DJ apps (e.g. djay, Serato, rekordbox) can already: • Display detailed scrolling waveforms of Apple Music songs • Scratch, loop or time-stretch those tracks in real time That implies they receive decoded PCM frames or at least high-resolution analysis data from Apple Music under a special entitlement. My questions: 1. Does MusicKit (or any public framework) expose real-time audio buffers, FFT bins, or beat markers for streaming Apple Music content? 2. If not, is there an Apple program or entitlement that developers can apply for—similar to the “DJ with Apple Music” initiative—to gain that deeper access? 3. Where can I find official documentation or a point of contact for this kind of request? I’ve sea
1
0
235
Oct ’25
APNs
{ aps: { content-available: 1 }, audio_file_name: ding.caf, audio_url: https://example.com/audio.mp3 } When the app is in the background or killed, it receives a remote APNs push. The data format is roughly as shown above. How can I play the MP3 audio file at the specified audio_url? The user does not need to interact with the device when receiving the APNs. How can I play the audio file immediately after receiving it?
1
0
238
Oct ’25
Reply to APNs or UNNotificationServiceExtension
{   aps : { alert : You got your emails., badge : 1, sound : default }, rc:{ cType:PR, fId:2121, oName:RC:TxtMsg, tId:3232, rId:3243, id:5FSClm2gQ9V9BZ-kUZn58B, rc-dlt-identifier:2FSClm2gQ }, appData:https://chushuo.com/sdhkguawggcguauw.mp3 } When the app is not running, as long as it receives APNS remote push. How to play the voice of 'appData'. Users do not require any additional actions
Oct ’25
Problems recording audio on Tahoe 26.0 (Intel only)
I have some tried-and-tested code that records and plays back audio via AUHAL which breaks on Tahoe on Intel. The same code works fine on Sequioa and also works on Tahoe on Apple Silicon. To start with something simple, the following code to request access to the Microphone doesn't work as it should: bool RequestMicrophoneAccess () { __block AVAuthorizationStatus status = [AVCaptureDevice authorizationStatusForMediaType: AVMediaTypeAudio]; if (status == AVAuthorizationStatusAuthorized) return true; __block bool done = false; [AVCaptureDevice requestAccessForMediaType: AVMediaTypeAudio completionHandler: ^ (BOOL granted) { status = (granted) ? AVAuthorizationStatusAuthorized : AVAuthorizationStatusDenied; done = true; }]; while (!done) CFRunLoopRunInMode (kCFRunLoopDefaultMode, 2.0, true); return status == AVAuthorizationStatusAuthorized; } On Tahoe on Intel, the code runs to completion but granted is always returned as NO. Tellingly, the popup to ask the user to grant microphone access is never displ
2
0
455
Oct ’25
Reply to Building macOS apps with Xcode 26 on macOS 26 VM
@RickMaddy I think that might be a little out of scope for this thread. But in any case, I was able to update my macOS 26.1b2 VM to 26.1b3 via my logged-in Apple Account, so I don't think it's universally broken. Although my host is still on Sequoia 15.7.1. The closest thing I've experienced to what you're describing is an issue that I've seen for a while where logging in with an Apple Account hasn't been working for me if I have physical security keys enabled on my account for 2FA (FB19614234). Unfortunately if you're in this situation I don't know of any workaround other than removing all the security keys (and even if you re-enable them after logging in with the VM, eventually the VM will start complaining with pop-ups asking you to login again, although beta updates will still work despite this). This isn't new to 26.1, though. If this isn't the issue you're experiencing you might find better luck creating a new thread since I don't think many people with similar issues are going to find this rep
Replies
Boosts
Views
Activity
Oct ’25
Reply to Matter Media Playback Cluster
Kevin and ssmith_c for your assistance. I fortunately managed to file in the bug: FB20677188 Thank you! Looking over the bug, there was one point I wanted to comment on: For the Matter Media Device Basic Video Player Type Apple Home does support the commissioning (adding the device to the list of accessories displayed in Home App), but no controls are displayed. Not even the mandatory required controls: ... This type of Matter Device, which is part of the standard, should at least expose the mandatory controls and commands. The idea that Apple (and general ecosystem vendors more broadly) should try and support the full range of the Matter specification is, IMHO, one of those ideas which sounds reasonable on the surface but is, in fact, a very bad idea. First off, quoting myself: The Matter specification itself is sufficiently broad and complex that, in practice, I'm not sure it will ever really be possible to create a high-quality controller app* that actually supports EVERY possible accessory config
Replies
Boosts
Views
Activity
Oct ’25
iOS26 UINavigationController swipe from middle on maps causing issue
I have a UINavigation controller which on the second viewController has a Google Map view. In iOS 18 you would navigate normally around the map, and swipe your finger from the left side of the screen to move the map, this worked correctly. However, in iOS 26 swiping from left to right from anywhere but the far right side of the screen will cause the UI to try to pop the viewController. This does not happen when I add Apple Maps or other map frameworks to the VC. Is there a way to disable the swipe from middle in iOS 26?
Topic: UI Frameworks SubTopic: UIKit Tags:
Replies
1
Boosts
0
Views
178
Activity
Oct ’25
Reply to XCode 18.6 not selectable
I've recently updated Xcode and it says I have to download ios 26 platform It sounds like you have Xcode 26 — to do development for iOS, you do need to download those additional components, as they are used as part of the compilation of your app, even if you're focused on debugging your app on an iOS 18 device. — Ed Ford,  DTS Engineer
Topic: Community SubTopic: Apple Developers Tags:
Replies
Boosts
Views
Activity
Oct ’25
Reply to [CarPlay] CPNowPlayingTemplate not being accepted when being passed to pushTemplate:animated:completion
Note: I have a max of 3 templates going on: Home screen Album screen with all songs NPS now playing screen I have the following entitlement also: com.apple.developer.carplay-audio Note: The issue is not reproducible on local xCode but when i deployed to appStore it started showing this exception.
Topic: Community SubTopic: Apple Developers Tags:
Replies
Boosts
Views
Activity
Oct ’25
Reply to iPAD: achieve UserClient's privileges in background to have robust IOServiceAddMatchingNotification(...) as if ~[Background modes][Enable external communications].
First off, I need to repeat what I asked here: Why? What are you trying to do? More specifically, why does your app need to be awake in the background? The system has multiple APIs for managing background work and it's entirely possible there IS an API that would meet your needs. Notably, we just introduced BGContinuedProcessingTask which will allow your app to stay awake in the background for an extended period of time (several minutes). If you're trying to complete work on the user's behalf, that API is the solution. In my tests 2 weeks ago, I (might be by my mistake) saw a correspondence between external-accessoryand the facts of callbacks from IOServiceAddMatchingNotification. Were you testing using Xcode? The issue here is that in order to debug your app, Xcode disables normal app suspension, which ends up keeping your app awake in the background forever. That then makes all sorts of things work with normally wouldn't. Having said that, the word correlation did raise a possibility I hadn't really conside
Topic: App & System Services SubTopic: Drivers Tags:
Replies
Boosts
Views
Activity
Oct ’25
Reply to When debugging with VisionPro in Xcode, the program will block and it will take a long time to execute
Hi @iOS-LI, Please file a feedback at https://feedbackassitant.apple.com and post the number back here. You should expect the first time you connect to any device to take a bit of extra time to load symbols but that should only happen once. From your description it sounds like it happens quite often. A feedback will help us track down and prioritize your findings. Thank you
Replies
Boosts
Views
Activity
Oct ’25
AlarmKit alarm UI strings are lost after updating OS
I have a working AlarmKit app, but I've noticed that after any iOS update (e.g. the 26.0.1 update from a few days ago), my scheduled alarms seem to lose their UI strings, so instead of the Stop button saying Stop, it says alarm_ui_stop_button (which is my localization key for the button text). If I delete the alarm and re-add it, then it works again... until the next software update. It seems like OS updates are interfering with the link between scheduled alarms and app localization strings, which I believe are dynamically looked up at alarm time (not at configuration time). I am settings the strings in the standard way like this: AlarmManager.AlarmConfiguration( schedule: .fixed(date), attributes: AlarmAttributes( presentation: AlarmPresentation( alert: AlarmPresentation.Alert( title: alarm_ui_title, stopButton: .init(text: alarm_ui_stop_button, textColor: .yellow, systemImageName: xmark) ) ), metadata: SDAlarmMetadata(title: title, subtitle: subtitle), tintColor: .yellow ), sound: .default ) Has an
Replies
3
Boosts
0
Views
194
Activity
Oct ’25
How can third-party iOS apps obtain real-time waveform / spectrogram data for Apple Music tracks (similar to djay & other DJ apps)?
Hi everyone, I’m working on an iOS MusicKit app that overlays a metronome on top of Apple Music playback. To line the clicks up perfectly I’d like access to low-level audio analysis data—ideally a waveform / spectrogram or beat grid—while the track is playing. I’ve noticed that several approved DJ apps (e.g. djay, Serato, rekordbox) can already: • Display detailed scrolling waveforms of Apple Music songs • Scratch, loop or time-stretch those tracks in real time That implies they receive decoded PCM frames or at least high-resolution analysis data from Apple Music under a special entitlement. My questions: 1. Does MusicKit (or any public framework) expose real-time audio buffers, FFT bins, or beat markers for streaming Apple Music content? 2. If not, is there an Apple program or entitlement that developers can apply for—similar to the “DJ with Apple Music” initiative—to gain that deeper access? 3. Where can I find official documentation or a point of contact for this kind of request? I’ve sea
Replies
2
Boosts
0
Views
469
Activity
Oct ’25
How can third-party iOS apps obtain real-time waveform / spectrogram data for Apple Music tracks (similar to djay & other DJ apps)?
Hi everyone, I’m working on an iOS MusicKit app that overlays a metronome on top of Apple Music playback. To line the clicks up perfectly I’d like access to low-level audio analysis data—ideally a waveform / spectrogram or beat grid—while the track is playing. I’ve noticed that several approved DJ apps (e.g. djay, Serato, rekordbox) can already: • Display detailed scrolling waveforms of Apple Music songs • Scratch, loop or time-stretch those tracks in real time That implies they receive decoded PCM frames or at least high-resolution analysis data from Apple Music under a special entitlement. My questions: 1. Does MusicKit (or any public framework) expose real-time audio buffers, FFT bins, or beat markers for streaming Apple Music content? 2. If not, is there an Apple program or entitlement that developers can apply for—similar to the “DJ with Apple Music” initiative—to gain that deeper access? 3. Where can I find official documentation or a point of contact for this kind of request? I’ve sea
Replies
1
Boosts
0
Views
235
Activity
Oct ’25
Long hang in SwiftUI Dev
Why is there always such a long hang when first switching tabs or popping up the keyboard in SwiftUI Development?
Replies
2
Boosts
0
Views
148
Activity
Oct ’25
Reply to APNs
Apps cannot start audio sessions in the background. There is no supported way for you to do this.
Topic: Media Technologies SubTopic: Audio Tags:
Replies
Boosts
Views
Activity
Oct ’25
APNs
{ aps: { content-available: 1 }, audio_file_name: ding.caf, audio_url: https://example.com/audio.mp3 } When the app is in the background or killed, it receives a remote APNs push. The data format is roughly as shown above. How can I play the MP3 audio file at the specified audio_url? The user does not need to interact with the device when receiving the APNs. How can I play the audio file immediately after receiving it?
Replies
1
Boosts
0
Views
238
Activity
Oct ’25
Reply to APNs or UNNotificationServiceExtension
{   aps : { alert : You got your emails., badge : 1, sound : default }, rc:{ cType:PR, fId:2121, oName:RC:TxtMsg, tId:3232, rId:3243, id:5FSClm2gQ9V9BZ-kUZn58B, rc-dlt-identifier:2FSClm2gQ }, appData:https://chushuo.com/sdhkguawggcguauw.mp3 } When the app is not running, as long as it receives APNS remote push. How to play the voice of 'appData'. Users do not require any additional actions
Replies
Boosts
Views
Activity
Oct ’25
Problems recording audio on Tahoe 26.0 (Intel only)
I have some tried-and-tested code that records and plays back audio via AUHAL which breaks on Tahoe on Intel. The same code works fine on Sequioa and also works on Tahoe on Apple Silicon. To start with something simple, the following code to request access to the Microphone doesn't work as it should: bool RequestMicrophoneAccess () { __block AVAuthorizationStatus status = [AVCaptureDevice authorizationStatusForMediaType: AVMediaTypeAudio]; if (status == AVAuthorizationStatusAuthorized) return true; __block bool done = false; [AVCaptureDevice requestAccessForMediaType: AVMediaTypeAudio completionHandler: ^ (BOOL granted) { status = (granted) ? AVAuthorizationStatusAuthorized : AVAuthorizationStatusDenied; done = true; }]; while (!done) CFRunLoopRunInMode (kCFRunLoopDefaultMode, 2.0, true); return status == AVAuthorizationStatusAuthorized; } On Tahoe on Intel, the code runs to completion but granted is always returned as NO. Tellingly, the popup to ask the user to grant microphone access is never displ
Replies
2
Boosts
0
Views
455
Activity
Oct ’25