Search results for

“Popping Sound”

20,039 results found

Post

Replies

Boosts

Views

Activity

Reply to After unholding CallKit, the audio does not restore.
@DTS Engineer But AVAudioSessionInterruptionTypeEnded does not get posted when the incoming cellular call is terminated from the Caller side. I think AVAudioSessionInterruptionTypeEnded only gets posted when there is an incoming call and user choose to not to receive the call. Even in Speakerbox, same issue is happening. Please refer https://developer.apple.com/forums/thread/762059 . And also why should the app need to activate the audio session separately. I think it should be done by iOS Callkit itself once we perform CXSetHeldAction with false right?
Topic: App & System Services SubTopic: General Tags:
Jan ’26
How to properly localize AppIntent dialogs for Siri?
Hi! I have defined the following app intent. It returns a result with a dialog to confirm that the intent has been executed. Naturally, that dialog needs to be localized properly. But the String interpolation with the provided format doesn't do that. I specified wide for the width parameter and expect spelled-out unit names. However, in the textual output, Siri always uses the abbreviated unit (e.g. min or s), in all languages I tested. In the audio output, Siri says minutes in English where the textual representation is min. In German, Siri says min, so it basically reads the textual representation aloud and that's not quite understandable to the user. struct StartTimerIntent: AppIntent { static let title: LocalizedStringResource = Start New Timer static var description = IntentDescription(Starts a timer with a custom duration.) @Parameter(title: Duration, description: The duration of the timer.) var duration: Measurement func perform() async throws -> some IntentResult & ProvidesDialog { //
3
0
553
Jan ’26
Reply to SwiftData with CloudKit Sync Issue
This might be late but I'm currently digging into a bug that popped up out of nowhere. I've had flawless syncing for 5 months, then a new user says his data is not syncing between iPad and iPhone, and now I'm experiencing the same on my devices. The debug builds worked fine, the production builds are intermittently not syncing and I have bad requests in my CloudKit log.
Jan ’26
Building a Full Space app that enables sharing a visionOS experience with nearby users.
Hello, I am currently considering developing a Full Space app that enables a shared visionOS experience with nearby users. Intended Features A Mixed Full Space app in which dozens of 3D models are placed in the space. These 3D models may play embedded animations when tapped, be programmatically moved or rotated, or be controlled via Reality Composer Pro timelines. The app also includes audio, spatial audio, videos with audio, and videos without audio, which are rendered as VideoTextures on planes and played back in the space. Some media elements play automatically, while others are triggered by user interaction. However, it is unclear whether AVPlaybackCoordinator supports shared playback across multiple types of media, such as: audio only spatial audio video without audio video with audio I am also unsure whether there are alternative or recommended approaches for synchronizing playback in this scenario. Questions Is it technically possi
1
0
370
Dec ’25
SoundAnalysis built-in classifier fails in background (SNErrorCode.operationFailed)
I’m seeing consistent failures using SoundAnalysis live classification when my app moves to the background. Setup iOS 17.x AVAudioEngine mic capture SNAudioStreamAnalyzer SNClassifySoundRequest(classifierIdentifier: .version1) UIBackgroundModes = audio AVAudioSession .record / .playAndRecord, active Audio capture + level metering continue working in background (mic indicator stays on) Issue As soon as the app enters background / screen locks: SoundAnalysis starts failing every second with domain:com.apple.SoundAnalysis, code:2(SNErrorCode.operationFailed) Audio capture itself continues normally When the app returns to foreground, classification immediately resumes without restarting the engine/analyzer Question Is live background sound classification with the built-in SoundAnalysis classifier officially unsupported or known to fail in background? If so, is a custom Core ML model the only supported approach for background detection? Or is there a required configuration I’m m
0
0
221
Dec ’25
Nasty problems in Xcode 26.2: Apple Intelligence crash & Source Code Control Failure
I know this post isn't going to give a lot of details, but what I experienced tonight was so completely weird that I wanted to get it posted here in case others run into it: FIRST: All was well until I made a trivial change to a large Objective-C++ module. I suddenly got the idea to look at that line in the code review pane, to see if that area of code had ever had recent modifications. But, the entire module showed up as modified -- one giant change bar, with nothing on the right side of the code review pane, no matter what commit I selected. Then I noticed that the two lines of code which had all of 4 characters edited were no longer showing any change bars. Yet, the file showed up as modified. Still, the exact line changes were not showing in the source code navigator, even though other files showed their changes. Note I'm connected to our remote repo on github. I did some command line git checks of the local repo, and the changes were there (as yet unstaged). So -- I figured, I'm gonna ask the Apple Codin
4
0
594
Dec ’25
AVAudioEngine obtains channel audio data
Currently, I have successfully used ChannelMap to map hardware input channels and obtained audio data from the hardware device's MIC and OTG inputs. Additionally, I have used ChannelMap to map output channels to freely feed data for playback to each output channel. However, I now have a problem. I have a hardware device that only has output channels (no input channels), and the system has set this hardware device as the default playback device. In this case, how can I obtain the audio data being played to the output channels for modification?
0
0
292
Dec ’25
Reply to Home App Intermittent Hub Not Responding Bug
Decided to start from scratch to see if it would resolve the issue. I removed all of my 100+ devices, deleted scenes/automation, and deleted the Home. I rebooted all devices (including network) and created a new home. I started off with adding Hue device and when I later launched the Home app I got the Home hub is unresponsive pop up. I did set everything back up again and will continue to monitor but so far the issue unfortunately still persists.
Topic: App & System Services SubTopic: General Tags:
Dec ’25
Reply to SwiftUI @State Updates Not Reflecting in UI Until View Reconstruction (Xcode Preview & Device)
Thanks for your reply! Here's the complete code and detailed information: Issue Description I'm experiencing a strange behavior with Xcode Preview in my SwiftUI app: On first Preview load: Button taps don't respond at all Temporary workaround: Click any other file, then click back to ContentView - buttons work perfectly Reproducibility: Happens 100% of the time after cleaning and restarting Preview Environment Xcode Version: Version 26.2 iOS Target: iOS 26.1 macOS Version: 26.1 Reproducibility: 100% consistent Relevant Code App Entry Point import SwiftUI import AVFoundation @main struct MyApp: App { init() { setupAudioSession() } private func setupAudioSession() { do { try AVAudioSession.sharedInstance().setCategory( .playback, mode: .default, options: [.mixWithOthers] ) try AVAudioSession.sharedInstance().setActive(true) } catch { print(Failed to setup audio session: (error)) } } var body: some Scene { WindowGroup { ContentView() } } }### ContentView - Simplified Version import SwiftUI struct Conten
Topic: UI Frameworks SubTopic: SwiftUI
Dec ’25
Reply to How do I provide developer documentation (4.1 - Copycats)
The answer to this question does not answer the question at all. I've setup an appeal three times. In all three instances, they just reply with the same 4.1 issue. Mind you, my app has already been APPROVED 7 times prior and now on the 8th time, all of a sudden, this 4.1 has popped up. I need answers to my questions above as my users are getting angry that my extension fixes are available on Chrome and Firefox, but no longer available on Apple (Safari) **1. Developer wants to know what needs to be provided? An email, A statement? How / what format would be required they are asking? How does such above documentation get submitted / included in subsequent updates to Not hinder approval?**
Dec ’25
SwiftUI @State Updates Not Reflecting in UI Until View Reconstruction (Xcode Preview & Device)
Issue Description I'm experiencing a bizarre SwiftUI state update issue that only occurs in Xcode development environment (both Canvas preview and device debugging), but does not occur in production builds downloaded from App Store. Symptom: User taps a button that modifies a @State variable inside a .sheet Console logs confirm the state HAS changed But the UI does not update to reflect the new state Switching to another file in Xcode and back to ContentView instantly fixes the issue The production build (same code) works perfectly fine Environment Xcode: 16F6 (17C52) iOS: 26.2 (testing on iPhone 13) macOS: 25.1.0 (Sequoia) SwiftUI Target: iOS 15.6+ Issue: Present in both Xcode Canvas and on-device debugging Production: Same code works correctly in App Store build (version 1.3.2) Code Structure Parent View (ContentView.swift) struct ContentView: View { @State private var selectedSound: SoundTheme = .none @State private var showSoundSheet = false var body: some View { VStack { // Display button shows current s
Topic: UI Frameworks SubTopic: SwiftUI
6
0
359
Dec ’25
Reply to After unholding CallKit, the audio does not restore.
@DTS Engineer But AVAudioSessionInterruptionTypeEnded does not get posted when the incoming cellular call is terminated from the Caller side. I think AVAudioSessionInterruptionTypeEnded only gets posted when there is an incoming call and user choose to not to receive the call. Even in Speakerbox, same issue is happening. Please refer https://developer.apple.com/forums/thread/762059 . And also why should the app need to activate the audio session separately. I think it should be done by iOS Callkit itself once we perform CXSetHeldAction with false right?
Topic: App & System Services SubTopic: General Tags:
Replies
Boosts
Views
Activity
Jan ’26
How to properly localize AppIntent dialogs for Siri?
Hi! I have defined the following app intent. It returns a result with a dialog to confirm that the intent has been executed. Naturally, that dialog needs to be localized properly. But the String interpolation with the provided format doesn't do that. I specified wide for the width parameter and expect spelled-out unit names. However, in the textual output, Siri always uses the abbreviated unit (e.g. min or s), in all languages I tested. In the audio output, Siri says minutes in English where the textual representation is min. In German, Siri says min, so it basically reads the textual representation aloud and that's not quite understandable to the user. struct StartTimerIntent: AppIntent { static let title: LocalizedStringResource = Start New Timer static var description = IntentDescription(Starts a timer with a custom duration.) @Parameter(title: Duration, description: The duration of the timer.) var duration: Measurement func perform() async throws -> some IntentResult & ProvidesDialog { //
Replies
3
Boosts
0
Views
553
Activity
Jan ’26
Why is there no pop-up prompt when using the NEHotspot Configuration Manager app's Configuration interface to connect to WiFi failed
When using the NEHotspotConfigurationManager applyConfiguration interface for network connection.Sometimes there is a pop-up prompt when the connection fails, and sometimes there is no pop-up prompt. Why is there no pop-up prompt when the connection fails
Replies
1
Boosts
0
Views
152
Activity
Jan ’26
Reply to SwiftData with CloudKit Sync Issue
This might be late but I'm currently digging into a bug that popped up out of nowhere. I've had flawless syncing for 5 months, then a new user says his data is not syncing between iPad and iPhone, and now I'm experiencing the same on my devices. The debug builds worked fine, the production builds are intermittently not syncing and I have bad requests in my CloudKit log.
Replies
Boosts
Views
Activity
Jan ’26
Building a Full Space app that enables sharing a visionOS experience with nearby users.
Hello, I am currently considering developing a Full Space app that enables a shared visionOS experience with nearby users. Intended Features A Mixed Full Space app in which dozens of 3D models are placed in the space. These 3D models may play embedded animations when tapped, be programmatically moved or rotated, or be controlled via Reality Composer Pro timelines. The app also includes audio, spatial audio, videos with audio, and videos without audio, which are rendered as VideoTextures on planes and played back in the space. Some media elements play automatically, while others are triggered by user interaction. However, it is unclear whether AVPlaybackCoordinator supports shared playback across multiple types of media, such as: audio only spatial audio video without audio video with audio I am also unsure whether there are alternative or recommended approaches for synchronizing playback in this scenario. Questions Is it technically possi
Replies
1
Boosts
0
Views
370
Activity
Dec ’25
SoundAnalysis built-in classifier fails in background (SNErrorCode.operationFailed)
I’m seeing consistent failures using SoundAnalysis live classification when my app moves to the background. Setup iOS 17.x AVAudioEngine mic capture SNAudioStreamAnalyzer SNClassifySoundRequest(classifierIdentifier: .version1) UIBackgroundModes = audio AVAudioSession .record / .playAndRecord, active Audio capture + level metering continue working in background (mic indicator stays on) Issue As soon as the app enters background / screen locks: SoundAnalysis starts failing every second with domain:com.apple.SoundAnalysis, code:2(SNErrorCode.operationFailed) Audio capture itself continues normally When the app returns to foreground, classification immediately resumes without restarting the engine/analyzer Question Is live background sound classification with the built-in SoundAnalysis classifier officially unsupported or known to fail in background? If so, is a custom Core ML model the only supported approach for background detection? Or is there a required configuration I’m m
Replies
0
Boosts
0
Views
221
Activity
Dec ’25
Reply to App Store distribution question
Is it possible? Sure. It's always possible for Apple to approve an app as long as it complies with all Developer and App Review requirements. If you add new features, then you'll need to release an update. If you're concerned that your app might not have enough real functionality, then your judgement might be sound.
Replies
Boosts
Views
Activity
Dec ’25
Nasty problems in Xcode 26.2: Apple Intelligence crash & Source Code Control Failure
I know this post isn't going to give a lot of details, but what I experienced tonight was so completely weird that I wanted to get it posted here in case others run into it: FIRST: All was well until I made a trivial change to a large Objective-C++ module. I suddenly got the idea to look at that line in the code review pane, to see if that area of code had ever had recent modifications. But, the entire module showed up as modified -- one giant change bar, with nothing on the right side of the code review pane, no matter what commit I selected. Then I noticed that the two lines of code which had all of 4 characters edited were no longer showing any change bars. Yet, the file showed up as modified. Still, the exact line changes were not showing in the source code navigator, even though other files showed their changes. Note I'm connected to our remote repo on github. I did some command line git checks of the local repo, and the changes were there (as yet unstaged). So -- I figured, I'm gonna ask the Apple Codin
Replies
4
Boosts
0
Views
594
Activity
Dec ’25
AVAudioEngine obtains channel audio data
Currently, I have successfully used ChannelMap to map hardware input channels and obtained audio data from the hardware device's MIC and OTG inputs. Additionally, I have used ChannelMap to map output channels to freely feed data for playback to each output channel. However, I now have a problem. I have a hardware device that only has output channels (no input channels), and the system has set this hardware device as the default playback device. In this case, how can I obtain the audio data being played to the output channels for modification?
Replies
0
Boosts
0
Views
292
Activity
Dec ’25
Reply to Home App Intermittent Hub Not Responding Bug
Decided to start from scratch to see if it would resolve the issue. I removed all of my 100+ devices, deleted scenes/automation, and deleted the Home. I rebooted all devices (including network) and created a new home. I started off with adding Hue device and when I later launched the Home app I got the Home hub is unresponsive pop up. I did set everything back up again and will continue to monitor but so far the issue unfortunately still persists.
Topic: App & System Services SubTopic: General Tags:
Replies
Boosts
Views
Activity
Dec ’25
Reply to SwiftUI @State Updates Not Reflecting in UI Until View Reconstruction (Xcode Preview & Device)
Thanks for your reply! Here's the complete code and detailed information: Issue Description I'm experiencing a strange behavior with Xcode Preview in my SwiftUI app: On first Preview load: Button taps don't respond at all Temporary workaround: Click any other file, then click back to ContentView - buttons work perfectly Reproducibility: Happens 100% of the time after cleaning and restarting Preview Environment Xcode Version: Version 26.2 iOS Target: iOS 26.1 macOS Version: 26.1 Reproducibility: 100% consistent Relevant Code App Entry Point import SwiftUI import AVFoundation @main struct MyApp: App { init() { setupAudioSession() } private func setupAudioSession() { do { try AVAudioSession.sharedInstance().setCategory( .playback, mode: .default, options: [.mixWithOthers] ) try AVAudioSession.sharedInstance().setActive(true) } catch { print(Failed to setup audio session: (error)) } } var body: some Scene { WindowGroup { ContentView() } } }### ContentView - Simplified Version import SwiftUI struct Conten
Topic: UI Frameworks SubTopic: SwiftUI
Replies
Boosts
Views
Activity
Dec ’25
Reply to How do I provide developer documentation (4.1 - Copycats)
The answer to this question does not answer the question at all. I've setup an appeal three times. In all three instances, they just reply with the same 4.1 issue. Mind you, my app has already been APPROVED 7 times prior and now on the 8th time, all of a sudden, this 4.1 has popped up. I need answers to my questions above as my users are getting angry that my extension fixes are available on Chrome and Firefox, but no longer available on Apple (Safari) **1. Developer wants to know what needs to be provided? An email, A statement? How / what format would be required they are asking? How does such above documentation get submitted / included in subsequent updates to Not hinder approval?**
Replies
Boosts
Views
Activity
Dec ’25
SwiftUI @State Updates Not Reflecting in UI Until View Reconstruction (Xcode Preview & Device)
Issue Description I'm experiencing a bizarre SwiftUI state update issue that only occurs in Xcode development environment (both Canvas preview and device debugging), but does not occur in production builds downloaded from App Store. Symptom: User taps a button that modifies a @State variable inside a .sheet Console logs confirm the state HAS changed But the UI does not update to reflect the new state Switching to another file in Xcode and back to ContentView instantly fixes the issue The production build (same code) works perfectly fine Environment Xcode: 16F6 (17C52) iOS: 26.2 (testing on iPhone 13) macOS: 25.1.0 (Sequoia) SwiftUI Target: iOS 15.6+ Issue: Present in both Xcode Canvas and on-device debugging Production: Same code works correctly in App Store build (version 1.3.2) Code Structure Parent View (ContentView.swift) struct ContentView: View { @State private var selectedSound: SoundTheme = .none @State private var showSoundSheet = false var body: some View { VStack { // Display button shows current s
Topic: UI Frameworks SubTopic: SwiftUI
Replies
6
Boosts
0
Views
359
Activity
Dec ’25
Motiv Audio app crashes while opening Motiv Video app
Install both Motiv Audio and Motiv Video apps from the App Store. Open and use the Motiv Audio app. Remove audio app from background. Open Motiv Video app. Observe that the Motiv Audio app crashes. Both apps are checking Free available space. Both app are using Audio Engine
Replies
1
Boosts
0
Views
69
Activity
Dec ’25
Reply to iOS 26.2 RC DeviceActivityMonitor.eventDidReachThreshold regression?
Just got a user email from iOS 26.2 that looks like this bug (Again, I'm not sure if this is any different than the bug here but sounds different due to the explicit start on iOS 26.2 and that its strictly a beginning-of-the-day-thing: https://developer.apple.com/forums/thread/793747)
Topic: App & System Services SubTopic: General Tags:
Replies
Boosts
Views
Activity
Dec ’25