Search results for

“Popping Sound”

20,032 results found

Post

Replies

Boosts

Views

Activity

Reply to Wi-Fi Raw Socket Disconnection Issue on iPhone 17 Series
Thanks for the clarifications. Just FYI, in networking, raw socket usually refers to raw IP sockets, hence my confusion. So, lemme summarise the issue: You have an iOS app. It talks to a Wi-Fi based accessory over TCP. With iPhone 16, this all works as expected. With iPhone 17, things go wrong, such that your TCP socket disconnects after 20 seconds. After that, your app is unable to reconnect. And the only way to resolve the issue is to restart the Wi-Fi subsystem on your accessory. Is that right? If so, this sounds eminently bugworthy. While there are lots of ways that local networking can go wrong, the fact that this fails when the only variable you change is moving from iPhone 16 to 17 suggests that it’s caused by a Wi-Fi level issue on iPhone 17. DevForums focuses on Apple APIs and tools, and isn’t really set up to help with Wi-Fi level issues. Given your description of the problem, I think it makes sense for you to file a bug so that our Wi-Fi engineering team can investigate. In your bug, make
Nov ’25
Reply to PWA video playback stopped working after updating iOS to 26.0.1
I'm encountering the same bug. I suspect it's related to the state of the AudioContext. From what I've observed, when first entering the page, the state of AudioContext is suspended, and calling resume() within a click event fails to restore it. If I first call suspend() followed by resume() in the developer tools console, the following error message appears. So far, I haven't found a workaround—please share one if you have it. error message: InvalidStateError: Failed to start the audio device
Topic: Safari & Web SubTopic: General
Nov ’25
How do I run an arbitrary subprocess from an Xcode extension?
I'd like to start a subprocess from an Xcode extension. My prototype performCommandWithInvocation method looks like this: - (void)performCommandWithInvocation:(XCSourceEditorCommandInvocation *)invocation completionHandler:(void (^)(NSError * _Nullable nilOrError))completionHandler { NSTask*task= [[NSTask alloc]init]; [task setExecutableURL:[NSURL fileURLWithPath:@/bin/sh]]; //option A: [task setArguments:@[@-c,@echo hello from sh]]; //option B: [task setArguments:@[@-c,@/usr/bin/python3 -c 'print(hello from python3)']]; [task launch]; [task waitUntilExit]; int status=[task terminationStatus]; printf(status: %dn,status); completionHandler(nil); } So there's two options there: option A, or option B. (Uncomment to taste.) Option A does work. /bin/sh starts, and runs echo, and it prints hello to its stdout. I was foolish enough to let this convince me that this whole thing might be a goer. Option B does not work. I get an error: xcrun error: cannot be used with in an App Sandbox. Which is annoying, as virtually
0
0
79
Nov ’25
Reply to Hosting x86 Audio Units on Silicon Mac
Related to my last question, I forgot the key: kAudioUnitConfigurationInfo_AvailableArchitectures that is present in the audio component's ConfigurationInfo dictionary. I'm not 100% sure, but it seems that the plugins built with a Universal architecture fill this array with [x86_64, arm64] while the x86 plugins leave it empty. If confirmed, it could be a way to infer that they run under Rosetta.
Topic: Media Technologies SubTopic: Audio Tags:
Nov ’25
Hosting x86 Audio Units on Silicon Mac
My app encountered problems when trying to open an x86 audioUnit v2 on a Silicon Mac (although Rosetta is installed). There seems to be a XPC connection issue with the AUHostingService that I don't know how to fix. I observed other host apps opening the same plugins without problem, so there is probably something wrong or incompatible in my codes. I noticed that: The issue occurs whether or not the app is sandboxed. The issue does no longer occur when the app itself runs under Rosetta. There is no error reported by CoreAudio during allocation and initialization of the audio unit. The first notified errors appears when the unit calls AudioUnitRender from the rendering callback. With most x86 plugins, the error is on first call: kAudioUnitErr_RenderTimeout and on any subsequent call: kAudioComponentErr_InstanceInvalidated On the UI side, when the Cocoa View is loaded, it appears shortly, then disappears immediately leaving its superview empty. With another x86 plugin, the Cocoa View is loaded normally,
2
0
712
Nov ’25
ScreenCaptureKit recording output is corrupted when captureMicrophone is true
Hello everyone, I'm working on a screen recording app using ScreenCaptureKit and I've hit a strange issue. My app records the screen to an .mp4 file, and everything works perfectly until the .captureMicrophone is false In this case, I get a valid, playable .mp4 file. However, as soon as I try to enable the microphone by setting streamConfig.captureMicrophone = true, the recording seems to work, but the final .mp4 file is corrupted and cannot be played by QuickTime or any other player. This happens whether capturesAudio (app audio) is on or off. I've already added the Privacy - Microphone Usage Description (NSMicrophoneUsageDescription) to my Info.plist, so I don't think it's a permissions problem. I have my logic split into a ScreenRecorder class that manages state and a CaptureEngine that handles the SCStream. Here is how I'm configuring my SCStream: ScreenRecorder.swift // This is my main SCStreamConfiguration private var streamConfiguration: SCStreamConfiguration { var streamConfig = SCStreamConfi
2
0
697
Nov ’25
Device Activity monitor extension Not working
anyone has the same problem which is that your device activity extension ain't working even tho all the code work perfectly in the console, I setup it in the right way , tried to make schedule and it did the same exact thing when I tried to create usage threshold. anyone know the reason for this bug? here is my extension code import ManagedSettings import FamilyControls import Foundation import OSLog import UserNotifications class MonitoringExtension: DeviceActivityMonitor { private let defaults = UserDefaults(suiteName: group.com.William.app) private let logger = Logger(subsystem: com.William.app, category: MonitoringExtension) override func eventDidReachThreshold(_ event: DeviceActivityEvent.Name, activity: DeviceActivityName) { let activityRaw = activity.rawValue logger.info(Limite atteinte: (activityRaw)) scheduleNotification(title: Limite dépassée, body: Tu as utilisé trop de temps sur (activityRaw).) guard let data = defaults?.data(forKey: (activityRaw)_selection), let selection = try? JSONDecoder().dec
0
0
352
Nov ’25
UIBarButtonItem-The background color of UIBarButtonItem is missing?
UIButton *backBtn = [UIButton buttonWithType:UIButtonTypeCustom]; backBtn.frame = CGRectMake(0, 0, 44, 44); [backBtn setImage:[UIImage imageNamed:@back] forState:UIControlStateNormal]; [backBtn addTarget:self action:@selector(gotoBack) forControlEvents:UIControlEventTouchUpInside]; UIBarButtonItem *backItem = [[UIBarButtonItem alloc] initWithCustomView:backBtn]; // backItem.hidesSharedBackground = YES; // backItem.tintColor = [UIColor redColor]; self.navigationItem.leftBarButtonItem = backItem; I am using Xcode 26, which is compatible with iOS 26. After I set up the back button of the navigation bar, why are the background colors of the back buttons on some pages gray? When I pushed the new page, after popping back to the current page, the buttons with the gray background turned back to the white background. This issue will also affect the rightBarButtonItem.
Topic: UI Frameworks SubTopic: UIKit
1
0
228
Nov ’25
Setting to screen share virtual macOS 15+ machine?
I want to know all settings that need to be set in a macOS 15+ virtual mac in order to successfully screen share into it from the physical machine. Also any virtual machine configuration settings I may not understand. I have successfully set up screen sharing into an macOS 15 M2 mac studio that sits in the corner, on which I am running a cooked up variant of the sample code from apple about virtualizing a mac on a mac. I'm screen sharing into it from my macOS 15 M2 laptop. So I know I have at least one set of screen sharing settings correct. I have also successfully set up a macOS 14 virtual machine in that app into which I am able to screen share into from the mac studio. However, when I set up macOS 15 (and macOS 26) virtual machines, connections for screen sharing and file access don't succeed after 1 minute (I get the aqua-colored cylon / knight-rider -style progress bar for a minute, and then a connection failure dialog). I don't get the prompt to choose standard or enhanced. and I don't get asked for my
2
0
243
Nov ’25
Reply to Failed to verify code signature when trying to install on iPad from XCode
i've had this same exact issue. default xcode Audio Unit Extension app project won't build on my devices but will run fine on simulator. Seeing the same A valid provisioning profile for this executable was not found. error, despite having tried letting xcode automatically managing signing or do so manually. trying to set enabled hardened runtime to false but still getting the same error unfortunately
Oct ’25
Video Audio + Speech To Text
Hello, I am wondering if it is possible to have audio from my AirPods be sent to my speech to text service and at the same time have the built in mic audio input be sent to recording a video? I ask because I want my users to be able to say CAPTURE and I start recording a video (with audio from the built in mic) and then when the user says STOP I stop the recording.
2
0
834
Oct ’25
AVAudioEngine installTap stops working after phone call interruption on iPhone 16e
Environment Device: iPhone 16e iOS Version: 18.4.1 - 18.7.1 Framework: AVFoundation (AVAudioEngine) Problem Summary On iPhone 16e (iOS 18.4.1-18.7.1), the installTap callback stops being invoked after resuming from a phone call interruption. This issue is specific to phone call interruptions and does not occur on iPhone 14, iPhone SE 3, or earlier devices. Expected Behavior After a phone call interruption ends and audioEngine.start() is called, the previously installed tap should continue receiving audio buffers. Actual Behavior After resuming from phone call interruption: Tap callback is no longer invoked No audio data is captured No errors are thrown Engine appears to be running normally Note: Normal pause/resume (without phone call interruption) works correctly. Steps to Reproduce Start audio recording on iPhone 16e Receive or make a phone call (triggers AVAudioSession interruption) End the phone call Resume recording with audioEngine.start() Result: Tap callback is not invoked T
0
0
208
Oct ’25
help getting audio verbs macOS Tahoe Deep Debug Logs
Im running macOS Tahoe and I have the proper nvram boot-args , however when I try to poke the log stream im not getting any verb information related to the card im using. The audio system im using is AppleHDA.kext from the Beta 1 KDK. I've tried asking AI it doesn't make a difference what it suggests to me..... In the meantime of while im asking for assistance what ill do is go ahead and let it template me a kernel extension that I guess just traffics it to the Log for me and hopefully this isn't filtered out as what I suspect is it saying is happening is is that it actually masks some of the information. Why am I doing this? not For the Linux Driver its so I can see from the Log where it came from as this is what the developer said he did GitHub/davidjo/snd_hda_macbookpro is the kabylake iMac.
0
0
155
Oct ’25
Reply to PWA video playback stopped working after updating iOS to 26.0.1
Same here for all audio tags!
Topic: Safari & Web SubTopic: General
Replies
Boosts
Views
Activity
Nov ’25
Reply to Wi-Fi Raw Socket Disconnection Issue on iPhone 17 Series
Thanks for the clarifications. Just FYI, in networking, raw socket usually refers to raw IP sockets, hence my confusion. So, lemme summarise the issue: You have an iOS app. It talks to a Wi-Fi based accessory over TCP. With iPhone 16, this all works as expected. With iPhone 17, things go wrong, such that your TCP socket disconnects after 20 seconds. After that, your app is unable to reconnect. And the only way to resolve the issue is to restart the Wi-Fi subsystem on your accessory. Is that right? If so, this sounds eminently bugworthy. While there are lots of ways that local networking can go wrong, the fact that this fails when the only variable you change is moving from iPhone 16 to 17 suggests that it’s caused by a Wi-Fi level issue on iPhone 17. DevForums focuses on Apple APIs and tools, and isn’t really set up to help with Wi-Fi level issues. Given your description of the problem, I think it makes sense for you to file a bug so that our Wi-Fi engineering team can investigate. In your bug, make
Replies
Boosts
Views
Activity
Nov ’25
Reply to PWA video playback stopped working after updating iOS to 26.0.1
I'm encountering the same bug. I suspect it's related to the state of the AudioContext. From what I've observed, when first entering the page, the state of AudioContext is suspended, and calling resume() within a click event fails to restore it. If I first call suspend() followed by resume() in the developer tools console, the following error message appears. So far, I haven't found a workaround—please share one if you have it. error message: InvalidStateError: Failed to start the audio device
Topic: Safari & Web SubTopic: General
Replies
Boosts
Views
Activity
Nov ’25
How do I run an arbitrary subprocess from an Xcode extension?
I'd like to start a subprocess from an Xcode extension. My prototype performCommandWithInvocation method looks like this: - (void)performCommandWithInvocation:(XCSourceEditorCommandInvocation *)invocation completionHandler:(void (^)(NSError * _Nullable nilOrError))completionHandler { NSTask*task= [[NSTask alloc]init]; [task setExecutableURL:[NSURL fileURLWithPath:@/bin/sh]]; //option A: [task setArguments:@[@-c,@echo hello from sh]]; //option B: [task setArguments:@[@-c,@/usr/bin/python3 -c 'print(hello from python3)']]; [task launch]; [task waitUntilExit]; int status=[task terminationStatus]; printf(status: %dn,status); completionHandler(nil); } So there's two options there: option A, or option B. (Uncomment to taste.) Option A does work. /bin/sh starts, and runs echo, and it prints hello to its stdout. I was foolish enough to let this convince me that this whole thing might be a goer. Option B does not work. I get an error: xcrun error: cannot be used with in an App Sandbox. Which is annoying, as virtually
Replies
0
Boosts
0
Views
79
Activity
Nov ’25
Reply to Hosting x86 Audio Units on Silicon Mac
Related to my last question, I forgot the key: kAudioUnitConfigurationInfo_AvailableArchitectures that is present in the audio component's ConfigurationInfo dictionary. I'm not 100% sure, but it seems that the plugins built with a Universal architecture fill this array with [x86_64, arm64] while the x86 plugins leave it empty. If confirmed, it could be a way to infer that they run under Rosetta.
Topic: Media Technologies SubTopic: Audio Tags:
Replies
Boosts
Views
Activity
Nov ’25
Hosting x86 Audio Units on Silicon Mac
My app encountered problems when trying to open an x86 audioUnit v2 on a Silicon Mac (although Rosetta is installed). There seems to be a XPC connection issue with the AUHostingService that I don't know how to fix. I observed other host apps opening the same plugins without problem, so there is probably something wrong or incompatible in my codes. I noticed that: The issue occurs whether or not the app is sandboxed. The issue does no longer occur when the app itself runs under Rosetta. There is no error reported by CoreAudio during allocation and initialization of the audio unit. The first notified errors appears when the unit calls AudioUnitRender from the rendering callback. With most x86 plugins, the error is on first call: kAudioUnitErr_RenderTimeout and on any subsequent call: kAudioComponentErr_InstanceInvalidated On the UI side, when the Cocoa View is loaded, it appears shortly, then disappears immediately leaving its superview empty. With another x86 plugin, the Cocoa View is loaded normally,
Replies
2
Boosts
0
Views
712
Activity
Nov ’25
ScreenCaptureKit recording output is corrupted when captureMicrophone is true
Hello everyone, I'm working on a screen recording app using ScreenCaptureKit and I've hit a strange issue. My app records the screen to an .mp4 file, and everything works perfectly until the .captureMicrophone is false In this case, I get a valid, playable .mp4 file. However, as soon as I try to enable the microphone by setting streamConfig.captureMicrophone = true, the recording seems to work, but the final .mp4 file is corrupted and cannot be played by QuickTime or any other player. This happens whether capturesAudio (app audio) is on or off. I've already added the Privacy - Microphone Usage Description (NSMicrophoneUsageDescription) to my Info.plist, so I don't think it's a permissions problem. I have my logic split into a ScreenRecorder class that manages state and a CaptureEngine that handles the SCStream. Here is how I'm configuring my SCStream: ScreenRecorder.swift // This is my main SCStreamConfiguration private var streamConfiguration: SCStreamConfiguration { var streamConfig = SCStreamConfi
Replies
2
Boosts
0
Views
697
Activity
Nov ’25
Device Activity monitor extension Not working
anyone has the same problem which is that your device activity extension ain't working even tho all the code work perfectly in the console, I setup it in the right way , tried to make schedule and it did the same exact thing when I tried to create usage threshold. anyone know the reason for this bug? here is my extension code import ManagedSettings import FamilyControls import Foundation import OSLog import UserNotifications class MonitoringExtension: DeviceActivityMonitor { private let defaults = UserDefaults(suiteName: group.com.William.app) private let logger = Logger(subsystem: com.William.app, category: MonitoringExtension) override func eventDidReachThreshold(_ event: DeviceActivityEvent.Name, activity: DeviceActivityName) { let activityRaw = activity.rawValue logger.info(Limite atteinte: (activityRaw)) scheduleNotification(title: Limite dépassée, body: Tu as utilisé trop de temps sur (activityRaw).) guard let data = defaults?.data(forKey: (activityRaw)_selection), let selection = try? JSONDecoder().dec
Replies
0
Boosts
0
Views
352
Activity
Nov ’25
UIBarButtonItem-The background color of UIBarButtonItem is missing?
UIButton *backBtn = [UIButton buttonWithType:UIButtonTypeCustom]; backBtn.frame = CGRectMake(0, 0, 44, 44); [backBtn setImage:[UIImage imageNamed:@back] forState:UIControlStateNormal]; [backBtn addTarget:self action:@selector(gotoBack) forControlEvents:UIControlEventTouchUpInside]; UIBarButtonItem *backItem = [[UIBarButtonItem alloc] initWithCustomView:backBtn]; // backItem.hidesSharedBackground = YES; // backItem.tintColor = [UIColor redColor]; self.navigationItem.leftBarButtonItem = backItem; I am using Xcode 26, which is compatible with iOS 26. After I set up the back button of the navigation bar, why are the background colors of the back buttons on some pages gray? When I pushed the new page, after popping back to the current page, the buttons with the gray background turned back to the white background. This issue will also affect the rightBarButtonItem.
Topic: UI Frameworks SubTopic: UIKit
Replies
1
Boosts
0
Views
228
Activity
Nov ’25
Setting to screen share virtual macOS 15+ machine?
I want to know all settings that need to be set in a macOS 15+ virtual mac in order to successfully screen share into it from the physical machine. Also any virtual machine configuration settings I may not understand. I have successfully set up screen sharing into an macOS 15 M2 mac studio that sits in the corner, on which I am running a cooked up variant of the sample code from apple about virtualizing a mac on a mac. I'm screen sharing into it from my macOS 15 M2 laptop. So I know I have at least one set of screen sharing settings correct. I have also successfully set up a macOS 14 virtual machine in that app into which I am able to screen share into from the mac studio. However, when I set up macOS 15 (and macOS 26) virtual machines, connections for screen sharing and file access don't succeed after 1 minute (I get the aqua-colored cylon / knight-rider -style progress bar for a minute, and then a connection failure dialog). I don't get the prompt to choose standard or enhanced. and I don't get asked for my
Replies
2
Boosts
0
Views
243
Activity
Nov ’25
Reply to Failed to verify code signature when trying to install on iPad from XCode
i've had this same exact issue. default xcode Audio Unit Extension app project won't build on my devices but will run fine on simulator. Seeing the same A valid provisioning profile for this executable was not found. error, despite having tried letting xcode automatically managing signing or do so manually. trying to set enabled hardened runtime to false but still getting the same error unfortunately
Replies
Boosts
Views
Activity
Oct ’25
Video Audio + Speech To Text
Hello, I am wondering if it is possible to have audio from my AirPods be sent to my speech to text service and at the same time have the built in mic audio input be sent to recording a video? I ask because I want my users to be able to say CAPTURE and I start recording a video (with audio from the built in mic) and then when the user says STOP I stop the recording.
Replies
2
Boosts
0
Views
834
Activity
Oct ’25
AVAudioEngine installTap stops working after phone call interruption on iPhone 16e
Environment Device: iPhone 16e iOS Version: 18.4.1 - 18.7.1 Framework: AVFoundation (AVAudioEngine) Problem Summary On iPhone 16e (iOS 18.4.1-18.7.1), the installTap callback stops being invoked after resuming from a phone call interruption. This issue is specific to phone call interruptions and does not occur on iPhone 14, iPhone SE 3, or earlier devices. Expected Behavior After a phone call interruption ends and audioEngine.start() is called, the previously installed tap should continue receiving audio buffers. Actual Behavior After resuming from phone call interruption: Tap callback is no longer invoked No audio data is captured No errors are thrown Engine appears to be running normally Note: Normal pause/resume (without phone call interruption) works correctly. Steps to Reproduce Start audio recording on iPhone 16e Receive or make a phone call (triggers AVAudioSession interruption) End the phone call Resume recording with audioEngine.start() Result: Tap callback is not invoked T
Replies
0
Boosts
0
Views
208
Activity
Oct ’25
help getting audio verbs macOS Tahoe Deep Debug Logs
Im running macOS Tahoe and I have the proper nvram boot-args , however when I try to poke the log stream im not getting any verb information related to the card im using. The audio system im using is AppleHDA.kext from the Beta 1 KDK. I've tried asking AI it doesn't make a difference what it suggests to me..... In the meantime of while im asking for assistance what ill do is go ahead and let it template me a kernel extension that I guess just traffics it to the Log for me and hopefully this isn't filtered out as what I suspect is it saying is happening is is that it actually masks some of the information. Why am I doing this? not For the Linux Driver its so I can see from the Log where it came from as this is what the developer said he did GitHub/davidjo/snd_hda_macbookpro is the kabylake iMac.
Replies
0
Boosts
0
Views
155
Activity
Oct ’25
Reply to .glassProminent toolbar buttons are glitchy in iOS 26.1 RC
That sounds like a bug. Do you have a feedback report yet? If not, would you mind to file one and share your report ID here. Thanks! Best, —— Ziqiao Chen  Worldwide Developer Relations.
Topic: UI Frameworks SubTopic: SwiftUI
Replies
Boosts
Views
Activity
Oct ’25