Overview

Post

Replies

Boosts

Views

Created

Notarization stuck at "In Progress"
I have been waiting well over 24 hours for my notarization to occur, and nothing - no "we might take a while if it is your first app" or any mention of that in the docs. So why is it taking this long? What's the hold up??? If this is part of the process, Apple should officially document it, so developers aren't left resubmitting and wondering. This is not a good first experience with the developer program. Successfully received submission history. history -------------------------------------------------- createdDate: 2026-01-07T05:22:34.038Z name: URSAMajorSpaceStationSST206_v1.0.0.zip status: In Progress -------------------------------------------------- createdDate: 2026-01-06T01:55:05.144Z name: URSAMajorSpaceStationSST206_v1.0.0.zip status: In Progress -------------------------------------------------- createdDate: 2026-01-05T20:55:50.624Z name: test.zip status: Invalid -------------------------------------------------- createdDate: 2026-01-05T20:32:52.944Z name: URSAMajorSpaceStationSST206.vst3.zip status: In Progress -------------------------------------------------- createdDate: 2026-01-05T19:37:15.426Z name: URSAMajorSpaceStationSST206.component.zip status: In Progress -------------------------------------------------- createdDate: 2026-01-05T18:37:43.101Z name: URSAMajorSpaceStationSST206.component.zip status: In Progress
2
0
129
5d
nable to Distribute iOS App to App Store – Archive Shows Only 'Distribute Content'
Hello Apple Developer Support Team, I am writing to request assistance with an issue preventing me from distributing my iOS app to the App Store. I am able to successfully build and archive my app in Xcode without errors. However, after archiving, Xcode only shows the option “Distribute Content” instead of “Distribute App”. Because of this, I am unable to upload the app to App Store Connect. App & Account Details • App Bundle Identifier: com.yesbuy.yesbuyapp • Team Name: Fathima Sayed Mohamed • Team ID: Z327MG963D • Xcode Version: 26.2 • macOS Version: macOS 15.6.1 Signing & Certificates • Signing Method: Automatically manage signing (enabled) • Signing Certificate (Release): Apple Distribution: Fathima Sayed Mohamed • Provisioning Profile: Xcode Managed / App Store Distribution • Archive Status: Build succeeded What I’ve Already Verified • The archive is created using Release configuration • Skip Install is set to No for the main app target (Runner) • The app target is selected correctly during Archive • A valid Apple Distribution certificate exists • The app is registered correctly in App Store Connect • App Groups and capabilities are configured correctly • Notification Service Extension is present and signed Despite all of this, Xcode still treats the archive as a Generic Xcode Archive and only allows Distribute Content, not Distribute App. I have attached screenshots showing: 1. Xcode Archives screen (showing “Distribute Content” only) 2. Signing & Capabilities for the app target 3. Certificates and provisioning profiles 4. Team ID and bundle identifier configuration Could you please review this issue and let me know: • Why the archive is not recognized as an App Store–distributable app • Whether there is any account, certificate, or configuration issue on Apple’s side • Any steps I should take to resolve this and successfully upload the app to App Store Connect Thank you very much for your time and support. I look forward to your guidance.
1
0
71
5d
Clarification Needed: histogrammedTimeToFirstDraw vs histogrammedOptimizedTimeToFirstDraw in MetricKit
Hi Apple Developer Community, I'm implementing MetricKit launch performance tracking in our iOS app and need clarification on two properties: histogrammedTimeToFirstDraw histogrammedOptimizedTimeToFirstDraw The Documentation Problem: The official MetricKit documentation provides minimal explanation of these properties beyond their names. Based on naming conventions, I initially assumed: histogrammedTimeToFirstDraw = cold launches histogrammedOptimizedTimeToFirstDraw = warm/optimized launches Based on our measurements: The “optimized” metric appears only in a small fraction of launches The "optimized" metric is actually slower The naming suggests the opposite behavior Questions: What specific launch conditions does each metric measure? Why would "optimized" launches be slower and less frequent? Is histogrammedOptimizedTimeToFirstDraw related to iOS app pre-warming or prediction features? If these metrics don’t correspond to cold vs. warm launch times, is there an alternative way to measure them accurately? Any clarification from Apple or insights from developers who've tracked these metrics would be greatly appreciated.
0
0
62
5d
NSHostingView stops receiving mouse events when layered above another NSHostingView (macOS Tahoe 26.2)
I’m running into a problem with SwiftUI/AppKit event handling on macOS Tahoe 26.2. I have a layered view setup: Bottom: AppKit NSView (NSViewRepresentable) Middle: SwiftUI view in an NSHostingView with drag/tap gestures Top: Another SwiftUI view in an NSHostingView On macOS 26.2, the middle NSHostingView no longer receives mouse or drag events when the top NSHostingView is present. Events pass through to the AppKit view below. Removing the top layer immediately restores interaction. Everything works correctly on macOS Sequoia. I’ve posted a full reproducible example and detailed explanation on Stack Overflow, including a single-file demo: Stack Overflow post: https://stackoverflow.com/q/79862332 I also found a related older discussion here, but couldn’t get the suggested workaround to apply: https://developer.apple.com/forums/thread/759081 Any guidance would be appreciated. Thanks!
2
0
235
5d
nable to Distribute iOS App to App Store – Archive Shows Only 'Distribute Content'
Hello Apple Developer Support Team, I am writing to request assistance with an issue preventing me from distributing my iOS app to the App Store. I am able to successfully build and archive my app in Xcode without errors. However, after archiving, Xcode only shows the option “Distribute Content” instead of “Distribute App”. Because of this, I am unable to upload the app to App Store Connect. App & Account Details • App Bundle Identifier: com.yesbuy.yesbuyapp • Team Name: Fathima Sayed Mohamed • Team ID: Z327MG963D • Xcode Version: 26.2 • macOS Version: macOS 15.6.1 Signing & Certificates • Signing Method: Automatically manage signing (enabled) • Signing Certificate (Release): Apple Distribution: Fathima Sayed Mohamed • Provisioning Profile: Xcode Managed / App Store Distribution • Archive Status: Build succeeded What I’ve Already Verified • The archive is created using Release configuration • Skip Install is set to No for the main app target (Runner) • The app target is selected correctly during Archive • A valid Apple Distribution certificate exists • The app is registered correctly in App Store Connect • App Groups and capabilities are configured correctly • Notification Service Extension is present and signed Despite all of this, Xcode still treats the archive as a Generic Xcode Archive and only allows Distribute Content, not Distribute App. I have attached screenshots showing: 1. Xcode Archives screen (showing “Distribute Content” only) 2. Signing & Capabilities for the app target 3. Certificates and provisioning profiles 4. Team ID and bundle identifier configuration Could you please review this issue and let me know: • Why the archive is not recognized as an App Store–distributable app • Whether there is any account, certificate, or configuration issue on Apple’s side • Any steps I should take to resolve this and successfully upload the app to App Store Connect Thank you very much for your time and support. I look forward to your guidance.
2
0
51
5d
CLMonitor API Missing Geofence Entry Events After Initial Registration
We are experiencing a failure in CLMonitor event delivery when the application is launched into the background via an APNS (Remote Push Notification). Even when a CLBackgroundActivitySession is instantiated immediately upon background launch, CLCircularGeographicCondition "Enter" events are suppressed. The system fails to deliver these events until the user manually brings the application to the Foreground. This indicates that CLBackgroundActivitySession does not correctly maintain background persistence when the session begins in a background state rather than transitioning from the foreground. Comparison of API Behavior (Background State) Launch via APNS: CLMonitor: Fails to trigger "Enter" events until the app is manually brought to the foreground. Legacy API: Successfully triggers and delivers "Enter" events immediately upon background launch. Exit Event Reliability: CLMonitor: Reliably triggers exit events even in the background. Legacy API: Reliably triggers exit events. Foreground Dependency: CLMonitor: Requires a foreground transition to "flush" or activate the delivery of pending entry events. Legacy API: No foreground transition required; events are delivered directly to the background process. Event Recovery: CLMonitor: Relies on the developer re-instantiating the CLMonitor and awaiting the events stream, which appears to "stall" during warm-starts. Legacy API: Relies on the CLLocationManagerDelegate which remains active as long as the manager instance exists. Steps to Reproduce Preconditions: Location Permissions: Set to "Always Allow". Background Modes: "Location updates" and "Remote notifications" enabled. App State: Terminated or Killed (by the user or the OS). Reproduction Path: Trigger Background Launch: Send a silent push notification (APNS) to wake the app in the background. Initialize Session: Within the background launch sequence (e.g., didFinishLaunchingWithOptions), immediately create and hold a strong reference to a CLBackgroundActivitySession. Register Monitor: * Initialize CLMonitor using requestMonitorWithConfiguration. Add a geofence using addConditionForMonitoring with a CLCircularGeographicCondition. Simulate Entry: Move the physical device (or simulate location) into the geofence boundary while the app remains in the background state. Observe: No "Enter" event is received in the CLMonitor event stream. Foreground Transition: Bring the app to the foreground. Actual Result: The "Enter" event is only delivered the moment the app enters the Foreground. Expected Result: The CLBackgroundActivitySession should enable CLMonitor to deliver "Enter" events immediately in the background, parity with the deprecated startMonitoringForRegion API.
1
0
79
5d
iOS React Native: Can two WebRTC stacks (Wazo & Jitsi) share media?
Hi everyone, I’m building a React Native iOS app where I’m integrating Wazo (native WebRTC) and Jitsi (WebView / WebRTC). Use case: Wazo is used to maintain a background call session (mainly signaling + audio keep-alive). Jitsi is used in the foreground for video calls. Problem: When Jitsi starts, it takes control of the microphone and camera. The Wazo call disconnects after ~5 minutes (likely due to media / audio session conflict). Even if Wazo audio/video is muted or tracks are disabled, the session still drops. My questions: Is it officially supported or recommended to run two WebRTC stacks (Wazo + Jitsi) simultaneously on iOS? Can Wazo stay connected without active audio/video tracks while Jitsi uses mic/camera? Is there a way to release Wazo media streams temporarily (but keep signaling alive) while Jitsi is loading or active? Are there any AVAudioSession / background mode limitations on iOS that make this impossible by design? If this is not supported, what is the recommended architecture (single WebRTC pipeline, switching media ownership, etc.)? Environment: iOS (React Native) Wazo SDK (native WebRTC) Jitsi Meet (WebView) CallKit + PushKit enabled Any guidance, documentation, or real-world experience would be greatly appreciated. Thanks in advance 🙏
1
0
301
5d
XCTest resources aren't being copied into place correctly
So I have a project with an Objective-C framework which has unit tests. One of the unit tests has a file 'sample.emlx' in its 'Resources' folder, which has its target membership set to the tests target. I expect this file to be copied into the correct place, such that I can find it (in the 'setup') method with: NSBundle *bundle = [NSBundle bundleForClass:[self class]]; NSString *path = [bundle pathForResource:@"sample" ofType:@"emlx"]; NSData *content = [NSData dataWithContentsOfFile:path]; NSString *entire = [NSString stringWithUTF8String:content.bytes]; _sample = [entire componentsSeparatedByString:@"\n"]; NSLog(@"sample lines: %d", (int)_sample.count); The idea is to make an array where I can pull various subsections of lines out to test various parts of the code. This works every time, if I set a breakpoint at the "create the NSBundle" line, and single-step through the code after right-clicking on the specific test method I want to run It never works if I just right-click on the specific test method I want to run and allow it to run freely. I will always see "sample lines: 0" This is with the latest public Xcode, I haven't tried any beta variant.
0
0
41
5d
Is Picture in Picture supported for WebRTC / WebView video on iOS (outside app)?
Hello, I am implementing video calling on iOS and need to support Picture in Picture (PiP) behavior similar to FaceTime or WhatsApp. What works Audio continues correctly in background CallKit UI works as expected Video works correctly while the app is in the foreground What I’m trying to achieve When the user presses the Home button or switches apps, I want to show a system Picture in Picture window (floating video outside the app). Current setup Video is rendered via WebRTC The video is displayed inside a WKWebView (HTML / JavaScript) PiP works only while the app is foregrounded When the app backgrounds, the video disappears (only audio remains) Questions Does iOS support system Picture in Picture for: WebRTC video WKWebView / HTML video 2 Is AVPictureInPictureController limited only to: AVPlayerLayer AVSampleBufferDisplayLayer 3 If PiP requires native rendering: Is it mandatory to render WebRTC frames natively using AVSampleBufferDisplayLayer? Is PiP explicitly unsupported for WebView / HTML video? 📌 Clarification Apps like FaceTime and WhatsApp are able to show PiP outside the app. I want to understand whether this behavior is achievable only with native video pipelines, or if WebView-based video is fundamentally restricted by iOS. Any official clarification or documentation reference would be appreciated. Thank you.
Topic: Safari & Web SubTopic: General Tags:
0
0
54
5d
Unable to download iOS simulator runtime 26.x on Xcode
Hi everyone, I'm experiencing a persistent issue for months now where I'm unable to download the iOS 26 simulator Runtime. I've tried reinstalling Xcode and also Xcode Beta but same issue. iOS 26 Simulator is also not on developers download page, so manual installation is impossible. And sadly I can't compile any code without having iOS 26 simulator installed. Anyone able to get passed this? Hardware: M1 Pro OS: Tahoe 26.1 Heres the error Download failed. Domain: DVTDownloadableErrorDomain Code: 41 User Info: { DVTErrorCreationDateKey = "2026-01-07 11:35:35 +0000"; } -- Download failed. Domain: DVTDownloadableErrorDomain Code: 41 -- Failed fetching catalog for assetType (com.apple.MobileAsset.iOSSimulatorRuntime), serverParameters ({ RequestedBuild = 23C54; }) Domain: DVTDownloadsUtilitiesErrorDomain Code: -1 -- Download failed due to not being able to connect to the host. (Catalog download for com.apple.MobileAsset.iOSSimulatorRuntime) Domain: com.apple.MobileAssetError.Download Code: 60 User Info: { checkNetwork = 1; } -- System Information macOS Version 26.1 (Build 25B78) Xcode 26.2 (24553) (Build 17C52) Timestamp: 2026-01-07T12:35:35+01:00
2
1
232
5d
Activating a Container App from a Custom Keyboard Extension to Enable Continuous Voice Input While Preserving the Original Typing Context
Project Background: I am developing a third-party custom keyboard for iOS whose primary feature is real-time voice input. In my current design, responsibilities are split as follows: 1. The container (main) app is responsible for: Audio recording Speech recognition (ASR) 2. The keyboard extension is responsible for: Providing the keyboard UI Initiating the voice input workflow Receiving transcription results via an App Group Inserting recognized text into the active text field using textDocumentProxy.insertText(_:) Intended User Flow The intended workflow is: The user is typing in a third-party app (for example, WeChat) using my custom keyboard. The user taps a “Voice Input” button in the keyboard extension. The keyboard extension activates the container app so that audio recording and ASR can begin. After recording has started, control returns to the original app where the user was typing. The container app continues running in the background, maintaining active audio recording and ASR. Recognized text is continuously streamed back to the keyboard extension and inserted into the current cursor position in real time. Observed Industry Behavior Some popular third-party keyboards on iOS, such as WeChat Keyboard and Doubao Keyboard, appear to provide a similar user experience in which: Voice input can be initiated directly from the keyboard while typing in another app. The user remains (or returns) in the original typing context after voice input starts. Speech recognition continues and text is streamed into the active text field without interrupting the typing experience. I would like to better understand how this type of workflow aligns with iOS platform capabilities and supported APIs. My Questions Is it supported by iOS public APIs for a custom keyboard extension to activate its container app to start audio recording and ASR, then return to the original host app while the container app continues recording and performing ASR in the background? If this workflow is not supported, are there any Apple-recommended or supported alternative architectures for achieving a similar user experience, especially when audio recording and ASR logic are currently implemented in the container app rather than in the keyboard extension? Goal My goal is to design a solution that is fully compliant with iOS public APIs and platform constraints, while providing a real-time voice input experience comparable to existing third-party keyboards on the platform. Any guidance on supported APIs, recommended architectures, or relevant documentation would be greatly appreciated.
3
0
105
5d
Xcode 16 issues with mac virtualisation
I have created a virtualised Mac machine where I am running the latest Xcode version 16.4 with macOS Sequoia as the base image. While running XCTest tests, I am observing a clear difference in behaviour between virtualised and non-virtualised setups. Below is the command I am using to run the tests: /Applications/Xcode.app/Contents/Developer/usr/bin/xcodebuild -xctestrun /Users/ltuser/290103613/xcui.xctestrun -destination "platform=iOS Simulator,id=79A5F8CC-4A2E-4C91-895D-AE97E25CB824,arch=arm64,OS=18.0" test-without-building "-only-testing:ClassName/MethodName/testName" -derivedDataPath /Users/ltuser/290103613/Archive -verbose IDELogLevel=debug IDETestLogLevel=debug ONLY_ACTIVE_ARCH=YES Observed Issue On virtualised hosts: Every XCTest is executed twice This happens for both passed and failed tests The behaviour is consistent across all virtualised machines On non-virtualised (bare-metal) Macs: Tests execute only once, as expected Debugging Steps Taken Multiple destination suspicion Initially, I noticed the following warning: --- xcodebuild: WARNING: Using the first of multiple matching destinations: { platform:iOS Simulator, id:0FFB9082-94AB-4D0A-A731-CC4452BDD4E7, OS:17.0, name:iPhone 15 } { platform:iOS Simulator, id:0FFB9082-94AB-4D0A-A731-CC4452BDD4E7, OS:17.0, name:iPhone 15 } This raised suspicion that tests might be running twice due to multiple matching destinations. Destination narrowed down I then explicitly restricted the destination to a single simulator UDID, after which: The warning disappeared However, the issue still persists Tests continue to run twice on virtualised hosts Version Comparison Xcode 16.4 + macOS Sequoia (virtualised) → Tests run twice Xcode 15.x (same setup otherwise) → Tests run once Questions Has anyone else faced this issue with Xcode 16.x on virtualised macOS? Is this a known issue with Xcode 16 + Virtualisation? Is there any recommended workaround or configuration change to avoid duplicate test execution? I am currently blocked on this issue and would appreciate confirmation on whether this is an Xcode or virtualisation-related problem. Thank you.
1
1
72
5d
Why doesn’t AVPlayer / AVFoundation support MPEG-DASH (MPD)? Any public rationale?
Hi, I understand that AVPlayer/AVFoundation doesn’t natively play MPEG-DASH manifests (.mpd) today, while HLS is supported and widely documented by Apple. I’m not asking for roadmap commitments, but I’d like to understand whether there is any publicly documented rationale for not supporting DASH/MPD in AVFoundation (e.g., technical constraints, platform integration, DRM ecosystem, power/performance considerations, etc.). Questions: Is there any Apple statement / documentation explaining why DASH (MPD) isn’t supported in AVFoundation? Is Apple’s recommended approach still “provide HLS for Apple clients” (potentially sharing CMAF segments and generating separate manifests)? If there’s no public rationale, is filing Feedback Assistant the best channel for requesting MPD playback support? Thanks!
1
1
309
5d
Trying to enroll, getting an error
My apple account region was set to US. I am not in US and don't have an American phone number. When I first tried to enroll with non-American number and Region set to US, I got an error. I have changed my region after that and when I am trying to change the details, I am landing on "Select your entity type." page and not on the page where I have to input my details. Contacted support and no answer. Can anyone advise what to do?
2
0
83
5d
What's Wrong with Apple Review?
I submitted a new app in the morning of December 30th, last year. A reviewer picked it up 2 days ago and then rejected it by saying that the Send button remains disabled. Of course, it remains disabled because the user is supposed to select a message out of 300 of them. The app comes with a how-to-use guide, which states that you need to select a message by tapping the Select button. Now, the reviewer is MIA. I don't necessarily blame the reviewer for not having read the user guide thoroughly. But I wonder why they don't just ask instead of jumping to a conclusion? They often reject the app for simple reasons these days. And they are often wrong. I understand that everybody makes a mistake. But they are wrong 4 out of 5 times. Last year, a reviewer rejected the app by saying that the user is not able to ask a question at the support site when in fact it was just the matter of scrolling down further. This is the type of reasons I strongly suggest they ask instead of jumping to a conclusion and then rejecting the app. It's quite difficult for me to make plans while dealing with reviewers if they are this sloppy.
4
1
204
5d
Confidential Innovation Proposal.Inquiry About Formal Communication Channel
I am writing to express interest in engaging with Apple regarding a highly original and commercially relevant concept related to future iPhone innovation. Given the confidential and proprietary nature of this idea, I am not in a position to share details through an open inquiry or standard feedback form. I would welcome the opportunity to present this concept through an official and formal communication channel that ensures appropriate confidentiality and professional evaluation, should Apple have an established process for external innovation or partnership discussions. Please advise if there is a suitable point of contact or procedure for initiating such a conversation in accordance with Apple’s policies. Thank you for your time and consideration. Please feel free to contact me though my email or phone Regards Tahmeed Hossain Contact: +880 1781882730
Topic: Design SubTopic: General
1
0
686
5d
Notarization is taking forever
I have recently enrolled in the Apple Developer to get my app notarized, and submitted an Archive for notarization, but it is taking forever. It has almost been a whole day, but the status is still in progress, whereas I have seen other developers say that the same takes 10-15 mins to an hour for them. Am I doing anything wrong? Please guide me through this.
1
0
56
5d
Problem Apple Developer, Something Went Wrong issue track order number
Hello, I am experiencing an issue with the Apple Developer Program. Yesterday, I enrolled in the Apple Developer Program and entered my Visa card details. However, later that night, I received a notification from my bank that the payment was charged but failed because my balance was insufficient. After that, I added funds to my bank account, withdrew/canceled my previous enrollment, and then enrolled again using the same Visa card. After re-enrolling, I received a notification from my bank and an email confirming that the payment was successful, with the order ID: “We're processing your order – D00****” However, when I try to track the order by clicking the link in the email or checking the order status, it shows the following message: Something went wrong. There was a problem with your request. Please try again. Could you please explain why this issue is happening and help me check the status of my Apple Developer Program enrollment?
0
0
42
5d
It crashes when AVAssetReader is released
Thread 5 Crashed: 0 libobjc.A.dylib 0x19af7b038 objc_msgSend + 56 1 CoreFoundation 0x19dfdb618 cow_cleanup + 135 2 CoreFoundation 0x19dfdb6fc -[__NSDictionaryM dealloc] + 147 3 MediaToolbox 0x1b167636c FigRemotePropertyCacheTeardown + 31 4 MediaToolbox 0x1b1c5b648 remoteXPCAsset_Finalize + 107 5 CoreMedia 0x1b1e9166c FigBaseObjectFinalize + 275 6 CoreFoundation 0x19dfcc5ec _CFRelease + 295 7 AVFCore 0x1b1054d64 -[AVFigAssetTrackInspector dealloc] + 151 8 AVFCore 0x1b0f818d8 -[AVAssetTrack dealloc] + 63 9 CoreFoundation 0x19dfdba28 RELEASE_OBJECTS_IN_THE_ARRAY + 115 10 CoreFoundation 0x19dfdb7e0 -[__NSArrayM dealloc] + 147 11 AVFCore 0x1b0f52e04 -[AVURLAsset dealloc] + 167 12 libobjc.A.dylib 0x19af821f8 object_cxxDestructFromClass(objc_object*, objc_class*) + 115 13 libobjc.A.dylib 0x19af7df20 objc_destructInstance_nonnull_realized(objc_object*) + 75 14 libobjc.A.dylib 0x19af7d4a4 _objc_rootDealloc + 71 15 AVFCore 0x1b0fef988 -[AVAssetReaderOutput dealloc] + 415 16 AVFCore 0x1b0ff11ec -[AVAssetReaderTrackOutput dealloc] + 127 17 CoreFoundation 0x19dfe20a4 -[__NSSingleObjectArrayI dealloc] + 63 18 libobjc.A.dylib 0x19af7d3f8 AutoreleasePoolPage::releaseUntil(objc_object**) + 203
1
0
264
5d