Search results for

DTiPhoneSimulatorErrorDomain Code 2

158,487 results found

Post

Replies

Boosts

Views

Activity

Reply to Extend Measurement to support inverse units
Dimension can be subclassed to create the custom units. You probably don't want to subclass Dimension for reciprocal units since there's no good way to convert between measurements of different Dimension types. Instead, you can extend the existing classes to add the new reciprocal units, and then conversion will work like usual. subclass [UnitConverter] to something like UnitConverterInverse Yep you'll need that. Interestingly, there's already a reciprocal unit buried in the existing measurement API. It uses a reciprocal converter that is exactly what you need, but unfortunately isn't part of the public API. Check this out: Welcome to Apple Swift version 6.2.3 (swiftlang-6.2.3.3.21 clang-1700.6.3.2). Type :help for assistance. 1> import Foundation 2> print(UnitFuelEfficiency.litersPer100Kilometers.converter) <_NSStatic_NSStaticUnitConverterLinear_NoConst: 0x20761bfc8> coefficient = 1.000000, constant = 0.000000 3> print(UnitFuelEfficiency.milesPerGallon.converter) reciprocalValue = 23
Topic: App & System Services SubTopic: General Tags:
3d
Reply to visionOS Bluetooth LE limited to 2 connections?
Is there a 2-device limit for CoreBluetooth on visionOS 2.1? Yes, that's the limit. Note that watchOS has the same limit. Is there any documentation I can refer to that states something like this? No. This is one of those technical details we've basically never documented. Note that I'm not aware of any documentation saying that 4 devices work on iOS either. __ Kevin Elliott DTS Engineer, CoreOS/Hardware
Topic: App & System Services SubTopic: Core OS Tags:
3d
Reply to Best practice for centralizing SwiftData query logic and actions in an @Observable manager?
A SwiftData query (@Query) is currently tied to a SwiftUI view and relies on the modelContext environment value of the view. There is currently no way to use @Query outside of a SwiftUI view. To implement the query logics in a controller class, you basically need to to do the following with your own code: Fetch data using FetchDescriptor + ModelContext.fetch. Update the result set when relevant changes happen. To do step 2, assuming you are using the SwiftData default store, which is based on Core Data, you need to observe the notifications triggered by a store change (basically NSManagedObjectContextObjectsDidChange and .NSPersistentStoreRemoteChange), check if the change is relevant, and re-fetch the data set as needed. All these will be a bit involved, and yet have already implemented (as Query) on the framework side. I'd hence suggest that you file a feedback report to request an API that does what Query does but doesn't rely on SwiftUI – If you do so, please share your report ID here fo
3d
SCK/replayd behaviors and delays
We're troubleshooting SCK issues. They occur with a relatively small amount of sessions, but lack of context and/or ability to advise the customer on how they could make behavior more predictable and reliable is problematic. Generally, there is 2 distinct issues which may or may not have the same root cause: Failure to establish SCK session. Usually manifests within the app as SCShareableContent.getWithCompletionHandler call either never invoking the completion handler, or taking prohibitively long time (we usually give it 3-10 sec before giving up). In the system log it may look like this: (log omitted - suspecting it triggers the content filter) Note the 6 seconds delay to completion of fetchShareableContentWithOption (normally it's a 30-40ms operation). Sometime, we'd see the stream established, but some minutes (or even hours) into the recording we'd stop receiving frames. Both scenarios are likely to occur when the disk space is low, with reliable repro of the problem #2 at below 8gb of
0
0
425
3d
onWorldRecenter memory leak and duplicate callbacks in ImmersiveSpace
Posting this here in case this information is helpful to other developers: As of visionOS 26.3 beta 1, onWorldRecenter has two significant issues: (FB21557639) Memory Leak: When onWorldRecenter is assigned to a RealityView within an ImmersiveSpace, it appears to retain a strong reference to the view's internal SwiftUI context. When the immersive space is dismissed, the view's @State objects will not be deallocated. Also, each time the immersive space view's body is executed, additional state storage will be allocated and leaked. Multiple Callbacks: When the user long-presses the Digital Crown, the onWorldRecenter closure will be called multiple times, once for each past view body execution, including those of immersive space views that have been previously dismissed. Although these issues seem to be most prevalent when onWorldRecenter is used with an ImmersiveSpace, they may also occur in the context of a WindowGroup under certain circumstances. It's possible to work around this problem by moving onW
0
0
431
3d
visionOS Bluetooth LE limited to 2 connections?
Hello, Is there a 2-device limit for CoreBluetooth on visionOS 2.1? My app connects to 4 BLE peripherals on iOS but fails at the 3rd device on Vision Pro. The 3rd call to centralManager.connect() is successful and the peripheral enters .connecting state, but didConnect never fires and it stays in .connecting forever. No errors reported. First 2 devices work perfectly. Same code on iOS connects all 4. Has anyone else had this problem? Is there any documentation I can refer to that states something like this? Environment: visionOS 2.1, CoreBluetooth, Apple Vision Pro. My BLE Peripherals are running on nRF52840.
1
0
24
3d
Unable to profile Metal app on M2 Ultra (profiling works on M3 Pro)
On MacBook Pro M3 14 I can profile the Metal App performance by running it, then clicking on the M icon and choosing profile after replay. On Mac Studio M2 Ultra I cannot: the profiler starts and crashes. I have tried everything including reinstalling the OS, Xcode, the Metal SDK, you name it. The app uses the Metal 4 API. The content of the replayer errorinfo report is shown at the end. Any ideas what is going on here and/or what else I can do do root cause this and fix it? FWIW, it was worse on 26.1 (Xcode just reported Metal 4 profiling not available). In 26.2 Xcode attempts to profile and invariably crashes. === Error summary: === 1x DYErrorDomain (512) - guest app crashed (512) 1x com.apple.gputools.MTLReplayer (100) - Abort trap: 6 === First Error === Domain: DYErrorDomain Error code: 512 Description: guest app crashed (512) GTErrorKeyPID: 26913 GTErrorKeyProcessName: GPUToolsReplayService GTErrorKeyCrashDate: 2026-01-09 19:22:52 +0000 === Underlying Error #1 === Domain: com.apple.gputools.MTLR
1
0
292
3d
Unable to Open Apple Ads Account
Hi. I'm not sure this is relevant for developer forums. But I'm not sure where else to turn. I've been trying to open an Apple Ads account for nearly a month. Every time I try to submit the form, I get Server encountered an error processing your request followed by reason code: befe5482-e13e-4439-ae0d-badb3ca398a7_1767983516160.C67AC66. I can't contact Apple Ads, because every time I try to fill in the contact form I get Please correct the errors below. But there are no errors to correct. I've tried every possible combination of phone number formatting, special characters, spaces and all the usual suspects. Nothing works. I have a UK developer account in good standing. No issues, 2 live apps. I have reached out to Developer Support. No reply. I reached out to regular consumer support. They suggested Developer Support. I've got an app I'm trying to market and a budget ready to spend. Any ideas anyone?
1
0
73
3d
Reply to Small Business Program
Thanks for the post. This is normal given there was a long holiday 2 weeks ago. My modest recommendation is to contact them on their contact support page at: https://developer.apple.com/contact Posting in the developer forums when their support team is not actively monitoring may not yield any response. These forums are specifically designed for inquiries related to software development and accessories for Apple platforms. Your question appears to be related to Developer Account support once more. I recommend reaching out to the Developer Account support team at https://developer.apple.com/contact to ensure that they are aware of the issue. Hope this is helpful to you. Albert Pascual
  Worldwide Developer Relations.
3d
Family Controls Distribution Entitlement Request Taking Longer Than Expected - Any Tips?
Hi everyone, I'm hoping someone can share their experience or offer advice on entitlement request timelines. I previously had two bundle IDs approved for an app I'm testing via TestFlight - both were approved within a few days. I recently submitted a request for a third bundle ID (JMSHRM8W5J), and after realizing I may not have included enough detail, I submitted a follow-up request (XS2QYC59UU) with more context. It's now been almost three weeks, which is significantly longer than my earlier approvals - though I recognize some of that time included the holidays. A few questions for the community: Has anyone experienced longer wait times for additional entitlements on an existing project (with approved entitlements)? Did submitting a second request help or potentially slow things down? Is there anything I should include in a request to improve chances of quick approval? Any insight would be appreciated. Thanks!
2
0
409
3d
Reply to Inability to Communicate via APDU on iOS Despite NFC Tag Detection
It's probably not failing that silently. You should be checking for any errors returned from CoreNFC functions and see what they say. You could also be checking the Console logs for any errors displayed by the NFC system. You could check if your project configurations are correct by using the sample app and see if that works. For your code, you would want to seek for advice from the support channels for the development tools you are using.
Topic: App & System Services SubTopic: Core OS Tags:
3d
iOS 26: Underexposed image in exposure bracket appears clamped vs single capture
Hello, I have an iOS camera app that captures exposure brackets and performs custom HDR processing. On iOS 26, I’m observing a visual difference between: a single photo captured at –2 EV, and the –2 EV frame from an exposure bracket (–2 / 0 / +2 EV). On iOS 26: The single –2 EV image looks natural and consistent. The –2 EV image from the bracket appears clamped / distorted, most noticeably in high dynamic range scenes (highlight compression and loss of detail). On iOS 18, both approaches produce visually identical and correct –2 EV images. The issue only appears for bracketed captures on iOS 26. Attachments (examples) iOS 26 Single capture –2 EV (JPEG): /Users/danilobudimir/Downloads/ios26SingleImage/JPEG image-4006-8B77-51-0.jpeg Single capture –2 EV — Capture report (dumped settings): /Users/danilobudimir/Downloads/ios26SingleImage/UnderExposureDebug_CaptureReport_2026-01-09T15-59-20Z.md Bracket capture –2 EV frame (JP
2
0
396
3d