Apple Developers

RSS for tag

This is a dedicated space for developers to connect, share ideas, collaborate, and ask questions. Introduce yourself, network with other developers, and foster a supportive community.

Learn More

Posts under Apple Developers subtopic

Post

Replies

Boosts

Views

Activity

Camera's settings at calibration time
Hi everyone, I am wondering under which settings the camera(s) were set by the time they were calibrated. For instance, one aspect that is easy to find is the reference resolution of the images taken when calibrating the intrinsics, this is by retrieving intrinsicMatrixReferenceDimensions. Making sure that the principal point is referenced to the by the time resolution used when the calibration was ongoing. However, recently I saw that there are focusing modes that potentially displace the lens' physical position. Settings like: AutoFocusRangeRestriction: none, near, far setFocusModeLocked: Locks the lens position at the specified value, and sets the focus mode to a locked state. My concern lies the impact this focusing lens displacements have on the intrinsic matrix parameters, if the lens is displaced, these parameters no longer describe the camera since the lens position has changed w.r.t. the lensPosition set when they were calibrated [0-1]. If my understanding is correct the AutoFocusRangeRestriction is just a range freedom the system is allowed to auto-focus and not a specific lens position. Conversely, the setFocusModeLocked does indeed fix the lensPosition to a certain value [0 - 1]. In simple words, what is the focus lensPosition the cameras were set when calibrating them for intrnisics?
0
0
455
Jan ’25
App development agency bankruptcy, what about our apps?
First, I apologize if I post to the wrong forum, but this is kinda new to me. Backstory: I recently joined a company as CTO and the first topic on my agenda is the mobile presence. The android app was built in-house, but the iOS app was outsourced to an app/web development agency. The problem: the aforementioned agency ceased operations in mid-2024 without notifying us. They cancelled their dev program enrollment and our apps are no longer available on the App Store. And the agency reps are not responding to emails and phone calls. My question: is there a way to get access to our application? We plan in-house redesign and refactoring in 2025, but we want access to the bundle ID, the app groups, the iCloud container, etc. We can present legal documents, company website and branding, everything to prove that app is technically ours. Thanks in advance! Have a good and productive 2025!
0
0
294
Jan ’25
Please support device mirroring with cable and without iCloud login
Hi all, if anyone from apple Dev team is seeing this, please do consider supporting device mirroring without the need to login to icloud and with cable connection in addition to wi-fi. I am working in a corporate setting which does not allow us to login to icloud, thus I am unable to use device mirroring when my work really needs is, I am stuck with Quick time preview. I am quite sure, that this does not only apply to me.
1
0
293
Jan ’25
Globe key cannot switch to iOS emoji keyboard
Scenario: hardware keyboard is attached to iPhone Before iOS 18.2, if I setup only two keyboards (e.g., english and emoji) in the system, click Globe key can switch to iOS emoji keyboard without problem. Since iOS 18.2, it stop working, and emoji keyboard is never displayed. I need to add one more keyboard, then I can switch to emoji keyboard (via the keyboard selector).
0
0
171
Jan ’25
Beta update 18.3 on IOS
Beta Update Feedback: iOS 18.3 The iOS 18.3 beta update has been a deeply frustrating experience due to numerous critical issues that severely impact the device's functionality. Below is a detailed report of the problems encountered: All major social applications, including WhatsApp, Instagram, and Photos, consistently crash. When launching these apps, the screen often turns dark before reverting to the home screen. This issue persists across other third-party and native apps, making the device unusable for basic tasks. The system frequently hangs, failing to process basic operations effectively. Restarting the device is often the only temporary solution; however, even after multiple restarts, the issues remain unresolved. Significant charging drops are observed, and the device intermittently fails to recognize the charger. Charging efficiency has drastically decreased, creating inconvenience for daily usage. The Photos app is riddled with bugs, making it nearly impossible to use. Basic functions such as viewing, editing, or sharing photos do not work properly. Every aspect of the update feels unstable, with issues cropping up in every corner of the system. The phone’s functionality has deteriorated, so it no longer performs as expected. Device Affected: iPhone 13 Impact Level: Severe This update has caused immense frustration and has made the iPhone 13 difficult.
2
1
2.3k
Jan ’25
Can't swap camera with AVCaptureMultiCamSession
I have a PIP camera that is streaming from the front and back based on AVCaptureMultiCamSession. It works fine, but when i go to swap the camera it crashes. This is code that works with a single camera, so not sure what is wrong. Also the object appears valid in the debugger. This is the snippit where the camera is swapped private func updateSessionConfiguration() { guard isCaptureSessionConfigured else { return } captureSession.beginConfiguration() defer { captureSession.commitConfiguration() } // Remove all current inputs for input in captureSession.inputs { if let deviceInput = input as? AVCaptureDeviceInput { captureSession.removeInput(deviceInput) app_log("removing input for \(input)") } } // Add the primary device input if let deviceInput = deviceInputFor(device: captureDevice) { app_log("device input \(deviceInput)") if !captureSession.inputs.contains(deviceInput), captureSession.canAddInput(deviceInput) { captureSession.addInput(deviceInput) } } if let secondaryDeviceInput = deviceInputFor(device: secondaryCaptureDevice) { app_log("Secondary device input \(secondaryDeviceInput)") if !captureSession.inputs.contains(secondaryDeviceInput), captureSession.canAddInput(secondaryDeviceInput) { captureSession.addInput(secondaryDeviceInput) } } updateVideoOutputConnection() } It crashes at: captureSession.addInput(deviceInput) with: Thread 10: EXC_BAD_ACCESS (code=1, address=0xcaeb36b964f0) which is strange because canAdd is checked prior to this call. Totally stumped here. Please help. Not sure if this is an AVCaptureMuliCamSession issue or something.
4
0
425
Jan ’25
Can a Default Messaging App Handle Legacy SMS Messages on iOS
I have been reviewing the documentation on preparing your app to be the default messaging app, and I am wondering about the following: If I set my app as the default messaging app, will it be able to send and receive legacy SMS messages? Does iOS provide any SMS-specific framework or APIs to facilitate handling SMS messages directly within my app? I’d appreciate any clarification on whether iOS supports such functionality and, if so, how to integrate it into my app. Thank you!
0
0
198
Jan ’25
Third party music apps crashing when trying to open them
I am having some serious issues with all of my music apps on the latest beta of iPadOS 18.3. when I try to open them , the splash page pops open , freezes then crashes. i can open Logic Pro and it opens fine, but if I try to open a project with any of my third party music apps in that project, Logic gives me an error message stating that it can’t access coremidi. It gives me the option to continue and then proceeds to play the project without issue, with all of the third party apps functioning within Logic. If I try to open a third party program outside of Logic ( that just played in Logic) the 3rd party app crashes. Anyone else having issues with their 3rd party music apps? I have never had so many apps in a particular genre that just wouldn’t function in any of the recent betas.
6
2
661
Jan ’25
Using txt files with searching strings for an iOS app
When dealing with SwiftUI and searchable modifier, I know you can use URL or hard code data to search when building projects in Xcode. I am looking to see if you can use a txt file as a way or storing string data of lists to search from when your device is offline when using the search modifier. Then when connected to internet you can update the search with url connection that then updates the txt file so you can do new searches the next time you are offline again. Is this something that is possible?
0
0
248
Jan ’25
Need clarification on using Apple MapKit in CarPlay
I have developed a mobile app using SwiftUI that supports GoogleMaps. Now I am in the process of building a CarPlay application. I assume CarPlay only supports Apple MapKit, as I could not find any way to integrate the Google Maps. Below are few queries, Could you please guide me on how I can obtain the user's current location on the CarPlay app launch? Is there a way CarPlay can get the details from the mobile app(not pretty sure as its using Google Maps)? If the user is logged out from the mobile app, what is the flow in CarPlay? Do we have any standard login page asking user to login to the mobile app first? Is there any UI asking the user to capture the location in CarPlay? This is my first CarPlay app. Kindly guide me to a document or so that covers these details. Thanks a ton!!
1
0
457
Jan ’25
Firebase Code Help
Hello Developer Community. I need help with Firebase. I have Problems with a Code, so if you can help me, simply reply on this Post and then I share the Code. Thanks for the Help :D
0
0
188
Jan ’25
iOS 18.3 beta
I would like to report a bug issues with the sound on the speakers I notice there is no sound coming out when watching a video and also that YouTube videos freezes and does not play I would appreciate it if you fix the bug and put you fix this and put a re-release beta software- Thank you apple for your time
1
0
454
Jan ’25
CoreAudio: Audio Output Device Stopped Error since macOS 15
Hello! I used the Apple CA Playthrough example code that pipes audio between devices. It uses AudioUnit callbacks to pipe the input to an output device, and I created a system equalizer with it - however users reported it stopped working in macOS 15. I am getting the error HALPlugIn.cpp:552 HALPlugIn::DeviceGetCurrentTime: got an error from the plug-in routine, Error: 1937010544 (stop) for the output device and no sound coming out of the speakers. The error only occurs when using a virtual device as an input, not using the microphone. First I thought the problem was in the loopback driver, but it also does not work with other loopback drivers like Blackhole. STEPS TO REPRODUCE Install a virtual device, for example "brew install blackhole-2ch" and run the CAPlayThrough example code (you need to add Mic Permission in the info.plist). Then set your system audio output to the virtual device, select the device as input in CAPlayThrough and hit start. You should see the error in console. My question: What did change in macOS 15 that could cause this? Is it something with the new permission handling maybe?
0
1
574
Jan ’25
Question on Developer Membership?
I am about to buy the yearly membership. My good friend was going to work on a project with me. If I start the project with the membership, would he need to also have a membership to collaborate on the project together? Really just seeing if we need two memberships to get started or just one.
1
0
267
Jan ’25