Hello, Starting in iOS 17, our application started having some issue publishing to our video session. More specifically the video capture seems to be broken in some, but not all sessions. What's troubling is that we're seeing that it fails consistently every 4 sessions. It also fails silently, without reporting any problems to the app. We only notice that there are no frames being rendered or sent to the remote devices. Here's what shows-up in the console: <<<< FigCaptureSourceRemote >>>> Fig assert: ! storage->connectionDied at bail (FigCaptureSourceRemote.m:235) - (err=0) <<<< FigCaptureSourceRemote >>>> Fig assert: err == 0 at bail (FigCaptureSourceRemote.m:253) - (err=-16453) Anyone seeing this? Any idea what could be the cause? Our sessions work perfectly on iOS16 and below. Thanks
Search results for
Visual Studio Maui IOS
105,637 results found
Selecting any option will automatically load the page
Post
Replies
Boosts
Views
Activity
Hello, I am currently implementing a biometric authentication registration flow using WebAuthn. I am using ASAuthorizationPlatformPublicKeyCredentialRegistrationRequest, and I would like to know if there is a way to hide the Save to another device option that appears during the registration process. Specifically, I want to guide users to save the passkey only locally on their device, without prompting them to save it to iCloud Keychain or another device. If there is a way to hide this option or if there is a recommended approach to achieve this, I would greatly appreciate your guidance. Also, if this is not possible due to iOS version or API limitations, I would be grateful if you could share any best practices for limiting user options in this scenario. If anyone has experienced a similar issue, your advice would be very helpful. Thank you in advance.
I'm working on integrating Passkey functionality into my iOS app (targeting iOS 16.0+), and I'm facing an issue where the system dialog still shows the Save to another device option during Passkey registration. I want to hide this option to force users to create Passkeys only on the current device. 1. My Current Registration Implementation Here’s the code I’m using to create a Passkey registration request. I’ve tried to use ASAuthorizationPlatformPublicKeyCredentialProvider (which is supposed to target platform authenticators like Face ID/Touch ID), but the Save to another device option still appears: `// Initialize provider for platform authenticators let provider = ASAuthorizationPlatformPublicKeyCredentialProvider(relyingPartyIdentifier: domain) // Create registration request let registrationRequest = provider.createCredentialRegistrationRequest( challenge: challenge, name: username, userID: userId ) // Optional configurations (tried these but no effect on another device option) registrat
Topic:
Privacy & Security
SubTopic:
General
iOS 26.1, the isSourceTypeAvailable: UIImagePickerControlSourceTypeCamera method keeps returning true when the camera is unavailable. Just to be sure I'm clear on the issue, the camera itself is disabled, but your app ends up showing the camera UI (because we said the camera was there)? Is it coming up with a black screen, or does something else happen? Do you have any good solution? Or can we just wait for the bug to be fixed? You might try checking AVCaptureDevice.authorizationStatus(for:), as it's possible that the lower-level API will give you a different answer, particularly if you're getting a black screen from UIImagePicker. If that doesn't work, then I think your only option is to wait for a fix. __ Kevin Elliott DTS Engineer, CoreOS/Hardware
Topic:
App & System Services
SubTopic:
Hardware
Tags:
I have an odd issue I'm trying to troubleshoot. I have an app that is deployed to our enterprise and works on almost all of our devices. Lately, I have devices (very small number) where the app installs from our MDM and it crashes upon launch, it does not get past the launch screen. If I remove and reinstall the app from the MDM it still crashes. I decided to put the devices in developer mode and if I run the App from Xcode on the device it works fine. If I stop it and run it right from the device it also works. If I do a final build of the app and install it using Devices and Simulators, the app crashes upon launch. Using Devices and Simulators I check for crash log for the app and no file. I am stumped as to whats going on.
While using the Camera app (in the camera modes described in the preferences under the setting you referenced) there is a control that allows you to select ProRAW format in the upper left corner of the screen (iOS 26.0.1). With that said, thank you for asking, but, these forums are for developers talking about making software programs like the Camera and that's a very different subject than using and setting up the preferences for the camera. You'll probably get more and better answers to this type of question in the Apple Support Community forums. For help with using the Camera, please post in the Apple Support Community. The Apple Developer Forums are for questions about APIs and features intended specifically for developers.
Topic:
App & System Services
SubTopic:
General
I am using iPhone 16 pro, and I am running on the IOS 26.1 (23B5064e) in settings, going to Camera, then to formats, then in “Pro Raw and Resolution Control”, I am only getting “Pro Raw Format”(JPEG Lossless Most Compatible), but despite after numerous attempts I am unable to find the option for “Pro Raw Max”(upto 48 MP) or HEIF Max (upto 48 MP); please help me anyone I desperately want “Pro Raw Max”, moreover I cannot find the “Raw Max” option in the camera app, but only getting the “Raw” option; thanks a lot in advance…🙏🏻🙏🏻🙏🏻🙏🏻
Topic:
App & System Services
SubTopic:
General
Thank you for your post. There are a few details that need clarification. Could you please provide the iOS version you are using? Additionally, how do you ascertain that the app is not receiving the token? A more detailed description of how to initiate the activity or if the user initiates the activity would be greatly appreciated. Do you have logs on these method to see if gets called? On the TASK I would add some logs to see when it gets called. https://developer.apple.com/documentation/activitykit/starting-and-updating-live-activities-with-activitykit-push-notifications Additionally, it is crucial to trace the issue if possible. Kindly provide me with the following information for a failed notification: The precise time and date (including the time zone) of the push notification. The public IPv4 address of the server that sent the push notification. The push topic (apns-topic header value). The app’s Bundle ID. The device token. The apns-id if it is set. The full contents of the headers and the pa
Topic:
App & System Services
SubTopic:
Notifications
Tags:
Hello, As I've been tinkering with the new URL filtering API I've been struggling to create bloom filters. I have been referencing this developer's post heavily: https://developer.apple.com/forums/thread/791352?answerId=851527022#851527022 He suggests that the expected FNV-1a implementation has the multiply and xor operations inverted, while an Apple engineer suggests that this will be updated on the offical iOS26 release (which has already happened). What is the correct FNV-1a implementation? In this sample project (which is from july, so pre-release ios 26): https://developer.apple.com/documentation/networkextension/filtering-traffic-by-url Is the included bloom filter data correct? I can only assume the other developer used this as a reference to then conclude that multiplying and xor-ing should be inverted, so I'm really not sure if this bloom filter bitVectorData here is trustworthy. Is there any reference implementation for this hashing procedure? How do we generate a bloom filter properly and
I use ARKit for motion tracking. I get the skeleton joint coordinates and use them for animation. I didn't make any changes to the code, but I updated the iOS version from 18 to 26, and modelTransform now always returns nil. https://developer.apple.com/documentation/arkit/arskeleton3d/modeltransform(for:) For example bodyAnchor.skeleton.modelTransform(for: .init(rawValue: head_joint)) bodyAnchor is ARBodyAnchor. I see the default skeleton on the screen, but now I can't get the coordinates out of it. I'm using an example from Apple's WWDC presentation. https://developer.apple.com/documentation/arkit/capturing-body-motion-in-3d Are there any changes in the API? Or just bug?
The value shown there is the LLVM/clang/swiftc target triple that Swift Testing used when Apple compiled it. Swift Testing's minimum deployment target (the lowest version of the OS you can run it on) is iOS 13, so the target triple shows iOS 13.
Topic:
Programming Languages
SubTopic:
Swift
Tags:
I am working on a Gaussian Splatting App using the UnityGaussianSplatting package and building for Apple Vision Pro. I want to be able to load splat (.ply / .splat) files from the Vision Pro's storage into the application. OK. So the main issue to be aware of here is the difference between works fine for a developer and works well for a user. For development purposes, you can basically just set the two keys (UIFileSharingEnabled/LSSupportsOpeningDocumentsInPlace) and then use your Documents directory as your working storage. There could be issues if you edited or deleted those files while your app was actively running, but you will just avoid those issues... by not doing that. On the other hand, if you're planning to ship this to end users then you'll need to use things like file coordination to avoid those issues. Similarly, apps that are more viewer oriented (meaning, they don't edit the files they're working with) often use an import model where they copy (actually clone, so they don't use additional space
Topic:
Spatial Computing
SubTopic:
General
Tags:
Hi. I‘m making an app with SwiftUI for iOS 26, and found that the old ways of changing unselected icons color seem not working as they did in old versions of iOS. I tried these methods: TabView() { Tab { // some view here } label: { Label(title: { Text(something) }, icon: { Image(systemName: checkmark.seal.fill) .foregroundStyle(.blue) } } } I wrapped Image with an if-else, but itisn't able to change any color even without if; struct ParentView: View { init() { UITabBar.appearance() .unselectedItemTintColor = UIColor.red } var body: some View { TabView() { // some tabs here } } } and an extension of above struct ParentView: View { init() { let tabBarAppearance = UITabBarAppearance() tabBarAppearance.configureWithOpaqueBackground() tabBarAppearance.backgroundColor = UIColor.green tabBarAppearance.stackedLayoutAppearance .selected.titleTextAttributes = [.foregroundColor: UIColor.red] tabBarAppearance.stackedLayoutAppearance.normal .titleTextAttributes = [.foregroundColor: UIColor.black] tabBar
With macOS Tahoe, Launchpad has been replaced by an App Library–style mode within Spotlight. While the alleged intention is UX consistency across the Apple ecosystem, the result is both a catastrophic usability regression and a radical break in consistency with iOS and iPadOS. Predefined App Library categorization is functionally incoherent: On iOS and now macOS, Apple’s predefined App Library categories place apps with seemingly identical functionality into unrelated groups—for example, 3D scanning tools scattered across Education, Utilities, and Productivity. Instead of making apps easier to find, this effectively creates a labyrinth that users must traverse to locate apps whose names and icons they may not recall. However Apple defines its app categories, they are not only inconsistent but also hopelessly inadequate for the long tail of real-world applications and user workflows. Loss of user control: Launchpad enabled users to group and organize applications according to their workflows.
Hello I am developing an iOS app and would like to read an EMV card that is issued by me and want my customers to tap to activate simply to validate the possession of the card at the time of activation. Any suggestions welcome on how can i achieve this using either NFC Framework or secure elements