Search results for

“Visual Studio Maui IOS”

109,084 results found

Post

Replies

Boosts

Views

Activity

Reply to SpeechTranscriber/SpeechAnalyzer being relatively slow compared to FoundationModel and TTS
I've been optimizing a similar STT-to-action pipeline on macOS 26 and found a few additional tricks beyond prepareToAnalyze that helped bring the finalization latency down: Use volatileResults aggressively for UI feedback, but trigger your downstream action (FoundationModel call) on the volatile transcript as soon as it stabilizes — don't wait for the finalized event. In my testing, the volatile transcript matches the final one ~95% of the time for short utterances. You can always correct if the final differs. Audio format matters more than you'd expect. If your input is coming through at 48kHz (common from ScreenCaptureKit or external mics), the internal resample to 16kHz adds measurable overhead. Setting up your AVAudioEngine tap at 16kHz mono from the start shaves ~200ms off the pipeline. The large variance Bersaelor observed with prepareToAnalyze (0.05s to 3s) likely correlates with whether the ANE was already warm. If other CoreML workloads are running concurrently (even system ones like Visual
Topic: Media Technologies SubTopic: Audio Tags:
1w
CoreML regression between macOS 26.0.1 and macOS 26.1 Beta causing scrambled tensor outputs
We’ve encountered what appears to be a CoreML regression between macOS 26.0.1 and macOS 26.1 Beta. In macOS 26.0.1, CoreML models run and produce correct results. However, in macOS 26.1 Beta, the same models produce scrambled or corrupted outputs, suggesting that tensor memory is being read or written incorrectly. The behavior is consistent with a low-level stride or pointer arithmetic issue — for example, using 16-bit strides on 32-bit data or other mismatches in tensor layout handling. Reproduction Install ON1 Photo RAW 2026 or ON1 Resize 2026 on macOS 26.0.1. Use the newest Highest Quality resize model, which is Stable Diffusion–based and runs through CoreML. Observe correct, high-quality results. Upgrade to macOS 26.1 Beta and run the same operation again. The output becomes visually scrambled or corrupted. We are also seeing similar issues with another Stable Diffusion UNet model that previously worked correctly on macOS 26.0.1. This suggests the regression may affect multiple diffusion-style ar
8
0
2.2k
1w
Reply to iOS 18 new RecognizedTextRequest DEADLOCKS if more than 2 are run in parallel
I've been working with the new Swift Vision API's RecognizeTextRequest on iOS 18 and hit this exact deadlock. After profiling with Instruments, I found that the Vision framework internally uses a limited thread pool for its neural engine requests — on most devices this caps at 2 concurrent ANE inference sessions. The workaround I'm using is a semaphore-based concurrency limiter that queues requests: actor OCRPipeline { private let maxConcurrent = 2 private var running = 0 private var pending: [CheckedContinuation] = [] func recognizeText(in image: CGImage) async throws -> [String] { await acquireSlot() defer { Task { await releaseSlot() } } let request = RecognizeTextRequest() let handler = ImageRequestHandler(image) let observations = try await handler.perform(request) return observations.compactMap { $0.topCandidates(1).first?.string } } } This keeps throughput high while never exceeding the 2-concurrent-request limit. In my testing across iPhone 15 Pro and iPad Air M2, this processes ~40 images
Topic: Machine Learning & AI SubTopic: General Tags:
1w
iOS 18 new RecognizedTextRequest DEADLOCKS if more than 2 are run in parallel
Following WWDC24 video Discover Swift enhancements in the Vision framework recommendations (cfr video at 10'41), I used the following code to perform multiple new iOS 18 `RecognizedTextRequest' in parallel. Problem: if more than 2 request are run in parallel, the request will hang, leaving the app in a state where no more requests can be started. -> deadlock I tried other ways to run the requests, but no matter the method employed, or what device I use: no more than 2 requests can ever be run in parallel. func triggerDeadlock() {} try await withThrowingTaskGroup(of: Void.self) { group in // See: WWDC 2024 Discover Siwft enhancements in the Vision framework at 10:41 // ############## THIS IS KEY let maxOCRTasks = 5 // On a real-device, if more than 2 RecognizeTextRequest are launched in parallel using tasks, the request hangs // ############## THIS IS KEY for idx in 0.. [RecognizedText] { // Create request var request = RecognizeTextRequest() // Single request: no need for ImageRequestHandler // Co
8
0
504
1w
Reply to How to uncheck "Connect via network" on physical device in Xcode 15
I guess that's nice of them to intend to break debugging over USB. I'm still hoping they'll intend to, perhaps, add a new feature to allow unchecking the box to Connect via network, which disables using TCP/IP and re-enables debugging over USB as it used to work. Edit: I now see Xcode 26 has removed the checkbox for devices with iOS >17. I guess my request would now be to put the checkbox back, and enable to to be used with recent iOS releases.
1w
Xcode 26.4 breaks compilation
With the latest Xcode 26.4 my project fails to compile due to a compiler unable to type-check in reasonable time error with a library I've been using for years and has not changed. The library is https://github.com/ngageoint/mgrs-ios and is specifically around this line https://github.com/ngageoint/mgrs-ios/blob/master/mgrs-ios/utm/UTM.swift#L96. Since this isn't my library I can't change it, but again we've been using this library for years now without issue - it was only when Xcode updated this morning that our compilation workflow broke. (Compilation machine is an M3 Pro with 18GB of RAM, so I don't think that's the problem here)
3
0
216
1w
Reply to Advanced App Clip Experiences stuck in "Received" — never transition to "Published"
Thanks for the post, interesting that's something in the process. TN3168: Making your App Clip available in the App Store | Apple Developer Documentation https://developer.apple.com/forums/thread/804858 There is something I always like to check when deploying AASA files. When the full app is installed, iOS handles the URL invocation locally via your apple-app-site-association (AASA) file and entitlements, bypassing the need to query Apple's App Clip CDN. The fact that this works proves your code, AASA file, and associated domains are configured correctly. The failure for fresh devices is happening because Apple's CDN has not yet been updated with your App Clip's metadata? Can you provide me a link to your AASA file in your server so we can see why is not syncing into the Apple servers? Is your server preventing or blocking IP address or user agents from requesting the AASA file? Sometimes, the job queue in App Store Connect simply drops the task. You can often force it to re-queue by making a trivial
1w
Xcode 26.4 and 26.3: Swift compiler crashes during archive with iOS 17 deployment target
I submitted this to Feedback Assistant as FB22331090 and wanted to share a minimal reproducible example here in case others are seeing the same issue. Environment: Xcode 26.4 or Xcode 26.3 Apple Swift version 6.3 (swiftlang-6.3.0.123.5 clang-2100.0.123.102) Effective Swift language version 5.10 Deployment target: iOS 17.0 The sample app builds successfully for normal development use, but archive fails. The crash happens during optimization in EarlyPerfInliner while compiling the synthesized deinit of a nested generic Coordinator class. The coordinator stores a UIHostingController. Minimal reproducer: import SwiftUI struct ReproView: UIViewRepresentable { let content: Content func makeUIView(context: Context) -> UIView { context.coordinator.host.view } func updateUIView(_ uiView: UIView, context: Context) { context.coordinator.host.rootView = content } func makeCoordinator() -> Coordinator { Coordinator(content: content) } final class Coordinator { let host: UIHostingController init(content: Con
1
0
69
1w
The Proliferation of DarkSword: iOS Exploit Chain Adopted by Multiple Threat Actors
Dear Team, Can you please check and confirm for this security Google Threat Intelligence Group (GTIG) has identified a new iOS full-chain exploit that leveraged multiple zero-day vulnerabilities to fully compromise devices. Based on toolmarks in recovered payloads, we believe the exploit chain to be called DarkSword. Since at least November 2025, GTIG has observed multiple commercial surveillance vendors and suspected state-sponsored actors utilizing DarkSword in distinct campaigns. These threat actors have deployed the exploit chain against targets in Saudi Arabia, Turkey, Malaysia, and Ukraine. https://cloud.google.com/blog/topics/threat-intelligence/darksword-ios-exploit-chain?hl=en
1
0
223
1w
Family Controls entitlement request submitted on March 9, 2026 — no response or status update
Hi, I submitted a Family Controls (Distribution) entitlement request on March 9, 2026 for my iOS app, but I still have not received any approval, rejection, or other status update. At this point, I’m mainly trying to understand: whether this waiting time is currently normal, whether there is any way to check if the request is actually under review, and whether Apple provides any follow-up if more information is needed. This is blocking my progress, because the app depends on the Screen Time / Family Controls APIs. Has anyone recently experienced similar delays, and is there any recommended next step besides waiting? Thanks. Imi
6
0
437
1w
Please, Apple. I am begging you. Fix the broken Text-To-Speech in macOS
Every new build of macOS 26 further breaks some part of text-to-speech or voice control. I have filed multiple bug reports on this, yet the situation gets worse, not better, with each new build. I am begging you to fix this! I am disabled and rely on these features to work. Accessibility in macOS is more than 50% of my reasons for choosing Macs over Windows. Here's some of what is currently broken: Speak announcements only works if the Samantha voice is selected. If other voices are selected either the announcements don't speak, or they default to the Samantha voice. This became broken with 26.1. Announce the time only works with some voices. I would like to use Australian English Siri 2 but with that voice selected it defaults to Samantha. This became broken with 26.4. With voice control enabled there are two menu bar icons. The blue voice control icon and an orange microphone an application is accessing the microphone icon. This orange icon started appearing with 26.3. For four years of macOS releases, the
9
0
2.5k
1w
Reply to SpeechTranscriber/SpeechAnalyzer being relatively slow compared to FoundationModel and TTS
I've been optimizing a similar STT-to-action pipeline on macOS 26 and found a few additional tricks beyond prepareToAnalyze that helped bring the finalization latency down: Use volatileResults aggressively for UI feedback, but trigger your downstream action (FoundationModel call) on the volatile transcript as soon as it stabilizes — don't wait for the finalized event. In my testing, the volatile transcript matches the final one ~95% of the time for short utterances. You can always correct if the final differs. Audio format matters more than you'd expect. If your input is coming through at 48kHz (common from ScreenCaptureKit or external mics), the internal resample to 16kHz adds measurable overhead. Setting up your AVAudioEngine tap at 16kHz mono from the start shaves ~200ms off the pipeline. The large variance Bersaelor observed with prepareToAnalyze (0.05s to 3s) likely correlates with whether the ANE was already warm. If other CoreML workloads are running concurrently (even system ones like Visual
Topic: Media Technologies SubTopic: Audio Tags:
Replies
Boosts
Views
Activity
1w
CoreML regression between macOS 26.0.1 and macOS 26.1 Beta causing scrambled tensor outputs
We’ve encountered what appears to be a CoreML regression between macOS 26.0.1 and macOS 26.1 Beta. In macOS 26.0.1, CoreML models run and produce correct results. However, in macOS 26.1 Beta, the same models produce scrambled or corrupted outputs, suggesting that tensor memory is being read or written incorrectly. The behavior is consistent with a low-level stride or pointer arithmetic issue — for example, using 16-bit strides on 32-bit data or other mismatches in tensor layout handling. Reproduction Install ON1 Photo RAW 2026 or ON1 Resize 2026 on macOS 26.0.1. Use the newest Highest Quality resize model, which is Stable Diffusion–based and runs through CoreML. Observe correct, high-quality results. Upgrade to macOS 26.1 Beta and run the same operation again. The output becomes visually scrambled or corrupted. We are also seeing similar issues with another Stable Diffusion UNet model that previously worked correctly on macOS 26.0.1. This suggests the regression may affect multiple diffusion-style ar
Replies
8
Boosts
0
Views
2.2k
Activity
1w
Reply to iOS 18 new RecognizedTextRequest DEADLOCKS if more than 2 are run in parallel
I've been working with the new Swift Vision API's RecognizeTextRequest on iOS 18 and hit this exact deadlock. After profiling with Instruments, I found that the Vision framework internally uses a limited thread pool for its neural engine requests — on most devices this caps at 2 concurrent ANE inference sessions. The workaround I'm using is a semaphore-based concurrency limiter that queues requests: actor OCRPipeline { private let maxConcurrent = 2 private var running = 0 private var pending: [CheckedContinuation] = [] func recognizeText(in image: CGImage) async throws -> [String] { await acquireSlot() defer { Task { await releaseSlot() } } let request = RecognizeTextRequest() let handler = ImageRequestHandler(image) let observations = try await handler.perform(request) return observations.compactMap { $0.topCandidates(1).first?.string } } } This keeps throughput high while never exceeding the 2-concurrent-request limit. In my testing across iPhone 15 Pro and iPad Air M2, this processes ~40 images
Topic: Machine Learning & AI SubTopic: General Tags:
Replies
Boosts
Views
Activity
1w
iOS 18 new RecognizedTextRequest DEADLOCKS if more than 2 are run in parallel
Following WWDC24 video Discover Swift enhancements in the Vision framework recommendations (cfr video at 10'41), I used the following code to perform multiple new iOS 18 `RecognizedTextRequest' in parallel. Problem: if more than 2 request are run in parallel, the request will hang, leaving the app in a state where no more requests can be started. -> deadlock I tried other ways to run the requests, but no matter the method employed, or what device I use: no more than 2 requests can ever be run in parallel. func triggerDeadlock() {} try await withThrowingTaskGroup(of: Void.self) { group in // See: WWDC 2024 Discover Siwft enhancements in the Vision framework at 10:41 // ############## THIS IS KEY let maxOCRTasks = 5 // On a real-device, if more than 2 RecognizeTextRequest are launched in parallel using tasks, the request hangs // ############## THIS IS KEY for idx in 0.. [RecognizedText] { // Create request var request = RecognizeTextRequest() // Single request: no need for ImageRequestHandler // Co
Replies
8
Boosts
0
Views
504
Activity
1w
Reply to Clipboard issues with simulators
This issue happens in the latest iOS 17, 18, and 26 simulators. I have not tested older simulators like iOS 15 or 16. I am using Xcode 26.4 (17E192), Simulator 16.0 (1063.2) (SimulatorKit 955.7, CoreSimulator 1051.49) and macOS Tahoe 26.4.
Replies
Boosts
Views
Activity
1w
Reply to iCloud Sync not working with iPhone, works fine for Mac.
For debugging purpose, would you mind to check if it's that the macOS app failed to export data to CloudKit, or that the iOS app failed to import from CloudKit? The detailed steps to do the check are documented here. Best, —— Ziqiao Chen  Worldwide Developer Relations.
Replies
Boosts
Views
Activity
1w
How to uncheck "Connect via network" on physical device in Xcode 15
It looks like the other thread got locked. I would still like to be able to uncheck the box for Connect via network so I can develop with my iOS 17+ devices normally with Xcode 15. Or a method to use Xcode 14 with iOS 17+ without CoreDevice or whatever the problematic API is.
Replies
4
Boosts
0
Views
691
Activity
1w
Reply to How to uncheck "Connect via network" on physical device in Xcode 15
I guess that's nice of them to intend to break debugging over USB. I'm still hoping they'll intend to, perhaps, add a new feature to allow unchecking the box to Connect via network, which disables using TCP/IP and re-enables debugging over USB as it used to work. Edit: I now see Xcode 26 has removed the checkbox for devices with iOS >17. I guess my request would now be to put the checkbox back, and enable to to be used with recent iOS releases.
Replies
Boosts
Views
Activity
1w
Xcode 26.4 breaks compilation
With the latest Xcode 26.4 my project fails to compile due to a compiler unable to type-check in reasonable time error with a library I've been using for years and has not changed. The library is https://github.com/ngageoint/mgrs-ios and is specifically around this line https://github.com/ngageoint/mgrs-ios/blob/master/mgrs-ios/utm/UTM.swift#L96. Since this isn't my library I can't change it, but again we've been using this library for years now without issue - it was only when Xcode updated this morning that our compilation workflow broke. (Compilation machine is an M3 Pro with 18GB of RAM, so I don't think that's the problem here)
Replies
3
Boosts
0
Views
216
Activity
1w
Reply to Xcode 26.4 breaks compilation
I used to be on the team that wrote that library. I have broken the crazy expression down just so that I can get my tests to compile, but I have not fully verified it. However, if this gets you further... https://github.com/danielbarela/mgrs-ios/blob/master/mgrs-ios/utm/UTM.swift
Replies
Boosts
Views
Activity
1w
Reply to Advanced App Clip Experiences stuck in "Received" — never transition to "Published"
Thanks for the post, interesting that's something in the process. TN3168: Making your App Clip available in the App Store | Apple Developer Documentation https://developer.apple.com/forums/thread/804858 There is something I always like to check when deploying AASA files. When the full app is installed, iOS handles the URL invocation locally via your apple-app-site-association (AASA) file and entitlements, bypassing the need to query Apple's App Clip CDN. The fact that this works proves your code, AASA file, and associated domains are configured correctly. The failure for fresh devices is happening because Apple's CDN has not yet been updated with your App Clip's metadata? Can you provide me a link to your AASA file in your server so we can see why is not syncing into the Apple servers? Is your server preventing or blocking IP address or user agents from requesting the AASA file? Sometimes, the job queue in App Store Connect simply drops the task. You can often force it to re-queue by making a trivial
Replies
Boosts
Views
Activity
1w
Xcode 26.4 and 26.3: Swift compiler crashes during archive with iOS 17 deployment target
I submitted this to Feedback Assistant as FB22331090 and wanted to share a minimal reproducible example here in case others are seeing the same issue. Environment: Xcode 26.4 or Xcode 26.3 Apple Swift version 6.3 (swiftlang-6.3.0.123.5 clang-2100.0.123.102) Effective Swift language version 5.10 Deployment target: iOS 17.0 The sample app builds successfully for normal development use, but archive fails. The crash happens during optimization in EarlyPerfInliner while compiling the synthesized deinit of a nested generic Coordinator class. The coordinator stores a UIHostingController. Minimal reproducer: import SwiftUI struct ReproView: UIViewRepresentable { let content: Content func makeUIView(context: Context) -> UIView { context.coordinator.host.view } func updateUIView(_ uiView: UIView, context: Context) { context.coordinator.host.rootView = content } func makeCoordinator() -> Coordinator { Coordinator(content: content) } final class Coordinator { let host: UIHostingController init(content: Con
Replies
1
Boosts
0
Views
69
Activity
1w
The Proliferation of DarkSword: iOS Exploit Chain Adopted by Multiple Threat Actors
Dear Team, Can you please check and confirm for this security Google Threat Intelligence Group (GTIG) has identified a new iOS full-chain exploit that leveraged multiple zero-day vulnerabilities to fully compromise devices. Based on toolmarks in recovered payloads, we believe the exploit chain to be called DarkSword. Since at least November 2025, GTIG has observed multiple commercial surveillance vendors and suspected state-sponsored actors utilizing DarkSword in distinct campaigns. These threat actors have deployed the exploit chain against targets in Saudi Arabia, Turkey, Malaysia, and Ukraine. https://cloud.google.com/blog/topics/threat-intelligence/darksword-ios-exploit-chain?hl=en
Replies
1
Boosts
0
Views
223
Activity
1w
Family Controls entitlement request submitted on March 9, 2026 — no response or status update
Hi, I submitted a Family Controls (Distribution) entitlement request on March 9, 2026 for my iOS app, but I still have not received any approval, rejection, or other status update. At this point, I’m mainly trying to understand: whether this waiting time is currently normal, whether there is any way to check if the request is actually under review, and whether Apple provides any follow-up if more information is needed. This is blocking my progress, because the app depends on the Screen Time / Family Controls APIs. Has anyone recently experienced similar delays, and is there any recommended next step besides waiting? Thanks. Imi
Replies
6
Boosts
0
Views
437
Activity
1w
Please, Apple. I am begging you. Fix the broken Text-To-Speech in macOS
Every new build of macOS 26 further breaks some part of text-to-speech or voice control. I have filed multiple bug reports on this, yet the situation gets worse, not better, with each new build. I am begging you to fix this! I am disabled and rely on these features to work. Accessibility in macOS is more than 50% of my reasons for choosing Macs over Windows. Here's some of what is currently broken: Speak announcements only works if the Samantha voice is selected. If other voices are selected either the announcements don't speak, or they default to the Samantha voice. This became broken with 26.1. Announce the time only works with some voices. I would like to use Australian English Siri 2 but with that voice selected it defaults to Samantha. This became broken with 26.4. With voice control enabled there are two menu bar icons. The blue voice control icon and an orange microphone an application is accessing the microphone icon. This orange icon started appearing with 26.3. For four years of macOS releases, the
Replies
9
Boosts
0
Views
2.5k
Activity
1w