Search results for

file uri scheme

79,827 results found

Post

Replies

Boosts

Views

Activity

Reply to Fileprovider Recycle Bin recovery does not trigger the create event, and dataless files should not be moved to the Recycle Bin
I am currently encountering a problem: during the process of uploading a large file, I have moved the file that was not successfully uploaded to the trash can. These two operations have been tested to be serial (triggering the 'create Item' callback first, followed by the 'modify Item' callback), which means that the file must be uploaded before it can be moved to the recycle bin (which can also result in the file being stored in the cloud recycle bin). I want to implement: directly interrupt this upload process and then do not complete the upload. How can I achieve this? Please help me. Thank you
Topic: App & System Services SubTopic: Core OS Tags:
6d
During the process of uploading a large file, I moved it to the trash can. How can I directly interrupt this upload process
I am currently encountering a problem: during the process of uploading a large file, I have moved the file that was not successfully uploaded to the trash can. These two operations have been tested to be serial (triggering the 'create Item' callback first, followed by the 'modify Item' callback), which means that the file must be uploaded before it can be moved to the recycle bin (which can also result in the file being stored in the cloud recycle bin). I want to implement: directly interrupt this upload process and then do not complete the upload. How can I achieve this? Please help me. Thank you
1
0
49
6d
AVPlayerViewController volume slider UI changes but audio output level remains constant
Facing an issue with audio playback using AVPlayerViewController in iOS application. We are using the native player to play recorded audio files. When the AVPlayerViewController appears, the native user interface is displayed correctly, including the playback controls and the volume slider. However, when the user interacts with the volume slider The slider UI moves and responds to touch events. The actual audio output volume does not change. The audio continues playing at the initial volume level regardless of the slider position. We initialize the player and present it modally using the following code: AVPlayerViewController *avController = [[AVPlayerViewController alloc] init]; avController.player = [AVPlayer playerWithURL:videoURL]; // Setting initial volume avController.player.volume = 1.0f; avController.modalPresentationStyle = UIModalPresentationOverFullScreen; avController.allowsPictureInPicturePlayback = NO; // Present the controller [self presentViewController:avController animated:YES compl
0
0
35
6d
VideoPlayer crashes on Archive build
I have found that following code runs without issue from Xcode, either in Debug or Release mode, yet crashes when running from the binary produced by archiving - i.e. what will be sent to the app store. import SwiftUI import AVKit @main struct tcApp: App { var body: some Scene { WindowGroup { VideoPlayer(player: nil) } } } This is the most stripped down code that shows the issue. One can try and point the VideoPlayer at a file and the same issue will occur. I've attached the crash log: Crash log Please note that this was seen with Xcode 26.2 and MacOS 26.2.
1
0
309
6d
Is Xcode's predictive code completion model has some censoring?
It seems Xcode's predictive code completion model is censored. Specifically, when typing the word torrent, the model stops working completely. It doesn't matter whether the word is written directly in the code or in a comment. It could also be part of another word, such as qBittorrent. In either case, the model stops working. Reproducing this issue is fairly simple. Create a Swift file and type the word torrent. The model will stop generating code. Xcode Version 26.2 (17C52) Predictive Code Completion Model: [com.apple.fm.code.generate_small_v2.base: 700.0.81600.13.202379,0] [com.apple.fm.code.generate_safety_guardrail.base: 1.6.81619.13.202072,0] [com.apple.gm.safety_deny.input.code_intelligence.base: 32025010.20251009.91600.100.1651,0] [com.apple.gm.safety_deny.output.code_intelligence.base: 32025010.20251009.91600.100.1651,0] (Installed)
0
0
80
6d
PhotogrammetrySession crashes after update from iOS 18 to iOS 26
After updating iPad/iPhone devices from iOS 18 to iOS 26, PhotogrammetrySession intermittently crashes during photogrammetry processing. The same workflow was stable on iOS 18 with no code changes to the app. Environment: OS versions: Works on OS 18, crashes on OS 26 Device: iPad/iPhone (reproducible across devices) Source images: ~170-200 JPG files at 2160 x 3840 resolution Reproduction: The crash occurs consistently on the second or third sequential run of the photogrammetry session with the same image set. First run typically succeeds. Crash details: Xcode shows an uncaught exception during image processing: terminating due to uncaught exception of type std::bad_alloc: std::bad_alloc VTPixelTransferSession 420f sid 269 (2160.00 x 3840.00) [0.00 0.00 2160 3840] rowbytes( 2160, 2160 ) Color( (null), 0x0, (null), (null), ITU_R_601_4 ) => 24 sid 19 (2160.00 x 3840.00) [0.00 0.00 2160 3840] rowbytes( 6528 ) Color( 0x0, (null), (null), (null) ) This appears to be a memory allocation failure in VTPixe
0
0
225
6d
Reply to watchOS architecture requirements
Does Apple suggest that developers transition to arm64-only and drop support for arm64_32 devices? No — Apple Watch Series 9 and later, and Apple Watch Ultra 2 and later all support the arm64 architecture. There are many Apple Watch devices that run watchOS versions your app likely supports beyond that list, so you need to keep the arm64_32 architecture around for those devices. As noted in the announcement, you'll want to use the Standard Architectures build setting for your watchOS app, and that will automatically build the right set of supported architectures for your app, which will include arm64_32 as well as arm64. If we add support for both arm64_32 and arm64, the binary will almost certainly exceed the 75 MB app size limit, and potentially violate the size constraints for each architecture slice as well. What size constraints per architecture are you thinking of? The Maximum build file sizes documentation lists such a requirement for iOS apps in the iOS 7 and 8 time frame, but that isn't rele
6d
Reply to My Notifications Message Extension doesn't seem to run after distributing my app via Enterprise IPA
Here is the code of my message extension, if that helps. I am indeed trying to write to the shared storage of the App Groups import os.log // Apple's modern, fast, privacy-safe logging system class NotificationService: UNNotificationServiceExtension { private let log = OSLog( subsystem: Bundle.main.bundleIdentifier!, category: pushnotificationsmessageextension ) var contentHandler: ((UNNotificationContent) -> Void)! // A mutable copy of the notification content — this is what we'll modify or save var bestAttemptContent: UNMutableNotificationContent? // Main entry point — called every time a push arrives with `mutable-content: 1` override func didReceive( _ request: UNNotificationRequest, withContentHandler contentHandler: @escaping (UNNotificationContent) -> Void ) { // Save the handler so we can call it later (required!) self.contentHandler = contentHandler // Make a mutable copy so we can modify title, body, attachments, etc. bestAttemptContent = request.content.mutableCopy() as? UNMutableNotification
6d
Reply to How to fix "Sample 0 missing LiDAR point cloud!" error?
This answer is somewhat covered here: https://developer.apple.com/forums/thread/741948?answerId=811104022#811104022 The essence is that this warning refers to custom metadata saved in the HEIC images when using ObjectCaptureSession UI on iOS devices with LiDAR. There is no public way to provide this data besides using our ObjectCaptureSession capture UI which adds it automatically. If you are using another capture approach, you may safely ignore these warnings. The benefit of using the ObjectCaptureSession front-end is that we may additionally use these raw points to improve reconstruction of textureless surfaces (e.g. white walls) where depth maps may not be sufficient. To be clear, this point cloud data is distinct from the depthDataMap in the sample. This depth map may be available when reading HEIC files. It may be automatically populated (even on non-LiDAR devices) from stereo disparity data generated by AVCaptureSession in a multi-camera session, or may be derived from LiDAR data. Regardless of
Topic: Spatial Computing SubTopic: General Tags:
6d
Reply to CIRAWFilter.outputImage first-time cost is huge (~3s), subsequent calls are ~3ms. Any official way to pre-initialize RAW pipeline (without taking a real photo)?
Our engineering teams need to investigate this issue, as resolution may involve changes to Apple's software. Please file a bug report, include a small Xcode project and some directions that can be used to reproduce the problem, and post the Feedback number here once you do. If you post the Feedback number here I'll check the status next time I do a sweep of forums posts where I've suggested bug reports. Bug Reporting: How and Why? has tips on creating your bug report.
6d
Reply to window.opener is null with WKWebview
Our engineering teams need to investigate this issue, as resolution may involve changes to Apple's software. Please file a bug report, include a small Xcode project and some directions that can be used to reproduce the problem, and post the Feedback number here once you do. If you post the Feedback number here I'll check the status next time I do a sweep of forums posts where I've suggested bug reports. Bug Reporting: How and Why? has tips on creating your bug report.
Topic: Safari & Web SubTopic: General Tags:
6d
How to fix "Sample 0 missing LiDAR point cloud!" error?
I'm trying to run a PhotogrammetrySession based on photos taken in an AVCaptureSession and stored as .heic files. When I load the files I'm always seeing the error Sample 0 missing LiDAR point cloud! showing up for each individual sample. Debugging shows that sample.depthDataMap is populated, also the .heic contains depth data which can be extracted using e.g. heif-convert on my Mac. Comparing the .heic I created to one of the ObjectCaptureSession which doesn't show the LiDAR warning, I noticed the only difference being the HEIC information here: So my questions are: Are these the missing information in my manual capture causing this warning? Can I somehow add these information in an AVCaptureSession? Do these information allow better photogrammetry results?
2
0
286
6d
Reply to Request to restore full ICC profile support (LUT-based display profiles) in macOS ColorSync
Hello @ricciohu. I talked to our engineers about this question. Here is the complete story. LUT-based ICC profiles have never been supported on Apple platforms. Previously you may have been able to provide LUT-based ICC profiles to our routines but it was a bug that our routines were not properly checking parameters and raising errors for LUT-based ICC profiles. This has been fixed in recent updates and our routines now produce errors and don't allow LUT-based ICC profiles. Previously, using LUT-based ICC profiles with our routines produced unpredictable and erroneous results as there was no officially supported and maintained code inside of any of our routines for handling them. Any functionality you may have observed in previous Apple products using our routines together with LUT-based ICC profiles is subject to error and change without notice at any time. I do not recommend using LUT-based ICC profiles in any production code including code targeting past system software versions. If you require LUT-based I
Topic: Graphics & Games SubTopic: General Tags:
6d
Reply to Disconnect from AirPlay device programmatically
Hello @erawq, There is no reliable way to disconnect your app's audio playback from an AirPlay route (aside from the user taking action to change the route in Control Center for example). From AVAudioSessionTypes.h: - AVAudioSessionCategoryPlayAndRecord: /// DefaultToSpeaker will default to false, but >can be set to true, routing to Speaker /// (instead of Receiver) when no other audio route is connected. The key phrasing here is (instead of Receiver) when no other audio route is connected. In other words: try audioSession.setCategory(.playAndRecord, options: .defaultToSpeaker) This call is only expected to have an effect when audio is currently being routed to the device's receiver (the speaker which on most devices is at the top near the front-facing camera). In your case, audio is being routed to an AirPlay route, so it is expected that this call should not have an effect. If you would like the capability of programmatically disconnecting from an AirPlay route, I suggest that you open a new Feedback to
Topic: Media Technologies SubTopic: Streaming Tags:
6d
Reply to How to properly use PermissionKit to ask permission
the communication permission work for me with xcode and ios 26.2 public release it needs sync contacts and enable communication limit on the child device. Thanks. I also got mine working. Not sure if it was the upgrade to 26.2 actual release from 26.2 RC or changing the Communications limits. I still don't know how to fake my location to texas to test significant app update Two thoughts: Use a VPN GPX file in your project (for testing) - https://techcommunity.microsoft.com/discussions/windows10space/can-i-spoof-gps-location-on-iphone--no-jailbreak-needed/4427403/replies/4427409
6d