Search results for

“Popping Sound”

20,035 results found

Post

Replies

Boosts

Views

Activity

Unable to capture only the cursor in macOS Tahoe
Precondition: In system settings, scale the pointer size up to the max. Our SCScreenshotManager code currently works in macOS 15 and earlier to capture the cursor at it's larger size, but broke in one of the minor releases of macOS Tahoe. The error it produces now is Failed to start stream due to audio/video capture failure. This only seems to happen with the cursor window, not any others. Another way to get the cursor is with https://developer.apple.com/documentation/appkit/nscursor/currentsystem, but that is now deprecated, which makes me think the capture of the cursor is being blocked deliberately. We see this as a critical loss of functionality for our apps, and could use guidance on what to use instead.
1
0
345
1w
AVAudioFile.read extremely slow after seeking in FLAC and MP3 files
I'm developing an audio player app that uses AVAudioFile to read PCM data from various formats. I'm experiencing severe performance issues when seeking in FLAC, while other compressed formats (M4A/AAC) work correctly. I don't intend to use them in my app, but I also tested mp3 files just by curiosity and they also have this issue. Environment: macOS 26 (Tahoe) Xcode 26.3 Apple Silicon (M1) The issue: After setting AVAudioFile.framePosition to a position mid-file, the subsequent call to AVAudioFile.read(into:frameCount:) blocks for an unreasonable amount of time for FLAC and MP3 files. The delay scales linearly with the seek target, seeking near the beginning is fast, seeking toward the end is proportionally slower, which suggests the decoder is decoding linearly from the beginning of the file rather than using any seek index. (My app deals with “images” of Audio CDs ripped as a single long audio file.) The issue is particularly severe when reading files from an SMB network share (se
0
0
123
1w
Do watchOS widget reloads in an active workout session count against the daily budget?
https://developer.apple.com/documentation/widgetkit/keeping-a-widget-up-to-date lists a number of exception including The widget’s containing app has an active audio or navigation session. https://developer.apple.com/videos/play/wwdc2021/10048/ mentions: However, there are a few situational exceptions that will make these reloads occur both immediately and budget-free. These are when your container app is foreground to the user or when your app is participating in a user session, like Navigation or Now Playing audio. Does an active workout session in a watchOS app count as your app is participating in a user session, so calls to WidgetCenter.shared.reloadTimelines(ofKind:) are budget-free?
2
0
248
1w
Reply to Unlock with Touch ID suggested despite system.login.screensaver being configured with authenticate-session-owner rule
[quote='819454021, oleksandr91, /thread/819454, /profile/oleksandr91'] Is this a known macOS bug? [/quote] I’m pretty sure I’ve seen this before. IIRC the problem was reported by a third-party developer and I found it super hard to reproduce, and thus I’m not sure if a bug got filed. Sorry to be vague about the details; I searched my records but was unable to track down the conversation. Anyway, it sounds like you can easily reproduce this. Given that, I think it make sense for you to file your own bug about it. Please post your bug number, just for the record. Share and Enjoy — Quinn “The Eskimo!” @ Developer Technical Support @ Apple let myEmail = eskimo + 1 + @ + apple.com
Topic: Privacy & Security SubTopic: General Tags:
1w
Building Real-Time Voice Input on macOS 26 with SpeechAnalyzer + ScreenCaptureKit
We built an open-source macOS menu bar app that turns speech into text and pastes it into the active app — using SpeechAnalyzer for on-device transcription, ScreenCaptureKit + Vision for screen-aware context, and FluidAudio for speaker diarization in meeting mode. Here's what we learned shipping it on macOS 26. GitHub: github.com/Marvinngg/ambient-voice Architecture The app has two modes: hotkey dictation (press to talk, release to inject) and meeting recording (continuous transcription with a floating panel). Dictation Mode Audio capture uses AVCaptureSession (more on why below). The captured audio feeds into SpeechAnalyzer via an AsyncStream: let transcriber = SpeechTranscriber( locale: locale, transcriptionOptions: [], reportingOptions: [.volatileResults, .alternativeTranscriptions], attributeOptions: [.audioTimeRange, .transcriptionConfidence] ) let analyzer = SpeechAnalyzer(modules: [transcriber]) let (inputSequence, inputBuilder) = AsyncStream.makeStream() try await analyzer.start(inpu
0
0
384
1w
Building Real-Time Voice Input on macOS 26 with SpeechAnalyzer + ScreenCaptureKit
We built an open-source macOS menu bar app that turns speech into text and pastes it into the active app — using SpeechAnalyzer for on-device transcription, ScreenCaptureKit + Vision for screen-aware context, and FluidAudio for speaker diarization in meeting mode. Here's what we learned shipping it on macOS 26. GitHub: github.com/Marvinngg/ambient-voice Architecture The app has two modes: hotkey dictation (press to talk, release to inject) and meeting recording (continuous transcription with a floating panel). Dictation Mode Audio capture uses AVCaptureSession (more on why below). The captured audio feeds into SpeechAnalyzer via an AsyncStream: let transcriber = SpeechTranscriber( locale: locale, transcriptionOptions: [], reportingOptions: [.volatileResults, .alternativeTranscriptions], attributeOptions: [.audioTimeRange, .transcriptionConfidence] ) let analyzer = SpeechAnalyzer(modules: [transcriber]) let (inputSequence, inputBuilder) = AsyncStream.makeStream() try await analyzer.start(inpu
0
0
366
1w
Reply to DriverKit Entitlement Model Has No Viable Path for Open Source and Community-Maintained Drivers
A free or reduced-cost entitlement path for non-commercial volunteer-maintained drivers If you haven't already, please file a bug asking for the dev-only entitlements to be enabled on free developer accounts, and post the bug number back here. I can't promise it will happen, but it's an interesting idea that I don't think we've really considered. Published approval criteria and timelines so projects can plan accordingly Unfortunately, publishing truly formal criteria is much more difficult than it sounds since, in practice, it's very difficult to come up with a criteria set that EXACTLY covers every reasonable use case while also excluding every possible unreasonable use case... and that assumes we could even come up with every use case. However, on this point: Even if the developer is a professional company, some legacy hardware will go unsupported because of a lack of support from the vendor. Providing a way for users who need access to older hardware would be needed. We understand that the nature
Topic: App & System Services SubTopic: Drivers Tags:
1w
Background upload issue in WatchOS
We are developing a watchOS application that records long audio sessions and uploads them to our backend in chunks (~5 MB each) using pre-signed URLs and URLSession background upload. Current behavior: While audio recording is active, uploads continue successfully even when the app is in the background. Once the recording stops, if multiple chunks (e.g., 10+) are still pending, the remaining uploads do not proceed in the background and appear to be suspended. We attempted to use WKExtendedRuntimeSession (mindfulness type) to allow sufficient time to enqueue background upload tasks, but the session is invalidated when the app goes to the background (e.g., wrist down or app inactive), which prevents reliable scheduling of uploads. Additionally, we added the entitlement: com.apple.developer.extended-runtime-session (mindfulness) in the Watch app entitlements file, but Xcode automatic signing fails with: “Provisioning profile does not include the com.apple.developer.extended-runtime-session enti
2
0
137
1w
Reply to WatchOS: Can a background metronome app coexist with both Runna workout and Spotify playback?
Thanks for the reply! I have also tested AVAudioSession with .playback + .mixWithOthers on watchOS. That works for mixing my metronome with Spotify while the metronome app is in the foreground, which matches the behavior I already see on iPhone. However, on watchOS the metronome stops once the app goes to the background, while longFormAudio keeps it alive but then conflicts with Spotify. So this seems to be a watchOS-specific limitation rather than a general AVAudioSession limitation on Apple platforms. It would be very useful if watchOS supported a way to combine background audio persistence (longFormAudio) with cooperative mixing behavior (mixWithOthers) for lightweight apps like metronomes, cues, or coaching audio. Is there any way I can formally request this feature other than the FB22273077 I have already submitted?
Topic: Media Technologies SubTopic: Audio Tags:
1w
Reply to SpeechAnalyzer.start(inputSequence:) fails with _GenericObjCError nilError, while the same WAV succeeds with start(inputAudioFile:)
I've been working with SpeechAnalyzer.start(inputSequence:) on macOS 26 and got streaming transcription working. A few things that might help: Make sure the AVAudioFormat you use to create AnalyzerInput buffers exactly matches what bestAvailableAudioFormat() returns. Even subtle mismatches (e.g., interleaved vs non-interleaved, different channel layouts) can cause the nilError without a descriptive message. I found that feeding buffers that are too small (< 4096 frames) occasionally triggers this error. Try using larger chunks â I settled on 8192 frames per buffer. The bufferStartTime parameter needs to be monotonically increasing and consistent with the actual audio duration. If there are gaps or overlaps in the timestamps, the stream mode can fail silently or throw nilError. Instead of replaying a WAV file as chunked buffers, I'd suggest testing with live audio from AVCaptureSession first. In my experience, live capture â AnalyzerInput works more reliably than simulated streaming from a
Topic: Media Technologies SubTopic: Audio Tags:
1w
Reply to DriverKit vs MFi for iPad custom hardware serial communication?
> I have a custom hardware board that I want to communicate serially with from an iPad. Should I use the DriverKit route or the MFi route? Depends on exactly what your needs are and the hardware you're working with. Starting with MFi, the biggest advantage is that the software side (ExternalAccessory framework) is far simpler to use and includes support for background access (DriverKit does not). The main downside is that it requires a certain amount of custom hardware and the certification process. Comparing it with DriverKit is... complicated. The software side of DriverKit is FAR more complicated. To start with, the DriverKit API is a somewhat odd derivative of the IOKit kernel API. For someone who's familiar with IOKit, it's similar enough to be understandable, but for a new developer, the learning curve can be quite significant. None of this is helped by the fact that DriverKit is poorly documented[1], to the point that my standard advice to new developers is to study the IOKit kernel API (which has o
Topic: App & System Services SubTopic: Drivers Tags:
1w
Reply to WatchOS: Can a background metronome app coexist with both Runna workout and Spotify playback?
If you are looking for a way to have multiple apps running their own active HKWorkoutSession on watchOS simultaneously, there is no way to do that. This is documented here: Apple Watch runs one workout session at a time. If a second workout starts while your workout is running, your HKWorkoutSessionDelegate object receives an HKError.Code.errorAnotherWorkoutSessionStarted error, and your session ends. Regarding audio, my limited knowledge is that, with AVAudioSession, you can tell the system how to mix your app's audio with the current now-playing app by setting up the session category. You can play with the categories and see if there is one fitting your use case. Best, —— Ziqiao Chen  Worldwide Developer Relations.
Topic: Media Technologies SubTopic: Audio Tags:
2w
WatchOS: Can a background metronome app coexist with both Runna workout and Spotify playback?
I’m building a standalone Apple Watch metronome app for running. My goal is for these 3 apps to work at the same time: Runna owns the workout session Spotify plays music my app plays a metronome click in the background So far this is what I've found: Using HKWorkoutSession in my metronome app works well with Spotify, but conflicts with Runna and other workout apps, so I removed that. Using watchOS background audio with longFormAudio allows my app run in the background, and it can coexist with Runna. However, it seems to conflict with Spotify playback, and one app tends to stop the other. Is there any supported watchOS audio/background configuration that allows all 3 at once? More specifically this is what I need: another app owns HKWorkoutSession Spotify keeps playing my app keeps generating metronome clicks in the background Or is this simply not supported by current watchOS session/background rules? My metronome uses AVAudioEngine / AVAudioPlayerNode with generated click audio. Th
4
0
319
2w
Reply to filecopy fails with errno 34 "Result too large" when copying from NAS
Have you filed a bug on this and, if so, what is the bug number? Not yet, but I do have another thread where I describe the same problem in regards to .DS_Store: https://developer.apple.com/forums/thread/819041 I will post the bug number (most likely there) when I carve out a little time for it. Currently if copyfile fails and errno is not set my code falls back to a kind of 'Unknown Error' which I hate but if the user is counting on my app to perform the copy it's kind of a gamble if the operation doesn't succeed to just hope not critical data was lost. A big part of this is that xattrs aren't actually safe“— the system doesn't have a documented system for defining how they should be handled, and app/copy engine/file format support for them is inconsistent, so storing critical data in them can end up being a good way to lose critical data. Importantly, that kind of data loss isn't really a bug in the same way other kinds of data loss would be. The general point of xattrs is to allow the attachment of arbitra
Topic: App & System Services SubTopic: Core OS Tags:
2w
Unable to capture only the cursor in macOS Tahoe
Precondition: In system settings, scale the pointer size up to the max. Our SCScreenshotManager code currently works in macOS 15 and earlier to capture the cursor at it's larger size, but broke in one of the minor releases of macOS Tahoe. The error it produces now is Failed to start stream due to audio/video capture failure. This only seems to happen with the cursor window, not any others. Another way to get the cursor is with https://developer.apple.com/documentation/appkit/nscursor/currentsystem, but that is now deprecated, which makes me think the capture of the cursor is being blocked deliberately. We see this as a critical loss of functionality for our apps, and could use guidance on what to use instead.
Replies
1
Boosts
0
Views
345
Activity
1w
AVAudioFile.read extremely slow after seeking in FLAC and MP3 files
I'm developing an audio player app that uses AVAudioFile to read PCM data from various formats. I'm experiencing severe performance issues when seeking in FLAC, while other compressed formats (M4A/AAC) work correctly. I don't intend to use them in my app, but I also tested mp3 files just by curiosity and they also have this issue. Environment: macOS 26 (Tahoe) Xcode 26.3 Apple Silicon (M1) The issue: After setting AVAudioFile.framePosition to a position mid-file, the subsequent call to AVAudioFile.read(into:frameCount:) blocks for an unreasonable amount of time for FLAC and MP3 files. The delay scales linearly with the seek target, seeking near the beginning is fast, seeking toward the end is proportionally slower, which suggests the decoder is decoding linearly from the beginning of the file rather than using any seek index. (My app deals with “images” of Audio CDs ripped as a single long audio file.) The issue is particularly severe when reading files from an SMB network share (se
Replies
0
Boosts
0
Views
123
Activity
1w
Do watchOS widget reloads in an active workout session count against the daily budget?
https://developer.apple.com/documentation/widgetkit/keeping-a-widget-up-to-date lists a number of exception including The widget’s containing app has an active audio or navigation session. https://developer.apple.com/videos/play/wwdc2021/10048/ mentions: However, there are a few situational exceptions that will make these reloads occur both immediately and budget-free. These are when your container app is foreground to the user or when your app is participating in a user session, like Navigation or Now Playing audio. Does an active workout session in a watchOS app count as your app is participating in a user session, so calls to WidgetCenter.shared.reloadTimelines(ofKind:) are budget-free?
Replies
2
Boosts
0
Views
248
Activity
1w
Reply to Unlock with Touch ID suggested despite system.login.screensaver being configured with authenticate-session-owner rule
[quote='819454021, oleksandr91, /thread/819454, /profile/oleksandr91'] Is this a known macOS bug? [/quote] I’m pretty sure I’ve seen this before. IIRC the problem was reported by a third-party developer and I found it super hard to reproduce, and thus I’m not sure if a bug got filed. Sorry to be vague about the details; I searched my records but was unable to track down the conversation. Anyway, it sounds like you can easily reproduce this. Given that, I think it make sense for you to file your own bug about it. Please post your bug number, just for the record. Share and Enjoy — Quinn “The Eskimo!” @ Developer Technical Support @ Apple let myEmail = eskimo + 1 + @ + apple.com
Topic: Privacy & Security SubTopic: General Tags:
Replies
Boosts
Views
Activity
1w
Building Real-Time Voice Input on macOS 26 with SpeechAnalyzer + ScreenCaptureKit
We built an open-source macOS menu bar app that turns speech into text and pastes it into the active app — using SpeechAnalyzer for on-device transcription, ScreenCaptureKit + Vision for screen-aware context, and FluidAudio for speaker diarization in meeting mode. Here's what we learned shipping it on macOS 26. GitHub: github.com/Marvinngg/ambient-voice Architecture The app has two modes: hotkey dictation (press to talk, release to inject) and meeting recording (continuous transcription with a floating panel). Dictation Mode Audio capture uses AVCaptureSession (more on why below). The captured audio feeds into SpeechAnalyzer via an AsyncStream: let transcriber = SpeechTranscriber( locale: locale, transcriptionOptions: [], reportingOptions: [.volatileResults, .alternativeTranscriptions], attributeOptions: [.audioTimeRange, .transcriptionConfidence] ) let analyzer = SpeechAnalyzer(modules: [transcriber]) let (inputSequence, inputBuilder) = AsyncStream.makeStream() try await analyzer.start(inpu
Replies
0
Boosts
0
Views
384
Activity
1w
Building Real-Time Voice Input on macOS 26 with SpeechAnalyzer + ScreenCaptureKit
We built an open-source macOS menu bar app that turns speech into text and pastes it into the active app — using SpeechAnalyzer for on-device transcription, ScreenCaptureKit + Vision for screen-aware context, and FluidAudio for speaker diarization in meeting mode. Here's what we learned shipping it on macOS 26. GitHub: github.com/Marvinngg/ambient-voice Architecture The app has two modes: hotkey dictation (press to talk, release to inject) and meeting recording (continuous transcription with a floating panel). Dictation Mode Audio capture uses AVCaptureSession (more on why below). The captured audio feeds into SpeechAnalyzer via an AsyncStream: let transcriber = SpeechTranscriber( locale: locale, transcriptionOptions: [], reportingOptions: [.volatileResults, .alternativeTranscriptions], attributeOptions: [.audioTimeRange, .transcriptionConfidence] ) let analyzer = SpeechAnalyzer(modules: [transcriber]) let (inputSequence, inputBuilder) = AsyncStream.makeStream() try await analyzer.start(inpu
Replies
0
Boosts
0
Views
366
Activity
1w
Reply to DriverKit Entitlement Model Has No Viable Path for Open Source and Community-Maintained Drivers
A free or reduced-cost entitlement path for non-commercial volunteer-maintained drivers If you haven't already, please file a bug asking for the dev-only entitlements to be enabled on free developer accounts, and post the bug number back here. I can't promise it will happen, but it's an interesting idea that I don't think we've really considered. Published approval criteria and timelines so projects can plan accordingly Unfortunately, publishing truly formal criteria is much more difficult than it sounds since, in practice, it's very difficult to come up with a criteria set that EXACTLY covers every reasonable use case while also excluding every possible unreasonable use case... and that assumes we could even come up with every use case. However, on this point: Even if the developer is a professional company, some legacy hardware will go unsupported because of a lack of support from the vendor. Providing a way for users who need access to older hardware would be needed. We understand that the nature
Topic: App & System Services SubTopic: Drivers Tags:
Replies
Boosts
Views
Activity
1w
Background upload issue in WatchOS
We are developing a watchOS application that records long audio sessions and uploads them to our backend in chunks (~5 MB each) using pre-signed URLs and URLSession background upload. Current behavior: While audio recording is active, uploads continue successfully even when the app is in the background. Once the recording stops, if multiple chunks (e.g., 10+) are still pending, the remaining uploads do not proceed in the background and appear to be suspended. We attempted to use WKExtendedRuntimeSession (mindfulness type) to allow sufficient time to enqueue background upload tasks, but the session is invalidated when the app goes to the background (e.g., wrist down or app inactive), which prevents reliable scheduling of uploads. Additionally, we added the entitlement: com.apple.developer.extended-runtime-session (mindfulness) in the Watch app entitlements file, but Xcode automatic signing fails with: “Provisioning profile does not include the com.apple.developer.extended-runtime-session enti
Replies
2
Boosts
0
Views
137
Activity
1w
Reply to WatchOS: Can a background metronome app coexist with both Runna workout and Spotify playback?
Thanks for the reply! I have also tested AVAudioSession with .playback + .mixWithOthers on watchOS. That works for mixing my metronome with Spotify while the metronome app is in the foreground, which matches the behavior I already see on iPhone. However, on watchOS the metronome stops once the app goes to the background, while longFormAudio keeps it alive but then conflicts with Spotify. So this seems to be a watchOS-specific limitation rather than a general AVAudioSession limitation on Apple platforms. It would be very useful if watchOS supported a way to combine background audio persistence (longFormAudio) with cooperative mixing behavior (mixWithOthers) for lightweight apps like metronomes, cues, or coaching audio. Is there any way I can formally request this feature other than the FB22273077 I have already submitted?
Topic: Media Technologies SubTopic: Audio Tags:
Replies
Boosts
Views
Activity
1w
Reply to SpeechAnalyzer.start(inputSequence:) fails with _GenericObjCError nilError, while the same WAV succeeds with start(inputAudioFile:)
I've been working with SpeechAnalyzer.start(inputSequence:) on macOS 26 and got streaming transcription working. A few things that might help: Make sure the AVAudioFormat you use to create AnalyzerInput buffers exactly matches what bestAvailableAudioFormat() returns. Even subtle mismatches (e.g., interleaved vs non-interleaved, different channel layouts) can cause the nilError without a descriptive message. I found that feeding buffers that are too small (< 4096 frames) occasionally triggers this error. Try using larger chunks â I settled on 8192 frames per buffer. The bufferStartTime parameter needs to be monotonically increasing and consistent with the actual audio duration. If there are gaps or overlaps in the timestamps, the stream mode can fail silently or throw nilError. Instead of replaying a WAV file as chunked buffers, I'd suggest testing with live audio from AVCaptureSession first. In my experience, live capture â AnalyzerInput works more reliably than simulated streaming from a
Topic: Media Technologies SubTopic: Audio Tags:
Replies
Boosts
Views
Activity
1w
Reply to Using StoreKit from an AUv3 plugin that can be loaded in-process
BTW, however I solve this, if I do, it needs to work on both Mac OS and iOS - I have no idea if there's much market for iOS Audio Units, but it was easy enough to build for both, so I will release both. It's at least likely to be a less saturated market :-)
Topic: Media Technologies SubTopic: Audio Tags:
Replies
Boosts
Views
Activity
1w
Reply to DriverKit vs MFi for iPad custom hardware serial communication?
> I have a custom hardware board that I want to communicate serially with from an iPad. Should I use the DriverKit route or the MFi route? Depends on exactly what your needs are and the hardware you're working with. Starting with MFi, the biggest advantage is that the software side (ExternalAccessory framework) is far simpler to use and includes support for background access (DriverKit does not). The main downside is that it requires a certain amount of custom hardware and the certification process. Comparing it with DriverKit is... complicated. The software side of DriverKit is FAR more complicated. To start with, the DriverKit API is a somewhat odd derivative of the IOKit kernel API. For someone who's familiar with IOKit, it's similar enough to be understandable, but for a new developer, the learning curve can be quite significant. None of this is helped by the fact that DriverKit is poorly documented[1], to the point that my standard advice to new developers is to study the IOKit kernel API (which has o
Topic: App & System Services SubTopic: Drivers Tags:
Replies
Boosts
Views
Activity
1w
Reply to WatchOS: Can a background metronome app coexist with both Runna workout and Spotify playback?
If you are looking for a way to have multiple apps running their own active HKWorkoutSession on watchOS simultaneously, there is no way to do that. This is documented here: Apple Watch runs one workout session at a time. If a second workout starts while your workout is running, your HKWorkoutSessionDelegate object receives an HKError.Code.errorAnotherWorkoutSessionStarted error, and your session ends. Regarding audio, my limited knowledge is that, with AVAudioSession, you can tell the system how to mix your app's audio with the current now-playing app by setting up the session category. You can play with the categories and see if there is one fitting your use case. Best, —— Ziqiao Chen  Worldwide Developer Relations.
Topic: Media Technologies SubTopic: Audio Tags:
Replies
Boosts
Views
Activity
2w
WatchOS: Can a background metronome app coexist with both Runna workout and Spotify playback?
I’m building a standalone Apple Watch metronome app for running. My goal is for these 3 apps to work at the same time: Runna owns the workout session Spotify plays music my app plays a metronome click in the background So far this is what I've found: Using HKWorkoutSession in my metronome app works well with Spotify, but conflicts with Runna and other workout apps, so I removed that. Using watchOS background audio with longFormAudio allows my app run in the background, and it can coexist with Runna. However, it seems to conflict with Spotify playback, and one app tends to stop the other. Is there any supported watchOS audio/background configuration that allows all 3 at once? More specifically this is what I need: another app owns HKWorkoutSession Spotify keeps playing my app keeps generating metronome clicks in the background Or is this simply not supported by current watchOS session/background rules? My metronome uses AVAudioEngine / AVAudioPlayerNode with generated click audio. Th
Replies
4
Boosts
0
Views
319
Activity
2w
Reply to filecopy fails with errno 34 "Result too large" when copying from NAS
Have you filed a bug on this and, if so, what is the bug number? Not yet, but I do have another thread where I describe the same problem in regards to .DS_Store: https://developer.apple.com/forums/thread/819041 I will post the bug number (most likely there) when I carve out a little time for it. Currently if copyfile fails and errno is not set my code falls back to a kind of 'Unknown Error' which I hate but if the user is counting on my app to perform the copy it's kind of a gamble if the operation doesn't succeed to just hope not critical data was lost. A big part of this is that xattrs aren't actually safe“— the system doesn't have a documented system for defining how they should be handled, and app/copy engine/file format support for them is inconsistent, so storing critical data in them can end up being a good way to lose critical data. Importantly, that kind of data loss isn't really a bug in the same way other kinds of data loss would be. The general point of xattrs is to allow the attachment of arbitra
Topic: App & System Services SubTopic: Core OS Tags:
Replies
Boosts
Views
Activity
2w