Issue: Background downloads using the flutter_downloader package work perfectly in debug mode and release mode when run directly from Xcode (plugged in). However, when I create an archive build and install the app separately (via TestFlight or direct IPA install), the background download stops working as soon as the app is minimized. ✅ What I’ve already done Info.plist UIBackgroundModes remote-notification fetch processing audio push-to-talk AppDelegate.swift import UIKit import Flutter import Firebase import flutter_downloader import BackgroundTasks @main @objc class AppDelegate: FlutterAppDelegate { static let backgroundChannel = com.example.app/background_service private var backgroundCompletionHandler: (() -> Void)? override func application( _ application: UIApplication, didFinishLaunchingWithOptions launchOptions: [UIApplication.LaunchOptionsKey: Any]? ) -> Bool { FirebaseApp.configure() GeneratedPluginRegistrant.register(with: self) FlutterDownloaderPlugin.setPluginRegistrantCallback(r
Search results for
Popping Sound
19,599 results found
Selecting any option will automatically load the page
Post
Replies
Boosts
Views
Activity
Hi everyone! I’ve been working with AVFoundation and trying to use the AVMetricEventStreamPublisher to discover media performance metrics, as described in the Apple documentation. https://developer.apple.com/cn/videos/play/wwdc2024/10113/?time=508 However, when following the example code, I’m not getting the expected results. The performance metrics for both audio and video don’t seem to be captured properly. Has anyone successfully used this example code? If so, could you share your experience or any solutions you’ve found? Any tips or insights would be greatly appreciated. Thanks in advance! Ps. the example code: AVPlayerItem *item = ... AVMetricEventStream *eventStream = [AVMetricEventStream eventStream]; id subscriber = [[MyMetricSubscriber alloc] init]; [eventStream setSubscriber:subscriber queue:mySerialQueue] [eventStream subscribeToMetricEvent:[AVMetricPlayerItemLikelyToKeepUpEvent class]]; [eventStream subscribeToMetricEvent:[AVMetricPlayerItemPlaybackSummaryEvent class]]; [eventStream addPu
@StevenPeterson Thanks for chiming in and thanks for the video itself, much appreciated. No, the view bodies are only run if the body uses the environment value for the key(s) used in that view and the value changes. That makes a lot more sense to me and aligns with the simple demo app I made to test out my original question. Consider: struct ContentView: View { @Environment(.name) var name @Environment(.age) var age var body: some View { Text(name) } } If .name is updated anywhere, then ContentView.body will be run but if .age is updated anywhere, then body is not run because .age is not referenced in the body. What worried me was that if somewhere else a view wrote to .address, then ContentView would be re-evaluated even though it doesn't refer to .address in any way. That's how I (incorrectly) interpreted the slide above. But in SwiftUI, an update doesn't always cause the view body to run again, but there is still a cost associated with these updates. That's the real clarifying statement, for me. There's
Topic:
UI Frameworks
SubTopic:
SwiftUI
Tags:
Hi! I'll try to provide some clarification: Does writing any value to the environment really cause all views in the entire SwiftUI view hierarchy that read any environment key to have their body re-evaluated? No, the view bodies are only run if the body uses the environment value for the key(s) used in that view and the value changes. But in SwiftUI, an update doesn't always cause the view body to run again, but there is still a cost associated with these updates. In Instruments, these updates are labeled as skipped. The demo in the video shows that the cumulative cost of all the skipped updates (seen in the Consequences detail view) is significant, even without their associated view bodies running (the view that demonstrates this in the video is called CardBackView). Do environment writes only affect child views, or do they propagate through the entire SwiftUI hierarchy? Great question! This only applies to child views of the .environment modifier that you're writing with. So if View B uses an .environment m
Topic:
UI Frameworks
SubTopic:
SwiftUI
Tags:
There are various reports of this issue scattered about. Any time I start to hear the crackle I run the following command from a terminal window: sudo pkill coreaudiod That fixes the issue for a while. If you run that while playing music or other audio, the sound will stop for a couple of seconds as the process restarts and then the audio will come back without the crackle.
Topic:
Developer Tools & Services
SubTopic:
Xcode
Tags:
[Submitted as FB20950954] Xcode Simulator causes crackling and distortion in audio playback across all apps (Apple Podcasts, Music, third-party). REPRO STEPS Open any audio app and start playback Note the audio quality Launch Xcode Simulator After a few seconds, note audio quality again Quit Xcode Simulator Audio returns to normal CURRENT Audio has crackling and distortion while Simulator is running. EXPECTED Clean audio playback regardless of whether Simulator is running. SYSTEM INFO macOS 26.1 (25B78) Xcode 26.1 (17B55) Simulator 26.0 (1058)
First, let me explain my understanding of this API. I believe its purpose is to allow an app to trigger tasks that can be represented with progress indicators and require a certain amount of time to complete. I put things much more simply than that and say that the goal is simply to provide an API that gives apps an extended period of time to complete user-critical work after they enter the background. The progress requirement exists as a practical UI compromise so that the user can see something and the system has some assurance that the app is still functioning. I'd also push back on the trigger tasks part of this. Nothing requires your app to actually bind its activity to the continuing task and, in fact, I think there are many situations where your app should simply start the work immediately, with the continuing task acting as a backup in case your app is sent to the foreground. See this forum thread for an extended discussion of this. After I called register and submit, the BGContinuedProcessingTask cou
Topic:
App & System Services
SubTopic:
Processes & Concurrency
Tags:
My app uses CallKit and PushKit for real-time voice and video calls. I want to configure VoIP push notifications so incoming calls can wake the app, even when it’s in the background or terminated. Right now, I only see the normal Push Notifications option in my App ID settings and no VoIP Services section. I'm not entirely sure of what you're looking at here, so I'm going to give you a general rundown of how VoIP apps are configured and hope that covers your questions: The app itself doesn't actually have any special configuration on the developer portal. In Xcode, you do need to add both the voip and audio background categories, but both of those are actually modifying the UIBackgroundModes Info.plist key, so that doesn't actually impact the portal. Most VoIP apps also implement standard push, both to support text messaging and for things like notifying the user of missed calls. However, receiving VoIP pushes does NOT actually require authorization from the user. The modern approach to sending VoIP
Topic:
App & System Services
SubTopic:
General
Tags:
Could anyone play a sound from the Library/Sounds directory using alarmkit? Same file can play in app bundle, but not available in Library/Sounds directory. Is this by design?
Topic:
App & System Services
SubTopic:
Notifications
[quote='806537021, jay2013, /thread/806537, /profile/jay2013'] This has to be a bug. [/quote] Well, that depends on your point of view (-: When dealing with managed capabilities like this, the system allows for various restrictions, and it seems that whoever set this up decided to apply this restriction (or made a mistake). I’m not involved in that process, so I can’t offer any insight into that. [quote='806537021, jay2013, /thread/806537, /profile/jay2013'] How do I get Apple to fix this? [/quote] My general advice for entitlement problems like this is to start a support request via Apple > Developer > Contact Us > Development and Technical > Entitlements. It sounds like you’ve already done that. You should’ve got back a case number. Please post that here. Also, can you post the first and last letters of your Team ID. For example, the Team ID of my individual test team is SKMME9E2Y8, and the redacted version is S________8. Share and Enjoy — Quinn “The Eskimo!” @ Developer Technical Suppo
Topic:
Developer Tools & Services
SubTopic:
Apple Developer Program
Tags:
Thanks for filing the feedback report (FB20915258) for us. Given that complications render correctly on the watch, it does sound like an issue on the Watch app in iOS. Best, —— Ziqiao Chen Worldwide Developer Relations.
Topic:
App & System Services
SubTopic:
Widgets & Live Activities
Tags:
We require assistance in resolving a critical audio design conflict within our Push-to-Talk (PTT) application. Our current volume amplification strategy—which relies on applying a GAIN factor to PCM samples in conjunction with setting the AVAudioSession category to Playback—is working successfully when PTT is used independently. However, upon integrating and reporting the same PTT call through the CallKit framework, this amplification effect is lost. The CallKit integration appears to be forcing a different, non-amplifying audio session category or configuration, negatively impacting the user's perceived call volume. We need guidance on how to maintain the AVAudioSessionCategoryPlayback setting, or an equivalent high-volume configuration, while operating under the control of CallKit.
Could you please advise on which framework or API Apple recommends for a third-party application to intercept and process the audio stream—specifically, to apply a GAIN factor to PCM samples—during an active CallKit call? AVAudioEngine is the API we generally recommend for lower level audio processing. __ Kevin Elliott DTS Engineer, CoreOS/Hardware
Topic:
Media Technologies
SubTopic:
Audio
Tags:
For #1 I think our recommendation would be to never rely on the main menu bar existing but I can certainly see challenges in building a UI that way at times! For #2 that sounds buggy to me but I don't know for sure. If nothing else it would probably make good feedback, since I would hope that we would behave consistently in this respect regardless of modes. I'll ping some folks and see if there is any additional insight they can provide here.
Topic:
UI Frameworks
SubTopic:
UIKit
Tags:
Thank you for the clear explanation. We must emphasize that, as a Mission critical Push-to-Talk (PTT) application, the perceived volume and clarity of the audio are fundamental user requirements. We currently achieve the necessary amplification and clarity using a custom PCM gain factor. Therefore, the ability for our application to apply a specific gain factor to PCM samples for amplification is essential. Could you please advise on which framework or API Apple recommends for a third-party application to intercept and process the audio stream—specifically, to apply a GAIN factor to PCM samples—during an active CallKit call?
Topic:
Media Technologies
SubTopic:
Audio
Tags: