Search results for

Popping Sound

19,350 results found

Post

Replies

Boosts

Views

Activity

Reply to Notification Service Extension and the main thread
Question, if I am writing async code in the notification service extension, I understand it terminates after 30 seconds. Correct, though I always recommend that anyone setting up things like timeouts use a shorter value just in case. So I'd probably build around ~25s, not 30s. If I want to wait until these async methods finish before calling the content handler, I believe an option I have is to use dispatch groups. However, I am open to other solutions if there are better options. What are you actually waiting on? In general, I've become very nervous anytime I see code that uses dispatch groups because they seem to be used as a slightly awkward band-aid trying to make something work that doesn't really want to work. Case in point here, the main reason an NSE would be waiting is network activity, in which case the simpler solution would be to simply set the right timeout on that network activity. Having said that.... My question is, if I use dispatch groups, is there any issue in using the main queue here? Or
Jun ’25
Phonetic vs Pronunciation contacts fields and Siri
The contacts app has fields for Phonetic and Pronunciation. My app adds phonetic data to the phonetic field to help Siri better understand contacts stored in Greek, Cyrillic, or Georgian. However, using the phonetic field causes the sorting order of contacts to be messed up. For example, Greek B (beta) is represented as a phonetic sound of V, resulting in a completely incorrect sorting order. The pronunciation field doesn’t seem to affect the sorting order, but I’m not sure what it does or should do. My questions are: Do we understand the difference between phonetic and pronunciation, and how Siri actively uses them? If the phonetic field is the correct one to use, how can we raise a feature request with Apple to add an option to sort contacts based on phonetic fields or not? Here’s a test you can try: Create a new contact with the following details: First name: test Last name: test Phonetic first name: Billy Phonetic last name: Idol Ask Siri to show the contact Billy Idol. It will return the “test t
5
0
96
May ’25
Permission requirements for LAContext's canEvaluatePolicy
Hi, I am developing an app that checks if biometric authentication capabilities (Face ID and Touch ID) are available on a device. I have a few questions: Do I need to include a privacy string in my app to use the LAContext's canEvaluatePolicy function? This function checks if biometric authentication is available on the device, but does not actually trigger the authentication. From my testing, it seems like a privacy declaration is only required when using LAContext's evaluatePolicy function, which would trigger the biometric authentication. Can you confirm if this is the expected behavior across all iOS versions and iPhone models? When exactly does the biometric authentication permission pop-up appear for users - is it when calling canEvaluatePolicy or evaluatePolicy? I want to ensure my users have a seamless experience. Please let me know if you have any insights on these questions. I want to make sure I'm handling the biometric authentication functionality correctly in my app. Thank you!
2
0
99
Jun ’25
Issue with Keyboard Avoidance for Sheet View in SwiftUI
I am currently working on a comments section modeled after TikTok's/Instagram's comment sections for a media app. The view is a sheet view that is presented as follows: .sheet(isPresented: $showChat) { TakesChatView(viewModel: viewModel) .presentationDetents([.medium, .large]) .presentationDragIndicator(.hidden) .overlay( VStack { RoundedRectangle(cornerRadius: 2) .fill(Color.gray) .frame(width: 40, height: 5) .padding(.top, 15) .opacity(0.8) Label(Chat, systemImage: message.badge) .lineLimit(nil) .padding(.top, 5) .padding([.leading, .trailing], 16) Divider() .padding(.top, 5) .padding([.leading, .trailing], 16) Spacer() } .frame(maxWidth: .infinity, alignment: .top) ) }.ignoresSafeArea(.keyboard, edges: .bottom) However, some issues arise regarding keyboard avoidance. Currently, when the user taps on the TextField to type a comment, the keyboard shifts the entire view upwards as it pops up. Instead, I need it where I can still view the comments without the keyboard affecting their placement when it pop
1
0
661
Sep ’24
Reply to Map Switcher MapKit iOS 14 and up
The main point here is that to be able to change the style of a map in iOS 14, you must use MKMapView. Another thing to note is that SwiftUI's Map and MKMapView actually rely on the same underlying technologies. The key difference is that MKMapView (UIKit) gives you much more granular control, whereas the SwiftUI Map is a higher-level, declarative wrapper that, in iOS 14, simply doesn't expose the needed map style API. Remember, SwiftUI was only a year old at that point, so it didn't have half the features it does today, including proper integration with other frameworks. Regarding tracking and user-following: you can definitely replicate this behaviour using MKMapView. It just takes a bit more code, but a quick skim through the documentation will point you in the right direction. You can make use of view properties and delegate methods to handle your tracking logic as needed, as well as any additional features you want to implement. For example, you can set userTrackingMode to .follow or .followWithHeading,
Topic: UI Frameworks SubTopic: SwiftUI Tags:
Jun ’25
Reply to Removing Matter device artefacts.
I’m a newbie on this and playing with Arduino Nano with Matter examples, which, by default, generate Matter Bridges named Matter Accessory. I’ve found that the best way to delete test matter devices is to delete the Bridge within the devices. Long click the accessory, choose Accessory Settings from the pop-up menu, scroll down to Bridge and select that. Now you’ll be in the bridge setting and can scroll down there to get to delete bridge. What I don’t know how to do is delete bridges, like those shown here, when you don’t have an associated accessory (device) interface.
Topic: App & System Services SubTopic: Hardware Tags:
Jun ’25
Reply to Verifying braille output in an iOS app without a physical braille device?
Okay, after more research and reading these articles: Use a braille display with VoiceOver on iPhone Change your VoiceOver settings on iPhone It sounds like braille displays just use the VoiceOver output and whether it reads parentheses is part of the device's Verbosity › Punctuation setting. Then looking into whether I can set that programmatically, I came across this .speechAlwaysIncludesPuncturation() modifier, which may do exactly what I want. I'll give that a try report back! 🤞
Jun ’25
Reply to Request authorization for the notification center crash iOS app on Swift 6
At the time of posting, this problem still persist. I had to downgrade the Swift Version Language that the Swift Compiler uses from Swift 6 to Swift 5. As soon as I invoke this function, I get hosed and the app crashes: UNUserNotificationCenter.current().requestAuthorization(options: [.alert, .sound, .badge]) { granted, error in // ... } I hear that this is a known Apple Bug. Is this true?
Topic: Programming Languages SubTopic: Swift Tags:
Jun ’25
Request authorization for the notification center crash iOS app on Swift 6
Hey all! During the migration of a production app to swift 6, I've encountered a problem: when hitting the UNUserNotificationCenter.current().requestAuthorization the app crashes. If I switch back to Language Version 5 the app works as expected. The offending code is defined here class AppDelegate: NSObject, UIApplicationDelegate { func application(_ application: UIApplication, didFinishLaunchingWithOptions launchOptions: [UIApplication.LaunchOptionsKey: Any]?) -> Bool { FirebaseApp.configure() FirebaseConfiguration.shared.setLoggerLevel(.min) UNUserNotificationCenter.current().delegate = self let authOptions: UNAuthorizationOptions = [.alert, .badge, .sound] UNUserNotificationCenter.current().requestAuthorization(options: authOptions) { _, _ in } application.registerForRemoteNotifications() Messaging.messaging().delegate = self return true } } The error is depicted here: I have no idea how to fix this. Any help will be really appreciated thanks in advance
19
0
5.4k
Jan ’25
Unable to install Audio App Extension on iOS Device
I keep getting this error when trying to install Audio app extension. Everything is reviewed from certificates to profiles, for some reason CreatingCustomAudioEffects sample is deployed correctly but when Creating new Project (Audio Extension App) from new project Option, it doesn't work at all. If I remove Extension from Frameworks and deploy app, then no problem but then App crashes as extension is missing. Something wrong with Xcode? I am pretty sure it used to build new projects but not anymore. Failed to verify code signature of /var/installd/Library/Caches/com.apple.mobile.installd.staging/temp.0k1RXy/extracted/AudioUnitsAppExtensionTest.app : 0xe8008015 (A valid provisioning profile for this executable was not found.) Please ensure sure that your app is signed by a valid provisioning profile. If this issue persists, please attach the following when sending a report to Apple: A sysdiagnose from this Mac A sysdiagnose from the device failing installation An IPA of the app failing instal
4
0
572
Feb ’25
Popping sound when running XCode simulator
Hello, im getting popping / crackling sounds from my Macbook Pro (M4 2024) speakers. This happens when you do many certain tasks like click buttons or toggling switches when xcode has a simulator open and any background audio is playing, like spotify. The speakers go crazy especially when starting the simulator in xcode with music in background. Ive tried: Using blackhole, and changing audio output in the simulator app Deleting both .plist files form preferences file. coreaudiod trick in terminal restarting many times different xcode versions and simulators and swift files Nothing has worked. Any help?
Topic: UI Frameworks SubTopic: SwiftUI Tags:
1
0
130
Jun ’25
Reply to ASAF Panner Pro Tools Plug In
Bumping this because I haven't found where access these tools either! I'm surprised the announcement only mentioned AAX/Pro Tools and Davinci Resolve. (And no mention of Logic Pro?) Please add support AU/VST3 versions to supports DAWs like REAPER, which has become a go-to audio editor for higher order ambisonic and spatial work.
Topic: Media Technologies SubTopic: Audio Tags:
Jun ’25
Microphone Recording interrupts when phone ringing
I'm developing an iOS app that requires continuous audio recording. Currently, when a phone call comes in, the AVAudioSession is interrupted and recording stops completely during the ringing phase. While I understand recording should stop if the call is answered, my app needs to continue recording while the phone is merely ringing. I've observed that Apple's Voice Memos app maintains recording during incoming call rings. This indicates the hardware and iOS are capable of supporting this functionality. Request Please advise on any available AVAudioSession configurations or APIs that would allow my app to: Continue recording during an incoming call ring Only stop recording if/when the call is actually answered Impact This interruption significantly impacts the user experience and core functionality of my app. Workarounds like asking users to enable airplane mode are impractical and create a poor user experience. Questions Is there an approved way to maintain microphone access during call rings? If not
2
0
91
Jun ’25