Siri and Voice

RSS for tag

Help users quickly accomplish tasks related to your app using just their voice.

Posts under Siri and Voice tag

51 Posts
Sort by:
Post not yet marked as solved
0 Replies
72 Views
Hi, As the title, I figure out the throughput of iAP2(BT) is bad while launching siri. So that I want to reduce the data during using siri. Is any way to get that event? Search a lot I find a workaround solution https://stackoverflow.com/a/54822234/4562870. But in some case, it will be failed. Is any idea for getting start & stop siri event? Please help or try to give some ideas how to achieve this. Thanks!
Posted
by
Post not yet marked as solved
0 Replies
108 Views
I have a custom intent that allows a user to start a meeting - they don’t have to pick up their phone at all to start/join the meeting. This works perfectly on device - the intent is created, confirmed, and goes to the custom intent handle(intent:completion:) function where I call completion(StartMeetingIntentResponse(code: .continueInApp, userActivity: userActivity)). This opens the app and starts the meeting - all hands-free. When my device is connected via CarPlay, this intent can’t be completed. It gets all the way to where I call completion(StartMeetingIntentResponse(code: .continueInApp, userActivity: userActivity)), then Siri responds with “Sorry, I can’t do that while you’re driving.” It never makes it to my app delegate, where I have implemented application(_:userActivity:restorationHandler:) and would expect my app to pick up the activity to start the meeting. Why is this happening? I have implemented the INStartCallIntent similarly - calling completion(INStartCallIntentResponse(code: .continueInApp, userActivity: userActivity)) and that intent works just fine when connected to CarPlay, but my app has a distinction between calling and meeting so I really need both to work.
Posted
by
Post not yet marked as solved
0 Replies
99 Views
Hi, I have a question regarding the integration of the speech to text library called SFSpeechRecognizer. I need SFSpeechRecognizer to recognize terms that are not present in the iOS dictionary like medication names, chemistry terms, etc. I would have to add them, somehow, for SFSpeechRecognizer to be able to recognise them. Is this possible? Thanks
Posted
by
Post not yet marked as solved
7 Replies
433 Views
i tried to call my girlfriend using siri and well theres a bug with that she keeps trying to do it but she always fail with every command i gave her, also while using airpods siri doesnt appear at all. im currently on iphone 13 pro
Posted
by
Post not yet marked as solved
1 Replies
164 Views
I have implemented a custom intent that has several parameters. One of the parameters is called "dimensions". I would like to give the user a hint on what to say to respond to the Siri Dialog Prompt for the dimensions parameter. For example, if the Siri Dialog Prompt is "What are the dimensions?", I would like to have a one time prompt that says "What are the dimensions? You can say something like 2 x 4 x 35 inches". Thereafter, I would like to fall back to having Siri only say "What are the dimensions?". Is this possible?
Posted
by
Post not yet marked as solved
1 Replies
156 Views
My Siri announce notifications arent working good I have my AirPods Pro connected while listening to music but when they announce a message it sounds like the volume is really low to where you can barley hear Siri talk then once siri is done it goes back to the volume I had it to how do I fix it?
Posted
by
Post not yet marked as solved
0 Replies
166 Views
I am adding Siri Shortcuts to my navigation app. I understood in order to use Siri voice to trigger shortcuts, for custom intents, the users have to add voice phrases or record voice for shorctus by Add to Siri button or built-in Shortcuts app; for system intents, the users don't need to do that because Siri already knows all the trigger voice phrases. But when I say "Navigate to the station using Google Maps", the whole shortcut works without adding to Siri manullay. And based on https://developer.apple.com/documentation/sirikit, I couldn't find any system intents related to Navigation Domain. Did I misunderstand anything here? How could Siri and Google Maps exchange intents here without system navigation intents or adding to Siri manually?
Posted
by
Post not yet marked as solved
0 Replies
165 Views
I am working on a Siri integration for a VoIP application. I've added values for both INAlternativeAppName and INAlternativeAppNamePronunciationHint under INAlternativeAppNames in the app target's info.plist. On iOS 14, the phrase "Call [number] using [alternate app name]" launches my app and initiates a VoIP call. on iOS 15, Siri responds with "I don't see an app for that. You'll need to download one." to the same phrase. Is this functionality broken in iOS 15? Here is INAlternativeAppName in my info.plist: <key>INAlternativeAppNames</key> <array> <dict> <key>INAlternativeAppName</key> <string>dialer</string> <key>INAlternativeAppNamePronunciationHint</key> <string>dial er</string> </dict> </array>
Posted
by
Post not yet marked as solved
1 Replies
246 Views
Very recently my Apple TV started pausing anytime I turn off [HomeKit] connected lights in my house using Siri. It happens every time I tell Siri to turn off a light - whether using Siri on my HomePod, Apple Watch, or iPhone. I'm using the most current operating systems for all devices. It happens every time - no matter which light or lights I control. I have Hue lights and Sengled lights - both are affected the same way. Very very very frustrating. Any suggestions on how to solve this?
Posted
by
Post not yet marked as solved
1 Replies
176 Views
I cant seem to find anywhere else to submit this so hopefully I can get an answer here. For some reason, voice control has been randomly popping up over the past few months for seemingly no reason. I recently got so fed up with it, I spent a while researching how to get this to stop and I found a few sources that said I could disable voice control by going into my "accessibility" settings etc... AND IT STILL DOES IT. I even went in a disabled every toggle I could find. What do I need to do to prevent voice control from popping up on my phone randomly.
Posted
by
Post not yet marked as solved
0 Replies
194 Views
Where do I find and how do I install Accessory Developer Profile for CarPlay/Enhanced Siri? I do not see it anywhere in developer.apple.com/download. I am running the (MFI) CarPlay Tests 2.7 app on my 2 iPhone 7's, each with iOS 15.4.0 installed. Main and Accessory are paired and Main shows it's connected to my Head Unit. When I start the Enhanced Siri Detector Quality test on Main, it shows a popup: Feature not supported Make sure you have the Accessory Developer Profile installed with Enhanced Siri enabled, and you're running on the latest iOS version. Please, also check if your HeadUnit supports this feature. I know my HeadUnit supports this feature because I am able to run it when an iPhone 8 with iOS 15.4.0 is set up as Main. I now need to run it on my iPhone 7. I didn't have to set up this profile on the iPhone 8. I have Siri enabled on this iPhone 7 Main.
Posted
by
Post not yet marked as solved
1 Replies
243 Views
Hello, My app is facing a crasher with users albeit not a lot just a couple here and there. I've been trying to figure out the root cause but I'm not getting much information. Has anyone experienced this before? This is where the crash happens: 1 IntentsUI -[INUIEditVoiceShortcutViewController remoteViewControllerDidDeleteVoiceShortcutWithIdentifier:] + 88 2 IntentsUI -[INUIVoiceShortcutHostViewController remoteViewControllerDidDeleteVoiceShortcutWithIdentifier:] + 100 Not sure if it's related to something on my end as there are no memory leaks and I tested on device. Any input would be appreciated. Thanks.
Posted
by
pud
Post not yet marked as solved
0 Replies
158 Views
I think it is good that you were giving Siri another voice however I don’t think Siri should lose her old voice. Equity is about giving everyone what they need some people have trouble with changes, some people find it comforting to have a female voice some people find it comforting to have a male voice, some people find it comforting to have a neutral voice I love 💕 Siri just the way she is! whoever programs her jokes are very funny and makes me smile. I hope you are you considering giving options of Siri voice or are you going to be changing Siri’s voice for everyone?
Posted
by
Post not yet marked as solved
0 Replies
265 Views
For iOS, is there any way to add words into the dictionary that voice dictation recognizes and to the dictionary that the bar on top of the keyboard uses, that shows suggestions for words? I would like to make an iOS app that uses Biblical words such as "thee" and "thou". Any ideas will be appreciated.
Posted
by
Post not yet marked as solved
0 Replies
281 Views
I was trying to add multiple sender avatars to a group conversation communication notifications but it's not reflecting in UI. Every Sender object has an image property which has an image value but not all images are showing up in the notification avatar. Also, tried to replicate this locally by creating a custom collage and assign it as image but the InImage doesn't take the local images to render and just showing empty avatar. How can we add multiple avatars to communication notifications? And how can we read InImage from local?
Posted
by
Post not yet marked as solved
0 Replies
297 Views
I develop my an IM app for support Siri suggestion feature. I donate a INSendMessageIntent when user send message in group. I create INPerson array by users in that group. But the photo share sheet only show max 20 persons, the trust is I set persons parameters count is 56. Who can explain why limit show max group users count, or my anything wrong? I circle the key point in picture by red pencil. The second and the fourth chat that users in group count more than 20.
Posted
by
Post not yet marked as solved
0 Replies
289 Views
I would like to allow users to run an Apple Watch App action by using automations on the iPhone. The action can only be done on the watch and not the companion iPhone app? I noticed that it's possible to start a workout on the default Apple watch app with automations, and also possible to start a workout on the 3rd party Dawn Patrol app using automations. Is it possible to do this with custom intents or NSUserActivities or only the system start workout intent?
Posted
by
Post not yet marked as solved
1 Replies
341 Views
I have an app that lets users log drinks. Users create CoreData entries that have attributes like drink type and quantity. I have a view that lets users the quantity they have consumed in the current day. This is done using a FetchRequest and a calendar predicate that only fetches entries from the current day. I am using @FetchRequest property wrapper in the view, then initializing it with init() {}. Inside the initializer, I set the predicate, sort descriptors, and use self._todayIntakeEvents ... to set the value for the fetch request variable. Then, each time the view appears, I use .onAppear to sum the quantity of the fetched items by using todayIntakeEvents.map, then using .reduce to sum the array. The problem: when users add drinks with Siri, the drink information doesn't show up when opening the app from background. However, it will appear if I quit and relaunch the app. What I want to happen is for the drink to appear immediately when the user opens the app again. My code for the view: struct TodaySummaryView: View { @Environment(\.managedObjectContext) private var viewContext     @FetchRequest var todayIntakeEvents: FetchedResults<WaterIntakeEvent> init() {         let calendar = Calendar.current         let dateFrom = calendar.startOfDay(for: Date())         let dateTo = calendar.date(byAdding: .day, value: 1, to: dateFrom)         let predicateTodayEvents = NSPredicate(format: "timeOfConsumption <= %@ AND timeOfConsumption >= %@", dateTo! as CVarArg, dateFrom as CVarArg)         self._todayIntakeEvents = FetchRequest(             entity: WaterIntakeEvent.entity(),             sortDescriptors: [NSSortDescriptor(keyPath: \WaterIntakeEvent.timeOfConsumption, ascending: false)],             predicate: predicateTodayEvents)     } var body: some View { // some UI .onAppear { totalWaterQuantityToday = todayIntakeEvents.map {Int($0.waterQuantity)}             sumOfWaterQuantityToday = totalWaterQuantityToday.reduce(0, +) print(sumOfWaterQuantityToday) // even after Siri creates new CoreData entries, this still prints the quantity from before it added } My code for the Siri Intent Handler: public func handle(intent: LogDrinkIntent, completion: @escaping (LogDrinkIntentResponse) -> Swift.Void) {         let context = PersistenceController.shared.container.viewContext         let newDrink = WaterIntakeEvent(context: context)         newDrink.drinkType = intent.drinkType         newDrink.quantity = intent.quantity as! Float         newDrink.waterQuantity = intent.quantity as! Float         newDrink.timeOfConsumption = Date()         newDrink.id = UUID()         do {             try context.save()             let response = LogDrinkIntentResponse(code: LogDrinkIntentResponseCode.success, userActivity: nil)             completion(response)             print("Successfully saved.")         } catch {             completion(LogDrinkIntentResponse(code: LogDrinkIntentResponseCode.failure, userActivity: nil))             print("Error saving.")         }     } I suspect the problem is not Siri (because the data appears after relaunching the app) but that because I initialized the view with the FetchRequest, it's not responding to changes. Can someone help please?
Posted
by