I thought Apple was all about inclusivity. Is this not harming their brand more than they’d realize?
Explore best practices for creating inclusive apps for users of Apple accessibility features and users from diverse backgrounds.
Post
Replies
Boosts
Views
Activity
I have pop-ups on the top row keys randomly that prevent me from typing. Since that I cannot answer comments in Youtube and Creator Apps. That kills my work process.
I don't wanna use Beta any update anymore, because it stops me from producing anymore
You guys killed me
Hi, I noticed that holding a guitar pick while wearing the headset is a constant input since you hold it between your thumb and index fingers.
it makes playing guitar very difficult. Piano is a bit better but I feel like being able to block hand inputs is necessary.
Hello everyone :)
For a project application we try to create a QR code that leads the user to the Software Update settings on iPhone after scanning to improve the accessibility for people who are not experienced in using smartphones.
On iOS 18.0.1 I see that using App-prefs: opens the settings app which is a good starting point. Furthermore, using e.g. App-prefs:com.apple.MobileSMS directs the user to the Settings of the Messages app.
But things like App-prefs:root=SoftwareUpdate don't work. That's why I wanted to kindly ask, if someone may have other ideas or knows what the exact URL might be for opening the Software Update settings or at least the General settings (App-prefs:root=General doesn't work).
If anyone has insights or can share the exact URL scheme for this purpose, I would greatly appreciate it!
Thank you so much!
Hi, is there any way to interact with the eye tracking accessibility feature in iOS and iPadOS? I want to be able to trigger an eye tracker recalibration programmatically. Also if possible, I would like to be able to retrieve gaze location data.
These are not intended to be features on an app being published to the app store but rather a custom made accessibility app.
I have a record button that either starts or stops a recording using the default action. When the user is recording, I want to add a custom action to discard the recording instead of saving it. That all works fine with the following code:
if isRecording {
recordButton.accessibilityCustomActions = [
.init(name: String(localized: "discard recording"), actionHandler: { [weak delegate] _ in
delegate?.discardRecording()
return true
})
]
recordButton.accessibilityLabel = String(localized: "stop recording", comment: "accessibility label")
} else {
recordButton.accessibilityCustomActions = []
recordButton.accessibilityLabel = String(localized: "start recording", comment: "accessibility label")
}
The problem I have is that when a user chose "discard recording", it becomes the default selected action again the next time the user records, and instead of stopping and saving the recording, the user might accidentally discard the next one as well.
How can I programmatically reset the selected action on this recordButton to the default action?
I am developing an application for macOS that runs exclusively in the menu bar, and in this application, I am using a ColorPicker to allow the user to choose a color. However, the ColorPicker is not working properly in the menu bar. What should I do to resolve this issue?
Hello,
after installing iOS 18 on our smartphones we experience a dramatic CPU load of the thread "AccessibilityUIServer" after installing one of our apps. This results in a drastic battery drainage and heats up the iPhone very fast. Under previous iOS versions this was no problem.
Do you have any ideas how this could happen?
I want to create a utility to import a list of words to the Voice Control user custom vocabulary. Is there an API to do this? I noticed if you use the built-in export vocabulary functionality (Settings > Accessibility > Voice Control > ...) the file that gets exported is a plist document type. If there is no API to add words programmatically should I just create a utility that generates a plist file and import it using the built-in import vocabulary functionality (Settings > Accessibility > Voice Control > ...)?
When the native info panel (which displays the title, subtitle, description, and custom buttons) opens, the focus immediately shifts to the first button. As a result, VoiceOver skips the description, which is crucial for users relying on accessibility features.
I haven’t found a way to detect when it opens. Knowing this would allow me to trigger custom VoiceOver announcements or adjust the focus order dynamically.
Are any other people experiencing this issue, and how do we solve it?
It’s impossible to enjoy video playback with this update. 1) Starting with the play/pause location, not friendly with your hand; 2) How are we supposed to back and forth just a few frames? I was thinking there is a hidden button somewhere to enable the old playback, because that’s waaay absurd.
3) Video playback with buttons enable makes video playback smaller, that’s ridiculous
Is it possible to spoof mac adress of Ventura 13.6.9?
I'm experiencing a heating issue with my iPhone 15 Pro Max. The device heats up even during light usage, such as browsing, checking social media, or using simple apps. It becomes noticeably warm, even though I'm not running any demanding tasks like gaming or video editing. I’ve tried restarting the phone, closing background apps, and ensuring that iOS is fully updated, but the problem persists.
Has anyone else encountered this issue with the iPhone 15 Pro Max?
What could be causing my device to heat up during regular use? Are there specific settings or steps I can take to reduce the heating? Is it more likely to be a software-related issue or a hardware problem? Any suggestions or solutions would be appreciated!
Has anyone else encountered this issue with the iPhone 15 Pro Max?
What could be causing my device to heat up during regular use? Are there specific settings or steps I can take to reduce the heating? Is it more likely to be a software-related issue or a hardware problem? Any suggestions or solutions would be appreciated!
I frequently use my iPad to develop remotely in VSCode or prototype designs in Figma. This is currently all done in the browser. Given that many of these experiences rely heavily on the keyboard I was hoping there would be a solution to make a keyboard persistent on the screen, or at least a few hot keys.
Is it possible for me to develop an accessibility tool that could stay persistent on the screen? Perhaps something that would talk with assistive touch? Or is that in Apple’s no no square?
Hi i have an app built by Cordova and Vue.js, and IOS cannot recognize the new prompt, when the old prompt still exist, and in android it works good.
Facing issue in earphone not working in iPhone 12 after updating iOS 18.1. It is so frustrating. Can anyone have any solution to this? Nothing is happening after resetting or rebooting the phone many times.
Does giving the ability for blocked users to still view the users who blocked them, despite not being able to reply, retweet, or otherwise directly engage with that user go against Apple‘s TOS for user created content? It feels a bit ambiguous.
Hello, I have a question regarding the voice and sound recognition features on the iPhone 15 Pro.
The iPhone 15 Pro is equipped with four microphones, and I understand that for features like Apple’s sound recognition and when invoking Siri, the microphone(s) must always be active. My question is whether the device uses a single microphone (mono channel) for these functions or if multiple microphones are activated simultaneously.
I would appreciate clarification on how the microphones are utilized in sound and voice recognition features.
Thank you for your assistance.
Best regards.
Hello, I have a question regarding the voice and sound recognition features on the iPhone 15 Pro.
The iPhone 15 Pro is equipped with four microphones, and I understand that for features like Apple’s sound recognition and when invoking Siri, the microphone(s) must always be active. My question is whether the device uses a single microphone (mono channel) for these functions or if multiple microphones are activated simultaneously.
I would appreciate clarification on how the microphones are utilized in sound and voice recognition features.
Thank you for your assistance.
Best regards.
Apple's testers tell me that "the app does not meet all requirements for apps that offer highly regulated services or handle sensitive user data," (however, the APP does not offer any type of regulated service), which is why which indicate that it must be published with a user of the Apple Developer program as an organization and not as an individual.
How can I correct this problem?