Many of us Bangladeshi iPhone users were upset when Apple changed the font to Bangla in the most recent iOS version (18.4.1). We prefer the old Bangla typeface. I want the old Bangla typeface to return, and so do we. Please consider this.
Posts under iPhone tag
193 Posts
Selecting any option will automatically load the page
Post
Replies
Boosts
Views
Activity
Hello,
A quick background:
I am developing an App that receives a data stream from a device through its Wi-Fi network. The device itself is not connected to the internet, so the app won't be either.
Now, I am adding a new feature to the App that would require internet connection during the data stream. Consequently, my users would need to use their cellular data.
On later versions of iPhone, the phone would occasionally detect the lack of internet connection and asks the user via a pop-up if they want to use their cellular data. However, this behavior is not consistent.
So my question is- can we programmatically invoke this pop-up so the user can connect to the internet?
Or even better- can we program the App to use cellular data while still being connected to a Wi-Fi network?
Note:
I have seen mixed answers on the internet whether this is doable or not, and I know that users are able do it themselves by manually configuring their IP in their WiFi settings page, but I doubt this operation can be done through the App for security reasons.
Thanks!
If I want to display my corporations digital position would it be considered legal to use an iPhone SIM card reader as it seems to have a “data” point that hints at the idea of stabilization of intelligence that be found both off line and online.
I did a little research and it said Apples proprietary patents are referring to the size, ejection, magnetic trigger etc which leads me to want to know if the idea of using a SIM card reader to position the corporation as it uses a $USD specimen data file to position the idea of the corporation into a tangible format.
The part I’m referring to is the iPhone 11 sim tray and reader. The picture refers to the idea of the point on the back being similar to the ***** eye as something similar to the dollar seal.
I want to be a winner in the industry, I’ve got 50 plus “winner web domains” to bring to life and it’s time to buckle down and evolve the digital economy but first I want to validate the legality of me using disassembled Apple products to configure a “digital unit” that displays a “cyber analog”.
DYLD, symbol '_CTRadioAccessTechnologyNR' not found, expected in '/System/Library/Frameworks/CoreTelephony.framework/CoreTelephony'
In the public release of iOS 18.4.1 and iPadOS 18.4.1, external input support for keyboards and mice is critically degraded. This issue affects both Apple-branded and third-party HID-compliant devices, over both wired USB-C and Bluetooth.
Tested Hardware:
• iPhone 16 Pro Max (256GB)
• iPad Pro (USB-C, latest gen), last gen iPad as well
Affected Devices:
• Apple Magic Mouse and Keys (wired USB-C/Bluetooth)
• Redragon K580RGBPRO (Bluetooth/wired USB-C)
• Razer Naga V2 Pro (Bluetooth/USB-C)
Symptoms:
• Severe keystroke delay and dropped input
• Modifier keys (Shift, Command, Option) fail intermittently
• Input degrades further with multiple HID devices connected
• Mouse input via Bluetooth exhibits pointer lag and jitter
• Occurs in all apps: Notes, Safari, Mail, text fields, password entries, etc.
• Identical results using Apple USB-C cables
Reproducibility:
100%. Clean boots, minimal background activity, and isolated environments (including Airplane Mode) do not resolve the issue. Identical behavior across both iPhone and iPad.
Expected Behavior:
All HID-compliant external input devices — particularly Apple-branded ones — should provide low-latency, reliable, and consistent input over both USB-C and Bluetooth, especially in a production iOS/iPadOS release.
Actual Behavior:
External keyboards and mice exhibit:
• Lag
• Dropped characters
• Failed modifiers
• Degraded mouse tracking
Even with the latest hardware and clean configurations.
Severity:
Critical.
This is a platform-level failure affecting I/O at the user interaction layer. Input reliability is non-negotiable — especially on $2000+ flagship devices using Apple’s own peripherals.
Closing Note (for Apple engineering & peer devs):
This is not a beta regression — it’s a public release flaw that undermines iOS and iPadOS usability for power users, professionals, and accessibility communities alike. That Apple Magic Keyboard, Redragon, and Razer gear all fail equally and consistently should be a wake-up call.
Apple: this needs to be escalated. Now.
External input — one of the most basic subsystems in any OS — is broken on your highest-end devices.
In our web application some functionalities will allow user to upload multiple images (More than 25 images) in a single page
It is working find in all OS and browsers except iOS
When user try to upload images directly from camera there will be some overlaps, duplication, missing etc.
This is happening in both Safari and Chrome, we had a thorough check in our application and found every thing is working fine from our end
You can reproduce the issue by creating a web page which accept more than 50 images (we tried the same in ASP MVC Core & PHP) and showing the images in order
access the page through your iPhone using Safari or Chrome
Try to upload images directly from your camera, try sequential images (Image of a stop watch, or some thing like that) so that you can easily identify the order of files uploaded
and check the listing page of uploaded image (Try these steps multiple times)
You can find some images are duplicated and some are missing
I’d love to see Apple implement a Bionic Reading feature as a system-wide accessibility option. This type of reading aid highlights the first part of each word in bold to help guide the eyes and improve comprehension.
It’s been shown to be especially helpful for people with ADHD, dyslexia, and other neurodivergent needs. Having a toggle in Settings > Accessibility would be life-changing.
Ideally, it could be:
• Enabled system-wide, or per-app
• Allow customization of how much of the word is bolded
• Available in Safari, Messages, Books, News, etc.
Dear Apple Engineers,
Recently, I have switched from the iPhone 13 Pro to the iPhone 14 Pro, and every single notification has not lit up my screen even though I heard the notification sounds as well as the vibrations. I have made sure dnd is off, no other focus mode is on, attention aware features are off, no Apple Watch connected, forced restart twice, normal restart multiple times.
Upon more research, I have discovered that it has been a prominent issue since September 2023 as you can see from this Apple discussion forum.
https://discussions.apple.com/thread/255133907?answerId=260178222022&sortBy=newest_first#260178222022
I find it deeply concerning that, despite reports from multiple users, Apple has yet to provide a permanent solution or acknowledge the defect publicly. This lack of action undermines consumer trust. I hope to resolve this issue amicably and continue supporting Apple as a brand I trust.
Thank you for your immediate attention to this matter.
I am developing a video streaming app for iPhone.
Minimum version is IOS 13.
I want to connect an external USB camera to the iPhone app and stream from it.
I have looked through a lot of information and have not found how to do this.
Is it possible to do this? Is there any documentation on this?
Hi all,
I’m running into a persistent linker error when building my Unity 6 project (IL2CPP, iOS target) that calls a Swift method via an Objective-C++ wrapper. Despite following all known steps, I keep getting:
Undefined symbols for architecture arm64:
"_placeGeoAnchor", referenced from:
_GeoAnchorTrigger_placeGeoAnchor in libGameAssembly.a
...
ld: symbol(s) not found for architecture arm64
I’m trying to place a persistent AR anchor at real-world GPS coordinates (so that the same asset can appear at the same location for a returning user). Since I’m targeting iOS, I can’t use Google’s geospatial anchors (but I sooo wish I could--please apple I beg of you stop being so selfish lol).
I've already done these things:
Swift file is added to Unity-iPhone target.
.mm and .h files are in Unity-iPhone target under Compile Sources.
Bridging header is set to Unity-iPhone-Bridging-Header.h.
Generated header name is correct (GeoTest-Swift.h).
Build Active Architecture Only set to No.
Function has attribute((visibility("default"))).
Unity project uses IL2CPP scripting backend.
Yet I'm still getting the same linker error — it appears Unity (via IL2CPP) references the function, but Xcode doesn't link it.
It’s something small that’s being missed with how IL2CPP links native symbols? Or maybe I need to explicitly include something in Link Binary With Libraries? I’ve verified symbol visibility and targets repeatedly.
I’ve built AR features in Unity before (for Quest), but this is my first time trying to bridge C# → Objective-C++ → Swift in this way for a geolocation-based AR anchor for an iphone.
I'm going crazy, I’ve been stuck on this for 12+ hours now, so any insight or nudge would be deeply appreciated.
SPECS:
Macbook Pro M4 Pro--Sequoia 15.4
Unity 6000.0.45f1
IPhone 11 iOS 18.4
Xcode 15
We have identified an issue when using NumberFormatter with the locale set to it_IT. Specifically, when formatting numbers with exactly four integer digits, the grouping separator is not applied: for example, the number is displayed as 4000,00 instead of the expected 4.000,00. This behavior occurs only with four-digit integers; for instance, 40.000,00 is formatted correctly. The issue appears to affect only iOS 18.4 and later versions.
I am trying to track a user's real-time sleep state using heart rate data, but I have encountered several issues:
When using HKSampleQuery on the phone to fetch heart rate data, I can only retrieve data recorded before the app comes to the foreground or before it is terminated and restarted (see related issue: https://developer.apple.com/forums/thread/774953).
I attempted to get data on the Apple Watch and send updates to the phone via Watch Connectivity. However, if I use WKExtendedRuntimeSession, although I can obtain data on the watch, once the watch screen goes off, it can no longer transmit data via Watch Connectivity to the phone (since I cannot guarantee the app will remain in the foreground when lying in bed).
On the other hand, using HKWorkoutSession results in interference with the activity rings and causes the heart rate sensor to run too frequently, which I worry may affect the battery life of the watch.
Is there an elegant solution for tracking a user's heart rate data for sleep monitoring?
14Pro can not delete any application after upgrade system18.4
I will give some suggestions and action buttons about dynamic islands on iPhonesAfter dragging and dropping an application or file on dynamic islands, when I switch to any application or site, I can switch to that file or application by holding down the Dynamic Island As for the action button, we can assign features to the action button. A feature like this can come to it. When we press it twice, it will open the camera, when we press it once, it will open the flash. If such a feature comes, it will be easier to use
Hello,
I uploaded a new iOS build (1.0.1 - Build 180) to App Store Connect for my app "Benimle Konus". The build has been processed successfully and is marked as "Ready to Test". However, when I try to assign this build to any internal TestFlight group, the group checkboxes are greyed out and cannot be selected.
I’ve made sure that:
All required metadata is filled out.
What’s New and compliance information is complete.
Previous builds worked with the same TestFlight groups.
To fix this, I have uploaded 3 separate builds, but the issue still persists with all of them.
There are no error messages, but I can't continue testing because I can't assign the build to testers.
App ID: 6742759002
Bundle ID: com.benimlekonus.app
Build Version: 1.0.1 (180)
Is this a known issue or something I'm missing? Any help would be appreciated.
Topic:
App Store Distribution & Marketing
SubTopic:
TestFlight
Tags:
iPhone
Xcode
TestFlight
Testing
Hello, aspiring programmer here.
I am developing a StepCounter APP, which keeps track of how many steps I have taken and sends to an MQTT server. I am trying to make this happen even while the app is not in focus, but so far I have not been able to get this working.
First tried with silent background music, which seemed pretty inconsistent and inpractical, since I usually play youtube videoes while walking, making the app stop with its silent audio. Then tried GPS, which didnt really do anything (could be implementation problem).
Has anyone made background processing work for their apps?
I am trying to open items saved to my files and keep receiving this notification every time I do so, “The operation couldn't be completed. (NSFileProvider-
ErrorDomain error -2005.)” What does this mean and how do I fix it? I have a 256 gb phone so I know storage is not an issue. Please help as I have files saved to my phone and not anywhere else.
Hello, aspiring programmer here.
I am developing a StepCounter APP, which keeps track of how many steps I have taken and sends to an MQTT server. I am trying to make this happen even while the app is not in focus, but so far I have not been able to get this working.
First tried with silent background music, which seemed pretty inconsistent and inpractical, since I usually play youtube videoes while walking, making the app stop with its silent audio. Then tried GPS, which didnt really do anything (could be implementation problem).
Has anyone made background processing work for their apps?
I ran into a problem. When using Storekit1 to purchase an SKU, the user payment was successful, but StoreKit1 did return paymentCancelled to my App. I would like to know under what circumstances this problem may occur? How do I fix it? Thank you
Hello, I have to create an app in Swift that it scan NFC Identity card. It extract data and convert it to human readable data. I do it with below code
import CoreNFC
class NFCIdentityCardReader: NSObject , NFCTagReaderSessionDelegate {
func tagReaderSessionDidBecomeActive(_ session: NFCTagReaderSession) {
print("\(session.description)")
}
func tagReaderSession(_ session: NFCTagReaderSession, didInvalidateWithError error: any Error) {
print("NFC Error: \(error.localizedDescription)")
}
var session: NFCTagReaderSession?
func beginScanning() {
guard NFCTagReaderSession.readingAvailable else {
print("NFC is not supported on this device")
return
}
session = NFCTagReaderSession(pollingOption: .iso14443, delegate: self, queue: nil)
session?.alertMessage = "Hold your NFC identity card near the device."
session?.begin()
}
func tagReaderSession(_ session: NFCTagReaderSession, didDetect tags: [NFCTag]) {
guard let tag = tags.first else {
session.invalidate(errorMessage: "No tag detected")
return
}
session.connect(to: tag) { (error) in
if let error = error {
session.invalidate(errorMessage: "Connection error: \(error.localizedDescription)")
return
}
switch tag {
case .miFare(let miFareTag):
self.readMiFareTag(miFareTag, session: session)
case .iso7816(let iso7816Tag):
self.readISO7816Tag(iso7816Tag, session: session)
case .iso15693, .feliCa:
session.invalidate(errorMessage: "Unsupported tag type")
@unknown default:
session.invalidate(errorMessage: "Unknown tag type")
}
}
}
private func readMiFareTag(_ tag: NFCMiFareTag, session: NFCTagReaderSession) {
// Read from MiFare card, assuming it's formatted as an identity card
let command: [UInt8] = [0x30, 0x04] // Example: Read command for block 4
let requestData = Data(command)
tag.sendMiFareCommand(commandPacket: requestData) { (response, error) in
if let error = error {
session.invalidate(errorMessage: "Error reading MiFare: \(error.localizedDescription)")
return
}
let readableData = String(data: response, encoding: .utf8) ?? response.map { String(format: "%02X", $0) }.joined()
session.alertMessage = "ID Card Data: \(readableData)"
session.invalidate()
}
}
private func readISO7816Tag(_ tag: NFCISO7816Tag, session: NFCTagReaderSession) {
let selectAppCommand = NFCISO7816APDU(instructionClass: 0x00, instructionCode: 0xA4, p1Parameter: 0x04, p2Parameter: 0x00, data: Data([0xA0, 0x00, 0x00, 0x02, 0x47, 0x10, 0x01]), expectedResponseLength: -1)
tag.sendCommand(apdu: selectAppCommand) { (response, sw1, sw2, error) in
if let error = error {
session.invalidate(errorMessage: "Error reading ISO7816: \(error.localizedDescription)")
return
}
let readableData = response.map { String(format: "%02X", $0) }.joined()
session.alertMessage = "ID Card Data: \(readableData)"
session.invalidate()
}
}
}
But I got null. I think that these data are encrypted. How can I convert them to readable data without MRZ, is it possible ?
I need to get personal informations from Identity card via Core NFC.
Thanks in advance.
Best regards