I work on a MacBook Pro M3 Max 36GB RAM
I am running macOS Tahoe Version 26.1 Beta (23B5042k)
The last few days I had a couple of situations where Blender 4.5 LTS Froze and I had to forcequit, I noticed Blender would get Deleted with no notification after the forcequit, it was nowhere to be found on the Application List, Finder.
Thought it was a blender issue because I run a special module with it.
However around 30min ago I was using Mem and the App froze and I had to ForceQuit, The app also got removed, is not on the list of application and nowhere to be found however the icon on the dock which remained shows a Question Mark.
I think this is an OS Behavior I just dont know what to attribute it to.
Posts under Beta tag
200 Posts
Selecting any option will automatically load the page
Post
Replies
Boosts
Views
Activity
After installing my notarized 3rd party app in a Tahoe VM, its embedded Automator actions can not be configured in Automator while defining a workflow: After adding the actions (enabling 3rd party extensions), their views / UI elements do not respond to any mouse event.
When enabling „show this action when running“, the options can be changed during execution of the workflow. Needless to say: Adjusting these action settings in Automator was working for years, macOS 12 - 15 and before.
Reported via Feedback Assistent (FB19015185).
Can anybody confirm this issue with Automator actions?
Ok, so I’ve downloaded iOS 26 for the iPhone and iPad and there are a couple of points bugging me.
Why oh why would somebody think that it was a good idea to remove the ability to scroll through photos in a folder in files?
Where in their tiny minds did they think that opening a photo, viewing said photo, then closing photo to then open the next photo in sequence was a better or more productive way than just swiping left and right?
I mean, seriously!!!
And why the need for a separate Preview app when it worked fine opening files in the Files app?!
I deleted the Preview app from my iPhone to see what difference it would make, but alas it made things worse. Now, because there’s no Preview app, Files wants to open my photos in an image editing app called Afterlite!
I checked settings for the files app, and the only options I have for opening photos are Afterlite (which it’s decided to make default), the Photoshop app, or iMovie!
Does anyone have any ideas how to get photos stored in the Files app to open IN the Files app or at least the Photos app?
Just when I thought they’d come to their senses and sorted out that god awful iOS 18 Photos app, they go and do something like this!
Also, can someone explain to me what the point is of having a month and year view in the Photos app?
When I sort by month and pick a particular month or date in a month, instead of showing just those images taken that month, it just takes you back to the view with ALL of your photos?
Surely if you have thousands of photos on your device and you want to see only photos on a particular date, it would make sense to show only the photos from that date.
Even if it generated automatic albums under the album view for each month and year that would be something!
I am a developer working on iOS apps.
I would like to report an issue occurring in iOS 26 beta 2.
Our company has Enterprise account, and we are developing apps.
When we distribute these apps, and install them on a device running iOS 26 beta2, apps install successfully, but apps crashed immediately after being launched.
MDM Install Application
When I install the app via Xcode and trust it, apps will run.
Launchd job spawn failed
This issue does not occur on versions prior to iOS 26. I would like to know if this is a problem that will be resolved in future updates, or if it is a policy change.
When trying to use a UISearchController setup with a UISearchBar that has scope buttons, the search controller's scopeBarActivation property is set to .onSearchActivation, the navigation item's preferredSearchBarPlacement property is set to .integrated. or .integratedButton, and the search bar/button appears in the navigation bar, then the scope buttons never appear. But space is made for where they should appear.
Some relevant code in a UIViewController shown as the root view controller of a UINavigationController:
private func setupSearch() {
let sc = UISearchController(searchResultsController: UIViewController())
sc.delegate = self
sc.obscuresBackgroundDuringPresentation = true
// Setup search bar with scope buttons
let bar = sc.searchBar
bar.scopeButtonTitles = [ "One", "Two", "Three", "Four" ]
bar.selectedScopeButtonIndex = 0
bar.delegate = self
// Apply the search controller to the nav bar
navigationItem.searchController = sc
// BUG - Under iOS/iPadOS 26 RC, using .onSearchActivation results in the scope buttons never appearing at all
// when using integrated placement in the nav bar.
// Ensure the scope buttons appear immediately upon activating the search controller
sc.scopeBarActivation = .onSearchActivation
// This works but doesn't show the scope buttons until the user starts typing - that's too late for my needs
//sc.scopeBarActivation = .automatic
if #available(iOS 26.0, *) {
// Under iOS 26 put the search icon in the nav bar - same issue for .integrated and .integratedButton
navigationItem.preferredSearchBarPlacement = .integrated // .integratedButton
// My toolbar is full so I need the search in the navigation bar
navigationItem.searchBarPlacementAllowsToolbarIntegration = false // Ensure it's in the nav bar
} else {
// Under iOS 18 put the search bar in the nav bar below the title
navigationItem.preferredSearchBarPlacement = .stacked
}
}
I need the search bar in the navigation bar since the toolbar is full. And I need the scope buttons to appear immediately upon search activation.
This problem happens on any real or simulated iPhone or iPad running iOS/iPadOS 26 RC.
I just recently updated to iOS 26 beta (23A5336a) to test an app I am developing
I running an MLModel loaded from a .mlmodelc file.
On the current iOS version 18.6.2 the model is running as expected with no issues.
However on iOS 26 I am now getting error when trying to perform an inference to the model where I pass a camera frame into it.
Below is the error I am seeing when I attempt to run an inference.
at the bottom it says "Failed with status=0x1d : statusType=0x9: Program Inference error status=-1 Unable to compute the prediction using a neural network model. It can be an invalid input data or broken/unsupported model " does this indicate I need to convert my model or something? I don't understand since it runs as normal on iOS 18.
Any help getting this to run again would be greatly appreciated.
Thank you,
processRequest:model:qos:qIndex:modelStringID:options:returnValue:error:: Could not process request ret=0x1d lModel=_ANEModel: { modelURL=file:///var/containers/Bundle/Application/04F01BF5-D48B-44EC-A5F6-3C7389CF4856/RizzCanvas.app/faceParsing.mlmodelc/ : sourceURL=(null) : UUID=46228BFC-19B0-45BF-B18D-4A2942EEC144 : key={"isegment":0,"inputs":{"input":{"shape":[512,512,1,3,1]}},"outputs":{"var_633":{"shape":[512,512,1,19,1]},"94_argmax_out_value":{"shape":[512,512,1,1,1]},"argmax_out":{"shape":[512,512,1,1,1]},"var_637":{"shape":[512,512,1,19,1]}}} : identifierSource=1 : cacheURLIdentifier=01EF2D3DDB9BA8FD1FDE18C7CCDABA1D78C6BD02DC421D37D4E4A9D34B9F8181_93D03B87030C23427646D13E326EC55368695C3F61B2D32264CFC33E02FFD9FF : string_id=0x00000000 : program=_ANEProgramForEvaluation: { programHandle=259022032430 : intermediateBufferHandle=13949 : queueDepth=127 } : state=3 :
[Espresso::ANERuntimeEngine::__forward_segment 0] evaluate[RealTime]WithModel returned 0; code=8 err=Error Domain=com.apple.appleneuralengine Code=8 "processRequest:model:qos:qIndex:modelStringID:options:returnValue:error:: ANEProgramProcessRequestDirect() Failed with status=0x1d : statusType=0x9: Program Inference error" UserInfo={NSLocalizedDescription=processRequest:model:qos:qIndex:modelStringID:options:returnValue:error:: ANEProgramProcessRequestDirect() Failed with status=0x1d : statusType=0x9: Program Inference error}
[Espresso::handle_ex_plan] exception=Espresso exception: "Generic error": ANEF error: /private/var/containers/Bundle/Application/04F01BF5-D48B-44EC-A5F6-3C7389CF4856/RizzCanvas.app/faceParsing.mlmodelc/model.espresso.net, processRequest:model:qos:qIndex:modelStringID:options:returnValue:error:: ANEProgramProcessRequestDirect() Failed with status=0x1d : statusType=0x9: Program Inference error status=-1
Unable to compute the prediction using a neural network model. It can be an invalid input data or broken/unsupported model (error code: -1).
Error Domain=com.apple.Vision Code=3 "The VNCoreMLTransform request failed" UserInfo={NSLocalizedDescription=The VNCoreMLTransform request failed, NSUnderlyingError=0x114d92940 {Error Domain=com.apple.CoreML Code=0 "Unable to compute the prediction using a neural network model. It can be an invalid input data or broken/unsupported model (error code: -1)." UserInfo={NSLocalizedDescription=Unable to compute the prediction using a neural network model. It can be an invalid input data or broken/unsupported model (error code: -1).}}}
Hello everyone,
I uploaded my first app to the App Store for TestFlight.
Since I read that TestFlight reviews are usually completed within a few hours, I am a bit unsure whether I might have done something wrong in the review process. I submitted the app to TestFlight on September 10, and after some back and forth, I received a message on September 13 saying:
“Your submission is still in review but is requiring additional time. We will provide further status updates as soon as we are able.”
On Tuesday, September 16, I sent an email to Support asking if I had done something wrong and mentioning that I uploaded a new build.
Apple then put the new build into review, and on the same day I received another message that the review would take some more time. Since then, I haven’t heard anything further. The older build is still shown as “In Review” (since September 12).
Is there anything I can do at this point?
Hi.
We are writing to report a critical issue we've encountered following the recent release of iOS 26 beta 6.
After updating our test devices, we discovered that our application is no longer able to establish HTTPS connections to several of our managed FQDNs. This issue was not present in beta 5 and appears to be a direct result of changes introduced in beta 6.
The specific FQDNs that are currently unreachable are:
d.socdm.com
i.socdm.com
tg.scodm.com
We have reviewed the official iOS & iPadOS 26 Beta 6 Release Notes, particularly the updates related to TLS. While the notes mention changes, we have confirmed that our servers for all affected FQDNs support TLS 1.2, so we believe they should still be compliant.
We have also investigated several of Apple's support documents regarding TLS connection requirements (e.g., HT214774, HT214041), but the information does not seem to apply to our situation, and we are currently unable to identify the root cause of this connection failure.
https://support.apple.com/en-us/102028
https://support.apple.com/en-us/103214
Although we hope this issue might be resolved in beta 7 or later, the official release is fast approaching, and this has become a critical concern for us.
Could you please provide any advice or insight into what might be causing this issue? Any guidance on potential changes in the networking or security frameworks in beta 6 that could affect TLS connections would be greatly appreciated.
We have attached the relevant code snippet that triggers the error, along with the corresponding Xcode logs, for your review.
Thank you for your time and assistance.
#import "ViewController.h"
@interface ViewController ()
@end
@implementation ViewController
- (void)viewDidLoad {
[super viewDidLoad];
NSURL *url = [NSURL URLWithString:@"https://i.socdm.com/sdk/js/adg-script-loader-b-stg.js"];
NSMutableURLRequest *req = [NSMutableURLRequest requestWithURL:url
cachePolicy:NSURLRequestReloadIgnoringLocalCacheData
timeoutInterval:30.0];
[self sendWithRequest:req completionHandler:^(NSData *_Nullable data,
NSHTTPURLResponse *_Nonnull response,
NSError *_Nullable error) {
if (error){
NSLog(@"Error occurred: %@", error.localizedDescription);
return;
}else{
NSLog(@"Success! Status Code: %ld", (long)response.statusCode);
}
}];
}
- (void) sendWithRequest:(NSMutableURLRequest *)request
completionHandler:(void (^ _Nullable)(NSData *_Nullable data,
NSHTTPURLResponse *response,
NSError *_Nullable error))completionHandler {
NSURLSessionConfiguration *configuration = [NSURLSessionConfiguration defaultSessionConfiguration];
NSURLSession *session = nil;
session = [NSURLSession sessionWithConfiguration:configuration
delegate:self
delegateQueue:nil];
NSURLSessionTask *task = [session dataTaskWithRequest:request
completionHandler:^(NSData *data, NSURLResponse *response, NSError *error) {
[session finishTasksAndInvalidate];
NSHTTPURLResponse *httpResponse = (NSHTTPURLResponse *) response;
if (error) {
if (completionHandler) {
completionHandler(nil, httpResponse, error);
}
}
else {
if (completionHandler) {
completionHandler(data, httpResponse, nil);
}
}
}];
[task resume];
}
@end
error
Connection 1: default TLS Trust evaluation failed(-9807)
Connection 1: TLS Trust encountered error 3:-9807
Connection 1: encountered error(3:-9807)
Task <C50BB081-E1DA-40FF-A1E5-A03A2C4CB733>.<1> HTTP load failed, 0/0 bytes (error code: -1202 [3:-9807])
Task <C50BB081-E1DA-40FF-A1E5-A03A2C4CB733>.<1> finished with error [-1202] Error Domain=NSURLErrorDomain Code=-1202 "The certificate for this server is invalid. You might be connecting to a server that is pretending to be “i.socdm.com” which could put your confidential information at risk." UserInfo={NSLocalizedRecoverySuggestion=Would you like to connect to the server anyway?, _kCFStreamErrorDomainKey=3, NSErrorPeerCertificateChainKey=(
"<cert(0x10621ca00) s: *.socdm.com i: GlobalSign RSA OV SSL CA 2018>",
"<cert(0x106324e00) s: GlobalSign RSA OV SSL CA 2018 i: GlobalSign>"
), NSErrorClientCertificateStateKey=0, NSErrorFailingURLKey=https://i.socdm.com/sdk/js/adg-script-loader-b-stg.js, NSErrorFailingURLStringKey=https://i.socdm.com/sdk/js/adg-script-loader-b-stg.js, NSUnderlyingError=0x1062bf960 {Error Domain=kCFErrorDomainCFNetwork Code=-1202 "(null)" UserInfo={_kCFStreamPropertySSLClientCertificateState=0, kCFStreamPropertySSLPeerTrust=<SecTrustRef: 0x10609d140>, _kCFNetworkCFStreamSSLErrorOriginalValue=-9807, _kCFStreamErrorDomainKey=3, _kCFStreamErrorCodeKey=-9807, kCFStreamPropertySSLPeerCertificates=(
"<cert(0x10621ca00) s: *.socdm.com i: GlobalSign RSA OV SSL CA 2018>",
"<cert(0x106324e00) s: GlobalSign RSA OV SSL CA 2018 i: GlobalSign>"
)}}, _NSURLErrorRelatedURLSessionTaskErrorKey=(
"LocalDataTask <C50BB081-E1DA-40FF-A1E5-A03A2C4CB733>.<1>"
), _kCFStreamErrorCodeKey=-9807, _NSURLErrorFailingURLSessionTaskErrorKey=LocalDataTask <C50BB081-E1DA-40FF-A1E5-A03A2C4CB733>.<1>, NSURLErrorFailingURLPeerTrustErrorKey=<SecTrustRef: 0x10609d140>, NSLocalizedDescription=The certificate for this server is invalid. You might be connecting to a server that is pretending to be “i.socdm.com” which could put your confidential information at risk.}
Error occurred: The certificate for this server is invalid. You might be connecting to a server that is pretending to be “i.socdm.com” which could put your confidential information at risk.
折りたたむ
Hello community,
After unable to use Metal Toolchain for Xcode Beta 1 and 2 even after following the workarounds, and using Xcode Beta 3, I am still unable to use the downloaded Metal Compiler on my device.
Building any project with metal file results in
warning: Could not read serialized diagnostics file: error("Failed to open diagnostics file") (in target '<target>' from project '<project>')
error: error: cannot execute tool 'metal' due to missing Metal Toolchain; use: xcodebuild -downloadComponent MetalToolchain
Command CompileMetalFile failed with a nonzero exit code
Here is my build environment,
$ xcodebuild -version
Xcode 26.0
Build version 17A5276g
I have also checked the downloaded metal toolchain.
$ xcodebuild -downloadComponent metalToolchain -exportPath /tmp/MyMetalExport/
2025-07-13 13:16:17.293 xcodebuild[2153:34019] IDEDownloadableMetalToolchainCoordinator: Failed to remount the Metal Toolchain: The file “com.apple.MobileAsset.MetalToolchain-v17.1.5276.7.3KEJwX” couldn’t be opened because you don’t have permission to view it.
Beginning asset download...
2025-07-13 13:16:17.427 xcodebuild[2153:34022] IDEDownloadableMetalToolchainCoordinator: Failed to remount the Metal Toolchain: The file “com.apple.MobileAsset.MetalToolchain-v17.1.5276.7.3KEJwX” couldn’t be opened because you don’t have permission to view it.
Downloaded asset to: /System/Library/AssetsV2/com_apple_MobileAsset_MetalToolchain/47af11e2964f385d510c6a9d1a49c8165f334a51.asset/AssetData/Restore/022-19457-052.dmg
Beginning asset export...
Done exporting: /tmp/MyMetalExport/MetalToolchain-17A5276g.exportedBundle
Done downloading: Metal Toolchain 17A5276g.
The ExportMetadata.plist has a buildUpdateVersion of 17A5276g.
Any suggestions are greatly appreciated.
This is related to the WWDC presentation, What's new in Metal rendering for immersive apps..
Specifically, the macOS spatial streaming to visionOS feature: For reference: the page in the docs.
The presentation demonstrates it using a full immersive space and Metal rendering using compositor services.
I'd like clarity on a few things:
Is the remote device wireless, or must the visionOS device be connected via a wired connected?
Is there a limit to the number of remote devices, and if not, could macOS render different things per remote device simultaneously?
Can I also use mixed mode with passthrough enabled, instead of just a fully-immersive mode?
Can I use RealityKit instead of Metal? If so, may I have an example, or would someone point to an example?
Hello there!
I’m currently stuck with Xcode 26.0 beta 6 (17A5305f), as I can’t download the latest SDK/Simulator for iOS 26.0 beta 6 (23A5324a). The download constantly fails for days now with the following error message:
(-67061 invalid signature (code or signature have been modified)
Domain: SimDiskImageErrorDomain
Code: 5
User Info: {
DVTErrorCreationDateKey = "2025-08-26 21:11:30 +0000";
unusableErrorDetail = "";
}
--
(-67061 invalid signature (code or signature have been modified)
Domain: SimDiskImageErrorDomain
Code: 5
User Info: {
unusableErrorDetail = "";
}
--
System Information
macOS Version 15.6.1 (Build 24G90)
Xcode 26.0 (24208.7) (Build 17A5305f)
Timestamp: 2025-08-26T23:11:30+02:00
I’ve already tried re-downloading the Xcode beta 6 from the Apple Developer site. I’ve also tried restarting my Mac and even tried to use the command line to install iOS 26 beta SDK via xcodebuild -downloadPlatform iOS -exportPath ~/Downloads (as noted here). This lead to the same error message.
Can someone help? I’m unable to develop my apps right now! 🫠
I’ve also filed a feedback: FB19915834
Hello, is it anyone here ( or on this planet ) that can help with this issue on Catalina Mac OS 10.15.5 Beta 3 (19F72f).I am using MacBook Pro 16 inches. 😟
I have two views I've applied Liquid Glass to in Swift UI. I've noticed that depending on the height of the view the material changes and I'm not sure why. See the attached screenshot. Both views add the liquidGlass style in the same way but behave very differently on the same background.
Ideally I'd like them to look the same as the bottom one. Is that the same as the clear style?
Hi! Short time lurker, and first time poster, but I will try to be as descriptive as possible.
I have had my HomePod mini since 2021, and have been loving it. Recently, though, I am facing issues which have paralyzed it.
Background: Running two HomePod Minis under a stereo pair on the same network in a room. Devices worked well, with the odd dropping of the sound once a while, not too big of an issue. Started running HomePod OS26 beta in July to see what the new updates have, and maybe if the audio issues would be resolved.
Issue: Two weeks ago, the after the (then) latest update, there was a no response issue in the app. No big deal, let's just restart the HomePod pair. After restart, no bueno. I also got a new beta software update notification, so decided if I updated them both, the issue will probably be resolved.
No dice. The issue persists. So I decided to pair the HomePods and factory reset them. Both HomePods disappeared from my home app. Regular behavior. Going to reset them with the old unplug for 30 s, replug into power, hold finger on the display had no issues for the first one.
For the second HomePod, unplug, hold, and plug in. Problems. No display. Tried it a bunch of times, nothing. Waited a day, in case the capacitors needed to be discharged for it to do a cold boot. Nothing. Looking across the interwebs, I found that we can connect the HomePods to Macs, have them show up on Finder, and update from there. Connecting into the Mac, nothing. Waited for the caps to discharge before engaging again, nothing.
Now, I fear I may have problems. So I scheduled a meeting at the Genius Bar. Genius Bar individual said no support as you are running beta software. We are not allowed to touch or diagnose those devices. They also said that there are no external facing endpoints for us to diagnose or check. I would have better luck with online Apple Support or over call with Apple Support.
Okay, sounds good. Contacting Apple Support, they said that we will have to wait for the official release of HomePodOS 26 for us to support you, to which I agreed. During this time, I thought about seeing if I could troubleshoot and see if anything was coming out of the HomePod when connected to the Mac. So I wrote a script, querying and logging different aspects of the USB controller on the Mac and seeing what device connects. I removed all other USB devices connected to the Mac, and started the script. First, I connected and disconnected the HomePod a couple of times, and on the last connect, I left it in, and let the script record the readings.
Reading the output told me that there was indeed a device that was connecting and disconnecting. Here are some of my findings:
Port Readout Summary
Apple's UVDM stack lights up on the Port-USB-C@2 (CC path, SOP)
AppleUVDM driver tried to read identity VDOs, initially rejected, then carried on and the transport finished powering on successfully. Furthermore, the VDM read attempts itself are rejected, typical when the accessory is gating Apple specific commands
The CC transport merged metadata for SOP with a 3-VDO identity set. The CC property change during the run indicates an active PD negotiations.
USB link states also toggled on that port with driver status Ready
Port-USB-C@2/SOP shows Vendor ID and Product ID (not sure if this is private information, so not sharing)
Port-USB-C@1/SOP, the e-marker in the HomePod cable shows the same, with product type "Passive Cable"
Only CC is flagged "TransportsActive"
Apple's accessory authentication stack is present
The USB speeds on the port flicker between Gen 3 and Gen 2, indicating to me some sort of handshake failing in the stack, which is preventing enumeration in the System Report and Finder
This likely is indicating to me that the target isn't enumerating as a normal USB device, but rather as a USB-PD Vendor Defined Messaging (UVDM) protocol on the CC line. Furthermore, the signature is in a "hidden" service/update modes, which does not enumerate standard USB functions.
So, having all this information available to me, I got on a call with Apple Support today, and they went through some of the troubleshooting steps and told me that they could neither escalate the issue (due to it being on a beta software that THEY DO NOT SUPPORT) nor remediate or troubleshoot the issue (as it is an accessory) and the only thing their support options say is to wipe the device, as it is something that is not supported. Upon my suggestion of this my USB logs being used to review the case, they said my personal USB logs would be detrimental towards my case. Their final recommendation was for me to reach out to Apple Developer support as I am running beta software on this HomePod Mini.
So, here I am! I highly believe this is a Software issue, contrary to what Apple Consumer Support call representative indicated. All and any help/guidance and support is appreciated. All and any help in boosting the visibility of this issue on the forums would also be greatly appreciated. I have had these bad boys since my undergrad, and I would hate to lose one of them.
For Apple Support, before you lock and close this post down and recommend I go to Apple Consumer Support, they said that they cannot help me, and consumer support steps start and end at resetting the device.
After my phone updated to the latest version of iOS 26 Beta overnight, I found out this morning that I cannot get past the setup process as it fails to recognise it's connected to the WiFi and is then unable to activate Apple Intelligence.
I've tried bypassing this via iPhone mirroring on my MacBook, and while I can access the home screen that way, all native apps are inaccessible and 3rd party apps are only accessible via app switcher. The 3rd party apps (Slack, Instagram, etc) work fine, meaning the device has an internet connection.
Unless someone has another suggestion, it looks like my only option is to remove the beta software by restoring my device - although I'm going to wait until iOS 26 is fully released later today (how convenient) as I don't want to use an iOS 18 backup and lose a bunch of stuff.
Durante o teste Uso do iOS 26 houve Build erro.
Foi ativado o recurso de Adiantamento de senha por seguranca. Bloqueio com tela azul e CRUZ AZUL e solicitando Recorvey.
O iPhone foi Resetado para IOS 18.6.2 pela Revendedora próxima
solicito apoio ao usuário
My iPhone is on iOS 26 RC and when I tried adding music to my library, the app crashed and when I opened it again my whole Apple Music library including playlists disappeared but it still shows that the app is showing 8 gb of storage. I’ve tried restarting my phone, closed the app and reopened it, and that didn’t work, what do I do
I have an app where the user can choose if the dark / light mode is enabled automatically during the night / day (dusk and dawn times are calculated according to to the date and location of the user).
I also have an UITabBarController, with two tabs which have a GMSMapView (Google Map View from their SDK), and three tabs with UITableViews.
I have noticed that when the automatic change of dark / light mode is performed from dark to light mode, and the user is in a tab with GMSMapView, the tabBar does not change to light model, it remains in dark mode.
I perform the switch by setting the self.window.overrideUserInterfaceStyle of the UIWindow of the UIScene. All the views change to light mode, including the navigation bar, but not the tab bar. If I move between the tabs with GMSMapView, the tabBar stays in dark mode. As soon as I move to a tab which contains a UITableView, the tabBar switches to light mode.
If the switch to light mode is performed while the user is in tab with the UITableViews, the tabBar switches to light mode immediately, as expected. When moving to the tabs with the GMSMapViews, the tabBar stays in light mode, also as expected.
I have also noticed the same problem in CarPlay, in CPMapTemplate (with a UIViewController whose view is GMSMapView), in the buttons with are set with the property “mapButtons”, and with the panning buttons. When the switch from dark mode to light mode is performed by setting the self.window.overrideUserInterfaceStyle of the UIWindow of CarPlay scene, the mapButtons keeps the glyphs in white, and set the background of the buttons to almost white, so the glyphs are barely visible unless you put png images on them instead of SF Symbols.
With the panning buttons there is not workaround, since you cannot set custom images for the panning buttons. This happens even if the panning buttons are not being displayed when the switch is performed to light mode, and you enable them later.
Is this a bug? Why does this only happen with view over GMSMapView (Google Maps view for a map from their SDK)? Can you think of any viable workaround?
Thank you.
Environment
MacOC 26
Xcode Version 26.0 beta 7 (17A5305k)
simulator: iPhone 16 pro
iOS: iOS 26
Problem
NLContextualEmbedding.load() fails with the following error
In simulator
Failed to load embedding from MIL representation: filesystem error: in create_directories: Permission denied ["/var/db/com.apple.naturallanguaged/com.apple.e5rt.e5bundlecache"]
filesystem error: in create_directories: Permission denied ["/var/db/com.apple.naturallanguaged/com.apple.e5rt.e5bundlecache"]
Failed to load embedding model 'mul_Latn' - '5C45D94E-BAB4-4927-94B6-8B5745C46289'
assetRequestFailed(Optional(Error Domain=NLNaturalLanguageErrorDomain Code=7 "Embedding model requires compilation" UserInfo={NSLocalizedDescription=Embedding model requires compilation}))
in #Playground
I'm new to this embedding model. Not sure if it's caused by my code or environment.
Code snippet
import Foundation
import NaturalLanguage
import Playgrounds
#Playground {
// Prefer initializing by script for broader coverage; returns NLContextualEmbedding?
guard let embeddingModel = NLContextualEmbedding(script: .latin) else {
print("Failed to create NLContextualEmbedding")
return
}
print(embeddingModel.hasAvailableAssets)
do {
try embeddingModel.load()
print("Model loaded")
} catch {
print("Failed to load model: \(error)")
}
}
I'm trying to understand how to best position my iOS apps in the event that—for whatever reason—I'm unable to release another version after submitting a version built with Xcode 26 that uses UIDesignRequiresCompatibility = YES.
For the sake of this discussion, assume that my apps require updates for proper UI presentation with the Liquid Glass UI, and that I'm submitting an initial version with UIDesignRequiresCompatibility = YES.
I need to make an initial update for iOS 26 right away, because running the app linked with the iOS 18.x SDK (as currently shipping on the App Store) on iOS 26 introduces UI usability problems.
Now, I have actually revised the UI for Liquid Glass and this requires slightly different behaviors for the legacy UI path (iOS 18.6 or iOS 26.0 with UIDesignRequiresCompatibility = YES) versus Liquid Glass UI path.
My Liquid Glass adoption UI may be "good enough," and it's certainly better than the results obtained when rebuilding with Xcode 26 and running without any updates at all (which produces a very poor, if not unusable, user experience), but I am not satisfied with these updates at present and I hope to make additional improvements before enabling the revised UI for users.
However, in the event that I cannot make another update for whatever reason, I need to ensure that users will never be exposed to the UI behavior that results from running with UIDesignRequiresCompatibility = NO with the code supporting Liquid Glass adoption disabled.
As long as Apple will guarantee going forward that an app built with Xcode/iOS 26.0 SDK including UIDesignRequiresCompatibility = YES in its plist will always use the old rendering mode in any future OS, then perhaps everything is fine.
However, if that's not the case, then I think I need some way to detect the enablement state of the new Liquid Glass UI at runtime (which goes beyond a simple @available check, e.g. iOS 26+ with "compatibility mode" disabled) so that I can provision to have my code supporting Liquid Glass adoption be automatically enabled in the event that my app never receives another update and a future OS ignores the UIDesignRequiresCompatibility key in its plist. I'm assuming that testing the value of the key in the plist at runtime would be insufficient, since it would still be present but ignored.
This sort of continuity planning seems like something that many developers would be concerned about.
What is Apple's guidance on this? So far I haven't found any clear discussion of this concern.
Note: My app is UIKit-based, mostly Objective-C with a bit of Swift only for TipKit.