Search results for

“A Summary of the WWDC25 Group Lab”

10,640 results found

Post

Replies

Boosts

Views

Activity

DriverKit DEXT Logs Only Once, Subsequent Logs Are Lost Despite Normal Functionality
Hello everyone, I've encountered a very strange and persistent logging issue with my DriverKit DEXT and would appreciate any insights from the community. [Problem Summary] My DriverKit DEXT, along with its companion Swift app, is functionally working perfectly. I can repeatedly call methods in the DEXT from the app (e.g., a Ping-Pong test and a StaticProcessInbandTask call) and receive the correct response every time. However, the os_log messages within my IOUserClient subclass are only successfully recorded for the very first set of interactions. After that, all subsequent logs are completely missing. What's even stranger is that all successfully recorded logs are attributed to the kernel: process, even for purely user-space methods like ExternalMethod. [Development Environment] macOS: 15.7.1 Xcode: 16.4 Hardware: MacBook Pro M1 DEXT Logging Macro (Log.h): #include #define Log(fmt, ...) do { os_log(OS_LOG_DEFAULT, [%{public}s] fmt, __FUNCTION__, ##__VA_ARGS__); } while (0) [Steps to Reproduce &
3
0
199
Oct ’25
Reply to DriverKit DEXT Logs Only Once, Subsequent Logs Are Lost Despite Normal Functionality
Hi Kevin, Thank you for your reply. It gave us a clearer understanding of how the DEXT Log mechanism works. You were right: the logs were indeed always there; our filtering methods were just missing them. Here is a summary of how we solved the silent log issue: We tried using sysdiagnose. However, for our specific need to capture IOUserClient interaction logs in real-time, the amount of data in the report was too large to be effectively analyzed. The core of the problem was that our DEXT generates a continuous and massive amount of logs, which drowned out the logs from the UserClient methods. Therefore, we created a new, dedicated logging macro, uLog(), for the exclusive use of our UserClient. Next, we modified Log.h to add a compile-time switch (#ifdef), which allows us to completely compile-out the Log() macro from the core engine at compile time, leaving only the new uLog(). All the noise from the DEXT disappeared, and the logs from ExternalMethod appeared clearly and reliably. We have now switche
Topic: App & System Services SubTopic: Drivers Tags:
Oct ’25
[iOS 26] iOS App Does Not Receive Deep Link from Widget When Using widgetAccentedRenderingMode on Image
Summary When a SwiftUI widget uses a Link containing an Image modified with .widgetAccentedRenderingMode (using any mode except .fullColor), tapping the image on the widget launches the app but does not pass the expected deep link URL to the SceneDelegate. Steps to Reproduce Create a SwiftUI Widget View with a Link: struct ImageView: View { var image: UIImage var body: some View { Link(URL(string: myapp://image)!) { Image(uiImage: image) .resizable() .widgetAccentedRenderingMode(.accentedDesaturated) // or any mode other than .fullColor .scaledToFill() .clipped() } } } Add Custom URL Scheme in Info.plist: CFBundleURLName com.company.myapp CFBundleURLSchemes myapp Implement Deep Link Handling in SceneDelegate: func scene(_ scene: UIScene, willConnectTo session: UISceneSession, options connectionOptions: UIScene.ConnectionOptions) { if let url = connectionOptions.urlContexts.first?.url { handle(url) } } func scene(_ scene: UIScene, openURLContexts urlContexts: Set) { if let url = urlContexts.first?
2
0
341
Oct ’25
Reply to visionOS – Starting GroupActivity FaceTime Call dismisses Immersive Space
Hi @arthurfromberlin and @scavengar It is unexpected for a visionOS app to behave differently when deployed to TestFlight. The feedback @arthurfromberlin filed will help us investigate that issue. In the meantime, I spoke with the SharePlay team and want to share their recommendations: Starting SharePlay from immersive spaces: There's no reliable way for a person to initiate a SharePlay session while remaining in an immersive space, since the system may background the immersive space to present the FaceTime UI during call initiation. Recommended approach: We recommend that SharePlay experiences start in a shared space, then allow participants to transition to the immersive experience once everyone has joined the activity and is ready to participate. To keep the group synchronized as participants enter and leave the immersive space, use SystemCoordinator.groupImmersionStyle. For additional guidance, see Creating fully immersive experiences in your app.
Topic: Spatial Computing SubTopic: General Tags:
Oct ’25
UIMenuElementAttributesKeepsMenuPresented not working when Glass Grouping changes
I'm working on a UIBarButtonItem that is supposed to be a filter button - it shows a menu in which the user can (un)check menu items. Using the UIMenuElementAttributesKeepsMenuPresented attribute on the UIActions prevents the menu to hide after the user clicks an item. The button is put alongside another button as the leftBarButtonItems of my navigation item. In iOS 26 they are grouped into a single glass container automatically. Now when the user starts filtering, I want to highlight the UIBarButton item to signal to the user that the filter is active (similar to what Apple does in the Mail app). I do that by setting the UIBarButtonItemStyle to prominent. Now that works too, but it causes the automatic glass grouping to break up and the menu to hide. I'm fine with the grouping to break up, but the menu shouldn't hide. Is this a bug or can this be prevented somehow?
Topic: UI Frameworks SubTopic: UIKit Tags:
2
0
153
Oct ’25
Data used for MLX fine-tuning
The WWDC25: Explore large language models on Apple silicon with MLX video talks about using your own data to fine-tune a large language model. But the video doesn't explain what kind of data can be used. The video just shows the command to use and how to point to the data folder. Can I use PDFs, Word documents, Markdown files to train the model? Are there any code examples on GitHub that demonstrate how to do this?
2
0
358
Oct ’25
[Bug] iPadOS 26: 4-Finger Fast Tap/Swipe Gesture Not Detected (Multitouch Issue)
Hi everyone I'm experiencing an issue with iPadOS 26 regarding multi-touch gesture detection. When performing a quick four-finger gesture (tap and swipe), the system often fails to recognize the input. This especially affects multi-touch gestures, such as rhythm games with difficult levels. Steps to Reproduce: Place four fingers on the screen. Perform a quick tap or a quick horizontal swipe (like the one used to switch apps). Observe whether the gesture is ignored or detected inconsistently. Expected Behavior: 4-finger multitouch gestures should be recognized regardless of gesture speed, just like previous iPadOS versions. Actual Behavior: Gestures fail to be detected when executed quickly—same gestures still work, and miss notes in rhythm games. You can check out my posts on Twitter/x and Facebook: [https://x.com/kokona_fwa/status/1978131164104728949?s=61] Facebook: [https://m.facebook.com/groups/idipad/permalink/24438964899058806/?]
2
0
1.2k
Oct ’25
Profile doesn't include the com.apple.application-identifier entitlement.
I have tried everything and still I am getting this. Just for a test I created a new app (Master-Detail template Xcode 11.5) I have created an entry in the iTunes Connect to receive the app upon archiving and uploading. I regenerated all new certificates for iOS Development and Distribution. I created all new Provisioning profiles. The Dev profile builds deploys and runs on my device The Dist profile builds but when I select the distribution profile I get the Profile doesn't include the com.apple.application-identifier entitlement. error. When I download the profile within Xcode all looks good for the distribution profile: App ID: matches correctly Certificated: 1 Included includes the new signing certificate iPhone Distribution.... Capabilities: 3 Included Includes Game Center, In-App Purchase, and Keychain Sharing Entitlements: 5 Included Includes application-identifier, keychain-access-groups, beta-reports-active, get-task-allow, and com.apple.developer.team-identifier. Im not sure what is going o
22
0
15k
Oct ’25
[iOS26] Widget Exntension configuration uable to load
I have developed a Widget Extension with editable dynamic options. `struct ModelQuery: EntityStringQuery { public var allModels:[ModelEntity] { // from App Groups UserDefaults let models = SharedDataManager.getModelList() // 检查原始数据是否为空,避免转换后的数据异常 guard !models.isEmpty else { return [] } let entites = models.map{ModelEntity(from: $0)} return entites } func entities(for identifiers: [ModelEntity.ID]) async throws -> [ModelEntity] { let models = allModels if models.isEmpty { return [] } // 尝试匹配ID let matchedEntities = identifiers.compactMap { id in models.first { $0.id == id } } // 如果没有匹配到任何实体,返回默认的第一个站点 if matchedEntities.isEmpty && !models.isEmpty { return [models[0]] } return matchedEntities } func entities(matching string: String) async throws -> [ModelEntity] { let stations = allModels if stations.isEmpty { return [] } if string.isEmpty { return stations } let lowercasedString = string.lowercased() let filteredStations = stations.filter { station in station.name.lowercased().contains(
2
0
161
Oct ’25
Request: Restore Launchpad Functionality or Allow Customizable App Organization in macOS Tahoe
With macOS Tahoe, Launchpad has been replaced by an App Library–style mode within Spotlight. While the alleged intention is UX consistency across the Apple ecosystem, the result is both a catastrophic usability regression and a radical break in consistency with iOS and iPadOS. Predefined App Library categorization is functionally incoherent: On iOS and now macOS, Apple’s predefined App Library categories place apps with seemingly identical functionality into unrelated groups—for example, 3D scanning tools scattered across Education, Utilities, and Productivity. Instead of making apps easier to find, this effectively creates a labyrinth that users must traverse to locate apps whose names and icons they may not recall. However Apple defines its app categories, they are not only inconsistent but also hopelessly inadequate for the long tail of real-world applications and user workflows. Loss of user control: Launchpad enabled users to group and organize applications according to their workflows.
2
0
771
Oct ’25
'openAppWhenRun' property causing AppIntentsExtension to fail
I have added an App Intents Extension target to my main application in macOS. This generated the below two files: TWAppIntent.swift import AppIntents struct TWAppIntent: AppIntent { static var title: LocalizedStringResource = TWAppIntentExtension static var parameterSummary: some ParameterSummary { Summary(Get information on (.$TWType)) } //launch app on running action static var openAppWhenRun: Bool = true // we can have multiple parameter of diff types @Parameter(title: TWType) var TWType: String func perform() async throws -> some IntentResult & ReturnsValue & ProvidesDialog { return .result(value: TWType, dialog: Logged break.) } } TWAppIntentExtension.swift import AppIntents @main struct TWAppIntentExtension: AppIntentsExtension { } I m able to build the extension target and I my intent action is available in the shortcuts app. However, on launching a shortcut with the above created intent action. I m getting the below popups: I have identified what is causing this error. Setting the
3
0
1k
Oct ’25
Getting the world position of a QR code
Hi, would love for your help in that matter. I try to get the position in space of two QR codes to make an alignment to their positions in space. The detection shows that the QR codes position is always 0,0,0 and I don't understand why. Here's my code: import SwiftUI import RealityKit import RealityKitContent struct AnchorView: View { @ObservedObject var qrCoordinator: QRCoordinator @ObservedObject var coordinator: ImmersiveCoordinator let qrName: String @Binding var startQRDetection: Bool @State private var anchor: AnchorEntity? = nil @State private var detectionTask: Task? = nil var body: some View { RealityView { content in // Add the QR anchor once (must exist before detection starts) if anchor == nil { let imageAnchor = AnchorEntity(.image(group: QRs, name: qrName)) content.add(imageAnchor) anchor = imageAnchor print(📌 Created anchor for (qrName)) } } .onChange(of: startQRDetection) { enabled in if enabled { startDetection() } else { stopDetection() } } .onDisappear { stopDetection() } } privat
2
0
502
Oct ’25
UICollectionView list: leading swipe overshoots in expanded split view for .plain/.grouped; .insetGrouped OK (iPadOS 26 b5–b8) [FB19785883]
Hi all, Sharing a reproducible UIKit issue I’m seeing across multiple iPadOS 26 betas, with a tiny sample attached and a short video. Short video https://youtu.be/QekYNnHsfYk Tiny project https://github.com/yoasha/ListSwipeOvershootReproSwift Summary In a UISplitViewController (.doubleColumn), a UICollectionView using list layout shows a large leading-swipe overshoot when the split view is expanded (isCollapsed == false). The cell content translates roughly 3–4× the action width. Repros with appearance = .plain and .grouped Does not repro with .insetGrouped Independent of trailing provider (issue persists when trailing provider is nil) Collapsed split (compact width) behaves correctly Environment Devices: iPad Air (3rd gen), iPadOS 26.0 (23A5326a) → Repro Simulators: iPad Pro 11-inch (M4), iPadOS 26.0 beta 6 → Repro Also tested on device: iPadOS 26 beta 5, 6, 7, 8 Xcode: 26.0 beta 6 (17A5305f) Steps to reproduce Launch the sample; ensure the split is expanded (isCollapsed == false). In the s
1
0
195
Oct ’25
How to display SegmentControl in tabViewBottomAccessory inline mode like iOS 26 Photos app?
Question I'm trying to replicate the iOS 26 Photos app behavior where a segmented control appears in the tab bar area when scrolling. Specifically, when the tab bar minimizes to inline mode, I want to show a Picker with .segmented style between the minimized tab button and search button. Expected Behavior (iOS 26 Photos App) When scrolling up in the Photos app: Tab bar minimizes to inline mode A segmented control (Years/Months/All) appears in the center Layout: [Tab Button] [Segmented Control] [Search Button] Current Implementation I'm using tabViewBottomAccessory with tabBarMinimizeBehavior: struct PhotosMainView: View { @Environment(ModelData.self) private var modelData @State private var searchText: String = var body: some View { @Bindable var modelData = modelData TabView { Tab(Library, systemImage: photo.on.rectangle) { NavigationStack { PhotosGridView() .navigationTitle(Library) } } Tab(Albums, systemImage: square.grid.2x2) { NavigationStack { AlbumsGridView() .navigationTitle(Albums) } } Tab(Search, s
1
0
220
Oct ’25
Reply to How to display SegmentControl in tabViewBottomAccessory inline mode like iOS 26 Photos app?
Hi @377632523@qq.com, I'm unable to reproduce the issue, as reported, in my own Xcode project. Because the provided example above will not build successfully due to missing types, I replaced your implementation with the following: import SwiftUI struct PhotosMainView: View { @State private var searchText: String = var body: some View { TabView { Tab(Library, systemImage: photo.on.rectangle) { NavigationStack { scrollView(colors: [.red, .orange, .yellow]) .navigationTitle(Library) } } Tab(Albums, systemImage: square.grid.2x2) { NavigationStack { scrollView(colors: [.yellow, .green, .teal]) .navigationTitle(Albums) } } Tab(Search, systemImage: magnifyingglass, role: .search) { NavigationStack { scrollView(colors: [.blue, .purple, .pink]) .navigationTitle(Search) .searchable(text: $searchText) } } } .tabBarMinimizeBehavior(.onScrollUp) .tabViewStyle(.sidebarAdaptable) .tabViewBottomAccessory { TimelineAccessoryView() } } // Trivial helper view to visualize scrolling. @ViewBuilder func scrollView(colors: [Color]
Topic: UI Frameworks SubTopic: SwiftUI Tags:
Oct ’25
DriverKit DEXT Logs Only Once, Subsequent Logs Are Lost Despite Normal Functionality
Hello everyone, I've encountered a very strange and persistent logging issue with my DriverKit DEXT and would appreciate any insights from the community. [Problem Summary] My DriverKit DEXT, along with its companion Swift app, is functionally working perfectly. I can repeatedly call methods in the DEXT from the app (e.g., a Ping-Pong test and a StaticProcessInbandTask call) and receive the correct response every time. However, the os_log messages within my IOUserClient subclass are only successfully recorded for the very first set of interactions. After that, all subsequent logs are completely missing. What's even stranger is that all successfully recorded logs are attributed to the kernel: process, even for purely user-space methods like ExternalMethod. [Development Environment] macOS: 15.7.1 Xcode: 16.4 Hardware: MacBook Pro M1 DEXT Logging Macro (Log.h): #include #define Log(fmt, ...) do { os_log(OS_LOG_DEFAULT, [%{public}s] fmt, __FUNCTION__, ##__VA_ARGS__); } while (0) [Steps to Reproduce &
Replies
3
Boosts
0
Views
199
Activity
Oct ’25
Reply to DriverKit DEXT Logs Only Once, Subsequent Logs Are Lost Despite Normal Functionality
Hi Kevin, Thank you for your reply. It gave us a clearer understanding of how the DEXT Log mechanism works. You were right: the logs were indeed always there; our filtering methods were just missing them. Here is a summary of how we solved the silent log issue: We tried using sysdiagnose. However, for our specific need to capture IOUserClient interaction logs in real-time, the amount of data in the report was too large to be effectively analyzed. The core of the problem was that our DEXT generates a continuous and massive amount of logs, which drowned out the logs from the UserClient methods. Therefore, we created a new, dedicated logging macro, uLog(), for the exclusive use of our UserClient. Next, we modified Log.h to add a compile-time switch (#ifdef), which allows us to completely compile-out the Log() macro from the core engine at compile time, leaving only the new uLog(). All the noise from the DEXT disappeared, and the logs from ExternalMethod appeared clearly and reliably. We have now switche
Topic: App & System Services SubTopic: Drivers Tags:
Replies
Boosts
Views
Activity
Oct ’25
[iOS 26] iOS App Does Not Receive Deep Link from Widget When Using widgetAccentedRenderingMode on Image
Summary When a SwiftUI widget uses a Link containing an Image modified with .widgetAccentedRenderingMode (using any mode except .fullColor), tapping the image on the widget launches the app but does not pass the expected deep link URL to the SceneDelegate. Steps to Reproduce Create a SwiftUI Widget View with a Link: struct ImageView: View { var image: UIImage var body: some View { Link(URL(string: myapp://image)!) { Image(uiImage: image) .resizable() .widgetAccentedRenderingMode(.accentedDesaturated) // or any mode other than .fullColor .scaledToFill() .clipped() } } } Add Custom URL Scheme in Info.plist: CFBundleURLName com.company.myapp CFBundleURLSchemes myapp Implement Deep Link Handling in SceneDelegate: func scene(_ scene: UIScene, willConnectTo session: UISceneSession, options connectionOptions: UIScene.ConnectionOptions) { if let url = connectionOptions.urlContexts.first?.url { handle(url) } } func scene(_ scene: UIScene, openURLContexts urlContexts: Set) { if let url = urlContexts.first?
Replies
2
Boosts
0
Views
341
Activity
Oct ’25
Reply to visionOS – Starting GroupActivity FaceTime Call dismisses Immersive Space
Hi @arthurfromberlin and @scavengar It is unexpected for a visionOS app to behave differently when deployed to TestFlight. The feedback @arthurfromberlin filed will help us investigate that issue. In the meantime, I spoke with the SharePlay team and want to share their recommendations: Starting SharePlay from immersive spaces: There's no reliable way for a person to initiate a SharePlay session while remaining in an immersive space, since the system may background the immersive space to present the FaceTime UI during call initiation. Recommended approach: We recommend that SharePlay experiences start in a shared space, then allow participants to transition to the immersive experience once everyone has joined the activity and is ready to participate. To keep the group synchronized as participants enter and leave the immersive space, use SystemCoordinator.groupImmersionStyle. For additional guidance, see Creating fully immersive experiences in your app.
Topic: Spatial Computing SubTopic: General Tags:
Replies
Boosts
Views
Activity
Oct ’25
UIMenuElementAttributesKeepsMenuPresented not working when Glass Grouping changes
I'm working on a UIBarButtonItem that is supposed to be a filter button - it shows a menu in which the user can (un)check menu items. Using the UIMenuElementAttributesKeepsMenuPresented attribute on the UIActions prevents the menu to hide after the user clicks an item. The button is put alongside another button as the leftBarButtonItems of my navigation item. In iOS 26 they are grouped into a single glass container automatically. Now when the user starts filtering, I want to highlight the UIBarButton item to signal to the user that the filter is active (similar to what Apple does in the Mail app). I do that by setting the UIBarButtonItemStyle to prominent. Now that works too, but it causes the automatic glass grouping to break up and the menu to hide. I'm fine with the grouping to break up, but the menu shouldn't hide. Is this a bug or can this be prevented somehow?
Topic: UI Frameworks SubTopic: UIKit Tags:
Replies
2
Boosts
0
Views
153
Activity
Oct ’25
Data used for MLX fine-tuning
The WWDC25: Explore large language models on Apple silicon with MLX video talks about using your own data to fine-tune a large language model. But the video doesn't explain what kind of data can be used. The video just shows the command to use and how to point to the data folder. Can I use PDFs, Word documents, Markdown files to train the model? Are there any code examples on GitHub that demonstrate how to do this?
Replies
2
Boosts
0
Views
358
Activity
Oct ’25
[Bug] iPadOS 26: 4-Finger Fast Tap/Swipe Gesture Not Detected (Multitouch Issue)
Hi everyone I'm experiencing an issue with iPadOS 26 regarding multi-touch gesture detection. When performing a quick four-finger gesture (tap and swipe), the system often fails to recognize the input. This especially affects multi-touch gestures, such as rhythm games with difficult levels. Steps to Reproduce: Place four fingers on the screen. Perform a quick tap or a quick horizontal swipe (like the one used to switch apps). Observe whether the gesture is ignored or detected inconsistently. Expected Behavior: 4-finger multitouch gestures should be recognized regardless of gesture speed, just like previous iPadOS versions. Actual Behavior: Gestures fail to be detected when executed quickly—same gestures still work, and miss notes in rhythm games. You can check out my posts on Twitter/x and Facebook: [https://x.com/kokona_fwa/status/1978131164104728949?s=61] Facebook: [https://m.facebook.com/groups/idipad/permalink/24438964899058806/?]
Replies
2
Boosts
0
Views
1.2k
Activity
Oct ’25
Profile doesn't include the com.apple.application-identifier entitlement.
I have tried everything and still I am getting this. Just for a test I created a new app (Master-Detail template Xcode 11.5) I have created an entry in the iTunes Connect to receive the app upon archiving and uploading. I regenerated all new certificates for iOS Development and Distribution. I created all new Provisioning profiles. The Dev profile builds deploys and runs on my device The Dist profile builds but when I select the distribution profile I get the Profile doesn't include the com.apple.application-identifier entitlement. error. When I download the profile within Xcode all looks good for the distribution profile: App ID: matches correctly Certificated: 1 Included includes the new signing certificate iPhone Distribution.... Capabilities: 3 Included Includes Game Center, In-App Purchase, and Keychain Sharing Entitlements: 5 Included Includes application-identifier, keychain-access-groups, beta-reports-active, get-task-allow, and com.apple.developer.team-identifier. Im not sure what is going o
Replies
22
Boosts
0
Views
15k
Activity
Oct ’25
[iOS26] Widget Exntension configuration uable to load
I have developed a Widget Extension with editable dynamic options. `struct ModelQuery: EntityStringQuery { public var allModels:[ModelEntity] { // from App Groups UserDefaults let models = SharedDataManager.getModelList() // 检查原始数据是否为空,避免转换后的数据异常 guard !models.isEmpty else { return [] } let entites = models.map{ModelEntity(from: $0)} return entites } func entities(for identifiers: [ModelEntity.ID]) async throws -> [ModelEntity] { let models = allModels if models.isEmpty { return [] } // 尝试匹配ID let matchedEntities = identifiers.compactMap { id in models.first { $0.id == id } } // 如果没有匹配到任何实体,返回默认的第一个站点 if matchedEntities.isEmpty && !models.isEmpty { return [models[0]] } return matchedEntities } func entities(matching string: String) async throws -> [ModelEntity] { let stations = allModels if stations.isEmpty { return [] } if string.isEmpty { return stations } let lowercasedString = string.lowercased() let filteredStations = stations.filter { station in station.name.lowercased().contains(
Replies
2
Boosts
0
Views
161
Activity
Oct ’25
Request: Restore Launchpad Functionality or Allow Customizable App Organization in macOS Tahoe
With macOS Tahoe, Launchpad has been replaced by an App Library–style mode within Spotlight. While the alleged intention is UX consistency across the Apple ecosystem, the result is both a catastrophic usability regression and a radical break in consistency with iOS and iPadOS. Predefined App Library categorization is functionally incoherent: On iOS and now macOS, Apple’s predefined App Library categories place apps with seemingly identical functionality into unrelated groups—for example, 3D scanning tools scattered across Education, Utilities, and Productivity. Instead of making apps easier to find, this effectively creates a labyrinth that users must traverse to locate apps whose names and icons they may not recall. However Apple defines its app categories, they are not only inconsistent but also hopelessly inadequate for the long tail of real-world applications and user workflows. Loss of user control: Launchpad enabled users to group and organize applications according to their workflows.
Replies
2
Boosts
0
Views
771
Activity
Oct ’25
'openAppWhenRun' property causing AppIntentsExtension to fail
I have added an App Intents Extension target to my main application in macOS. This generated the below two files: TWAppIntent.swift import AppIntents struct TWAppIntent: AppIntent { static var title: LocalizedStringResource = TWAppIntentExtension static var parameterSummary: some ParameterSummary { Summary(Get information on (.$TWType)) } //launch app on running action static var openAppWhenRun: Bool = true // we can have multiple parameter of diff types @Parameter(title: TWType) var TWType: String func perform() async throws -> some IntentResult & ReturnsValue & ProvidesDialog { return .result(value: TWType, dialog: Logged break.) } } TWAppIntentExtension.swift import AppIntents @main struct TWAppIntentExtension: AppIntentsExtension { } I m able to build the extension target and I my intent action is available in the shortcuts app. However, on launching a shortcut with the above created intent action. I m getting the below popups: I have identified what is causing this error. Setting the
Replies
3
Boosts
0
Views
1k
Activity
Oct ’25
Getting the world position of a QR code
Hi, would love for your help in that matter. I try to get the position in space of two QR codes to make an alignment to their positions in space. The detection shows that the QR codes position is always 0,0,0 and I don't understand why. Here's my code: import SwiftUI import RealityKit import RealityKitContent struct AnchorView: View { @ObservedObject var qrCoordinator: QRCoordinator @ObservedObject var coordinator: ImmersiveCoordinator let qrName: String @Binding var startQRDetection: Bool @State private var anchor: AnchorEntity? = nil @State private var detectionTask: Task? = nil var body: some View { RealityView { content in // Add the QR anchor once (must exist before detection starts) if anchor == nil { let imageAnchor = AnchorEntity(.image(group: QRs, name: qrName)) content.add(imageAnchor) anchor = imageAnchor print(📌 Created anchor for (qrName)) } } .onChange(of: startQRDetection) { enabled in if enabled { startDetection() } else { stopDetection() } } .onDisappear { stopDetection() } } privat
Replies
2
Boosts
0
Views
502
Activity
Oct ’25
UICollectionView list: leading swipe overshoots in expanded split view for .plain/.grouped; .insetGrouped OK (iPadOS 26 b5–b8) [FB19785883]
Hi all, Sharing a reproducible UIKit issue I’m seeing across multiple iPadOS 26 betas, with a tiny sample attached and a short video. Short video https://youtu.be/QekYNnHsfYk Tiny project https://github.com/yoasha/ListSwipeOvershootReproSwift Summary In a UISplitViewController (.doubleColumn), a UICollectionView using list layout shows a large leading-swipe overshoot when the split view is expanded (isCollapsed == false). The cell content translates roughly 3–4× the action width. Repros with appearance = .plain and .grouped Does not repro with .insetGrouped Independent of trailing provider (issue persists when trailing provider is nil) Collapsed split (compact width) behaves correctly Environment Devices: iPad Air (3rd gen), iPadOS 26.0 (23A5326a) → Repro Simulators: iPad Pro 11-inch (M4), iPadOS 26.0 beta 6 → Repro Also tested on device: iPadOS 26 beta 5, 6, 7, 8 Xcode: 26.0 beta 6 (17A5305f) Steps to reproduce Launch the sample; ensure the split is expanded (isCollapsed == false). In the s
Replies
1
Boosts
0
Views
195
Activity
Oct ’25
How to display SegmentControl in tabViewBottomAccessory inline mode like iOS 26 Photos app?
Question I'm trying to replicate the iOS 26 Photos app behavior where a segmented control appears in the tab bar area when scrolling. Specifically, when the tab bar minimizes to inline mode, I want to show a Picker with .segmented style between the minimized tab button and search button. Expected Behavior (iOS 26 Photos App) When scrolling up in the Photos app: Tab bar minimizes to inline mode A segmented control (Years/Months/All) appears in the center Layout: [Tab Button] [Segmented Control] [Search Button] Current Implementation I'm using tabViewBottomAccessory with tabBarMinimizeBehavior: struct PhotosMainView: View { @Environment(ModelData.self) private var modelData @State private var searchText: String = var body: some View { @Bindable var modelData = modelData TabView { Tab(Library, systemImage: photo.on.rectangle) { NavigationStack { PhotosGridView() .navigationTitle(Library) } } Tab(Albums, systemImage: square.grid.2x2) { NavigationStack { AlbumsGridView() .navigationTitle(Albums) } } Tab(Search, s
Replies
1
Boosts
0
Views
220
Activity
Oct ’25
Reply to How to display SegmentControl in tabViewBottomAccessory inline mode like iOS 26 Photos app?
Hi @377632523@qq.com, I'm unable to reproduce the issue, as reported, in my own Xcode project. Because the provided example above will not build successfully due to missing types, I replaced your implementation with the following: import SwiftUI struct PhotosMainView: View { @State private var searchText: String = var body: some View { TabView { Tab(Library, systemImage: photo.on.rectangle) { NavigationStack { scrollView(colors: [.red, .orange, .yellow]) .navigationTitle(Library) } } Tab(Albums, systemImage: square.grid.2x2) { NavigationStack { scrollView(colors: [.yellow, .green, .teal]) .navigationTitle(Albums) } } Tab(Search, systemImage: magnifyingglass, role: .search) { NavigationStack { scrollView(colors: [.blue, .purple, .pink]) .navigationTitle(Search) .searchable(text: $searchText) } } } .tabBarMinimizeBehavior(.onScrollUp) .tabViewStyle(.sidebarAdaptable) .tabViewBottomAccessory { TimelineAccessoryView() } } // Trivial helper view to visualize scrolling. @ViewBuilder func scrollView(colors: [Color]
Topic: UI Frameworks SubTopic: SwiftUI Tags:
Replies
Boosts
Views
Activity
Oct ’25