When using a VStack containing two TextFields inside a ScrollView on an iPad running iOS 18.0, the FocusState of the topmost TextField does not trigger, while the second TextField's FocusState works correctly. Adding an invisible TextField on top resolves the issue, but it appears to be a bug specifically in iOS 18.0 on iPads. This issue does not occur on iOS versions below or above 18.0 (including iOS 18.1).
Code that is not working
struct ContentView: View {
@State var text: String = ""
@FocusState
private var focusState: Bool
var body: some View {
ScrollView(.vertical, content: {
VStack(spacing: 0) {
TextField(text: $text) {
Text("Hello, World!")
}
.border(focusState ? Color.red : Color.gray)
.focused($focusState)
.onChange(of: focusState) { oldValue, newValue in
print(newValue)
}
.onChange(of: text) { oldValue, newValue in
focusState = true
}
}
})
.padding()
}
}
Code that is working
struct ContentView: View {
@State var text: String = ""
@FocusState
private var focusState: Bool
var body: some View {
ScrollView(.vertical, content: {
VStack(spacing: 0) {
// Invisible Text Field
TextField("", text: .constant(""))
.frame(height: 0)
TextField(text: $text) {
Text("Hello, World!")
}
.border(focusState ? Color.red : Color.gray)
.focused($focusState)
.onChange(of: focusState) { oldValue, newValue in
print(newValue)
}
.onChange(of: text) { oldValue, newValue in
focusState = true
}
}
})
.padding()
}
}
This is a dedicated space for developers to connect, share ideas, collaborate, and ask questions. Introduce yourself, network with other developers, and join us in fostering a supportive community.
Selecting any option will automatically load the page
Post
Replies
Boosts
Views
Activity
Team, I am seeing following crash with no trace of evidence to reproduce the crash.
Any ideas how to reproduce this/ do potential fixes?
Application Specific Information:
*** Terminating app due to uncaught exception 'b'MACH_EXCEPTION_TYPE: Unknown type:999'', reason: 'b'Unknown Type:999/code:999''
Thread 0 Crashed:
0 libsystem_kernel.dylib 0x00001d9b951d4 0x1d9b89000 + 49620
1 libsystem_c.dylib 0x0000191587ad8 0x191510000 + 490200
2 libc++abi.dylib 0x0000211b0d5b8 0x211afa000 + 79288
3 libc++abi.dylib 0x0000211afbbac 0x211afa000 + 7084
4 libobjc.A.dylib 0x0000186ae6e14 0x186ab4000 + 208404
5 libc++abi.dylib 0x0000211b0c87c 0x211afa000 + 75900
6 libc++abi.dylib 0x0000211b100ac 0x211afa000 + 90284
7 libobjc.A.dylib 0x0000186ae2650 0x186ab4000 + 190032
8 CoreFoundation 0x00001897c8934 0x189776000 + 338228
9 GraphicsServices 0x00001d57a81c4 0x1d57a7000 + 4548
10 UIKitCore 0x000018c32eeb0 0x18bf5c000 + 4009648
11 UIKitCore 0x000018c3dd5b4 0x18bf5c000 + 4724148
12 Chase 0x00001032d98e0 main + 5789920 (main.swift:29)
13 ?-?-? 0x00001af1b6ec8 0x0 + 0
The rearrangemnet of the emoji menu is driving me crazy. I never use Memojis or the Apple animal emojis and now they're all squished together at the front of the emoji keyboard and it's difficult to select what I want, since the menu is so small.
I switch between language and emoji keyboards frequently and this is extremely frustrating.
ProfileConnection Error com.apple.ManagedConfiguration 12:33:23.432323+0900 ManagedConfiguration PerfPowerServices MDM profile installation check failed with error:NSError.
Desc: Connecting process lacks permission 'com.apple.managedconfiguration.profiled-access'
US Desc: Calling process lacks 'com.apple.managedconfiguration.profiled-access' entitlement
Domain: MCXPCErrorDomain
Code: 39000
Type: MCFatalError
Params : (
"com.apple.managedconfiguration.profiled-access"
)
Topic:
Community
SubTopic:
Apple Developers
when i send a http request with NSURLSession, app is crashed.
_quic_frame_write
_quic_frame_write
__quic_packet_builder_assemble
_quic_conn_assemble_and_encrypt
_quic_conn_send_frames_for_key_state
_quic_conn_send_internal
_quic_recovery_pto
_quic_recovery_timer_fired
_quic_timer_run
___nw_protocol_instance_schedule_wakeup_block_invoke
___nw_context_reset_timer_block_with_time_block_invoke
__dispatch_client_callout
__dispatch_continuation_pop
__dispatch_source_latch_and_call
__dispatch_source_invoke
__dispatch_workloop_invoke
__dispatch_root_queue_drain_deferred_wlh
__dispatch_workloop_worker_thread
__pthread_wqthread
Topic:
Community
SubTopic:
Apple Developers
Hi Everyone,
I’m currently working on an app that uses NFC functionality, and I’ve hit a roadblock during the App Store review process. The review team has requested a demo video that demonstrates the app’s NFC functionality in use, as per Guideline 2.1.
Here are the details I’ve received from the App Store team:
• The demo video should show the app running on a physical iOS device, not on a simulator.
• It needs to clearly document the NFC-related app features, user permission requests, and how the app interacts with NFC tags.
My Challenges:
1. I’m unsure about the best way to visually capture the NFC interaction, as it involves scanning physical tags.
2. Are there any recommended tools or techniques for recording a demo on an iOS device that complies with Apple’s guidelines?
3. Has anyone successfully submitted an NFC app with a similar requirement? If so, I’d appreciate tips or examples of how you structured your video.
I want to make sure I meet all the necessary requirements so the app passes review. Any advice or guidance would be greatly appreciated!
Thank you in advance for your help.
Best regards,
Edwin
Topic:
Community
SubTopic:
Apple Developers
Hi Apple Developers!
I’m using the DetectBarcodesRequest function to identify QR codes in some images and PDFs.
However, I’m facing an issue where the function doesn’t detect the barcode on certain documents and machines, while it works on others using the same document.
The only common factor I’ve noticed is that the machines that successfully identify the QR code of the “problematic” document are all heavy developer machines that have Xcode installed. Interestingly, this doesn’t seem to be related to processor type (Intel vs. Apple Silicon).
Could you please provide some guidance or leads on how to resolve this issue?
With CarPlay communication plugin R18.1, I followed these steps to integrate Enhance siri, the music sound was output from carplay and there is no option for output to Car.
============================================
Enhanced Siri
Declare supported audioFormats for the AuxIn and AuxOut streams
Since the AuxIn and AuxOut streams for Siri do not have to be both active at the same time, the accessory must claim audio formats support
for AuxIn Audio and AuxOut audio independently. The audio formats for each stream can differ from each other (48KHz for AuxOut and 16KHz
for AuxIn). The new audio types represent these new streams - AuxIn/speechRecognition & AuxOut/speechRecognition.
Check if connected device supports the feature
AirPlayReceiverSessionHasFeatureEnhancedSiri()
Claim support in the Setup Response message if device supports the feature
Add kAirPlayKeyAccessoryEnabledFeature_EnhancedSiri key through the AirPlayReceiverServer delegate AirPlayReceiverServerCopyProperty_f
function for the kAirPlayKey_AccessoryEnabledFeatures key.
Helper function: CFArrayAppendValue()
Add Enhanced Siri parameters dictionary in the INFO message
Add dictionary through the AirPlayReceiverServer delegate AirPlayReceiverServerCopyProperty_f
function for the kAirPlayKey_EnhancedSiriInfo key.
kAirPlayKey_EnhancedSiriInfo dictionary parameters:
Voice activation of Siri - kAirPlayKey_EnhancedSiriVoice
Current language of voice model - kAirPlayKey_VoiceModelCurrentLanguage
Supported languages of voice model - kAirPlayKey_VoiceModelSupportedLanguages
Enhanced Button activation of Siri - kAirPlayKey_EnhancedSiriButton
Supported zone(s) if any - kAirPlayKey_SupportedSiriTriggerZones
Update AudioStream
Get state of the AuxIn state by providing an implementation of AudioStreamUpdateState() - Off, local buffering, or streaming to device.
Decouple input streams from output streams. AuxIn is an independent input stream only. The property kAudioStreamProperty_Direction will
provide the necessary information if the stream is input, output or input & output.
Provide a handler for the AirPlayReceiverSessionDelegate setEnhancedSiriParams_f
This will provide additional information:
Activation type
Setting the language of the voice model
Invoke the Communication Plugin to start buffering
Once the activation type has been specified, the accessory can request the plugin to start buffering using
AirPlayReceiverSessionAuxInStart().
Use the new APIs to trigger Siri:
AirPlayReceiverSessionRequestSiriActionWithLatency()
AirPlayReceiverSessionRequestSiriVoiceActivationWithLatency()
AirPlayReceiverSessionRequestSiriVoiceActivationWithSample()
Button presses and voice activations should use this new APIs which adds a timestamp of the activation. These APIs allow
a choice of a latency or a sample for button and voice activations.
If there is a delay between the user pressing the button to notifying the device on the button press, this latency value
should represent this time.
If the accessory can determine which zone activated, it can provide the zone with the request.
Invoke the Communication Plugin to stop buffering
You may need to invoke the plugin to stop buffering (AirPlayReceiverSessionAuxInStop()) if exclusive access to the microphone is necessary.
Such instances may include but not limited to:
Native voice recognition session
Telephony
Another stream function which uses the microphone starts
modesChanged notification can be used to determine if a resource is being used
Note, if the session ends, the plugin will automatically stop buffering
My screen time report says that I’m on Google an average of 24 hrs and 4 minutes a day. How is this possible? I can’t be on Google more than 10 minutes a day let alone 24 hrs. I think it’s probably tracking whatever Google does when I’m not in the app or on the cite as part of screen time, but I can’t get it to stop. Are other people having this issue? How do I stop this so I can see my real screen time?
When I use the screenshot feature, I find that the screenshot becomes darker in color compared to the original image, this seems to be a bug in iOS 18. can anyone help me?
We are in the process of developing a mobile banking application and have a key requirement related to NFC-based peer-to-peer (P2P) transactions. Our goal is to enable secure and seamless money transfers between two devices using NFC.
Business Requirement:
The application will provide a feature where users can initiate money transfers via NFC.
The sender starts an NFC payment session from their device.
Simultaneously, the receiver starts an NFC session to accept the payment.
The sender's device should be able to detect the receiver’s NFC tag and process the transaction securely.
Hi there, with the latest update, the ability to click in the file name field at the top of a document (eg Pages, Numbers) to add or change the name of the file, then select filing location, no longer works. Previously, a user could do this and then hit 'return' and it would save to that location with that new file name. It no longer does this. This is a very useful functionality to have. Can it be returned in a future update?
Topic:
Community
SubTopic:
Apple Developers
Open an existing contact
Click Edit —> Add Birthday
Select Date and Save
Close the contact app
Come back to the same contact, the birthday is gone.
For 9 Glorious Months, My MacBook Pro & J5 Create JCD543 HUB Were Best Friends… Until I Updated macOS
For nearly a year, my MacBook Pro 14” (2023, M2 Pro, 16GB RAM) worked flawlessly with my J5 Create JCD543 HUB, allowing me to run:
✅ 4 External Displays (Samsung, LG, Insignia, Samsung)
A setup that worked perfectly… until I updated macOS.
Then Came the Update… and the Issues Began
First, I updated last week. Two of my four displays stopped working. No big deal, I thought—I’ve dealt with display quirks before.
Then, like an optimist ignoring red flags, I updated again last night to 15.4 Beta (24E5222f) and… well, now things are completely broken.
My Once-Perfect Setup (Now a Display Disaster)
JCD543 HUB → 2 HDMI displays (not working)
JUD380 HUB → 1 HDMI display (partially working, unstable)
MacBook HDMI port → 1 HDMI display (working, but feeling abandoned)
JCA365 USB-C to Dual HDMI Adapter (Not part of my setup, but tested—also not working)
What I’ve Tried (aka My Troubleshooting Journey)
✔ Restarted my Mac multiple times
✔ Uninstalled and reinstalled J5 drivers
✔ Swapped cables, unplugged/replugged everything, tested known-working HDMI cables (they work fine on other devices)
✔ Deleted /Library/Preferences/com.apple.windowserver.displays.plist (this fixed a similar issue in the past, but the file no longer exists in this update)
The Likely Culprit?
It seems to be a J5 driver issue triggered by the macOS update—and I’m not alone. Others are reporting similar problems.
Apple… Any Fixes?
I just want my four screens back. The setup worked flawlessly for 9 months, so it’s hard to believe it was never meant to work.
If anyone has a workaround, I’m open to testing any solution—whether it’s a driver update, system tweak, or hidden macOS setting that might bring my displays back to life.
Any advice would be greatly appreciated! Thanks in advance.
(P.S. Apple, if you’re reading this—please look into this issue!)
Scenario: hardware keyboard is attached to iPhone
Before iOS 18.2, if I setup only two keyboards (e.g., english and emoji) in the system, click Globe key can switch to iOS emoji keyboard without problem. Since iOS 18.2, it stop working, and emoji keyboard is never displayed. I need to add one more keyboard, then I can switch to emoji keyboard (via the keyboard selector).
Topic:
Community
SubTopic:
Apple Developers
ISAC Client is crashing on macOS 15.4 Beta 1 which is from from the WebKit engine the underlying framework of WKWebView. And the "ResourceLoadNotifier" is from WebKit's internal framework.It seems to be related to resource loading failure which is potentially triggered by changes in macOS 15.4 Beta.
Dear Apple:
The program we developed uses BLE broadcasting to discover devices. The discovered device sends three broadcasts, but when we debug the program, we found that the addresses reported by centralManager::didDiscoverPeripheral for the same device's three broadcasts are actually different CBPeripheral * addresses. I would like to ask how we can identify that these three broadcasts are from the same device? Are there any variables in the peripheral that can identify it as the same device? Thanks。
Hello, I am trying to enroll to apple developer program as solo developer. While trying to enroll via app my passport and drivers license are not validated. When trying to enroll via web, I fill in the details and try to make the payment, but payment itself is not deducted from my bank account. I receive email that payment is being processed, but nothing further happens.
Topic:
Community
SubTopic:
Apple Developers
Crashed Thread: 0 Dispatch queue: com.apple.main-thread
Exception Type: EXC_CRASH (SIGABRT)
Exception Codes: 0x0000000000000000, 0x0000000000000000
Termination Reason: Namespace SIGNAL, Code 6 Abort trap: 6
Terminating Process: Microsoft Excel [1618]
Application Specific Information:
stack buffer overflow
Thread 0 Crashed:: Dispatch queue: com.apple.main-thread
0 libsystem_kernel.dylib 0x19da2b720 __pthread_kill + 8
1 libsystem_pthread.dylib 0x19da63f70 pthread_kill + 288
2 libsystem_c.dylib 0x19d97099c __abort + 136
3 libsystem_c.dylib 0x19d961e74 __stack_chk_fail + 96
4 Microsoft Excel 0x106a89f88 0x104bdc000 + 32169864
5 Microsoft Excel 0x106a89b74 0x104bdc000 + 32168820
6 Microsoft Excel 0x106acc2fc 0x104bdc000 + 32441084
7 Microsoft Excel 0x106a98cc4 0x104bdc000 + 32230596
8 Microsoft Excel 0x106a97c04 0x104bdc000 + 32226308
9 Microsoft Excel 0x106a897f4 0x104bdc000 + 32167924
10 Microsoft Excel 0x106a8731c 0x104bdc000 + 32158492
11 Microsoft Excel 0x105d69ac8 0x104bdc000 + 18406088
12 Microsoft Excel 0x1061bea10 0x104bdc000 + 22948368
13 Microsoft Excel 0x104dbc1b4 0x104bdc000 + 1966516
14 Microsoft Excel 0x104d828a8 0x104bdc000 + 1730728
15 Microsoft Excel 0x104d80ff8 0x104bdc000 + 1724408
16 Microsoft Excel 0x104d809b4 0x104bdc000 + 1722804
17 Microsoft Excel 0x104d808e0 0x104bdc000 + 1722592
18 Foundation 0x19ed394e8 __NSFireDelayedPerform + 372
19 CoreFoundation 0x19db66384 CFRUNLOOP_IS_CALLING_OUT_TO_A_TIMER_CALLBACK_FUNCTION + 32
20 CoreFoundation 0x19db66028 __CFRunLoopDoTimer + 1012
21 CoreFoundation 0x19db65b38 __CFRunLoopDoTimers + 356
22 CoreFoundation 0x19db4b520 __CFRunLoopRun + 1856
23 CoreFoundation 0x19db4a724 CFRunLoopRunSpecific + 588
24 HIToolbox 0x1a90a2530 RunCurrentEventLoopInMode + 292
25 HIToolbox 0x1a90a8348 ReceiveNextEventCommon + 676
26 HIToolbox 0x1a90a8508 _BlockUntilNextEventMatchingListInModeWithFilter + 76
27 AppKit 0x1a16b5034 _DPSNextEvent + 660
28 AppKit 0x1a20192d4 -[NSApplication(NSEventRouting) _nextEventMatchingEventMask:untilDate:inMode:dequeue:] + 688
29 AppKit 0x1a16a8060 -[NSApplication run] + 480
30 AppKit 0x1a167e854 NSApplicationMain + 888
31 dyld 0x19d6e4274 start + 2840
Thread 1:
0 libsystem_kernel.dylib 0x19da22f54 mach_msg2_trap + 8
1 libsystem_kernel.dylib 0x19da35604 mach_msg2_internal + 80
2 libsystem_kernel.dylib 0x19da2baf8 mach_msg_overwrite + 480
3 libsystem_kernel.dylib 0x19da2329c mach_msg + 24
4 merp 0x108eb7f74 0x108eb4000 + 16244
5 libsys
Topic:
Community
SubTopic:
Apple Developers
I bought a Mac mini M4 and when I used it to sign up for a developer account, I was always encountered, it needed me to take a photo, but I didn't have a camera, so I used the iriun software to connect my phone to take a photo, it works fine in FaceTime, but when I take a photo in the Apple Developer app, the round photo frame is shown black.
How can I make the camera work and if I buy a third-party camera?