A sample app that demonstrates how to use the Vision framework and Core ML in iOS 12 to detect objects in live video.
Delivering an Exceptional Accessibility Experience
An app that demonstrates how to use grouped notifications, introduced in iOS 12.
Sample app that demonstrates the performance benefit of Core ML's batch prediction API in iOS 12.
Transmit ARKit world-mapping data between nearby devices with the MultipeerConnectivity framework to create a shared basis for AR experiences.
An app that demonstrates various iOS 12 notification features: provisional notifications, dynamic notification action buttons, interactive notification content, and notification management features.
An app that demonstrates 3D scanning of reference objects using ARKit.
Implement two-way communication between your Apple Watch app and the paired iPhone app with the Watch Connectivity framework.
An app that demonstrates how to work with Siri Shortcuts in iOS 12.
Apply Vision algorithms to track objects or rectangles throughout a video.
Manage Vision resources for efficient execution of a Core ML image classifier, and use SpriteKit to display image classifier output in AR.
A sample app that demonstrates how to use the Natural Language framework to recognize languages, tokenize text into words and sentences, and tag named entities and parts of speech.
SwiftShot is an AR game for two to six players, featured in the WWDC18 keynote. Use this sample code project to experience it on your own device, see how it works, and build your own customized version of the game.