Handheld AR App Development with Unity

325 次查看
  • 完成时间大约为 10 个小时
  • 中级
  • 英语


Augmented Reality, or AR, will transform how we see and interact with the world. And the hardware that makes AR possible is the smartphone that you may already have in your pocket. In this course, you’ll learn how to develop your own mobile AR applications in Unity for iOS and Android devices. You’ll learn about the features offered by Unity’s AR Foundation, and about additional features in ARKit and ARCore.

Using the very latest techniques recommended by Unity engineers, you’ll build a complete AR environment that you can continue to use after the course, while learning to apply best practices in user experience and interaction.

This is the last of three courses in Unity’s XR Specialization, which includes an Introduction to XR course as well as a course focused on developing VR applications with Unity.

Note that, to complete this course, you will need a smartphone capable of running ARKit or ARCore. Compatible devices are listed in the links below:

The course also assumes that you already have experience developing applications with Unity and that you are comfortable with basic C# programming. It will also be helpful if you are familiar with building other types of apps to run on mobile devices.


Laying the AR Foundation with Unity's AR Foundation Package

This module will give the you a brief history of augmented reality technologies and introduce you to the concept of SLAM (Simultaneous Localization and Mapping). We'll show you an overview of the technologies used for Mobile AR tracking and the Unity components that are used to work with Mobile AR devices. In the project work, you will create a Unity AR project from scratch using the Unity AR Foundation package and wire up enough functionality to get it running on either an ARKit or ARCore compatible smartphone. The app will allow you to pan your smartphone around to see live video of your environment on the display.

Architecting AR Space - Pose Tracking and Environment Detections

In this module, you will learn how to interpret and visualize the information generated by the AR subsystem with regard to the real world geometry it has detected. In the project work, you will add trackable managers and visualizers to your scene so that you can see what the AR vision system is detecting and tracking.

Designing the UX in AR - Raycast, Light Estimation, Physics and Occlusion

In this module, you will learn how to use lighting estimation, in addition to the geometry generated by the AR system, to create realistic and immersive occlusion effects. You will also learn techniques that allow Unity physics objects to interact with the detected geometry through screen touches or physics. The result will be that the robot in the virtual scene will match the environment more realistically as the scene lighting is adjusted to correspond to measured lighting conditions. You will also be able to move the robot around using the touchscreen, and place the robot behind surfaces so that it’s partially occluded.

Advanced AR

In this module, we will look at features of ARCore and ARKit that are not yet supported by AR Foundation. In our final lesson, we'll learn about AR design best practices recommended by Unity.

此外,您还能在所有设备上欣赏您的整个音乐资料库。免费畅听 3 个月,之后每月只需 ¥10.00。
Apple 广告
  • Coursera
  • edX
  • OpenLearning
  • FutureLearn
  • iversity
  • Udacity
  • NovoEd
  • Canvas
  • Open2Study
  • Google
  • ewant
  • FUN
  • IOC-Athlete-MOOC
  • World-Science-U
  • Codecademy
  • CourseSites
  • opencourseworld
  • ShareCourse
  • gacco
  • MiriadaX
  • openhpi
  • Stanford-Open-Edx
  • 网易云课堂
  • 中国大学MOOC
  • 学堂在线
  • 顶你学堂
  • 华文慕课
  • 好大学在线CnMooc
  • (部分课程由Coursera、Udemy、Linkshare共同提供)

© 2008-2020 MOOC.CN 慕课改变你,你改变世界