Introduction to AR – Rob Manson
- Milgram’s Reality – Virtual Worlds spectrum
- Azuma’s definition of AR in 1997
- ARToolkit released in 1999 (C++ Libraries)
- Wikipedia’s 2002 definition has been superseded by 2011 definition (current)
- 2008 Wikitude released the first mobile browser (phone had camera, GPS, mobile)
- 2009 FLARToolkit released in 2009 (Flash based, therefore worked in a web-browser). In Google traffic/trends, you can see a big spike.
- 2010 First meeting of the ARStandards workshop in Soul (they meet a few times a year). The API allows you to access a range of input devices, not just camera.
- 2011 JSARToolkit released, and it runs very well and can run in a browser.
- Rob proposes a much broader definition of AR, beyond graphics overlay. Towards “Augmented Cognition” (Thad Starner)
- “Animating the world with the human body” video http://blogs.technet.com/b/next/archive/2012/08/07/kin-202-tre-animating-the-world-with-the-human-body.aspx
Designing learning experiences using AR – Danny Munnerly
- Design Thinking approach
- Demonstrated some marker based AR apps
- Spacecraft 3D – amazing model of the Mars Curiosity.
- Transformer AR app hole in desk and shoot aliens.
- Nomenclature: “Imagining spaces”, “revealing data”
Some great apps:
- AR Studio
- String Augmented Reality
- SpaceCraft 3D
- Streetmuseum – Londinium
- Transparent Earth
- Skyview Free
- Plane Finder AR Free
- Sun Seeker Lite
- Around Me
- Peak AR
- Magic Plan
- Word Lens
- Qibla AR
- Aus Post
- Magic Camera
Rob Manson – AR Technology Overview and Conceptual Unpacking
- The four key modes of Augmented Reality:
Public – multiuser fixed space
Intimate – user alone potentially at a laptop
Personal – Field of View pointing out to the real world scene
Private – only you can see your Field of View (e.g. Google Glass)
- World Lens is an App that does Augmented Reality translation of signs
- For key user experiences of AR: http://ar-ux.com/the-4-key-user-experience-modes-of-augmented
- Three approaches to AR triggers: Marker based Object based, Location based. These become the ‘anchors’ of for the AR ‘overlays’.
- Uses multiple input devices: GPS, camera, gyroscope
- We we select an App like Junaio or Aurasma we create a channel and content – can be used to create a custom App. There are also AR Content Management Systems, notably BuildAR. In the next few weeks BuildAR will add a HTML layer on top of the 3D projection layer that is projected on the 2D video stream from the camera (or elsewhere).
- When interacting with location based content (e.g. walking tour) the world plane whereas for pictures on walls may have the vertical axis pointed out of the wall.
- Facial recognition is now excellent and here: http://www.youtube.com/watch?v=tb0pMeg1UN0