AR Symposium at Macquarie ICT Innovation Centre 14th August 2012

Introduction to AR – Rob Manson

  • Milgram’s Reality – Virtual Worlds spectrum
  • Azuma’s definition of AR in 1997
  • ARToolkit released in 1999 (C++ Libraries)
  • Wikipedia’s 2002 definition has been superseded by 2011 definition (current)
  • 2008 Wikitude released the first mobile browser (phone had camera, GPS, mobile)
  • 2009 FLARToolkit released in 2009 (Flash based, therefore worked in a web-browser). In Google traffic/trends, you can see a big spike.
  • 2010 First meeting of the ARStandards workshop in Soul (they meet a few times a year). The API allows you to access a range of input devices, not just camera.
  • 2011 JSARToolkit released, and it runs very well and can run in a browser.
    theAWEsomeweb.com.
  • Rob proposes a much broader definition of AR, beyond graphics overlay. Towards “Augmented Cognition” (Thad Starner)
  • “Animating the world with the human body” video http://blogs.technet.com/b/next/archive/2012/08/07/kin-202-tre-animating-the-world-with-the-human-body.aspx

Designing learning experiences using AR – Danny Munnerly

  • Design Thinking approach
  • Demonstrated some marker based AR apps
  • Spacecraft 3D – amazing model of the Mars Curiosity.
  • Transformer AR app hole in desk and shoot aliens.
  • Nomenclature: “Imagining spaces”, “revealing data”

Some great apps:

  • Aurasma
  • Junaio
  • Layar
  • AR Studio
  • String Augmented Reality
  • SpaceCraft 3D
  • Streetmuseum – Londinium
  • Transparent Earth
  • ARBasketball
  • Skyview Free
  • Plane Finder AR Free
  • Sun Seeker Lite
  • Around Me
  • Skinvaders
  • Peak AR
  • Magic Plan
  • Word Lens
  • seeLevel
  • Qibla AR
  • Aus Post
  • Magic Camera

Rob Manson – AR Technology Overview and Conceptual Unpacking

  • The four key modes of Augmented Reality:
    Public – multiuser fixed space
    Intimate – user alone potentially at a laptop
    Personal – Field of View pointing out to the real world scene
    Private – only you can see your Field of View (e.g. Google Glass)
  • World Lens is an App that does Augmented Reality translation of signs
  • For key user experiences of AR: http://ar-ux.com/the-4-key-user-experience-modes-of-augmented
  • Three approaches to AR triggers: Marker based Object based, Location based. These become the ‘anchors’ of for the AR ‘overlays’.
  • Uses multiple input devices: GPS, camera, gyroscope
  • We we select an App like Junaio or Aurasma we create a channel and content – can be used to create a custom App. There are also AR Content Management Systems, notably BuildAR. In the next few weeks BuildAR will add a HTML layer on top of the 3D projection layer that is projected on the 2D video stream from the camera (or elsewhere).
  • When interacting with location based content (e.g. walking tour) the world plane whereas for pictures on walls may have the vertical axis pointed out of the wall.
  • Facial recognition is now excellent and here: http://www.youtube.com/watch?v=tb0pMeg1UN0

 

About matthewbower

Associate Professor at Macquarie University.
This entry was posted in Uncategorized. Bookmark the permalink.