Estimating scale using depth from focus for mobile augmented reality

  • Authors:
  • Klen Čopič Pucihar;Paul Coulton

  • Affiliations:
  • Lancaster University, Lancaster, United Kingdom;Lancaster University, Lancaster, United Kingdom

  • Venue:
  • Proceedings of the 3rd ACM SIGCHI symposium on Engineering interactive computing systems
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

Whilst there has been a considerable progress in augmented reality (AR) over recent years, it has principally been related to either marker based or apriori mapped systems, which limits its opportunity for wide scale deployment. Recent advances in marker-less systems that have no apriori information, using techniques borrowed from robotic vision, are now finding their way into mobile augmented reality and are producing exciting results. However, unlike marker based and apriori tracking systems these techniques are independent of scale which is a vital component in ensuring that augmented objects are contextually sensitive to the environment they are projected upon. In this paper we address the problem of scale by adapting a Depth From Focus (DFF) technique, which has previously been limited to high-end cameras to a commercial mobile phone. The results clearly show that the technique is viable and adds considerably to the enhancement of mobile augmented reality. As the solution only requires an auto focusing camera, it is also applicable to other AR platforms.