SlideShare a Scribd company logo
Investigating integration between auditory and visual location
information presented in an augmented-reality (AR) device
Hiraku Okumura1,2, Sushrut Khiwadkar1, and Sungyoung Kim1
1 Rochester Institute of Technology, Rochester, NY, USA, 2 Yamaha Corporation, Hamamatsu, Japan
Background
Method
Devices Visual Auditory
VR HMD (Oculus Rift, HTC Vive, Sony Playstation VR, etc.) Immersive Immersive
AR
HMD (Microsoft, HoloLens, etc.) Immersive Immersive
Glasses (Sony SmartEyeGlass, Epson Moverio, etc.) Simple
Immersive?
Simple?
The leap of the Virtual Reality (VR) and Augmented
Reality (AR) technologies allows more people to
access head mounted displays (HMDs) and smart
glasses in daily uses. HMDs can provide users with
immersive visual information, while smart glasses
display relatively simple visual information.
We presented a target visual object, the green small square, in the AR
glasses, Sony SmartEyeGlass (Fig. 1), at the 0° (face-forward center), -
5° and -10° in a counter clockwise from the center. Auditory stimuli,
White noise, were presented through the headphone to have target
locations in the horizontal plane from +45° to -45° with a 5-degree
interval.
The subjects were forced to choose between Yes or No as a response to
the question: ”Is the sound coming from the direction of the video cue”.
Discussion & Future work
[1] Kaneko, S., et al., “Ear Shape
Modeling for 3D Audio and Acoustic
Virtual Reality: The Shape-Based
Average HRTF,” in Audio
Engineering Society 61st
International Conference on Audio
for Games, February 2016.
Reference
Result
Panning
HRTF
Video cue angle (Horizontal)
0º -5º -10º
App
Host Android device
(Android tablet)
Laptop PCMAX
Bluetooth
Video Cue
Wireless Network
AKG K550
−45°45°
0° −10°
Sony
SmartEyeGlass
Audio Cue
By integrating proper Auditory information, we can enhance perceived immersion as well as improve clarity (and thus understandability) of visual
information. In audio engineering, binaural techniques based on HRTF (Head Related Transfer Function) are very popular rendering methods for a
pair of headphones. HRTF-based techniques can generate virtual sound sources all around a user and generate more immersive auditory image.
However, they require more computational costs. On the other hand, conventional amplitude-based panning methods can generate virtual sound
sources between loudspeakers, which require less computational costs. How much different are those two approaches in the context of AR?
To answer this question, we have conducted a study that compared two audio rendering techniques–HRTF-based and panning-based one. The
objective of the study was to quantitatively investigate possible benefits of employing panning-based rendering methods for AR devices.
• Audio rendering
• HRTF Dataset: Shape-based average HRTF (30-people) [1]
• Panning Method: Sine-Cosine panning law
• AR Device: Sony SmartEyeGlass
• Screen size (Angle of view) ± 10º (horizontal) x ± 5º
(vertical)
• Display color Monochrome (green)
Eyewear Controller
Fig.2 System Configuration
Fig.3 The user interface for subjectsFig.1 Sony SmartEyeGlass
The results show that two methods–amplitude-based panning and HRTF-based methods–did not produce any
significant differences for the tested AR device. This implicates that we can employ simple panning methods for
AR applications that mainly display simple visual information and save the computational cost.
As the first step, the current study has concentrated on audio/visual interaction along the horizontal plane. We
plan to extend the current work to the integration in the median plane as well as integration of perceived
distance. Furthermore, other types of the AR device such as Microsoft Hololens are being tested for multimodal
integration for more complex visual cues.

More Related Content

PPTX
Aman
aman gaur
 
PPTX
holographic mobile
Divi Balasubramanian
 
PPTX
3D Holography
Vish Pepala
 
PPTX
Digital Hologram Image Processing
Conor Mc Elhinney
 
PDF
3D Holography: When Might it become Economically Feasible?
Jeffrey Funk
 
PPT
INTRODUCTION TO HOLOGRAPHY
Mr. Swapnil G. Thaware
 
PDF
Holography
mandeep kaur virk
 
PDF
Introduction to 3D Holographic Technology
Andrew Japar
 
Aman
aman gaur
 
holographic mobile
Divi Balasubramanian
 
3D Holography
Vish Pepala
 
Digital Hologram Image Processing
Conor Mc Elhinney
 
3D Holography: When Might it become Economically Feasible?
Jeffrey Funk
 
INTRODUCTION TO HOLOGRAPHY
Mr. Swapnil G. Thaware
 
Holography
mandeep kaur virk
 
Introduction to 3D Holographic Technology
Andrew Japar
 

What's hot (20)

PPTX
3 d holographic projection
Ranjan Dhar
 
PPTX
Holography
Krishna Teja
 
PDF
Holographic technologies
Gilbertaubin
 
PPTX
3 d holographic projection technology
swathi b
 
PDF
Seminar report of 3D Holographic Projection Technology (Hologram).
SafwanSadi
 
PPTX
Holographic Projection Technology COMPLETE DETAILS NEW PPT
 Abin Baby
 
DOC
Seminar report neelam
Neelam Chhipa
 
PPTX
Holographic projection technology
Janardhan Raju
 
PPT
Holographic projections vaibhavp
Vaibhav P
 
PPTX
HOLOGRAPHIC TECHNOLOGY
koneru bhavani
 
PPTX
3D Holography Projection
ROHIT PAWAR
 
PPTX
Holography
Rahul Baghla
 
PPTX
Holography
nakshatradev28
 
PPTX
PPT on 3D Holographic Projection Technology ( Hologram)
SafwanSadi
 
PPTX
3d holographic projection ppt
Shahid Shihabudeen
 
PDF
3D Holographic Projection Technology
Shahid Shihabudeen
 
PPTX
Technology of Holographic Projection
Bhanupriya
 
PPTX
3D Holographic Projection Technology
Kirantheja Kittu
 
PPT
Holography phy-fahad
Fahad Mohammed
 
PPTX
holography in future
Chaitanya Ram
 
3 d holographic projection
Ranjan Dhar
 
Holography
Krishna Teja
 
Holographic technologies
Gilbertaubin
 
3 d holographic projection technology
swathi b
 
Seminar report of 3D Holographic Projection Technology (Hologram).
SafwanSadi
 
Holographic Projection Technology COMPLETE DETAILS NEW PPT
 Abin Baby
 
Seminar report neelam
Neelam Chhipa
 
Holographic projection technology
Janardhan Raju
 
Holographic projections vaibhavp
Vaibhav P
 
HOLOGRAPHIC TECHNOLOGY
koneru bhavani
 
3D Holography Projection
ROHIT PAWAR
 
Holography
Rahul Baghla
 
Holography
nakshatradev28
 
PPT on 3D Holographic Projection Technology ( Hologram)
SafwanSadi
 
3d holographic projection ppt
Shahid Shihabudeen
 
3D Holographic Projection Technology
Shahid Shihabudeen
 
Technology of Holographic Projection
Bhanupriya
 
3D Holographic Projection Technology
Kirantheja Kittu
 
Holography phy-fahad
Fahad Mohammed
 
holography in future
Chaitanya Ram
 
Ad

Similar to Poster02 (20)

PPTX
Build Your Own VR Display Course - SIGGRAPH 2017: Part 4
StanfordComputationalImaging
 
PDF
Spatial Audio for Augmented Reality
Mark Billinghurst
 
PDF
COMP 4010 Lecture5 VR Audio and Tracking
Mark Billinghurst
 
PDF
Comp4010 lecture3-AR Technology
Mark Billinghurst
 
PDF
COMP 4010: Lecture8 - AR Technology
Mark Billinghurst
 
PDF
3° RALI-Lisboa. SEEing THROUGH: enabling Augmented Reality applications from ...
SaraMautino
 
PDF
Multimodal Multi-sensory Interaction for Mixed Reality
Mark Billinghurst
 
PDF
Novel Interfaces for AR Systems
Mark Billinghurst
 
DOCX
Augmented reality documentation
Bhargav Doddala
 
PPT
Augmented reality
Brijesh Kothari
 
PDF
2022 COMP 4010 Lecture 7: Introduction to VR
Mark Billinghurst
 
DOCX
Augmented reality
sahebsab
 
PPTX
Augmented reality
AartiHariramani
 
PDF
Head-Mounted Display Visualizations to Support Sound Awareness for the Deaf a...
Dhruv Jain
 
PPTX
U - 5 Emerging.pptx
MulukenTamrat2
 
PDF
Lecture 9 AR Technology
Mark Billinghurst
 
PPTX
Chapter 5 - Augmented Reality.pptx
AmanuelZewdie4
 
PDF
COMP 4010 - Lecture 8 AR Technology
Mark Billinghurst
 
PDF
Thesis Defense Presentation
brinehold
 
PDF
COMP 4010 Lecture12 - Research Directions in AR and VR
Mark Billinghurst
 
Build Your Own VR Display Course - SIGGRAPH 2017: Part 4
StanfordComputationalImaging
 
Spatial Audio for Augmented Reality
Mark Billinghurst
 
COMP 4010 Lecture5 VR Audio and Tracking
Mark Billinghurst
 
Comp4010 lecture3-AR Technology
Mark Billinghurst
 
COMP 4010: Lecture8 - AR Technology
Mark Billinghurst
 
3° RALI-Lisboa. SEEing THROUGH: enabling Augmented Reality applications from ...
SaraMautino
 
Multimodal Multi-sensory Interaction for Mixed Reality
Mark Billinghurst
 
Novel Interfaces for AR Systems
Mark Billinghurst
 
Augmented reality documentation
Bhargav Doddala
 
Augmented reality
Brijesh Kothari
 
2022 COMP 4010 Lecture 7: Introduction to VR
Mark Billinghurst
 
Augmented reality
sahebsab
 
Augmented reality
AartiHariramani
 
Head-Mounted Display Visualizations to Support Sound Awareness for the Deaf a...
Dhruv Jain
 
U - 5 Emerging.pptx
MulukenTamrat2
 
Lecture 9 AR Technology
Mark Billinghurst
 
Chapter 5 - Augmented Reality.pptx
AmanuelZewdie4
 
COMP 4010 - Lecture 8 AR Technology
Mark Billinghurst
 
Thesis Defense Presentation
brinehold
 
COMP 4010 Lecture12 - Research Directions in AR and VR
Mark Billinghurst
 
Ad

Poster02

  • 1. Investigating integration between auditory and visual location information presented in an augmented-reality (AR) device Hiraku Okumura1,2, Sushrut Khiwadkar1, and Sungyoung Kim1 1 Rochester Institute of Technology, Rochester, NY, USA, 2 Yamaha Corporation, Hamamatsu, Japan Background Method Devices Visual Auditory VR HMD (Oculus Rift, HTC Vive, Sony Playstation VR, etc.) Immersive Immersive AR HMD (Microsoft, HoloLens, etc.) Immersive Immersive Glasses (Sony SmartEyeGlass, Epson Moverio, etc.) Simple Immersive? Simple? The leap of the Virtual Reality (VR) and Augmented Reality (AR) technologies allows more people to access head mounted displays (HMDs) and smart glasses in daily uses. HMDs can provide users with immersive visual information, while smart glasses display relatively simple visual information. We presented a target visual object, the green small square, in the AR glasses, Sony SmartEyeGlass (Fig. 1), at the 0° (face-forward center), - 5° and -10° in a counter clockwise from the center. Auditory stimuli, White noise, were presented through the headphone to have target locations in the horizontal plane from +45° to -45° with a 5-degree interval. The subjects were forced to choose between Yes or No as a response to the question: ”Is the sound coming from the direction of the video cue”. Discussion & Future work [1] Kaneko, S., et al., “Ear Shape Modeling for 3D Audio and Acoustic Virtual Reality: The Shape-Based Average HRTF,” in Audio Engineering Society 61st International Conference on Audio for Games, February 2016. Reference Result Panning HRTF Video cue angle (Horizontal) 0º -5º -10º App Host Android device (Android tablet) Laptop PCMAX Bluetooth Video Cue Wireless Network AKG K550 −45°45° 0° −10° Sony SmartEyeGlass Audio Cue By integrating proper Auditory information, we can enhance perceived immersion as well as improve clarity (and thus understandability) of visual information. In audio engineering, binaural techniques based on HRTF (Head Related Transfer Function) are very popular rendering methods for a pair of headphones. HRTF-based techniques can generate virtual sound sources all around a user and generate more immersive auditory image. However, they require more computational costs. On the other hand, conventional amplitude-based panning methods can generate virtual sound sources between loudspeakers, which require less computational costs. How much different are those two approaches in the context of AR? To answer this question, we have conducted a study that compared two audio rendering techniques–HRTF-based and panning-based one. The objective of the study was to quantitatively investigate possible benefits of employing panning-based rendering methods for AR devices. • Audio rendering • HRTF Dataset: Shape-based average HRTF (30-people) [1] • Panning Method: Sine-Cosine panning law • AR Device: Sony SmartEyeGlass • Screen size (Angle of view) ± 10º (horizontal) x ± 5º (vertical) • Display color Monochrome (green) Eyewear Controller Fig.2 System Configuration Fig.3 The user interface for subjectsFig.1 Sony SmartEyeGlass The results show that two methods–amplitude-based panning and HRTF-based methods–did not produce any significant differences for the tested AR device. This implicates that we can employ simple panning methods for AR applications that mainly display simple visual information and save the computational cost. As the first step, the current study has concentrated on audio/visual interaction along the horizontal plane. We plan to extend the current work to the integration in the median plane as well as integration of perceived distance. Furthermore, other types of the AR device such as Microsoft Hololens are being tested for multimodal integration for more complex visual cues.