skip to main content
10.1145/2782782acmconferencesBook PagePublication PagessiggraphConference Proceedingsconference-collections
SIGGRAPH '15: ACM SIGGRAPH 2015 Emerging Technologies
ACM2015 Proceeding
Publisher:
  • Association for Computing Machinery
  • New York
  • NY
  • United States
Conference:
SIGGRAPH '15: Special Interest Group on Computer Graphics and Interactive Techniques Conference Los Angeles California August 9 - 13, 2015
ISBN:
978-1-4503-3635-2
Published:
31 July 2015
Sponsors:
Recommend ACM DL
ALREADY A SUBSCRIBER?SIGN IN

Reflects downloads up to 01 Jan 2025Bibliometrics
Skip Abstract Section
Abstract

Play with the latest interactive and graphics technologies before they transform the way we live and work. Emerging Technologies presents demonstrations of research from several fields, including displays, input devices, collaborative environments, and robotics.

Skip Table Of Content Section
abstract
A method for automatically aligning projection mapping systems

By creating an overlapped array of projected images, we can use projection technology to create truly seamless displays with resolutions that can scale up to tens of millions of pixels. Projection also lets us map these displays onto any surface, ...

abstract
A multi-projector display system of arbitrary shape, size and resolution

A long anticipated system in graphics environments is an inexpensive multi-projector display on immersive surfaces like cylinders or domes that can be easily deployed and maintained. The key feature to popularize its adoption is the capability to run ...

abstract
Acoustruments: passive, acoustically-driven, interactive controls for handheld devices

Smartphones and handheld devices are increasingly being used in interactive applications beyond their conventional touchscreens. For example, tangibles allow users to interact with mobile devices using physical objects both on screen and around the ...

abstract
AffectiveWear: toward recognizing facial expression

Facial expression is a powerful way for us to exchange information nonverbally. They can give us insights into how people feel and think. There are a number of works related to facial expression detection in computer vision. However, most works focus on ...

abstract
Air haptics: displaying feeling of contact with AR object using visuo-haptic interaction

"Air Haptics" is a visuo-haptic Augmented Reality (AR) system which gives us a sense of interacting with AR objects and characters like pinching and pulling, without any haptic devices, only using an effect of visuo-haptic interaction.

abstract
An automultiscopic projector array for interactive digital humans

Automultiscopic 3D displays allow a large number of viewers to experience 3D content simultaneously without the hassle of special glasses or head gear. Our display uses a dense array of video projectors to generate many images with high-angular density ...

abstract
CHILDHOOD: wearable suit for augmented child experience

Understanding and perceiving the world from a child's perspective is a very important key not only to design products and architecture, but also to remind staff who work closely with children, such as hospitals and kindergartens. Ida et al. investigated ...

research-article
Deformation lamps: a projection technique to make a static picture dynamic

We propose a new light projection technique named 'Deformation Lamps', which adds a variety of realistic movement impressions to a static projection target. While static pictures are good at expressing spatial information about objects and events, they ...

abstract
Doppler time-of-flight imaging

Over the last few years, depth cameras have become increasingly popular for a range of applications, including human-computer interaction and gaming, augmented reality, machine vision, and medical imaging. Many of the commercially-available devices use ...

abstract
Fairy lights in femtoseconds: aerial and volumetric graphics rendered by focused femtosecond laser combined with computational holographic fields

We envision a laser-induced plasma technology in general applications for public use. If laser-induced plasma aerial images were made available, many useful applications such as spatial aerial AR, aerial user interfaces, volumetric images could be ...

abstract
FlashTouch: touchscreen communication combining light and touch

FlashTouch is a new technology that enables data communication between touchscreen-based mobile devices. Touchscreen can be used as communication media using visible light and capacitive touch. We designed a stylus and small case modules for ...

abstract
High brightness HDR projection using dynamic phase modulation

We demonstrate a new, large screen projection technology that utilizes dynamic phase modulation for light steering to achieve both black levels indiscernible from the screen in dark environments and peak luminance levels some 20 times above what ...

abstract
HoloChat: 3D avatars on mobile light field displays

Holographic displays have long inspired science fiction writers and movie makers as the communication platform of the future. Why should we restrict ourselves to a two-dimensional screen if we can transmit our virtual selves in 3D? Despite striking ...

research-article
Laster technologies OmnivisiO project

There are around 285 million of visually impaired people in the world. Aged related Macular Degeneration (AMD) is for, example, one of the most important visual disease in occidental countries. The forecasts of the World Health Organization [WHO ICD ...

abstract
Making small spaces feel large: infinite walking in virtual reality

Over the past few years, virtual reality has experienced a resurgence. Fueled by a proliferation of consumer-level head-mounted display and motion tracking devices, an unprecedented quantity of immersive experiences and content are available for both ...

abstract
MEME: smart glasses to promote healthy habits for knowledge workers

As people use computing devices more and more -not only desktops and laptops but also smart phones and tablets- related health risks also increase. Two of the most common and severe issues related to computer work are (1) the computer vision syndrome (...

abstract
Midair touch display

We demonstrate a system which enables users with bare hands and naked eyes to tactually interact with midair three-dimensional objects. The displayed midair images are three-dimensional and can be seen and touched from multiple angles. By creating ...

abstract
Moving around in virtual space with spider silk

With the recent advances of wearable I/O devices, designers of immersive VR systems are able to provide users with many different ways to explore the virtual space. For example, Birdly [Rheiner 2014] is a flying simulator composed of visual, auditory, ...

abstract
PANORAMICAL

PANORAMICAL is a digital anthology of musical landscapes with analog controls. It is a collection of places that morph to your touch using a MIDI controller or a game controller. Initially conceived as an installation piece, players twist knobs on a ...

abstract
Po2: augmented haptics for interactive gameplay

User gestures are important in current gaming platforms, where users' movements are tracked and their actions trigger events in the games. We introduce Power of 2 (Po2), a new haptic technology that augments illusive tactile sensations on gesture-based ...

abstract
SemanticPaint: interactive segmentation and learning of 3D worlds

We present a real-time, interactive system for the geometric reconstruction, object-class segmentation and learning of 3D scenes [Valentin et al. 2015]. Using our system, a user can walk into a room wearing a depth camera and a virtual reality headset, ...

abstract
Shogyu Mujo

Shogyo Mujo utilizes 360° projection mapping to take artistic sculpture to the next level as temporal dimensional artwork that pulsates to music. The skull is a three-dimensional structure designed by Josh Harker, with projection and animation developed ...

abstract
The light field stereoscope

Over the last few years, virtual reality (VR) has re-emerged as a technology that is now feasible at low cost via inexpensive cellphone components. In particular, advances of high-resolution micro displays, low-latency orientation trackers, and modern ...

abstract
VibroSkate: a locomotion interface with the exact haptics and kinesthesia

Latest development of VR technology enabled people to enjoy varieties of immersive VR environment in small spaces. Locomotion interfaces are essential to explore extensive virtual worlds with more sense of immersion.

abstract
Wobble strings: spatially divided stroboscopic effect for augmenting wobbly motion of stringed instruments

When we snap strings playing with a CMOS camera, the strings seems to vibrate in a wobbly slow motion pattern. Because a CMOS sensor scans one line of video in sequence, fast moving objects are distorted during the scanning sequence. The morphing and ...

References

  1. EG 2007 Phlegmatic Dragon" Eurographics 2007.Google ScholarGoogle Scholar
  2. KUCHERA-MORIN, J., WRIGHT, M., WAKEFIELD, G., ROBERTS, C., ADDERTON, D., SAJADI, B., HÖLLERER, T., AND MAJUMDER, A. 2014. Immersive full-surround multi-user system design. Computers & Graphics 40, 10-21. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. SAJADI, B., AND MAJUMDER, A. 2009. Markerless view-independent registration of multiple distorted projectors on extruded surfaces using an uncalibrated camera. Visualization and Computer Graphics, IEEE Transactions on 15, 6, 1307-1316. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. SAJADI, B., AND MAJUMDER, A. 2010. Auto-calibration of cylindrical multi-projector systems. In Virtual Reality Conference (VR), 2010 IEEE, IEEE, 155-162. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. SAJADI, B., AND MAJUMDER, A. 2010. Automatic registration of multiple projectors on swept surfaces. In Proceedings of the 17th ACM Symposium on Virtual Reality Software and Technology, ACM, 159-166. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. SAJADI, B., AND MAJUMDER, A. 2011. Automatic registration of multi-projector domes using a single uncalibrated camera. In Computer Graphics Forum, vol. 30, Wiley Online Library, 1161-1170. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. SAJADI, B., LAZAROV, M., GOPI, M., AND MAJUMDER, A. 2009. Color seamlessness in multi-projector displays using constrained gamut morphing. Visualization and Computer Graphics, IEEE Transactions on 15, 6, 1317-1326. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. SAJADI, B., MEMBER, S., AND MAJUMDER, A., 2011. Autocalibrating projectors for tiled displays on piecewise smooth vertically extruded surfaces.Google ScholarGoogle Scholar
  9. Laput, G., Brockmeyer, E., Hudson, S.E., Harrison, C. Acoustruments: Passive, Acoustically-Driven, Interactive Controls for Handheld Devices. In Proc. CHI'15. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. FUKUMOTO, K., TERADA, T., TSUKAMOTO, M. A Smile/Laughter Recognition Mechanism for Smile-Based Life Logging. In Proceeding of AH 2013, ACM, pp. 213-220. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. LECUYER, A. et al. 2000. Pseudo-haptic feedback: can isometric input devices simulate force feedback? Virtual Reality 2000, Proc. IEEE, IEEE, 83-90. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. SCHAEFER, S. et al. 2006, Image Deformation Using Moving Least Squares Global illumination in production. ACM SIGGRAPH 2006 Papers, ACM, New York, NY, USA, 533-540. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. ARTSTEIN, R., TRAUM, D., ALEXANDER, O., LEUSKI, A., JONES, A., GEORGILA, K., DEBEVEC, P., SWARTOUT, W., MAIO, H., AND SMITH, S. 2014. Time-offset interaction with a holocaust survivor. In Proceedings of the 19th International Conference on Intelligent User Interfaces, ACM, New York, NY, USA, IUI '14, 163-168. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. JONES, A., NAGANO, K., LIU, J., BUSCH, J., YU, X., BOLAS, M., AND DEBEVEC, P. 2014. Interpolating vertical parallax for an autostereoscopic 3d projector array.Google ScholarGoogle Scholar
  15. IDA, S., AND YAMANAKA, T. 2000. A study in the usability of environment objects by the difference in eye level. In Proc. of the 47th Annual Conf. of JSSD, 396-397.Google ScholarGoogle Scholar
  16. AMANO, T. 2013. Projection based real-time material appearance manipulation. In Proceedings of 2013 IEEE Conference on Computer Vision and Pattern Recognition Workshops, 918-923. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. RASKAR, R., WELCH, G., AND LOW, K.-L. 2001. Shader lamps: Animating real objects with image-based illumination. In Proceedings of the 12th Eurographics Workshop on Rendering Techniques, 89-102. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. KIMURA, H., UCHIYAMA, T., AND YOSHIKAWA, H. 2006. Laser produced 3d display in the air. In ACM SIGGRAPH 2006 Emerging Technologies, ACM, New York, NY, USA, SIGGRAPH '06. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. SAITO, H., KIMURA, H., SHIMADA, S., NAEMURA, T., KAYAHARA, J., JARUSIRISAWAD, S., NOZICK, V., ISHIKAWA, H., MURAKAMI, T., AOKI, J., ASANO, A., KIMURA, T., KAKEHATA, M., SASAKI, F., YASHIRO, H., MORI, M., TORIZUKA, K., AND INO, K. 2008. Laser-plasma scanning 3d display for putting digital contents in free space. Proc. SPIE 6803, 680309-680309-10.Google ScholarGoogle ScholarCross RefCross Ref
  20. OGATA, M., SUGIURA, Y., OSAWA, H., IMAI, M. FlashTouch: Data Communication through Touchscreens. In Proceedings of SIGCHI 2013, ACM, 2321-22. Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. BLACKHAM, G., AND NEALE, A., 1998. Image display apparatus, Mar. 18. EP Patent App. EP19,970,306,624.Google ScholarGoogle Scholar
  22. DAMBERG, G., SEETZEN, H., WARD, G., HEIDRICH, W., AND WHITEHEAD, L. 2007. 3.2: High dynamic range projection systems. In SID Symposium Digest of Technical Papers, vol. 38.Google ScholarGoogle ScholarCross RefCross Ref
  23. HOSKINSON, R., STOEBER, B., HEIDRICH, W., AND FELS, S. 2010. Light reallocation for high contrast projection. ACM Transactions on Graphics (TOG) 29, 6, 165. Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. REINHARD, E., POULI, T., KUNKEL, T., LONG, B., BALLESTAD, A., AND DAMBERG, G. 2012. Calibrated image appearance reproduction. ACM Transactions on Graphics (TOG) 31, 6, 201. Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. FATTAL, D., PENG, Z., TRAN, T., VO, S., FIORENTINO, M., BRUG, J., AND BEAUSOLEIL, R. G. 2013. A multi-directional backlight for a wide-angle, glasses-free three-dimensional display. Nature 495, 7441, 348-351.Google ScholarGoogle ScholarCross RefCross Ref
  26. LI, H., YU, J., YE, Y., AND BREGLER, C. 2013. Realtime facial animation with on-the-fly correctives. ACM Transactions on Graphics (Proceedings SIGGRAPH 2013) 32, 4 (July). Google ScholarGoogle ScholarDigital LibraryDigital Library
  27. The international classification of diseases, 10th revision (icd-10) 2010. Technical report, Geneva Switzerland, 2010.Google ScholarGoogle Scholar
  28. World Health Organization. The use of residual vision by visually disabled person, euro reports and studies 41. Regional Office for Europe WHO, January 1981. Copenhagen, Denmark.Google ScholarGoogle Scholar
  29. World Health Organization. Global data on visual impairments, 2012. WHO/NMH/PBD/12.01.Google ScholarGoogle Scholar
  30. Benoît Froissard, Hubert Konik, Alain Trémeau, Éric Dinet 2014, Contribution of Augmented Reality Solutions to Assist Visually Impaired People in Their Mobility. HCI (7) 2014: 182-191.Google ScholarGoogle Scholar
  31. AZMANDIAN, M., BOLAS, M., AND SUMA, E. 2014. Countering user deviation during redirected walking. In ACM Symposium on Applied Perception, 129. Google ScholarGoogle ScholarDigital LibraryDigital Library
  32. AZMANDIAN, M., YAHATA, R., BOLAS, M., AND SUMA, E. 2014. An enhanced steering algorithm for redirected walking in virtual environments. In IEEE Virtual Reality, 65-66.Google ScholarGoogle Scholar
  33. RAZZAQUE, S., KOHN, Z., AND WHITTON, M. C. 2001. Redirected Walking. In Eurographics (Short Presentation).Google ScholarGoogle Scholar
  34. BONNEY, R., AND CORLETT, E. 2002. Head posture and loading of the cervical spine. Applied Ergonomics 33, 5, 415-417.Google ScholarGoogle ScholarCross RefCross Ref
  35. HEIDE, W., KOENIG, E., TRILLENBERG, P., KÖMPF, D., AND ZEE, D. 1999. Electrooculography: technical standards and applications. Electroencephalogr Clin Neurophysiol Suppl 52, 223-240.Google ScholarGoogle Scholar
  36. ISHIMARU, S., KUNZE, K., TANAKA, K., UEMA, Y., KISE, K., AND INAMI, M. 2015. Smart eyewear for interaction and activity recognition. In CHI Interacitity, ACM, 307-310. Google ScholarGoogle ScholarDigital LibraryDigital Library
  37. YAN, Z., HU, L., CHEN, H., AND LU, F. 2008. Computer vision syndrome: A widely spreading but largely unknown epidemic among computer users. Computers in Human Behavior 24, 5, 2026-2042. References. Google ScholarGoogle ScholarDigital LibraryDigital Library
  38. HOSHI, T., TAKAHASHI, M., IWAMOTO T., AND SHINODA, H., 2010, Noncontact tactile display based on radiation pressure of airborne ultrasound, In Trans. on Haptics, Vol. 3, No. 3, pp.155-165, IEEE. Google ScholarGoogle ScholarDigital LibraryDigital Library
  39. HASEGAWA, K., AND SHINODA, H., 2013, Aerial display of vibrotactile sensation with high spatial-temporal resolution using large-aperture airborne ultrasound phased array, In World Haptics Conference, IEEE.Google ScholarGoogle ScholarCross RefCross Ref
  40. CARTER, T, SEAH, S., LONG, B., DRINKWATER, B., SUBRAMANIAN S., 2013, UltraHaptics: multi-point mid-air haptic feedback for touch surfaces, In UIST, ACM. Google ScholarGoogle ScholarDigital LibraryDigital Library
  41. INOUE, S., Makino, Y., AND SHINODA, H., 2015, Active Touch Perception produced by airborne ultrasonic haptic hologram, In World Haptics, IEEE.Google ScholarGoogle Scholar
  42. ASUKANET, 2011. Aerial imaging. http://aerialimaging.tv.Google ScholarGoogle Scholar
  43. FAN, K., HUBER, J., NANAYAKKARA, S., AND INAMI, M. 2014, SpiderVision: extending the human field of view for augmented awareness. In Proc. AH '14, ACM, No. 49. Google ScholarGoogle ScholarDigital LibraryDigital Library
  44. RHEINER, M. 2014, Birdly an attempt to fly. In ACM SIGGRAPH 2014 Emerging Technologies, No. 3. Google ScholarGoogle ScholarDigital LibraryDigital Library
  45. ISRAR, A., AND POUPYREV, I. 2011. Tactile brush: drawing on skin with a tactile grid display. In Proc. SIGCHI ACM Conference on Human Factors in Computing Systems, ACM, 2019-2028. Google ScholarGoogle ScholarDigital LibraryDigital Library
  46. ZHAO, S., ISRAR, A., AND KLATZKY, R. 2015. Intermanual Apparent Tactile Motion on handheld tablets. In Proc. IEEE World Haptics Conference, IEEE.Google ScholarGoogle Scholar
  47. NEWCOMBE, R. A. et al. 2011. KinectFusion: Real-Time Dense Surface Mapping and Tracking. In ISMAR, IEEE. Google ScholarGoogle ScholarDigital LibraryDigital Library
  48. NIESSNER, M. et al. 2013. Real-time 3D Reconstruction at Scale using Voxel Hashing. ACM TOG 32, 6, 169. Google ScholarGoogle ScholarDigital LibraryDigital Library
  49. PRISACARIU, V. A., KÄHLER, O. et al. 2014. A Framework for the Volumetric Integration of Depth Images. ArXiv e-prints.Google ScholarGoogle Scholar
  50. VALENTIN, J. P. C. et al. 2015. SemanticPaint: Interactive 3D Labeling and Learning at your Fingertips. To appear in ACM TOG.Google ScholarGoogle Scholar
  51. CAKMAK, T., AND HAGER, H. 2014. Cyberith virtualizer: A locomotion device for virtual reality. In ACM SIGGRAPH 2014 Emerging Technologies, ACM, New York, NY, USA, SIGGRAPH '14, 6:1-6:1. Google ScholarGoogle ScholarDigital LibraryDigital Library
  52. CHRISTENSEN, R., HOLLERBACH, J. M., XU, Y., AND MEEK, S. 1998. Inertial force feedback for a locomotion interface. In Proc. ASME Dynamic Systems and Control Division, DSC, vol. 64, 119-126.Google ScholarGoogle Scholar
  53. FARKHATDINOV, I., OUARTI, N., AND HAYWARD, V. 2013. Vibrotactile inputs to the feet can modulate vection. In World Haptics Conference (WHC), 2013, IEEE, 677-681.Google ScholarGoogle ScholarCross RefCross Ref
  54. FUKAYA, T., IWAI, T., AND YAMANOUCHI, Y. 2006. Morphovision. In ACM SIGGRAPH 2006 Emerging Technologies, ACM, New York, NY, USA, SIGGRAPH '06. Google ScholarGoogle ScholarDigital LibraryDigital Library

Recommendations

Acceptance Rates

Overall Acceptance Rate 1,822 of 8,601 submissions, 21%
YearSubmittedAcceptedRate
SIGGRAPH '114328219%
SIGGRAPH '1039010326%
SIGGRAPH '094397818%
SIGGRAPH '085189017%
SIGGRAPH '0745510824%
SIGGRAPH '064748618%
SIGGRAPH '054619821%
SIGGRAPH '044788317%
SIGGRAPH '034248119%
SIGGRAPH '023586719%
SIGGRAPH '013006522%
SIGGRAPH '003045919%
SIGGRAPH '993205216%
SIGGRAPH '983034515%
SIGGRAPH '972654818%
SIGGRAPH '962475221%
SIGGRAPH '952575622%
SIGGRAPH '942425724%
SIGGRAPH '932254620%
SIGGRAPH '922134521%
SIGGRAPH '902104320%
SIGGRAPH '891903820%
SIGGRAPH '881613421%
SIGGRAPH '871403324%
SIGGRAPH '851753520%
SIGGRAPH '841184135%
SIGGRAPH '811323829%
SIGGRAPH '801405237%
SIGGRAPH '791104339%
SIGGRAPH '781206453%
Overall8,6011,82221%