default search action
VRST 2018: Tokyo, Japan
- Stephen N. Spencer, Shigeo Morishima, Yuichi Itoh, Takaaki Shiratori, Yonghao Yue, Rob Lindeman:
Proceedings of the 24th ACM Symposium on Virtual Reality Software and Technology, VRST 2018, Tokyo, Japan, November 28 - December 01, 2018. ACM 2018, ISBN 978-1-4503-6086-9
UI & display
- Alexandre Armengol-Urpi, Sanjay E. Sarma:
Sublime: a hands-free virtual reality menu navigation system using a high-frequency SSVEP-based brain-computer interface. 1:1-1:8 - Alexander Marquardt, Ernst Kruijff, Christina Trepkowski, Jens Maiero, Andrea Schwandt, André Hinkenjann, Wolfgang Stuerzlinger, Johannes Schöning:
Audio-tactile proximity feedback for enhancing 3D manipulation. 2:1-2:10 - Alexander Marquardt, Jens Maiero, Ernst Kruijff, Christina Trepkowski, Andrea Schwandt, André Hinkenjann, Johannes Schöning, Wolfgang Stuerzlinger:
Tactile hand motion and pose guidance for 3D interaction. 3:1-3:10 - Maarten Wijnants, Hendrik Lievens, Nick Michiels, Jeroen Put, Peter Quax, Wim Lamotte:
Standards-compliant HTTP adaptive streaming of static light fields. 4:1-4:12 - Dylan Brodie Fafard, Qian Zhou, Chris Chamberlain, Georg Hagemann, Sidney S. Fels, Ian Stavness:
Design and implementation of a multi-person fish-tank virtual reality display. 5:1-5:9 - Shotaro Ichikawa, Kazuki Takashima, Anthony Tang, Yoshifumi Kitamura:
VR safari park: a concept-based world building interface using blocks and world tree. 6:1-6:5 - Soichiro Toyohara, Toshiki Sato, Hideki Koike:
Balloonygen: extended tabletop display embedded with balloon-like deformable spherical screen. 7:1-7:5
AR / MR
- Anna Bolder, Stefan M. Grünvogel, Emanuel Angelescu:
Comparison of the usability of a car infotainment system in a mixed reality environment and in a real car. 8:1-8:10 - Jann Philipp Freiwald, Nicholas Katzakis, Frank Steinicke:
Camera time warp: compensating latency in video see-through head-mounted-displays for reduced cybersickness effects. 9:1-9:7 - Fouad Shoie Alallah, Ali Neshati, Yumiko Sakamoto, Khalad Hasan, Edward Lank, Andrea Bunt, Pourang Irani:
Performer vs. observer: whose comfort level should we consider when examining the social acceptability of input modalities for head-worn display? 10:1-10:9 - Ben J. Congdon, Tuanfeng Wang, Anthony Steed:
Merging environments for shared spaces in mixed reality. 11:1-11:8 - Satoshi Hashiguchi, Shohei Mori, Miho Tanaka, Fumihisa Shibata, Asako Kimura:
Perceived weight of a rod under augmented and diminished reality visual effects. 12:1-12:6 - Masashi Nitta, Tomohiro Sueishi, Masatoshi Ishikawa:
Tracking projection mosaicing by synchronized high-speed optical axis control. 13:1-13:5 - Junpei Miyamoto, Hideki Koike, Toshiyuki Amano:
Gaze navigation in the real world by changing visual appearance of objects using projector-camera system. 14:1-14:5
Human & machine visual perception
- Alexis D. Souchet, Stéphanie Philippe, Dimitri Zobel, Floriane Ober, Aurélien Léveque, Laure Leroy:
Eyestrain impacts on learning job interview with a serious game in virtual reality: a randomized double-blinded study. 15:1-15:12 - Daniel Medeiros, Rafael Kuffner dos Anjos, Daniel Mendes, João Madeiras Pereira, Alberto Raposo, Joaquim A. Jorge:
Keep my head on my shoulders!: why third-person is bad for navigation in VR. 16:1-16:10 - Yi Zhou, Mingjun Cao, Jingdi You, Ming Meng, Yuehua Wang, Zhong Zhou:
MR video fusion: interactive 3D modeling and stitching on wide-baseline videos. 17:1-17:11 - David R. Walton, Anthony Steed:
Dynamic HDR environment capture for mixed reality. 18:1-18:11 - Brendan John, Pallavi Raiturkar, Arunava Banerjee, Eakta Jain:
An evaluation of pupillary light response models for 2D screens and VR HMDs. 19:1-19:11 - Menandro Roxas, Tomoki Hori, Taiki Fukiage, Yasuhide Okamoto, Takeshi Oishi:
Occlusion handling using semantic segmentation and visibility-based rendering for mixed reality. 20:1-20:8
Presence
- Iana Podkosova, Hannes Kaufmann:
Co-presence and proxemics in shared walkable virtual environments with mixed colocation. 21:1-21:11 - Fares Moustafa, Anthony Steed:
A longitudinal study of small group interaction in social virtual reality. 22:1-22:10 - Mathias Parger, Joerg H. Mueller, Dieter Schmalstieg, Markus Steinberger:
Human upper-body inverse kinematics for increased embodiment in consumer-grade virtual reality. 23:1-23:10 - Richard Skarbez, Frederick P. Brooks Jr., Mary C. Whitton:
Immersion and coherence in a stressful virtual environment. 24:1-24:11 - Myungho Lee, Nahal Norouzi, Gerd Bruder, Pamela J. Wisniewski, Gregory F. Welch:
The physical-virtual table: exploring the effects of a virtual human's physical influence on social interaction. 25:1-25:11 - Jan Kolkmeier, Emiel Harmsen, Sander Giesselink, Dennis Reidsma, Mariët Theune, Dirk Heylen:
With a little help from a holographic friend: the OpenIMPRESS mixed reality telepresence toolkit for remote collaboration systems. 26:1-26:11 - Ceenu George, Michael Spitzer, Heinrich Hussmann:
Training in IVR: investigating the effect of instructor design on social presence and performance of the VR user. 27:1-27:5
VR environment
- Linjia He, Hongsong Li, Tong Xue, Deyuan Sun, Shoulun Zhu, Gangyi Ding:
Am I in the theater?: usability study of live performance based virtual reality. 28:1-28:11 - Anh Nguyen, Andreas M. Kunz:
Discrete scene rotation during blinks and its effect on redirected walking algorithms. 29:1-29:10 - Yang Hong, Andrew MacQuarrie, Anthony Steed:
The effect of chair type on users' viewing experience for 360-degree video. 30:1-30:11 - Aniket Bera, Tanmay Randhavane, Emily Kubin, Husam Shaik, Kurt Gray, Dinesh Manocha:
Data-driven modeling of group entitativity in virtual environments. 31:1-31:10 - Benjamin Nuernberger, Tobias Höllerer, Matthew A. Turk:
Hybrid orbiting-to-photos in 3D reconstructed visual reality. 32:1-32:10 - Sangyoon Han, Amit Bhardwaj, Seungmoon Choi:
Automatic transfer of musical mood into virtual environments. 33:1-33:5 - Congzhi Wang, Oana A. Dogaru, Patrick L. Strandholt, Niels C. Nilsson, Rolf Nordahl, Stefania Serafin:
Step aside: an initial exploration of gestural input for lateral movement during walking-in-place locomotion. 34:1-34:5
Modality
- Ping-Hsuan Han, Yang-Sheng Chen, Kong-Chang Lee, Hao-Cheng Wang, Chiao-En Hsieh, Jui-Chun Hsiao, Chien-Hsing Chou, Yi-Ping Hung:
Haptic around: multiple tactile sensations for immersive environment and interaction in virtual reality. 35:1-35:10 - Tanh Quang Tran, Thanh Ngoc-Dat Tran, Duy-Tam Nguyen, Holger Regenbrecht, Minh-Triet Tran:
Can we perceive changes in our moving speed: a comparison between directly and indirectly powering the locomotion in virtual environments. 36:1-36:10 - Fabien Danieau, Philippe Guillotel, Olivier Dumas, Thomas Lopez, Bertrand Leroy, Nicolas Mollet:
HFX studio: haptic editor for full-body immersive experiences. 37:1-37:9 - Dandan Pan, Qing Xu, Shiqiang Ma, Kunlong Zhang:
The impact of fear of the sea on working memory performance: a research based on virtual reality. 38:1-38:9 - Sharif Mohammad Shahnewaz Ferdous, Tanvir Irfan Chowdhury, Imtiaz Muhammad Arafat, John Quarles:
Investigating the reason for increased postural instability in virtual reality for persons with balance impairments. 39:1-39:7 - Michinari Kono, Takashi Miyaki, Jun Rekimoto:
In-pulse: inducing fear and pain in virtual experiences. 40:1-40:5 - Taizhou Chen, Yi-Shiun Wu, Kening Zhu:
Investigating different modalities of directional cues for multi-task visual-searching scenario in virtual reality. 41:1-41:5 - Sunghwan Shin, Seungmoon Choi:
Effects of haptic texture rendering modalities on realism. 42:1-42:5
Demo abstracts
- Ya-Kuei Chang, Jui-Wei Huang, Chien-Hua Chen, Chien-Wen Chen, Jian-Wei Peng, Min-Chun Hu, Chih-Yuan Yao, Hung-Kuo Chu:
A lightweight and efficient system for tracking handheld objects in virtual reality. 43:1-43:2 - Wataru Wakita, Tomoyuki Takano, Toshiyuki Hadama:
A low-cost motion platform with balance board. 44:1-44:2 - Wataru Wakita, Tomoyuki Takano, Toshiyuki Hadama:
A low-cost omni-directional VR walking platform by thigh supporting and motion estimation. 45:1-45:2 - Ryosuke Tanno, Takumi Ege, Keiji Yanai:
AR DeepCalorieCam V2: food calorie estimation with CNN and AR-based actual size estimation. 46:1-46:2 - Elijah Schwelling, Kyungjin Yoo:
Automatic 3D modeling of artwork and visualizing audio in an augmented reality environment. 47:1-47:2 - Andrew R. Lilja, Campbell W. Strong, Benjamin J. Bailey, Kristofer J. Thurecht, Zachary H. Houston, Nicholas L. Fletcher, John B. McGhee:
Design-led 3D visualization of nanomedicines in virtual reality. 48:1-48:2 - Ker-Jiun Wang, Quanbo Liu, Soumya Vhasure, Quanfeng Liu, Caroline Yan Zheng, Prakash Thakur:
EXG wearable human-machine interface for natural multimodal interaction in VR environment. 49:1-49:2 - Tomasz Bednarz, Daniel Filonik, Amy Buchan, Lucy Ogden-Doyle:
Future-mine VR as narrative decision making tool. 50:1-50:2 - Chih-Hao Cheng, Chia-Chi Chang, Ying-Hsuan Chen, Ying-Li Lin, Jing-Yuan Huang, Ping-Hsuan Han, Ju-Chun Ko, Lai-Chung Lee:
GravityCup: a liquid-based haptics for simulating dynamic weight in virtual reality. 51:1-51:2 - George Chernyshov, Kirill Ragozin, Cedric Caremel, Kai Kunze:
Hand motion prediction for just-in-time thermo-haptic feedback. 52:1-52:2 - Shiro Ise:
Immersive auditory display system 'sound cask': three-dimensional sound field reproduction system based on the boundary surface control principle. 53:1-53:2 - Taru Rustagi, Kyungjin Yoo:
Indoor AR navigation using tilesets. 54:1-54:2 - Saadiq K. Shaik, Kyungjin Yoo:
Interactive virtual exhibition: creating custom virtual art galleries using web technologies. 55:1-55:2 - Hirofumi Seo, Naoyuki Shono, Taichi Kin, Takeo Igarashi:
Real-time virtual brain aneurysm clipping surgery. 56:1-56:2 - Takumi Azai, Syunsuke Ushiro, Junlin Li, Mai Otsuki, Fumihisa Shibata, Asako Kimura:
Tap-tap menu: body touching for virtual interactive menus. 57:1-57:2 - Ying-Li Lin, Tsai-Yi Chou, Yu-Cheng Lieo, Yu-Cheng Huang, Ping-Hsuan Han:
TransFork: using olfactory device for augmented tasting experience with video see-through head-mounted display. 58:1-58:2 - Kelvin Cheng, Ichiro Furusawa:
Using mixed reality for promoting brand perception. 59:1-59:2 - Tomomi Takashina, Yuji Kokumai:
Virtual reality environment to support activity in the real world: a case of working environment using microscope. 60:1-60:2 - Jean-Philippe Farrugia:
VirtualHaus: a collaborative mixed reality application with tangible interface. 61:1-61:2 - Artur Baranowski, Peter Seipel, Andreas Schreiber:
Visualizing and exploring OSGi-based software architectures in augmented reality. 62:1-62:2 - Kevin Wennrich, Benjamin Tag, Kai Kunze:
VRTe do: the way of the virtual hand. 63:1-63:2
Poster abstracts
- Daewoon Kim, Sung-In Choi, Jung Seo Park, Kwang Hee Ko:
3D model augmentation using depth information in an AR environment. 64:1-64:2 - Yanxiang Zhang, Yirun Shen, Weiwei Zhang, Pengfei Ma:
A binocular stereo effect parameter calculator towards visual comfort. 65:1-65:2 - Yuka Tanaka, Homare Kon, Hideki Koike:
A real-time golf-swing training system using sonification and sound image localization. 66:1-66:2 - Takafumi Iwaguchi, Hiroyuki Kubo, Takuya Funatomi, Yasuhiro Mukaigawa, Srinivasa G. Narasimhan:
Acquiring short range 4D light transport with synchronized projector camera system. 67:1-67:2 - Jiajia Tan, Boyang Gao, Xiaobo Lu:
An AR system for artistic creativity education. 68:1-68:2 - Kristoffer Waldow, Martin Misiak, Ursula Derichs, Olaf Clausen, Arnulph Fuhrmann:
An evaluation of smartphone-based interaction in AR for constrained object manipulation. 69:1-69:2 - Taiki Yamada, Fumihisa Shibata, Asako Kimura:
Analysis of the R-V dynamics illusion behavior in terms of auditory stimulation. 70:1-70:2 - Taru Rustagi, Kyungjin Yoo:
AR navigation solution using vector tiles. 71:1-71:2 - Elijah Schwelling, Kyungjin Yoo:
Automatic 3D modeling of artwork and visualizing audio in an augmented reality environment. 72:1-72:2 - Dong-Yong Lee, Yong-Hun Cho, In-Kwon Lee:
Being them: presence of using non-human avatars in immersive virtual environment. 73:1-73:2 - Yishuo Liu, Yichuan Zhang, Shiliang Zuo, Wai-Tat Fu:
BoatAR: a multi-user augmented-reality platform for boat. 74:1-74:2 - Tomokazu Hayashizaki, Takuya Yamamoto, Yasuyuki Yanagida:
Chest compression simulator that presents vibrations at the moment of rib fracture: transition of learning effect of compression position over a month. 75:1-75:2 - Shohei Morikawa, Suguru Saito:
Deep face rotation in the wild. 76:1-76:2 - Yoshifumi Kitamura, Kazuki Takashima, Kazuyuki Fujita:
Designing dynamic aware interiors. 77:1-77:2 - Oana Balan, Gabriela Moise, Alin Moldoveanu, Florica Moldoveanu, Marius Leordeanu:
Does automatic game difficulty level adjustment improve acrophobia therapy?: differences from baseline. 78:1-78:2 - Chi-Jung Lee, Hung-Kuo Chu:
Dual-MR: interaction with mixed reality using smartphones. 79:1-79:2 - Hyeonah Choi, Jiwon Oh, Minwook Chang, Gerard J. Kim:
Effect of accompanying onomatopoeia to interaction sound for altering user perception in virtual reality. 80:1-80:2 - Jiwon Oh, Gerard J. Kim:
Effect of accompanying onomatopoeia with sound feedback toward presence and user experience in virtual reality. 81:1-81:2 - Hideki Kawai, Hiroki Hara, Yasuyuki Yanagida:
Effect of change of head angle on visual horizontal plane. 82:1-82:2 - Juno Kim, Matthew Moroz, Benjamin Arcioni, Stephen A. Palmisano:
Effects of head-display lag on presence in the oculus rift. 83:1-83:2 - Takuya Kadowaki, Michika Maruyama, Tomohiko Hayakawa, Naoki Matsuzawa, Kenichiro Iwasaki, Masatoshi Ishikawa:
Effects of low video latency between visual information and physical sensation in immersive environments. 84:1-84:2 - Takuma Shimizume, Takeshi Umezawa, Noritaka Osawa:
Estimation of distance between thumb and forefinger from hand dorsal image using deep learning. 85:1-85:2 - Tomi Nukarinen, Jari Kangas, Jussi Rantala, Olli Koskinen, Roope Raisamo:
Evaluating ray casting and two gaze-based pointing techniques for object selection in virtual reality. 86:1-86:2 - Junjian Zhang, Yaohao Chen, Satoshi Hashizume, Naoya Muramatsu, Kotaro Omomo, Riku Iwasaki, Kaji Wataru, Yoichi Ochiai:
EXController: enhancing interaction capability for VR handheld controllers using real-time vision sensing. 87:1-87:2 - Jie Zhang, Hengxin Chen, Jiahui Wang, Mingqi Gao:
Experience the dougong construction in virtual reality. 88:1-88:2 - Syed Fawad Mustafa Zaidi, Thomas Male:
Experimenting novel virtual-reality immersion strategy to alleviate cybersickness. 89:1-89:2 - Mamoru Hatanaka, Rei Hamakawa:
Extending recreational environments with a landscape-superimposed display using mixed reality. 90:1-90:2 - Jun'ichi Murakami, Takashi Morimoto, Ikuhisa Mitsugami:
Gaze and body capture system under VR experiences. 91:1-91:2 - Olli Koskinen, Ismo Rakkolainen, Roope Raisamo:
Gigapixel virtual reality employing live superzoom cameras. 92:1-92:2 - Krzysztof Pietroszek, Christian Eckhardt, Liudmila Tahai:
Hamlet: directing virtual actors in computational live theater. 93:1-93:2 - Tomi Nukarinen, Jari Kangas, Jussi Rantala, Toni Pakkanen, Roope Raisamo:
Hands-free vibrotactile feedback for object selection tasks in virtual reality. 94:1-94:2 - Yuqi Li:
High color-fidelity display using a modified projector. 95:1-95:2 - Ismo Rakkolainen, Roope Raisamo, Matthew A. Turk, Tobias Höllerer:
Illumination for 360 degree cameras. 96:1-96:2 - Kazuki Shimozawa, Masakazu Nakazawa, Hideki Koike, Ryoichi Miyanaga, Naoki Hosoe:
Image compensation and stabilization for immersive 360-degree videos from capsule endoscopy. 97:1-97:2 - Natsuki Sunda, Kensuke Tobitani, Atsushi Takemoto, Iori Tani, Yusuke Tani, Taishi Fujiwara, Noriko Nagata, Nobufumi Morita:
Impression estimation model and pattern search system based on style features and Kansei metric. 98:1-98:2 - Saadiq K. Shaik, Kyungjin Yoo:
Interactive virtual exhibition: creating custom virtual art galleries using web technologies. 99:1-99:2 - Akifumi Takahashi, Kenta Tanabe, Hiroyuki Kajimoto:
Investigation on the cutaneous/proprioceptive contribution to the force sensation induced by electrical stimulation above tendon. 100:1-100:2 - Souta Mizuno, Tetsuro Kitahara, Shun Shiramatsu, Shugo Ichinose:
JamGesture: an improvisation support system based on physical gesture observed with smartphone. 101:1-101:2 - Jean-Pierre Lomaliza, Hanhoon Park:
Learning-based word segmentation for reliable text document retrieval and augmentation. 102:1-102:2 - Keonhee Park, Seongah Chin:
Let's guide a smart interface for VR HMD and leap motion. 103:1-103:2 - Samir Benbelkacem, Abdelkader Bellarbi, Nadia Zenati-Henda, Ahmed Bentaleb, Ahmed Nazim Bellabaci, Samir Otmane:
Low-cost VR collaborative system equipped with haptic feedback. 104:1-104:2 - Yukiko Nakanishi, Hiroyuki Kubo, Takuya Funatomi, Yasuhiro Mukaigawa:
Mathematical model for pop-up effect of ChromaDepth. 105:1-105:2 - Nikesh Mishra, Eelke Folmer:
Measuring physical exertion in virtual reality exercise games. 106:1-106:2 - Kosuke Sato, Keita Watanabe, Hiroaki Yano, Hiroo Iwata:
Motion recognition for automatic control of a block machine. 107:1-107:2 - Aaron Hitchcock, Kelvin Sung:
Multi-view augmented reality with a drone. 108:1-108:2 - Zipeng Zheng, Zhuo Yang, Yinwei Zhan, Yuqing Li, Wenxin Yu:
Perceptual model optimized efficient foveated rendering. 109:1-109:2 - Yu-Chih Lin, Leon Hsu, Mike Y. Chen:
PeriTextAR: utilizing peripheral vision for reading text on augmented reality smart glasses. 110:1-110:2 - Kazuma Chiba, Yunosuke Nakayama, Tomoko Hashida:
Phantazuma: the stage machinery enabling the audience members to watch different contents depending on their position by vision control film and pepper's ghost. 111:1-111:2 - Emi Seki, Sho Kamei, Hisakazu Hada:
plARy: sound augmented reality system using video game background music. 112:1-112:2 - Yoshihiro Ishihara, Makio Ishihara:
Preliminary study on angular properties of spatial awareness of human in virtual space. 113:1-113:2 - Gen Hori:
Prototyping impossible objects with VR. 114:1-114:2 - Daichi Horita, Jaehyeong Cho, Takumi Ege, Keiji Yanai:
Ramen spoon eraser: CNN-based photo transformation for improving attractiveness of ramen photos. 115:1-115:2 - Ryo Sugawara, Jiawei Huang, Kazuki Takashima, Taku Komura, Yoshifumi Kitamura:
Random-forest-based initializer for solving inverse problem in 3D motion tracking systems. 116:1-116:2 - Simon Stock, Christina Erler, Wilhelm Stork:
Realistic simulation of progressive vision diseases in virtual reality. 117:1-117:2 - Erwin Wu, Hideki Koike:
Real-time human motion forecasting using a RGB camera. 118:1-118:2 - Qi Feng, Hubert P. H. Shum, Shigeo Morishima:
Resolving occlusion for 3D object manipulation with hands in mixed reality. 119:1-119:2 - Yuto Hirao, Hiroshi Kawasaki:
Simultaneous independent information display at multiple depths using multiple projectors and patterns created by epipolar constraint and homography transformation. 120:1-120:2 - Akihiro Miyata, Hironobu Uno, Kenro Go, Kyosuke Higuchi, Ryota Shinozaki:
Study on VR-based wheelchair simulator using vection-inducing movies and limited-motion patterns. 121:1-121:2 - Mashu Irie, Tatsuya Shibata:
System of delivering virtual object to user in remote place by handing gestures. 122:1-122:2 - Clément Glédel, Hideaki Uchiyama, Yuji Oyamada, Rin-Ichiro Taniguchi:
Texture synthesis for stable planar tracking. 123:1-123:2 - Sylvia Rothe, Boris Kegeles, Mathias Allary, Heinrich Hußmann:
The impact of camera height in cinematic virtual reality. 124:1-124:2 - Katherine Tarre, Adam S. Williams, Lukas Borges, Naphtali David Rishe, Armando B. Barreto, Francisco R. Ortega:
Towards first person gamer modeling and the problem with game classification in user studies. 125:1-125:2 - Peter Wozniak, Antonio Capobianco, Nicolas Javahiraly, Dan Curticapean:
Towards unobtrusive obstacle detection and notification for VR. 126:1-126:2 - Jeremy Albert, Kelvin Sung:
User-centric classification of virtual reality locomotion. 127:1-127:2 - Naoki Kimura, Michinari Kono, Jun Rekimoto:
Using deep-neural-network to extend videos for head-mounted display experiences. 128:1-128:2 - Tanvir Irfan Chowdhury, Sharif Mohammad Shahnewaz Ferdous, Tabitha C. Peck, John Quarles:
"Virtual ability simulation" to boost rehabilitation exercise performance and confidence for people with disability. 129:1-129:2 - Stevanus Kevin, Yun Suen Pai, Kai Kunze:
Virtual gaze: exploring use of gaze as rich interaction method with virtual agent in interactive virtual reality content. 130:1-130:2 - Julia Hayes, Kyungjin Yoo:
Virtual reality interactivity in a museum environment. 131:1-131:2 - Marcel Bock, Andreas Schreiber:
Visualization of neural networks in virtual reality using Unreal Engine. 132:1-132:2 - Lisa Nafeie, Andreas Schreiber:
Visualization of software components and dependency graphs in virtual reality. 133:1-133:2 - Dae Kyo Jeong, Sangbong Yoo, Yun Jang:
VR sickness measurement with EEG using DNN algorithm. 134:1-134:2 - Wei Ma, Qiuyuan Wang, Danqing Shi, Shuo Liu, Congxin Cheng, Qingyuan Shi, Qiong Li, Tong Li, Yingqing Xu:
Walking into ancient paintings with virtual candles. 135:1-135:2 - Donghao Ren, Bongshin Lee, Tobias Höllerer:
XRCreator: interactive construction of immersive data-driven stories. 136:1-136:2
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.