Guest Editors
Prof. Zhigeng Pan, Nanjing University of Information Science & Technology, China
Dr. Gustavo Marfia, University of Bologna, Italy
Prof. Zhihan Lv, Qingdao University, China
Summary
Interactive technologies, such as eye tracking, speech recognition, and gesture input, have made great progress with the deepening of research. Users can use these interactive technologies to obtain information from virtual reality systems through multi-channels to eliminate the boundaries between the human environment and computer systems. Currently, various sensory organs of humans participate in the process of interaction between humans and computer systems, which can be called multimodal interaction from the perspective of system and technology. Compared with single-channel interaction, multimodal interaction has more extensive application potential in the human-computer natural interaction system. In the study of multimodal interaction, user behavior is one of the main input methods. The effective natural interaction can be achieved through classifying user behavior which conveys user intention. Besides, there is a corresponding mathematical mapping between user cognition and user behavior. It is necessary to diversify into various interactive devices for the sensory interactive system to enhance users’ interactive experience. At present, the interactive devices are mainly divided into voice input and output devices, image input and image display devices, the TouchPad, input devices, and visual tracking devices according to different sensory attributes. These interactive devices are applied to various computer systems to improve the computer efficiency and the accuracy of human-computer interaction. Through reasonable application in augmented reality systems, they will greatly improve the interactivity and user experience of augmented reality. Relevant researches have proved that the multimodal interaction system has achieved good application effects in the fields of digital media, cultural communication, commodity sales, information communication and so on.
Although virtual reality technology has been greatly developed and promoted, there are still some limitations in the study of multimodal interaction. Limited by technology, the system cannot directly identify user cognition. Meanwhile, the imperfect design of virtual reality system will affect the accuracy and efficiency of information transmission, thereby increasing cognitive load and reducing user experience. The existing research of multimodal interaction is mainly in the two fields of software design and hardware equipment with the focus on the application of multimodal system. For current augmented reality systems and applications, most studies still focus on visual and gesture-based augmented reality, which lacks direct physical and sensory stimulation for users. Therefore, it is expected to explore more interactive channels for users to input information which is processed by the system into different forms of information feedback to the sensory organs of users participating in the interaction corresponding to the information form. Research in the future also should pay attention to how to combine the architecture of multimodal interaction with augmented reality.
Some research topics of multimodal interaction have evoked public interest, including user cognition, user recognition, and behavior coding. The research topic of this special issue aims to provide readers with a comprehensive overview of multimodal interaction research and work in virtual reality. This special issue particularly expects original comment articles, opinions, methods and modeling research. The excellent papers of ICVR2022 (2022 IEEE 8th International Conference on Virtual Reality) will be considered for inclusion in the Special Issue. All submitted papers will undergo the Journal's standard peer-review process.
The areas covered by this special issue may include but are not limited to the following:
• Construction and Optimization of the Multimodal Interaction
• Multimodal Interaction Realization for Virtual Reality
• Analysis of Application Scenario of Multimodal Interaction
• Comparative analysis of Multimodal Interaction and Single-channel Interaction
• Analysis of Application Values of Multimodal Interaction
• Cognitive Expression under Multichannel Integration
• Quantitative Description of User Cognition in Interactive Systems
• Mathematical logic Relationship between Multimodal Natural Interrelation and User Cognition
• Data Fusion of User Behavior in Multimodal Natural Interaction
• Research on Cognition of Multimodal Interaction
• Relationship between Cognitive Load and User Behavior in Interactive Systems
• Evaluation and Analysis of User Multisensory Cognitive Channels
• Analysis of Weighting of Multisensory Channels
• Research on Multimodal Interaction Based on User Cognition
Published Papers
-
Open Access
ARTICLE
UOUU: User-Object Distance and User-User Distance Combined Method for Collaboration Task
Xiangdong Li, Pengfei Wang, Hanfei Xia, Yuting Niu
CMES-Computer Modeling in Engineering & Sciences, Vol.136, No.3, pp. 3213-3238, 2023, DOI:10.32604/cmes.2023.023895
(This article belongs to this Special Issue:
Recent Advances in Virtual Reality)
Abstract Augmented reality superimposes digital information onto objects in the physical world and enables multi-user collaboration. Despite that previous proxemic interaction research has explored many applications of user-object distance and user-user distance in an augmented reality context, respectively, and combining both types of distance can improve the efficiency of users’ perception and interaction with task objects and collaborators by providing users with insight into spatial relations of user-task object and user-user, less is concerned about how the two types of distances can be simultaneously adopted to assist collaboration tasks across multi-users. To fulfill the gap, we present UOUU, the user-object distance…
More >
Graphic Abstract
-
Open Access
ARTICLE
The Flipping-Free Full-Parallax Tabletop Integral Imaging with Enhanced Viewing Angle Based on Space-Multiplexed Voxel Screen and Compound Lens Array
Peiren Wang, Jinqiang Bi, Zilong Li, Xue Han, Zhengyang Li, Xiaozheng Wang
CMES-Computer Modeling in Engineering & Sciences, Vol.136, No.3, pp. 3197-3211, 2023, DOI:10.32604/cmes.2023.024305
(This article belongs to this Special Issue:
Recent Advances in Virtual Reality)
Abstract Tabletop integral imaging display with a more realistic and immersive experience has always been a hot spot in
three-dimensional imaging technology, widely used in biomedical imaging and visualization to enhance medical
diagnosis. However, the traditional structural characteristics of integral imaging display inevitably introduce the
flipping effect outside the effective viewing angle. Here, a full-parallax tabletop integral imaging display without
the flipping effect based on space-multiplexed voxel screen and compound lens array is demonstrated, and two
holographic functional screens with different parameters are optically designed and fabricated. To eliminate the
flipping effect in the reconstruction process, the space-multiplexed voxel screen consisting…
More >
Graphic Abstract
-
Open Access
ARTICLE
ER-Net: Efficient Recalibration Network for Multi-View Multi-Person 3D Pose Estimation
Mi Zhou, Rui Liu, Pengfei Yi, Dongsheng Zhou
CMES-Computer Modeling in Engineering & Sciences, Vol.136, No.2, pp. 2093-2109, 2023, DOI:10.32604/cmes.2023.024189
(This article belongs to this Special Issue:
Recent Advances in Virtual Reality)
Abstract Multi-view multi-person 3D human pose estimation is a hot topic in the field of human pose estimation due to its wide range of application scenarios. With the introduction of end-to-end direct regression methods, the field has entered a new stage of development. However, the regression results of joints that are more heavily influenced by external factors are not accurate enough even for the optimal method. In this paper, we propose an effective feature recalibration module based on the channel attention mechanism and a relative optimal calibration strategy, which is applied to the multi-view multi-person 3D human pose estimation task to…
More >
Graphic Abstract
-
Open Access
ARTICLE
Aggregate Point Cloud Geometric Features for Processing
Yinghao Li, Renbo Xia, Jibin Zhao, Yueling Chen, Liming Tao, Hangbo Zou, Tao Zhang
CMES-Computer Modeling in Engineering & Sciences, Vol.136, No.1, pp. 555-571, 2023, DOI:10.32604/cmes.2023.024470
(This article belongs to this Special Issue:
Recent Advances in Virtual Reality)
Abstract As 3D acquisition technology develops and 3D sensors become increasingly affordable, large quantities of 3D point cloud data are emerging. How to effectively learn and extract the geometric features from these point clouds has become an urgent problem to be solved. The point cloud geometric information is hidden in disordered, unstructured points, making point cloud analysis a very challenging problem. To address this problem, we propose a novel network framework, called Tree Graph Network (TGNet), which can sample, group, and aggregate local geometric features. Specifically, we construct a Tree Graph by explicit rules, which consists of curves extending in all…
More >
-
Open Access
ARTICLE
Monocular Depth Estimation with Sharp Boundary
Xin Yang, Qingling Chang, Shiting Xu, Xinlin Liu, Yan Cui
CMES-Computer Modeling in Engineering & Sciences, Vol.136, No.1, pp. 573-592, 2023, DOI:10.32604/cmes.2023.023424
(This article belongs to this Special Issue:
Recent Advances in Virtual Reality)
Abstract Monocular depth estimation is the basic task in computer vision. Its accuracy has tremendous improvement in the decade with the development of deep learning. However, the blurry boundary in the depth map is a serious problem. Researchers find that the blurry boundary is mainly caused by two factors. First, the low-level features, containing boundary and structure information, may be lost in deep networks during the convolution process. Second, the model ignores the errors introduced by the boundary area due to the few portions of the boundary area in the whole area, during the backpropagation. Focusing on the factors mentioned above.…
More >