Skip to content
BY 4.0 license Open Access Published by De Gruyter November 10, 2024

Image modeling algorithm for environment design based on augmented and virtual reality technologies

  • Da Gong , Tong Chen , Shuxin Zhou , Jiaheng Han and Zhenghang Li EMAIL logo

Abstract

With the continuous development of computer vision technology, the image modeling of environmental design has also received more and more attention. This study proposed an image modeling algorithm for environment design based on augmented reality (AR) and virtual reality (VR) technologies. On the system architecture, the environment design image modeling algorithm based on AR and VR technologies has provided users with an immersive 3D simulation space environment. It also has provided users with a good interactive interface through AR technology, enabling users to interact with computers in a realistic virtual three-dimensional environment and interactive operation. This study verified the algorithm from three aspects: time, authenticity, and interactivity. When facing object A, the modeling time values of the traditional environmental design image modeling method and the environmental design image modeling method proposed in this study were 12.14 and 8.72 s, respectively. The modeling authenticity reached 87.51 and 98.57%, respectively, and the interactivity reached 79.14 and 96.45%, respectively. Some columns of data have proved that the environmental design image modeling method proposed in this study was superior to the traditional environmental design image modeling method. In addition, this study also verified the system model from three aspects: stability, running smoothness, and performance loss. When the number of image modeling in the environment design was 2, the stability, running smoothness, and performance loss of the system were 99.37, 99.25 and 0.75%, respectively. The experimental data again proved that the algorithm system model proposed in this study was feasible and worthy of further application.

1 Introduction

With the continuous development of information technology and the continuous expansion of application fields, the construction of a virtual environment using augmented reality (AR) technology has become an important topic in the field of computer graphics and digital image processing. It is a new simulation method that uses computer 3D graphics to simulate the real-world environment to realize the interaction between the virtual scene and real space. Among them, AR is to build a virtual and real interactive environment with the help of computer graphics technology and intelligent mobile devices. Currently, traditional image modeling methods for environmental design require a lot of manpower and time to complete, especially for complex environments and large-scale scenes. Moreover, traditional methods lack real-time interaction with users and cannot dynamically update the model when users operate. In order to solve the problem of rapid modeling of large-scale and complex virtual scenes, this study decomposes the scene reasonably and presents a multi-view-based image sequence model, which solves the problems of traditional 3D models such as long rendering cycles, rendering difficulties, and 3D reconstruction difficulties. This method can track the viewing position of the viewer in real-time and dynamically project the corresponding photos, which has a good visual effect in practical applications.

Environmental design is one of the most dynamic and research-oriented design disciplines in China, originating from interior decoration and design. Gharaveis et al. reviewed and discussed the literature on the impact of the building environment on team cooperation and communication in detail. The environmental design involves nurses, support staff and doctors, which is one of the key factors to improve the efficiency of team cooperation and collaborative communication [1]. The purpose of Bangwal and Tiwari was to study how the environmental design features of green buildings could promote the formation of employees’ organizational image by improving their environmental awareness [2]. de Gaulmyn and Dupre focused on architectural students in environmental design, aiming to show how to use a new sustainable performance simulation tool, namely a “simple method of sustainable and environmental design,” to help them learn innovative sustainable design [3]. Zadeh et al. synthesized and summarized the research evidence in the fields of medicine, environmental psychology, nursing, palliative treatment, environmental design, interior design, and evidence-based design through a comprehensive literature review [4]. Yang et al. aimed to investigate the role of strategic environmental orientation in the implementation of environmental design practice [5]. Although these studies have promoted environmental design to a certain extent, they have not been combined with the actual situation.

At the same time, virtual reality (VR) technology has gradually attracted widespread attention from the academic community. Lou advocated a new environmental design culture, which focused on the use of holistic, people-centered, and interdisciplinary methods to create and realize sustainable life/space ecosystems, including experience, communication, and place, to optimize the interaction between human beings and the surrounding environment, which provided ideas for the application of VR [6]. Joseph et al. proposed a method framework to use an immersive VR environment in headworn displays for environmental design research [7]. Yang and Dan reviewed the recent history of environmental design discipline, relevant major educational institutions, and the views of famous scholars. In addition, in the context of emerging design trends, he also proposed new collective and constructive ideas and methods of environmental design, which promoted the use of VR [8]. Eren and Yilmaz analyzed in detail the students’ attitudes toward digital and traditional painting methods in the environmental design studio of the Department of Landscape Architecture and the impact of these technologies on the academic success of the course, which had an impact on the depth of VR application [9]. Lawrenc et al. investigated the performance of environmental strategies in seven newly built or renovated university buildings in the United Kingdom. These buildings included a series of administrative spaces, classrooms, libraries, and studios, reflecting their complex, multipurpose, and heterogeneous nature, which laid the foundation for the use of VR in environmental design [10]. Although these research methods are innovative, a large number of experimental data are needed to prove the reliability of the methods.

This study introduced the image modeling method of environmental design in detail and introduced the implementation of the algorithm from three aspects: image modeling, AR and VR technologies, and the implementation of the image modeling algorithm of environmental design. On this basis, this study constructed a system model of environment design image modeling algorithm based on AR and VR technologies and proved the feasibility of the system model proposed in this study through algorithm experiments and system model test experiments.

2 Image modeling method for environmental design

2.1 Image modeling

Because image reconstruction needs to use different methods according to different image types, numbers, and reconstruction methods, there is no unified processing flow for image-based 3D reconstruction.

Figure 1 shows the basic flow of the image-based 3D reconstruction method.

Figure 1 
                  Basic flow of the image-based 3D reconstruction method.
Figure 1

Basic flow of the image-based 3D reconstruction method.

It can be seen from Figure 1 that, first, multiple sample images are used as input for calibration to obtain the direction and camera sampling parameters related to each sample image and 3D scene. Second, the point cloud data of the target surface is reconstructed in 3D. Third, the surface of the target is restored to obtain the reconstructed 3D geometric model. Finally, the texture of the object surface is processed to obtain the required 3D model.

In the method of image modeling based on the environment design, the geometric model of the scene or object should be reconstructed first, and then the visual effect of the scene should be obtained from the following two aspects. The image modeling method based on environment design carries out texture extraction and texture mapping on the reconstructed geometric model to generate a surface with high fidelity, thus avoiding the tedious lighting operation in the traditional drawing process and solving the problem that has been puzzling users’ rendering speed and realism for a long time. The image modeling method based on the environment design can obtain the reflection characteristics and light conditions of the surface, which provides the most intuitive basis for the rendering of the model, so as to achieve the real effect that cannot be achieved under the manually set reflection properties and light conditions. In addition, it can be rendered realistically in the case of light changes.

2.2 AR and VR technologies

In recent years, VR technology has been widely used in various fields but it also exposes its inherent problems, such as the isolation of the objective environment, which makes people feel unreal [11,12]. AR technology is an effective way to solve the above problems. It integrates virtual objects or other information generated by computers (such as text and illustration) with the objective environment to strengthen the objective world [13].

AR technology is the product of VR technology, but there are essential differences between them [14]. VR technology emphasizes that users are not in conflict with the real environment but are attracted by the virtual world created by computers. The characteristic of AR technology is that it is not isolated from the real environment but also can highlight the existence of the objective world and can improve the user’s perception ability [15]. AR technology pursues the unity of “virtual” and “real,” and “virtual” and “real” complement each other. At the same time, because the relationship between users and the real world has not been completely separated, their interaction is more harmonious and natural. AR technology builds a bridge between VR and reality and can realize many interactions that cannot be realized in reality. Its application prospect is very broad. Today, with the development of human society, this technology has been widely used in all walks of life.

VR technology is a virtual world created by people and computers. It can obtain information from multiple users in real time and track and interact with them in real time [16,17]. “Virtual” is generated by a computer using 3D software and programming language, and it does not exist. “Reality” is the physical space that can be felt and seen. VR technology is a comprehensive and cutting-edge technology that covers computer, human–computer interaction, image art, body feeling technology, and many other professional technologies [18,19]. With humans as the core, VR technology uses human perception technology to capture the user’s behavior, position and vision, compile them into data, and transmit them to the database [20]. The database re-enters the real-time user information into the virtual image and modifies it through real-time user data. This leads to changes according to the user’s actions when the user makes an action. Compared with the previous “Windows,” the computer image can be operated by a mouse, keyboard, and other tools [21].

The environment design image modeling algorithm model based on AR and VR technologies is composed of the scene interaction module, 3D model rendering module, and AR module. Figure 2 shows the main functions of the system [22].

Figure 2 
                  Main functions of the system.
Figure 2

Main functions of the system.

It can be seen from Figure 2 that the main functions of the system include simulation of real scenes, dynamic interaction, and real-time rendering. Among them, the 3D model building module models the shape of the virtual object [23]. Taking the building in the real scene as an example, the trees and plants outside the building, and the walls and ceilings inside the building are modeled [24]. Based on the 3D model, the virtual and real environment effect is established. By calling multiple layers in geographic information systems (such as topographic maps and building outline maps) and Line View technology, the fusion and interaction between geometry and virtual objects in real scenes are realized. This module can realize the digital representation of scenes in real scenes [25].

2.3 Implementation of image modeling algorithm for environmental design

In this study, the complex virtual scene is divided into two categories: regular shape and irregular shape. To solve the problem of modeling complex irregular objects, a virtual scene modeling method based on multi-view photo sequence is proposed. The basic flow of the algorithm is shown in Figure 3.

Figure 3 
                  Basic flow of the algorithm.
Figure 3

Basic flow of the algorithm.

It can be seen from Figure 3 that, first, according to the specific accuracy, the complex objects modeled are extracted from photos from multiple angles. On this basis, the positions of photos and observation points are combined to form a one-to-one “photo-view” relationship. Geometric virtual scene modeling technology is applied to rapidly generate regular shape models and other scene components included in the virtual environment. Finally, the composite virtual model generated by the photo sequence is used to integrate it into the virtual scene.

When the viewer roams in the virtual environment, it can track the position of the observation point in the world coordinate system in real time and convert it into the target coordinate system. By defining the corresponding relationship between “photo and perspective,” the picture corresponding to the current perspective is determined and displayed to the viewer until the user leaves the roaming scene.

During shooting, the photographic background corresponding to the surface material, color, and other characteristics of the object should be selected as much as possible, and the appropriate brightness and number should be set. The image sequence of the modeled object from different angles is obtained and then segmented. The background of the photo is separated from the outline of the object to form a series of transparent backgrounds, which is used for the next step of modeling. During the acquisition process, a background with obvious characteristics can be selected according to the surface material, color, and other characteristics of the modeled object, thus simplifying the image content and improving the image segmentation effect.

With the continuous development of image-matching technology, image-matching technology came into being. The current feature-based methods are basically based on the characteristics of the image itself, and the region-based matching algorithm is based on this [26].

Association degree operation is a commonly used matching algorithm at present, which is mainly realized by various correlation degree functions. Among them, the simplest correlation operation is completed by the average value of the brightness difference between pixels. Some researchers also proposed a correlation operation based on standard deviation correlation operation and divided it into low-level processing and high-level processing.

First, the gradient-based Plessey corner extraction algorithm is adopted. In the original gray image, the Prewitt operator is used to calculate the derivatives in the x and y directions, respectively. Second, according to the algorithm, the square of the gradient is calculated, and the Gaussian method is used to reduce the noise. Finally, the Plessey operator value is calculated.

Through relevant operations, diagonal points are processed accordingly. The pixel point in the first image is set to x, and the window with size ( 2 m + 1 ) × ( 2 n + 1 ) centered on x is used as the contrast window. In the second image, the correlation operation is performed in the search area of size ( 2 d u + 1 ) × ( 2 d b + 1 ) around the pixel. Related operations can be defined as:

(1) ϑ ( x , x ) = Cov ( x , x ) σ ( x ) σ ( x ) .

The gray mean, standard deviation, and covariance of the correlation operation window at point x = ( u , b ) are

(2) E ( x ) = o = m m k = n n O ( u + o , b + k ) ( 2 m + 1 ) ( 2 n + 1 ) ,

(3) σ ( x ) = o = m m k = n n [ O ( u + o , b + k ) E ( x ) ] 2 ( 2 m + 1 ) ( 2 n + 1 ) ,

(4) C o v ( x , x ) = o = m m k = n n [ O ( u + o , b + k ) E ( x ) ] [ O ( u + o , b + k ) E ( x ) ] ( 2 m + 1 ) ( 2 n + 1 ) ,

When the value of Cov ( x , x ) is −1, it means that the two regions are completely different, and when it is 1, it means that the two regions are completely the same.

A relaxation algorithm is used for matching optimization. The basic principle of this method is as follows: assuming that ( x , x ) is a candidate matching point, M ( x ) and M ( x ) are adjacent regions with a radius of R and a center of x and x , respectively. If ( x , x ) is a good match point, there are many match points ( y , y ) . Among them, y M ( x ) and y M ( x ) , or there are few matching points, or even no matching points.

In practice, due to the large number of feature points extracted in the low-level processing stage and the large amount of computation, it takes a long time to carry out the corresponding processing. In order to solve this problem, the threshold limit method can be used to select better matching points while maintaining certain feature points. However, the number of reserved feature points is difficult to calculate. If the retention rate is too high, the matching time cannot be effectively reduced. On the contrary, the accuracy of matching is greatly affected, and the corresponding image may not be found.

In addition, in order to speed up matching, some scholars suggest that feature extraction should be carried out before matching. The characteristics and objectives of this method are the same. It is to shorten the matching area by extracting feature points, thus speeding up the matching speed and improving the matching accuracy. However, if too few feature points are extracted, accuracy cannot be guaranteed, which is also a big problem.

Therefore, in order to improve the corresponding image processing speed and ensure that the corresponding image processing speed has strong robustness, more efficient feature extraction technology must be adopted to solve the aforementioned problems.

In this article, an edge detection algorithm based on image pixel weight is proposed to achieve fast panorama matching.

The edge point set extraction algorithm is as follows:

It is assumed that the resolution of the image is N × M and the weight of each pixel is u nm = 0 . The image is updated according to the weight of 8 adjacent pixels:

(5) u nm = u nm + 1 if O O ¯ o u nm if O < O ¯ o ( o = 1 , , 8 ) .

Among them, O ¯ o = 1 9 l = o 1 o + 1 z = k 1 k + 1 u l z represents the average intensity of the area adjacent to the pixel. Each pixel is detected. When the weight exceeds a certain threshold, the region is determined as a boundary pixel, and vice versa. The noise is reduced. If the weight of the measured pixel is non-zero, and the weight of its eight neighboring pixels is 0, its weight is 0.

The fast point correspondence algorithm is as follows.

Plessey’s corner extraction method is used to extract the corners of two images and obtain the set of feature points. The edges of the two images are extracted, and the set of boundary points is obtained. For the points in the point set, the area correlation method is used to establish a preliminary point correspondence. The relaxation algorithm is adopted to relax the corresponding points and further find out the corresponding relationship. By limiting conditions, corresponding points that do not meet the conditions are further filtered.

The image modeling algorithm is designed in combination with the environment of AR and VR technologies, which can transfer data to the server. After receiving the information, the data are analyzed, displayed, and stored, and the relevant design of the algorithm model is completed in combination with the system architecture to test the system. The image modeling algorithm model of environment design based on AR and VR technologies is shown in Figure 4.

Figure 4 
                  Environment design image modeling algorithm model.
Figure 4

Environment design image modeling algorithm model.

It can be seen from Figure 4 that the image modeling algorithm model designed in the AR and VR environment is mainly composed of three modules: visual interaction module, entity information acquisition module, and rendering system. The model is user-centric. The visual interaction module collects data through sensors, cameras, VR glasses, and other devices. The entity information acquisition module collects images through the graphics rendering system. The rendering system completes the rendering of images through three-dimensional (3D) reconstruction technology [27].

3 Environment design image modeling algorithm system for model verification experiment

Experimental environment: Hardware includes a desktop computer with Intel Core i7 processor (ninth generation), 16GB DDR4 memory and 1TB SSD storage, NVIDIA GeForce RTX 3080, Intel RealSense D435, Logitech C920 HD Pro, and Oculus Rift S head-mounted display. The software uses Windows 10 Professional (64-bit), OpenCV: version 4.5.2, MATLAB: version R2021b, Unity: version 2021.1.16f1, Vuforia: version 9.8.8, Oculus SDK: version 1.58, SteamVR: version 1.20.

3.1 Algorithm verification

In this study, different objects in the five environmental designs of object A, object B, object C, object D, and object E were selected as the environmental design content to be modeled, and five environments composed of at least two objects, environment 1, environment 2, environment 3, environment 4, and environment 5 were selected as the environmental design to be modeled. This experimental algorithm verification is divided into three categories: time verification, authenticity verification, and interactivity verification. For time verification, the experimental settings were designed by selecting different environmental objects and environments, and modeling using traditional image modeling methods and proposed image modeling methods, and the modeling of each method under different objects and environments was recorded time. The experimental method uses a timing tool to record the start and end time of each modeling and calculates the time used for modeling. Multiple experiments are conducted for each method on different objects and environments, and the average value is taken as the final result. The authenticity verification experiment also selected different environmental design objects and environments, and used traditional image modeling methods and proposed image modeling methods for modeling, recording the authenticity of the modeling results of each method. The experimental method evaluates the authenticity of the modeling results by comparing them with the actual environment, using objective evaluation criteria and including expert review. The interactive verification experiment settings are the same as the above two verification methods. The experimental method uses a questionnaire to evaluate the user’s interactive experience with the modeling results, including the user’s operational feedback and evaluation of the ease of use of the interactive interface. The method proposed in this study has been tested and verified. In order to test the correctness and feasibility of the whole algorithm, a virtual environment with certain regular shapes and virtual scenes was constructed using the geometric model based on the graphics model, and the required real-time performance was achieved.

Figure 5 shows the time-related experimental data of image modeling for different objects and different methods in different environments.

Figure 5 
                  Image modeling time for different objects and different methods: (a) object and (b) environment.
Figure 5

Image modeling time for different objects and different methods: (a) object and (b) environment.

It can be seen from Figure 5a that when facing object A, the modeling time values of the traditional image modeling method and the image modeling method proposed in this article were 12.14 and 8.72 s, respectively. When facing object B, the modeling time values of the two methods were 9.68 and 4.11 s, respectively. When facing object C, the modeling time values of the two methods were 17.75 and 11.92 s, respectively. When facing object D, the modeling time values of the two methods were 18.09 and 12.24 s, respectively. When facing object E, the modeling time values of the two methods were 14.67 and 9.86 s, respectively.

It can be seen from Figure 5b that when facing environment 1, the modeling time of the traditional image modeling method and the image modeling method proposed in this study were 75.16 and 42.20 s, respectively. When facing environment 2, the modeling time values of the two methods were 62.37 and 31.33 s, respectively. When facing environment 3, the modeling time values of the two methods were 89.15 and 46.98 s, respectively. When facing environment 4, the modeling time values of the two methods were 98.43 and 50.17 s, respectively. When facing environment 5, the modeling time values of the two methods were 79.12 and 44.54 s, respectively.

From the data in Figure 5, it can be seen that the modeling time of the image modeling method proposed in this study is better than that of the traditional image modeling method.

Figure 6 shows the experimental data related to the authenticity of image modeling for different objects and different methods in different environments.

Figure 6 
                  The authenticity of image modeling with different methods under different objects and environments: (a) different objects and (b) different environments.
Figure 6

The authenticity of image modeling with different methods under different objects and environments: (a) different objects and (b) different environments.

It can be seen from Figure 6a that when facing object A, the modeling authenticity of the traditional image modeling method and the image modeling method proposed in this study reached 87.51 and 98.57%, respectively. When facing object B, the modeling authenticity of the two methods reached 90.36 and 98.89%, respectively. When facing object C, the modeling authenticity of the two methods reached 82.62 and 96.28%, respectively. When facing object D, the modeling authenticity of the two methods reached 81.44 and 95.91%, respectively. When facing object E, the modeling authenticity of the two methods reached 86.63 and 97.77%, respectively.

It can be seen from Figure 6b that when facing environment 1, the modeling authenticity of the traditional image modeling method and the image modeling method proposed in this study reached 86.42 and 97.25%, respectively. When faced with environment 2, the modeling authenticity of the two methods reached 88.37 and 98.13%, respectively. When faced with environment 3, the modeling authenticity of the two methods reached 78.84 and 95.84%, respectively. When faced with environment 4, the modeling authenticity of the two methods reached 82.23 and 94.66%, respectively. When faced with environment 5, the modeling authenticity of the two methods reached 80.51 and 96.58%, respectively.

From the data in Figure 6, it can be seen that the modeling authenticity of the image modeling method proposed in this study is better than the traditional image modeling method.

Figure 7 shows the experimental data related to the interactivity of image modeling for different objects and different methods in different environments.

Figure 7 
                  Interactivity of image modeling with different methods under different objects and environments: (a) object data and (b) environmental data.
Figure 7

Interactivity of image modeling with different methods under different objects and environments: (a) object data and (b) environmental data.

It can be seen from Figure 7a that when facing object A, the modeling interactivity of the traditional image modeling method and the image modeling method proposed in this study reached 79.14 and 96.45%, respectively. When facing object B, the interactivity of the two methods reached 76.21 and 95.18%, respectively. When facing object C, the interactivity of the two methods was 80.98 and 98.35%, respectively. When facing object D, the interactivity of the two methods was 82.94 and 98.92%, respectively. When facing the object E, the interactivity of the two methods reached 80.35 and 97.12%, respectively.

It can be seen from Figure 7b that when faced with environment 1, the modeling interactivity of the traditional image modeling method and the image modeling method proposed in this study reached 77.61 and 98.06%, respectively. When facing environment 2, the interactivity of the two methods reached 75.15 and 98.61%, respectively. When facing environment 3, the interactivity of the two methods reached 79.57 and 96.32%, respectively. When facing environment 4, the interactivity of the two methods reached 81.38 and 95.68%, respectively. When faced with environment 5, the interactivity of the two methods reached 79.16 and 97.56%, respectively.

From the data in Figure 7, it can be seen that the interactivity of the image modeling method proposed in this study is better than the traditional image modeling method.

In order to better compare the performance differences between the experimental method in this article and the traditional method, object A is now taken as an example, and the comparison is shown in Table 1. It can also be seen from Table 1 that the interactivity of the image modeling method proposed in this article is better than the traditional image modeling method.

Table 1

Performance comparison of different methods

Modeling time (s) Modeling authenticity (%) Modeling interactivity (%)
Traditional image modeling method 12.14 87.51 79.14
Image modeling method in this study 8.72 98.57 96.45

One of the reasons why the proposed method performs better in terms of authenticity is that it adopts a modeling approach based on multi-view photo sequences. By collecting photos from different angles and combining them, one can more fully capture the details and features of a target object or scene. This multi-perspective data collection method can better reflect the appearance and form of objects in the real world, thereby improving the authenticity of modeling. Another reason is that the proposed method takes advantage of AR and VR technologies, which can provide a more immersive experience. By displaying and interacting with real-world objects and scenes in a virtual environment, users can feel the authenticity of the modeling results more intuitively. For example, in environmental design, users can roam in the modeled environment and interact with virtual objects through AR and VR technologies, which can enhance their perception of the authenticity of the modeling results.

One of the reasons why the proposed method performs better in terms of interactivity is that it adopts an image modeling algorithm model for environment design based on AR and VR technologies. This model allows users to directly interact with modeling objects in the virtual environment, allowing users to interact with modeling results more naturally and intuitively. In comparison, traditional image modeling methods may be more static and limited, and users cannot directly interact with the modeled objects as they can with AR and VR technologies. Another reason is that the proposed method incorporates real-time tracking technology of user behavior and location. By tracking a user’s behavior and location in real time and applying this information to virtual environments, a more personalized and interactive experience can be achieved. This gives users greater flexibility in interacting with modeled objects and the freedom to explore the virtual environment according to their needs.

In summary, the reasons why the proposed method performs better in terms of authenticity and interactivity include factors such as the modeling method of multi-view photo sequences, the utilization of AR and VR technologies, and real-time tracking of user behavior and location. These factors work together to enable the proposed approach to provide a more realistic and interactive modeling experience.

3.2 System model verification

Based on the algorithm verification above, this study verified and analyzed the whole image modeling algorithm system model of environment design based on AR and VR technologies. Stability testing evaluates the stability of the system under different conditions by running the system for a long time or processing a large amount of data. The indicators include the system error rate for display. The operation fluency test evaluates the fluency and response speed of the system during operation and uses user experience and system performance indicators to evaluate. The indicators are expressed by the system freeze rate. Performance loss assessment evaluates the performance of the system under different loads, including computing resource utilization, memory usage, power consumption, etc. The indicators are expressed by the performance degradation percentage of the system.

Table 2 shows the data related to the stability test of the system model.

Table 2

Stability test

Modeling quantity Stability (%)
2 99.37
4 98.66
6 98.05
8 97.24
10 96.80

It can be seen from Table 2 that when the number of image modeling in the environment design was 2, the stability of the system was 99.37%. When the number of image modeling in the environment design was 4, the stability of the system was 98.66%. When the number of image modeling in the environment design was 6, the stability of the system was 98.05%. When the number of image modeling in the environment design was 8, the stability of the system was 97.24%. When the number of image modeling in the environment design was 10, the stability of the system was 96.80%.

Table 3 shows the data related to the running smoothness test of the system model.

Table 3

Running smoothness test

Modeling quantity Operation smoothness (%)
2 99.25
4 98.48
6 97.61
8 97.02
10 96.23

It can be seen from Table 3 that when the number of image modeling in the environment design was 2, the system’s running smoothness was 99.25%. When the number of image modeling in the environment design was 4, the smoothness of the system was 98.48%. When the number of image modeling in the environment design was 6, the smoothness of the system was 97.61%. When the number of image modeling in the environment design was 8, the smoothness of the system was 97.02%. When the number of image modeling in the environment design was 10, the smoothness of the system was 96.23%.

Table 4 shows the data related to the performance loss test of the system model.

Table 4

Performance loss test

Modeling quantity Performance loss (%)
2 0.75
4 1.64
6 2.31
8 2.45
10 4.67

It can be seen from Table 4 that when the number of image modeling in the environment design was 2, the performance loss of the system was 0.75%. When the number of image modeling in the environment design was 4, the performance loss of the system was 1.64%. When the number of image modeling in the environment design was 6, the performance loss of the system was 2.31%. When the number of image modeling in the environment design was 8, the performance loss of the system was 2.45%. When the number of image modeling in the environment design was 10, the performance loss of the system was 4.67%.

To sum up, this study tested the image modeling algorithm system model of environment design based on AR and VR technologies from three aspects: stability, running smoothness, and performance loss. The test results show that the system model has excellent performance in image modeling.

For texture-based methods, the following challenges may be encountered.

  1. Accuracy of texture extraction and mapping: Ensure that the texture extracted from the image can be accurately mapped into the reconstructed surface model to maintain the authenticity and accuracy of the model. The strategy is to develop efficient texture extraction and mapping algorithms to increase speed and accuracy and minimize the need for computational resources.

  2. Noise and distortion processing: In the image modeling process, the noise and distortion that may exist in the image must be taken into account to avoid the impact of these factors on texture extraction and mapping. Use image processing techniques to reduce the effects of noise and distortion, such as using filters and denoising algorithms.

  3. Speed and efficiency: Ensure that the texture extraction and mapping process has sufficient speed and efficiency to achieve real-time or near-real-time image modeling and rendering. Leverage hardware acceleration technologies such as modern graphics processing units (GPUs) to increase the speed and efficiency of texture processing.

  4. Algorithm complexity and resource requirements: Ensure that texture-based methods have appropriate algorithm complexity and resource requirements to make them suitable for different types of hardware platforms and application scenarios. Use deep learning and machine learning methods to improve the accuracy and efficiency of texture extraction and mapping, such as using convolutional neural networks (CNN) to learn texture features.

4 Conclusions

In this study, based on AR and VR technologies, image modeling methods in traditional environment design were deeply studied. On this basis, an environment design image modeling algorithm model based on AR and VR technologies was proposed. At the same time, the proposed system was tested according to the actual situation. The results showed that the proposed system had stable performance and smooth operation. Users can interact and operate on the built environment through AR and VR technologies, and 3D simulation environment can be built through AR and VR technologies. Therefore, the research results have high practical value and development prospects. However, there are still deficiencies in many aspects, which need further exploration and improvement. In the future, a texture-based method can be proposed, that is, the texture of the target is extracted from the image and mapped to the reconstructed surface model so as to improve the fidelity of the model and the speed and accuracy and reduce the sensitivity of image modeling to noise, so as to further improve the accuracy of modeling.

  1. Funding information: Authors state no funding involved.

  2. Author contributions: Da Gong: Writing; Tong Chen: Editing; Shuxin Zhou: Data analysis; Jiaheng Han: Resources; Zhenghang Li: Review.

  3. Conflict of interest: The authors state that this article has no conflict of interest.

  4. Data availability statement: Data sharing is not applicable to this article as no new data were created or analyzed in this study.

References

[1] Gharaveis A, Hamilton DK, Pati D. The impact of environmental design on teamwork and communication in healthcare facilities: a systematic literature review. HERD: Health Environ Res Des J. 2018;11(1):119–37.10.1177/1937586717730333Search in Google Scholar PubMed

[2] Bangwal D, Tiwari P. Environmental design and awareness impact on organization image. Eng Constr Archit Manag. 2019;26(1):29–45.10.1108/ECAM-02-2017-0029Search in Google Scholar

[3] de Gaulmyn C, Dupre K. Teaching sustainable design in architecture education: critical review of easy approach for sustainable and environmental design (EASED). Front Archit Res. 2019;8(2):238–60.10.1016/j.foar.2019.03.001Search in Google Scholar

[4] Zadeh RS, Eshelman P, Setla J, Kennedy L, Hon E, Basara A. Environmental design for end-of-life care: An integrative review on improving the quality of life and managing symptoms for patients in institutional settings. J Pain Symptom Manag. 2018;55(3):1018–34.10.1016/j.jpainsymman.2017.09.011Search in Google Scholar PubMed PubMed Central

[5] Yang MG, Roh JJ, Kang M. The role of strategic environmental orientation in environmental design practices. Manag Decis. 2021;59(2):341–57.10.1108/MD-07-2019-0865Search in Google Scholar

[6] Lou Y. The idea of environmental design revisited. Des Issues. 2019;35(1):23–35.10.1162/desi_a_00518Search in Google Scholar

[7] Joseph A, Browning MHEM, Jiang S. Using immersive virtual environments (IVEs) to conduct environmental design research: A primer and decision framework. HERD: Health Environ Res Des J. 2020;13(3):11–25.10.1177/1937586720924787Search in Google Scholar PubMed

[8] Yang Y, Dan Z. Environmental design vs. Environmental art design: a Chinese perspective. J History Cult Art Res. 2020;9(4):122–33.10.7596/taksad.v9i4.2862Search in Google Scholar

[9] Eren ET, Yılmaz S. The student attitudes towards digital and conventional drawing methods in environmental design studios and the impact of these techniques on academic achievement in the course. Int J Technol Des Educ. 2022;32(1):617–44.10.1007/s10798-020-09605-xSearch in Google Scholar

[10] Lawrenc R, Elsayed M, Keime C. Evaluation of environmental design strategies for university buildings. Build Res & Inf. 2019;47(8):883–900.10.1080/09613218.2019.1652551Search in Google Scholar

[11] Abich IV, J, Parker J, Murphy JS, Eudy M. A review of the evidence for training effectiveness with virtual reality technology. Virtual Real. 2021;25(4):919–33.10.1007/s10055-020-00498-8Search in Google Scholar

[12] Bashabsheh AK, Hussain HA, Mostafa ZA. The application of virtual reality technology in architectural pedagogy for building constructions. Alex Eng J. 2019;58(2):713–23.10.1016/j.aej.2019.06.002Search in Google Scholar

[13] Altinpulluk H. Determining the trends of using augmented reality in education between 2006-2016. Educ Inf Technol. 2019;24(2):1089–114.10.1007/s10639-018-9806-3Search in Google Scholar

[14] Huang TK, Yang CH, Hsieh YH, Wang JC, Hung CC. Augmented reality (AR) and virtual reality (VR) applied in dentistry. Kaohsiung J Med Sci. 2018;34(4):243–8.10.1016/j.kjms.2018.01.009Search in Google Scholar PubMed

[15] Farshid M, Paschen J, Eriksson T, Kietzmann J. Go boldly!: Explore augmented reality (AR), virtual reality (VR), and mixed reality (MR) for business. Bus Horiz. 2018;61(5):657–63.10.1016/j.bushor.2018.05.009Search in Google Scholar

[16] Marks B, Thomas J. Adoption of virtual reality technology in higher education: An evaluation of five teaching semesters in a purpose-designed laboratory. Educ Inf Technol. 2022;27(1):1287–305.10.1007/s10639-021-10653-6Search in Google Scholar PubMed PubMed Central

[17] Shafer DM, Carbonara CP, Korpi MF. Factors affecting enjoyment of virtual reality games: a comparison involving consumer-grade virtual reality technology. Games Health J. 2019;8(1):15–23.10.1089/g4h.2017.0190Search in Google Scholar PubMed

[18] Lv Z. Virtual reality in the context of internet of things. Neural Comput Appl. 2020;32(13):9593–602.10.1007/s00521-019-04472-7Search in Google Scholar

[19] Spiegel JS. The ethics of virtual reality technology: Social hazards and public policy recommendations. Sci Eng ethics. 2018;24(5):1537–50.10.1007/s11948-017-9979-ySearch in Google Scholar PubMed

[20] Javaid M, Haleem A. Virtual reality applications toward medical field. Clin Epidemiol Glob Health. 2020;8(2):600–5.10.1016/j.cegh.2019.12.010Search in Google Scholar

[21] Zeng X, Wang Z, Hu Y. Enabling efficient deep convolutional neural network-based sensor fusion for autonomous driving. arXiv preprint arXiv:2202.11231; 2022.10.1145/3489517.3530444Search in Google Scholar

[22] Gao Z, Braud TC. VR-driven museum opportunities: digitized archives in the age of the metaverse. Artnodes. 2023;32:1–14.10.7238/artnodes.v0i32.402462Search in Google Scholar

[23] Xiaofeng. L, Jing W, Hongshuang J. Real-time tracking algorithm for aerial vehicles using improved convolutional neural network and transfer learning. IEEE Trans Intell Transp Syst. 2022;23(3):2296–305.10.1109/TITS.2021.3072872Search in Google Scholar

[24] Macias E. Design and implementation of intelligent fault diagnosis system for construction machinery supporting wireless communication network. Kinetic Mech Eng. 2020;1(3):17–24.10.38007/KME.2020.010303Search in Google Scholar

[25] Risser S. Dynamic analysis of landscape pattern in natural environment protection areas based on GIS. Nat Environ Prot. 2021;2(3):21–30.10.38007/NEP.2021.020303Search in Google Scholar

[26] Wang X, Liu M, Wang Y, Fan J, Meijering E. A 3D Tubular flux model for centerline extraction in neuron volumetric images. IEEE Trans Med Imaging. 2022;41(5):1069–79.10.1109/TMI.2021.3130987Search in Google Scholar PubMed

[27] Bhadoria RS, Chaudhari NS. Pragmatic sensory data semantics with service-oriented computing. J Organ End User Comput (JOEUC). 2019;31(2):22–36.10.4018/JOEUC.2019040102Search in Google Scholar

Received: 2023-12-06
Accepted: 2024-03-25
Published Online: 2024-11-10

© 2024 the author(s), published by De Gruyter

This work is licensed under the Creative Commons Attribution 4.0 International License.

Downloaded on 19.11.2024 from https://www.degruyter.com/document/doi/10.1515/jisys-2023-0296/html
Scroll to top button