Through scientific advancements, humans are now able to access and potentially work in uninhabited and hard-to-reach environments such as space, low Earth orbit, deep oceans, and polar regions with extreme visuospatial and gravitational conditions (Carroll 1993). Working in such unfamiliar conditions may pose significant challenges to human cognitive processing that may be accustomed to Earth’s familiar conditions (Newcombe 2002). In familiar environments on the earth, there are spatial cues offered by natural and manmade familiar objects such as roads, buildings, trees, people, cars, and animals (Ruddle, Volkova et al. 2011). These cues help people comprehend not just the relative position, size, and distance of stationary objects but also the speed of moving objects (Kearney, Gorzel et al. 2012). Mental representations of a space are based on these landmarks. Interacting with the world requires being able to perceive displacement after a visual distraction (Winter, 2014). These landmarks provide spatial cues that help people perceive displacement and create a clear mental representation of the environment (Ruddle, Volkova et al. 2011). These cues help people perceive, for instance, how tall or how far a building is. In addition, people can assess the size, distance, and position of various spatial elements and determine the speed of a moving object (Park et al., 2021). Visuospatial cues, therefore, influence human spatial perception, influencing our ability to carry out day-to-day vocational and recreational activities (Shatil, Korczyn et al. 2008).
While navigating or working in altered visuospatial conditions of low earth orbit, deep space, other planets, polar regions, and deep ocean, these cues may not be available (Oman 2007). For instance, on Moon and Mars, there are no familiar visual landmarks such as roads, cars, trees, buildings, and people, which may make determining relative size, position, and distance a challenging task (Plumert, Kearney et al. 2005). The lack of visuospatial cues, therefore, has a profound impact on how the perception of spatial space can be altered (Park et al., 2021). Worse, in deep space, not only these essential cues but also the base terrain (e.g., ground) is unavailable, which usually exists on other planets (Newman and McNamara 2022). This makes spatial perception even more difficult in deep space. Altered visuospatial conditions also exist on earth in regions such as North and South Poles and in hot and cold deserts, where roads, trees, cars, and buildings may be either limited or not available at all (Ruddle, Volkova et al. 2011). In fact, such settings are often used as analogs for training to operate in extreme conditions of space exploration (Kearney, Gorzel et al. 2012). Such environments with altered spatial conditions on earth and beyond earth may impair human spatial perception, leading to not only low work productivity but also compromised safety that may put the entire mission at risk.
The ability to perceive relative distance, height, and size is one of the critical cognitive abilities that people need to work safely and more productively in a work environment not to mention in an extreme environment like deep space (Kiefer, Giannopoulos et al. 2017). These abilities must be examined for any impact that the absence of visuospatial cues may have (). In altered work conditions, there have been a few studies conducted on the effect of visuospatial cues emanating from landmarks on people’s perceptions of size and distance (Kearney, Gorzel et al. 2012). However, comprehensive studies that assess how limited and no spatial cues impact human ability to perceive distance and size are rare (Fukushima, Fukushima et al. 2013, Ernst, Williams et al. 2017). The main goal of this study is to compare people’s size and distance perception using immersive Virtual Reality (VR) analogs of altered conditions having complete, limited, and no visuospatial cues. It is hypothesized that people would perceive object distance and size more accurately if familiar visual cues were more readily available.
1.2 Background
Over the past decade, emerging technologies have fundamentally changed the way we live, work, and interact, from how we make phone calls to how companies operate. As a result, the nature of work and the work environment are changing fast (Folkierska-Żukowska et al., 2020). The fact that humans are now able to explore the deep oceans, space, and polar regions that humans were never able to before should not be surprising. Such spaces may, however, alter the human brain’s ability to process spatial information (Fukushima, Fukushima et al. 2013, Harris, Jenkin et al. 2017). People’s ability to work productively and their understanding of spatial relationships in new environments can be adversely affected by changing visuospatial conditions in such environments (Turgut, 2015). Those working in such settings may be operating rovers and quadcopters and need to judge distance, sizes, and speeds more accurately. If their spatial perception is distorted, their safety and wellbeing may be at risk (Salehi, 2023).
1.3 Spatial Cognition
An important area of cognitive science is spatial cognition, which concerns understanding how humans perceive, mentally represent, and manipulate space and its characteristics (Newcombe, 2002; Newcombe et al., 2013). In addition to physical attributes like size and shape, spatial characteristics include depth, orientation, and position (Newcombe, 2002). Humans use egocentric and allocentric encodings to perceive these features (Tuena, Mancuso et al. 2021). Objects’ locations are encoded in egocentric systems with references to our body, and in allocentric systems with references to other objects (Guo et al., 2016a; Harris & Mander, 2014; Lipshits et al., 2001). Three types of spatial information contribute to creating a cognitive map of space: visual, vestibular, and somatosensory (Friederici & Levelt, 1987; Harris & Mander, 2014; Reschke et al., 1998). Objects within a scene can provide monocular or binocular spatial cues that are used in defining a visual Frame of Reference (FOR) (Harris & Mander, 2014). We detect motion, the position of the head, and orientation with the vestibular system to maintain our balance (Friederici & Levelt, 1987; Harris & Mander, 2014). In addition to providing spatial information, the somatosensory system is comprised of receptors for touch, heat, pain, vibrations, and pressure (Friederici & Levelt, 1987; Reschke et al., 1998). In addition to creating clear coordinates using spatial information, our cognitive system selects a FOR that can be used to reference objects in space (Ten Brink, Biesbroek et al. 2019). Important directional references such as up, down, left, and right hinge upon a subjective vertical that humans define using gravitational cues, body axis, or a visual frame (Du et al., 2015; Harris et al., 2017; Shepard & Metzler, 1971). Using the sensory information gathered from the visual, vestibular, and somatosensory systems, humans perceive their spatial environments and spatial objects.
1.4 Spatial Perception and Landmarks
Spatial perception relates to the ability and the process of identifying spatial relationships between objects based on their orientation and position in relation to the space around them (Carroll 1993, Brenner and van Damme 1999). In other words, it refers to the ability to perceive, measure, and comprehend outside spatial information such as features, properties, shapes, positions, sizes, and motions visually (Clark, Stimpson et al. 2010). A great deal of research has been conducted on spatial perception in recent years. According to Peretz (2011), spatial perception is a result of individuals’ awareness of their relationship with the environment within and outside their body (exteroception) and within their own minds (interception) (Peretz, Korczyn et al. 2011). We live in a space surrounded by manmade and natural objects, people, elements, etc. Our thinking is also formed by space since it is where all our experiences come together. Our spatial environment is characterized by these two systems of exteroception and interception that give us information about its characteristics (Peretz et al., 2011). Shatil (2008) noted that our understanding of a spatial environment and our relationship to it depends on having good spatial awareness, which involves grasping not only how objects are situated but also how objects change position in space (Shatil et al., 2008). Peretz (2011) added that we can perceive things around us with shapes, sizes, distances, etc., due to this cognitive ability of spatial perception. Spatial perception, therefore, not only helps us mentally represent a space in 2D and 3D but also anticipate any change that may occur (Peretz et al., 2011).
Another term that is often mentioned in spatial cognition literature is “visuospatial perception,” which involves processing and interpreting visual information about the position of objects in space (Höhler, Rasamoel et al. 2021). Visuospatial perception is responsible for daily functions involving a wide range of activities of daily living, which makes it an important aspect of cognitive functioning (Kolb & Whishaw, 2009). Our ability to perceive, mentally represent, manipulate a spatial environment and navigate through it is greatly impacted by visuospatial perception (Henry and Furness 1993). Visual perception also plays a role in the ability to grasp objects in our visual field and the ability to shift our gaze to different places in space (Bartlett & Dorribo Camba, 2023; De Luca et al., 2023; Guo et al., 2016a; He et al., 2019). According to studies, the association areas of the visual cortex are divided into two major component pathways: parieto-occipital regions to process information regarding both visual motion and spatial orientation in humans and inferotemporal region to process visual information about the form and color of objects (Kim et al., 1997; Kolb & Whishaw, 2009).
Mental representations of space are based on landmarks that serve as points of reference (Higgins and Wang 2010). A mental representation of an object has the function of locating other objects within the representation (Higgins and Wang 2010). In this function, a connection is established between place and place-mark, which is yet another geographical concept that is used to structure space and is perhaps even more elusive than landmarks in terms of establishing their boundaries (Winter, 2014). An essential skill in interaction with the world is the ability to perceive the displacement of an object following a visual distraction (Higgins & Wang, 2010; Ruddle et al., 2011). People follow routes between specific places in everyday life as a primary mode of navigation — such as from their homes to their workplaces, to and from coworkers’ offices, and to their own neighborhood shops (Ruddle et al., 2011). Route knowledge, at its simplest, is a set of decisions that lead from where you are to where you should be (Siegel & White, 1975). Aside from this, we can also add metric information (distances and turn angles) as well as information about landmarks perceived primarily visually (Lynch, 1960; Montello, 1998).
According to previous studies (Casey et al., 1995; Kyttälä & Björn, 2014), landmarks, or non-geometric spatial features such as roads, buildings, and trees, contribute to distance perception, size perception, and direction perception. According to Naceri and Hoinville (Fennema, 1974), familiar objects can give us a perspective of linear dimensions and a sense of scale (familiar size), thus allowing us to judge distance more accurately (Newcombe, 2002). According to wayfinding research, landmark is an essential feature of navigation toward a destination (Casey et al., 1995). Spatial cognitive processing can be hindered by the absence of spatial cues emanating from these landmarks. There is a cognitive challenge associated with calculating depth, height, location, and direction in a desert or polar region, in which landmarks provide context (Lunneborg & Lunneborg, 1986). In altered environments, limited and incomplete visual cues might affect spatial perception as they do on Earth by affecting spatial accuracy (Parker 2003, Allred, Kravets et al. 2023). Munich University conducted a study on the physiology and pathology of human navigation control and found that successful navigation depends on effective use of topological knowledge and metric representations of space (Schöberl, Zwergal and Brandt 2020). Moreover, spatial navigation can be strongly affected by the dimensions of space (2D vs. 3D), the scales of space (vista-scale vs. large-scale environment), and the abundance of visual landmarks (Harle & Towns, 2011).
1.5 Spatial Perception Research: Size and Distance Perception
An observer’s ability to judge distances between objects in any direction relative to their eye is called distance perception (Y. Sun et al., 2019). In object or surface perception domain, depth refers to the distance directly ahead of the observer’s eye. Depth is looking directly into a hole or tube and estimating forward distances (Kearney, Gorzel et al. 2012). Taking an accurate measurement of distances requires binocular stereoscopic vision (stereopsis) in addition to other cues, especially when measuring distances over long distances (Treisman 1962). The apparent distance to an object can sometimes be indirectly determined by measuring the size of the object at a certain distance. It has been suggested that people perceive the size of an object based on the retinal size of the object compared with the distance from it (Harle & Towns, 2011; Li & Wang, 2021). A size perception inaccuracy would lead to inaccuracies in distance perception. Several studies have shown a link between distance and size perception (Xinyue He et al., 2021). There is a tendency for observers to overestimate the size of objects that are located at a distance. Our tendency is to attribute a larger size to an object when we overestimate the distance at which it can be seen, and, inversely, if we underestimate the distance at which it can be seen, we tend to attribute a smaller size to the same object (Xinyu He et al., 2021) It is learned through repeated practice that distances and sizes can be perceived. Linear perspective is less relevant in microgravity because there is no gravitational reference and no visual horizon. In addition, astronauts whose eyes are free-floating have a varying distance from the floor; therefore, they cannot use the eye height scaling to determine their height (Xinyu He et al., 2021).
1.6 Virtual Reality and Its Application in Spatial Perception Research
Virtual Reality (VR) provides an intuitive and naturalistic experience that allows users to immerse themselves in real-time environments (Schulteis & Rothbaum, 2002) and interact with them. VR can offer a more efficient and effective way to experience, particularly environments and scenarios that cannot be experienced first-hand (R. Sun et al., 2019, Salehi et al. 2023). VR has been applied in spatial cognition research by several studies (Fujimoto; Tachibana, 2020). Using VR, Guzsvinecz (2020) looked at the time it took students to interact with each other (Guzsvinecz, Orbán-Mihálykó et al. 2020). A study conducted Papakostas (2021) examined the psychological functioning of older adults who play digital games and those who don’t, including well-being, affect, depression, and social functioning. They designed VR mockups for a training game intended to teach scuba techniques to participants. According to many studies, VR can help offer a variety of underwater sensations in a fully immersive manner, and VR settings have also been applied to the domain of space exploration. Over various years, diverse studies have demonstrated that VR serves as a potent tool for simulating, comprehending, and forecasting the work and navigation performance of astronauts in multi-module space stations (W. L. Shebilske et al., 2006, Guo et al., 2016b). In comparison with conventional parabolic flights and drop towers, VR simulations have been recommended for astronaut testing and training (Embry-Riddle, 2017; Nick Kanas, 2003). Astronauts can eliminate several problems associated with spatial orientation and navigation by creating VR-based crew training (Oman, 2007; Wayne L Shebilske et al., 2006).
Studies also suggested that VR technology can help simulate extreme environments with same or similar physiological responses as the real-world (Embry-Riddle, 2017; Jain et al., 2016a; W. L. Shebilske et al., 2006). Several studies were conducted in the mid to late 1990s in which fMRI was combined with stationary desktop VR settings to study the control of human spatial navigation in 2D space (Aguirre, Detre et al. 1996, Maguire, Burgess and O’Keefe 1999, He, Li et al. 2021). Later, advanced VR task designs and the development of fMRI techniques allowed for further differentiation in how the human brain processes visual scenes and cues. Located directly posterior to the parahippocampus, the parahippocampal place area (PPA) is important for recognizing visual scenes and for selecting visual landmarks within those scenes (Tracy, 1987).