• No results found

Investigation of spatial navigation between real and virtual environments among healthy individuals

N/A
N/A
Protected

Academic year: 2021

Share "Investigation of spatial navigation between real and virtual environments among healthy individuals"

Copied!
26
0
0

Bezig met laden.... (Bekijk nu de volledige tekst)

Hele tekst

(1)

                                               

Master Thesis

Master’s Specialization in Clinical Neuropsychology

Faculty of Social and Behavioural Sciences – Leiden University Section: Health, Medical and Neuropsychology Date: 27/03/2017 Student name: Kah Hui Yap Student number: s1852515 Daily Supervisor: Dr. C.J.M. Van der Ham CNP co‐evaluator: Milan van der Kuil

 

Investigation of spatial

navigation between real and

virtual environments among

healthy individuals

 

 

 

 

(2)

Abstract

Spatial navigation is a fundamental cognitive function responsible for survival‐essential goal‐ directed processes. The introduction of virtual reality technology has led to studies on navigation ability, which have shown that spatial navigation process differs between real and virtual environments. Spatial navigation utilizes sensory information such as proprioceptive and visual information, particularly immediately after the encoding phase where one formulates a plan to navigate an environment. These types of information supports spatial navigation differently according to the type of environment. Spatial dimensions of virtual environments are found to be consistently underestimated compared to the real world. Therefore, it is hypothesised that larger metric errors will be made in a virtual environment compared to the real world when visual input is provided during the encoding phase. Angular deviation and distance error were assessed in two navigation tasks: triangle completion task and distance reproduction task. This study used a 3 x 2 within‐subjects design between visual input condition (encoding only, responding only, and encoding + responding) and the type of environment (real world vs virtual environment). Results revealed that the hypothesis is partially supported; while the angular deviation was greater in the virtual environment than in the real world, no difference was found across visual input conditions. Distance error was greater in the virtual environment during the triangle completion task but not in distance reproduction task. The findings are discussed in relation to the difficulty level of the task (i.e. angle estimation is harder than distance estimation) and cognitive load theory. Avenues for further research are also suggested in relation to examining the effect of different triangle sizes and subjects’ subjective experiences across different conditions of the tasks.

(3)

Contents

 

Abstract ... 2

1. Introduction ... 4

2. Methods ... 7

2.1

Participants ... 7

2.2

Materials ... 8

2.2.1 HTC Vive... 8

2.2.2 Spatial Tasks ... 9

2.3

Procedure ... 11

2.4

Measure ... 12

2.5

Study Design ... 13

2.6

Statistical Analysis ... 13

3. Results ... 13

3.1 Angular Deviation... 14

3.2 Distance error – TCT ... 15

3.3 Distance error – DRT... 17

4. Discussion ... 19

References ... 24

     

(4)

1. Introduction

Spatial navigation refers to the cognitive function of maintaining a sense of direction and location while moving in an environment (Wolbers & Hegarty, 2010). It is a fundamental cognitive function responsible for survival‐essential processes; including food and shelter‐seeking behaviours as evidenced in a wide range of animals (Milford & Schulz, 2014). Spatial navigation ability has evolved in humans to accommodate more complex goal‐directed processes such as formulating plans to navigate complex environment and planning routes to distant locations, through utilization of the information about the environment. This ability is subjected to both individual differences and pathological processes (Wolbers & Hegarty, 2010). With the introduction of virtual reality (VR) technology, a computer technology that generates realistic sensations to replicate real environment (RE) (Linkenauger, Mohler, & Proffitt, 2011), it is suggested that VR is a useful tool to test navigation ability, particularly in assessing individual differences and pathological processes (Barrett & Craver‐Lemley, 2008; Tsirlin, Dupierrix, Chokron, Coquillart, & Ohlmann, 2009). However, it has also been shown that spatial perception differs between RE and virtual environments (VE) (Linkenauger et al., 2011). Therefore, the main goal of this study is to compare spatial navigation between RE and VE in healthy population.

While multiple brain regions are implicated, it is suggested that the hippocampus is largely responsible for spatial navigation ability (Chersi & Burgess, 2015). In turn, this ability is largely attributed to place cells, a group of pyramidal neurons in the hippocampus, with each group corresponds to a specific place field and work collectively to form the cognitive map (Muir & Bilkey, 2001), a mental representation of spatial environment; of which spatial navigation is largely dependent upon (Tolman, 1948). Place cells activate upon navigating through a specific part of the environment, the correspondent place field (Alme et al., 2014). The place cells are associated with the activity of head direction cells, which is responsible for egocentric map of locations (Taube, 2007). These hippocampal structures are subjected to pathological changes. It is further suggested that the degree of spatial navigation deficit is associated with the volumetric loss in the right hippocampus (Nedelska et al., 2012). Consistently, pathological processes that implicate the hippocampus often result in spatial navigation deficits among a wide range of cognitive impairment (Chersi & Burgess, 2015; Wolbers & Hegarty, 2010). Spatial navigation deficits result in disorientation, problems with goal‐oriented navigation, and spatial information‐dependent tasks such as recognizing salient landmarks (Hebert & Dash, 2004). These deficits can be attributed to the inabilities to access the cognitive map and to judge an object’s location relative to self (Stark, 1996). In addition, compromised place cells may also result in hemispatial neglect, the inability of a person to process and perceive one side of the body

(5)

and environment that is contralateral to the damaged brain hemisphere (Nico et al., 2008). Pathologies that are accompanied by spatial navigation deficits with hippocampal origins include Alzheimer’s disease (AD), anxiety, and stroke (Bannerman et al., 2014; Laakso, Lehtovirta, Partanen, Riekkinen, & Soininen, 2000; Li et al., 2009).

Several treatment modalities have been derived in order to improve spatial navigation ability in pathological conditions, including medications and physical rehabilitation (Alomari, Khabour, Alzoubi, & Alzubi, 2013; Li et al., 2009; Luo et al., 2007; Vasconcellos, Tabajara, Ferrari, Rocha, & Dalmaz, 2003; Wang et al., 2007). Despite excellent treatment outcome offered by these approaches, patients are often unmotivated and refuse rehabilitation. This might be attributed to the personality changes involved in the pathological processes (Maclean, Pound, Wolfe, & Rudd, 2000). In addition, medication might not be suitable for many patients due to contraindications and adverse effects (Alldredge et al., 2013). Therefore, an alternative approach in addressing these issues from the traditional rehabilitation settings is required. Due to the role of VE in influencing spatial perception, the use of VR technology as a potential alternative in the rehabilitation setting is growing (Henderson, Korner‐Bitensky, & Levin, 2007). At this juncture, it is necessary to look into the VR technology and its applications.

Major and rapid advancement in the development of VR technology in the twenty‐first century, most notably represented by the invention of head‐mounted displays (HMDs), has allowed its applications in various settings. HMDs are head‐mounted goggles with a projection screen in front of the eyes, complemented by sensory information through sound and haptic systems to simulate the RE that mimic physical interactions (Crison et al., 2005). In recent years, VR technology has been implemented in rehabilitation for medical conditions, especially in post‐ stroke rehabilitation (Henderson et al., 2007; Laver, George, Thomas, Deutsch, & Crotty, 2012; Saposnik et al., 2010). The effect of VR was particularly pronounced in terms of the patients’ motivation and effectiveness of the rehabilitation (Henderson et al., 2007). However, the mechanisms underlying these effects were not well‐understood. Furthermore, the vast majority of existing studies have been focusing on movement and physical rehabilitation (Henderson et al., 2007; Laver et al., 2012; Saposnik et al., 2010); studies utilizing VR on cognitive rehabilitation are relatively scarce. The potential difference in spatial perception between RE and VE need to be addressed in order to fully grasp how VR can be used and what generalizations of findings in VR are empirically supported. It is crucial to examine how VE are comparable to RE, in particular with regards to spatial navigation performance. The difference between RE and VE generally concerns spatial navigation (Ziemer, Plumert, Cremer, & Kearney, 2009). As mentioned earlier, VE generated by VR allows users’ interaction with the isolated virtual space. By simulating the user’s presence in the VE, the users are isolated

(6)

from the real world while immersed in a world that is completely artificial (Bricken, 1991). This is contrasted in augmented reality, where users are able to interact with virtual and real world contents, and are able to distinguish between the two (Tiantai & Yintong, 2003). With the complete isolation from the real world, it has been suggested that spatial processing differs in VE; thus necessitating discussion on its underlying mechanism.

Studies utilizing VR have found spatial dimensions of VE are found to be consistently underestimated compared to RE (Arthur, Hancock, & Chrysler, 1997; Sahm, Creem‐Regehr, Thompson, & Willemsen, 2005). Distances appeared to be more compressed in VE particularly through the use of HMDs (Ziemer et al., 2009), while a different degree of compression was found in different settings (Messing & Durgin, 2005; Richardson & Waller, 2007; Sahm et al., 2005). It is suggested that the compression effect is attributed to the alteration of the visual body parts perception, which in turn influences distance estimated in VE (Linkenauger et al., 2011). These findings suggest that spatial processing differs between RE and VE; as well as the potential of utilizing VR in rehabilitation through manipulation of the compression effect. To better understand these processes, the spatial tasks used in this study are the triangle‐completion task (TCT) and the distance reproduction task (DRT), which assess core aspects of navigation ability.

TCT is among the spatial navigation tasks that were used in conjunction with VR (Adamo, Briceño, Sindone, Alexander, & Moffat, 2012; Riecke, Van Veen, & Bülthoff, 2002). In TCT, the subjects are led along two sides of a triangle and have to find the shortest way back to the starting position by themselves (Klatzky, Loomis, & Golledge, 1997). Errors made i.e. deviation in angle or distance diverted from the supposed endpoint or both, can be considered a measure of lower spatial navigation ability (Adamo et al., 2012; Garcia Popov, Paquet, & Lajoie, 2013). DRT involves similar processes, where subjects are guided through a linear displacement and are required to rotate and travel towards the starting point without guidance (Adamo et al., 2012). At this juncture, it is necessary to discuss the underlying mechanism of these measures and their implication on spatial navigation.

Two cognitive processes are involved in TCT and DRT: 1) the encoding phase whereby the cognitive map is formed; and 2) the responding phase whereby the planned trajectory is performed (Fujita, Klatzky, Loomis, & Golledge, 1993). Larger angle and distance deviated from the designated location indicates an altered spatial navigation process (Adamo et al., 2012; Riecke et al., 2002). These measures require subjects to actively navigate the environment, suggesting the importance of proprioception, a primary sensory system that provides information about awareness of body and limb position (Goodwin, McCloskey, & Matthews, 1972), in spatial navigation.

(7)

Proprioception is crucial in mediating one’s conscious perception of movement as well as encoding body movements in the space. The proprioceptive information includes joint position, direction of movement, as well as floor texture, facilitates spatial navigation process (Blasier, Carpenter, & Huston, 1994). In the absence of proprioception, e.g. in an endless hallway without tactile input, visual information dominates the spatial navigation process (Kearns, Warren, Duchon, & Tarr, 2002). Optic flow allows the processing of the flow of visual information from movement of the observer (Rankin, Mucke, Miller, & Gorno‐Tempini, 2007). Visual information, in conjunction with proprioceptive information, is particularly crucial in the encoding phase of TCT and DRT (Fujita et al., 1993). In addition, these senses play a crucial role in distance compression due to the alteration of visual body parts as discussed previously (Linkenauger et al., 2011), thereby altering the optic flow. In summary, spatial navigation utilizes sensory information such as proprioceptive and visual information, particularly immediately after the encoding phase where one formulates plan to navigate an environment (Fujita et al., 1993). It is expected that these types of information support spatial navigation differently according to the type of environment. This might be due to the compression effect in VE which is attributed to the alteration of the visual body parts perception (Linkenauger et al., 2011). The present study aims to compare the spatial navigation process between RE and VE in healthy population. An understanding of spatial navigation performances of healthy participants during TCT and DRT in both RE and VE may provide crucial input for VR‐based intervention for the patients. It is hypothesised that: 1) Greater angle deviation will be made during TCT in VE compared to RE when visual input is provided during the encoding phase; 2) Greater distance error will be made during TCT and DRT in VE compared to RE when visual input is provided during the encoding phase. The expected results of this study will provide crucial input for VR‐ based intervention, including extension of boundaries in physical therapy as well as to motivate the patients under rehabilitation by perceiving the distance as shorter or an easier task.

2. Methods

2.1 Participants

A total of 24 participants aged between 18 and 30 year old were recruited through various channels: Leiden University Research Participation website, printed advertisements, social media, as well as friends and family method. Inclusion criteria include the mastery of English language

(8)

and having normal or corrected‐to‐normal vision. Exclusion criteria include visual and auditory impairment (Schinazi, Thrash, & Chebat, 2016), history of psychiatric diagnoses (Gould et al., 2007), and any other medical diagnoses affecting cognitive functioning, including stroke (Henderson et al., 2007) and history of significance trauma (Livingstone & Skelton, 2007), as well as motion sickness (So, Lo, & Ho, 2001). Each of these conditions, either individually or collectively, may affect the results of the study directly (i.e. inability to perform spatial navigation fully) or indirectly (i.e. inability to understand instruction); as well as ethical concerns in placing excessive burdens on the individuals (Iphofen, 2016). The demographic data of participants were summarized in Table 1. The study was approved by the Ethics Committee (Commissie Ethiek Psychologie) of the Faculty of Social and Behavioural Sciences of the Leiden University. Informed consent was obtained from all participants. Table 1. Demographic data of the participants. Gender (Male / Female), frequency 8 / 16 Age in years, mean (standard deviation) 22.13 (2.56) Handedness (Right / Left), frequency 21 / 3 Highest education level, frequency Prior completion of bachelor degree Bachelor degree Master degree 13 10 1 Experience with VR (Yes / No), frequency 7 / 17

2.2 Materials

2.2.1 HTC Vive

HTC Vive (HTC and Valve Co., 2016), a VR HMD, was used in projecting VE in this study. The HTC Vive was obtained from Triple through collaboration, where the context of the VE was designed. This device has a screen in which a tracking system is mounted. This tracking technology allows the user to turn the head up to 360°. A PC capable of providing refresh rate of 90 Hz is used to minimize the chance of triggering VR sickness, a VR‐induced motion sickness, including nausea, headache, and disorientation (Groen & Bos, 2008).

(9)

2.2.2 Spatial Tasks

TCT and DRT were used to assess the spatial ability of participants in both RE and VE. In the RE setting, participants performed the TCT in a lab setting with the following measurements: 3.5 m (width) x 4 m (length) x 2.7 m (height). In the VR setting, a VE was calibrated with the same measurements to match the borders of the RE. Figure 1 illustrates the TCT experimental set‐up in which the participants were guided on a 1.8 x 2.5 m right triangles before making the rotation themselves to complete the triangle. The length between starting point (S) and point 1, as well as between point 1 and point 2 were alternated between 1.8 or 2.5 m; and different starting points were used to minimize practice effect.

(a) Right turn (R1) (b) Right turn (R2)

(Start from S with 1.8 m followed by 2.5 m) (Start from S with 2.5 m followed by 1.8 m) 2.5 m 2.5 m 1.8 m 1.8 m 3.08 m 3.08 m

(10)

(c) Left turn (L1) (d) Left turn (L2) (Start from S with 1.8 m followed by 2.5 m) (Start from S with 2.5 m followed by 1.8 m) Figure 1. Triangle completion task combinations. (a) Right turn (R1). (b) Right turn (R2). (c) Left turn (L1). (c) Left turn (L2). Two metrics, angular deviation and distance error, were measured in TCT (Figure 2). The angular deviation was determined as the angle between the ideal trajectory and the straight line to the final foot position. The angular deviation was measured with a protractor. The side of deviation, of either left or right from the ideal trajectory, was ignored; and only the absolute value was taken into account. The distance travelled between point 2 and the End Target (i.e. back to the starting point S) was measured with a measuring tape. This ideal distance is subtracted by the linear distance travelled by the participants; only the absolute value was taken into account. 2.5 m 2.5 m 1.8 m 1.8 m 3.08 m 3.08 m

(11)

Figure 2. Example of TRT with measures or linear distance travelled and angular deviation illustrated. In DRT, participant were guided through a linear displacement of 2.5 m. In contrast to TCT, only one condition was used in DRT as no turning is required. In addition, only the distance error was measured in DRT. Similar to TCT, this distance error was measured with a measuring tape. This ideal distance is subtracted by the linear distance travelled by the participants; only the absolute value is taken into account.

2.3 Procedure

Participants were scheduled for a specific time slot through online registration on the university research participation website. The participants were required to read the instructions of the experiment prior signing the informed consent, as well as to complete a demographic form. This is followed by the assessment of inclusion and exclusion criteria. Prior to the experimental process, the participants were asked to remove their watches and to turn off their mobile phones.

Participants then undergone practice session of TCT and DRT in RE, as well as to familiarize themselves by putting on the HTC Vive and the VE. This is followed by the actual session where participants gone through TCT on both RE and VE condition. No feedback was provided during each task of the experiment. Visual input was provided at various time points according to the setting: 1) visual inputs were provided during both encoding and responding phases (Encoding + Responding); 2) participants were provided visual input during encoding

(12)

phase and blindfolded during the responding phase (Encoding); 3) participants were blindfolded during the encoding phase and provided visual input during the responding phase (Responding). A baseline condition without visual input was conducted in the RE prior other 3 visual input conditions: participants were blindfolded during both encoding and responding phases (Baseline). To avoid serial order carryover effects, participants were assigned to one of the 6 different combinations of visual input sequences (e.g. 1,2,3; 2,3,1).

A total of 4 trials were conducted in each conditions of the tasks to ensure reliability of the results; 2 trials of left and right turning each in TCT (Figure 3). Similar to the visual input conditions, participants were assigned to one of the 4 different combinations of TCT to avoid serial order carryover (e.g. R1,R2,L1,L2; L1,L2,R1,R2; R1,L1,R2,L2; R2,L2,R1,L1). Together with 6 different combinations of visual input sequences, each of the 24 participants were assigned a unique combinations of visual input sequences and TCT trial sequences. VE were presented last because of concerns that motion sickness or spatial bias induced by VE navigation may affect performance on RE (Adamo et al., 2012; Kearns et al., 2002).

The same setting is applied to DRT; 4 trials per visual input condition were required however there is no specific side of turning is required in this task. Upon completion, participants were debriefed regarding the actual aims of the present study and were given the opportunity to ask questions. Agreement was made on prohibition of revealing information of the study to the third parties to avoid potential impact on the data collection and subsequent results. All instructions in the experiment, both written and verbal, were demonstrated in English. The participants were tested individually in a small, quiet laboratory in the Department of Psychology at the Leiden University.

2.4 Measure

Firstly, the results of each participants (i.e. both angular deviation and distance error) from all 4 trials of TCT and DRT in each environments were averaged. Raw scores were converted into Z‐ scores to identify outliers from the TCT and DRT results. Data values with Z‐score of less than ‐3 or greater than +3 were considered as outliers. A final data set of raw scores without outliers was then obtained. These scores were converted into standardized scores: 1) angular deviation; 2) distance error‐TCT; 3) distance error‐DRT.

(13)

2.5 Study Design

The study used a 3 x 2 within‐subjects design. The within‐subject factors include type of environment (RE / VE) and the performances (angular deviation, distance error‐TCT, and distance error‐DRT). The between‐subject factors include the presence or absence of visual input (Encoding + Responding / Encoding only / Responding only) as well as a baseline without visual input.

2.6 Statistical Analysis

Software package SPSS was used to perform General Linear Model – Repeated Measures (GLM‐ RM) with two‐tailed alpha level of 0.05 to compare the performances. The performances were indicated by mean angular deviation (°) and mean distance error for both TCT and DRT (cm), across the type of environment and visual input conditions.

3. Results

Table 2 presents the average behavioural data of all participations on angular deviation, distance error‐TCT, and distance error‐DRT in each visual input condition across RE and VE. Table 2. Mean and standard error (SE) of average performance error in each visual input condition across real environment (RE) and virtual environment (VE). Angular deviation (°) Distance error ‐ TCT (cm) Distance error ‐ DRT (cm) Mean (SE)

RE VE RE VE RE VE

Baseline 16.08 (1.55) 51.30 (6.31) 29.11 (3.08) Encoding only 9.56 (0.66) 13.74 (1.17) 34.65 (4.79) 46.26 (4.44) 18.85 (2.59) 17.84 (0.18) Responding only 10.34 (0.88) 11.07 (0.99) 32.10 (4.58) 34.57 (3.55) 15.71 (2.24) 15.03 (1.44) Encoding + Responding 8.79 (0.40) 11.22 (1.20) 21.78 (2.10) 32.43 (3.28) 15.06 (2.29) 18.54 (2.74)

(14)

3.1 Angular Deviation

First of all, GLM‐RM was used to compare the angular deviation between RE and VE at each visual input condition. The interaction effect between the type of environment and visual input was at trend level, F (2, 21) = 4.40, p = .07, partial ƞ2 = .08. Figure 3 shows the mean angular deviation for each condition. As a follow up analysis, the main effect of the type of environment was analysed in terms of the difference in angular deviation between RE and VE at each visual input condition. Hedges’ g was used as an effect size measure. During the Encoding + Responding condition, angular deviation was significantly greater in VE compared to RE with small effect size (t (23) = 2.13, p < .05, g = .43). A similar effect was found during the Encoding only condition with large effect size (t (23) = 4.21, p < .001, g = .86). In contrast, angular deviation did not differ between RE and VE in Responding only condition (t (23) = .77, p = .45, g = .53).

Figure 3. Bar chart between type of environment and visual input condition (Angular deviation). Error bars represent standard error of the mean (SEM).

The main effect of visual input was analysed in terms of the difference in angular deviation across visual input conditions for each environment. In RE, the difference in angular deviation was not significant when comparing among the three visual input conditions: 1) Encoding +

(15)

Responding vs. Encoding only: (t (23) = 1.00, p = .33, g = .20); 2) Encoding + Responding vs. Responding only: (t (23) = 1.89, p = .07, g = .38); 3) Encoding only vs Responding only: (t (23) = .83, p = .42, g = .17). Angular deviation at each visual input condition in RE was compared to the baseline. Angular deviation is significantly greater in baseline as compared to all visual input condition with large effect size: 1) Encoding + Responding (t (23) = 4.81, p < .001, g = .98); 2) Encoding only (t (23) = 4.87, p < .001, g = .99); 3) Responding only (t (23) = 4.33, p < .001, g = .88). Similarly in VE, no difference in angular deviation was found across visual input conditions: 1) Encoding + Responding vs Encoding only: (t (23) = 1.97, p = .06, g = .40); 2) Encoding + Responding vs. Responding only: (t (23) = .17, p = .86, g = .03); 3) Encoding only vs. Responding only: (t (23) = 1.81, p = .08, g = .17). Angular deviation at each visual input condition in VE was compared to the baseline. Angular deviation at baseline is significantly greater as compared to Encoding + Responding condition (t (23) = 3.06, p < .05, g = .62) and Responding only condition (t (23) = 2.87, p < .05, g = .59), with medium effect size. No difference was found between baseline and Encoding only condition (t (23) = 1.58, p = .13, g = .32).

3.2 Distance error – TCT

GLM‐RM revealed that no interaction between the type of environment and visual input condition in distance error – TCT (F (2, 21) = 2.44, p = .26, partial ƞ2 = .04; Figure 4). Hypotheses justified the examination of main effects of the type of environment and visual input conditions on distance error ‐ TCT.

(16)

Figure 4. Bar Chart between type of environment and visual input condition (Distance error – TCT). Error bars represent SEM.

The main effect of the type of environment was analysed in terms of the difference in distance error ‐ TCT between RE and VE at each visual input condition. During the Encoding + Responding condition, distance error ‐ TCT is significantly greater in VE compared to RE with medium effect size (t (23) = 3.06, p < .05, g = .62). Similar finding was found in Encoding only condition (t (23) = 2.56, p < .05, g = .52). In contrast, distance error ‐ TCT did not differ between RE and VE in Responding only condition (t (23) = .52, p = .61, g = .11). The main effect of visual input was analysed in terms of the difference in distance error ‐ TCT across visual input conditions for each environment. In RE, distance error ‐ TCT was significantly greater in Encoding only than Encoding + Responding (t (23) = 3.51, p < .05, g = .71); as well as when Responding only condition was compared to Encoding + Responding condition (t (23) = 2.76, p < .05, g = .56). Hedges’ g revealed that these differences have medium effect size. In contrast, distance error – TCT did not differ between Encoding only condition and Responding only condition (t (23) = .54, p = .59, g = .11). Distance error ‐ TCT at each visual input condition in RE was compared to the baseline. Distance error ‐ TCT is significantly greater at baseline, as compared to all visual input condition: 1) Encoding + Responding (t (23) = 5.07, p < .001, g = 1.04);

(17)

2) Encoding only (t (23) = 3.29, p < .05, g = .67); 3) Responding only (t (23) = 3.60, p < .05, g = .74). The effect sizes of these differences range from medium to large. In VE, the main effect of visual input revealed that distance error – TCT was significantly greater during the Encoding only condition compared to the other two visual input conditions: 1) Encoding + Responding (t (23) = 3.53, p < .05, g = .72); 2) Responding only (t (23) = 3.37, p < .05, g = .69). The effect sizes for these differences are medium. In contrast, distance error – TCT did not differ between Encoding + Responding and Responding only condition (t (23) = .74, p = .47, g = .15). Distance error ‐ TCT at each visual input condition in VE was compared to the baseline. Distance error ‐ TCT at baseline is significantly greater as compared to when visual input is provided at Encoding + Responding (t (23) = 3.10, p < .05, g = .63) and Responding only (t (23) = 2.59, p < .05, g = .53); both of which have shown medium effect size. No difference was found between baseline and Encoding only condition (t (23) = .86, p = .40, g = .17).

3.3 Distance error – DRT

GLM‐RM revealed that no interaction was found between the type of environment and visual input condition in distance error – DRT (F (2, 21) = 1.12, p = .33, partial ƞ2 = .03; Figure 5). Hypotheses justified the examination of main effects of the type of environment and visual input conditions on distance error ‐ DRT.

(18)

Figure 5. Bar Chart between type of environment and visual input condition (Distance error – DRT). Error bars represent standard error of the mean (SEM). The main effect of the type of environment was analysed in terms of the difference in distance error ‐ DRT between RE and VE at each visual input condition. Distance error – DRT did not differ between RE and VE in each visual input condition: Encoding + Responding (t (23) = 1.34, p = .19, g = .27); Encoding (t (23) = .42, p = .68, g = .09); and Responding (t (23) = .33, p = .75, g = .07). The main effect of visual input was analysed in terms of the difference in distance error ‐ DRT across visual input conditions for each environment. In RE, no significant difference in distance error ‐ DRT was found in the following visual input conditions: 1) Encoding + Responding vs. Responding only (t (23) = 1.19, p = .24, g = .08); 2) Encoding vs Responding (t (23) = .37, p = .71, g = .24). Difference between Encoding + Responding and Encoding only condition was at trend level (t (23) = 1.77, p = .09, g = .36). Distance error ‐ DRT at each visual input condition in RE was compared to the baseline. Distance error ‐ DRT is significantly greater at baseline as compared to all visual input conditions with large effect size: 1) Encoding + Responding (t (23) = 4.55, p < .001, g = .93); Encoding (t (23) = 3.97, p = .001, g = .81); Responding (t (23) = 3.69, p = .001, g = .75).

Similarly in VE, no difference in distance error ‐ DRT was found across visual input conditions: 1) Encoding + Responding vs. Encoding (t (23) = .26, p = .80, g = .05); 2) Encoding + Responding vs. Responding (t (23) = 1.23, p = .21, g = .26); Encoding vs. Responding (t (23) = 1.29, Baseline 0 5 10 15 20 25 30 35

Encoding only Responding only Encoding +

Responding Distance error DRT (cm) Visual Input Condition RE VE Baseline

(19)

p = .21, g = .26). Distance error ‐ DRT at each visual input condition in VE was compared to the baseline. Distance error ‐ DRT at baseline is significantly greater as compared to all three visual input conditions: Encoding + Responding (t (23) = 3.15, p < .05, g = .64); Encoding (t (23) = 3.62, p = .001, g = .77); Responding (t (23) = 4.06, p < .001, g = .98). The effect sizes of these differences range from medium to large.

4. Discussion

The main purpose of this study was to compare the spatial navigation process in a real world and virtual environment with varying context of visual input in healthy population. As studies utilizing virtual reality in cognitive rehabilitation are relatively scarce, this study was therefore conducted to examine how virtual reality can be used and what generalizations of findings can be implemented into virtual reality‐based intervention. In this study, the ability of healthy subjects to estimate angle and distance travelled accurately were used to assess the differences in spatial navigation abilities in a real world and virtual environment; as well as under different visual input condition. The first hypothesis was that, when visual input is provided during the encoding phase, angle estimation will be less accurate during triangle completion task in a virtual environment compared to the real world. The results revealed that angle estimation was less accurate in a virtual environment when visual input is provided during both encoding and responding phase, as well as solely during the encoding phase, complemented by medium and large effect size respectively. This is consistent to the findings that spatial dimensions of virtual environments are found to be consistently underestimated compared to the real world (Arthur et al., 1997; Sahm et al., 2005; Ziemer et al., 2009). The difference in the effect size suggested that the difference between virtual environment and real world was more pronounced when visual input is provided solely during the encoding phase. In contrast, no difference was found when visual input is provided solely during the responding phase, suggesting that the effect of visual input at responding phase was not sufficient to enhance the effect of virtual environment on spatial navigation as compared to the encoding phase. These results supported the findings that visual information, in conjunction with proprioceptive information, is particularly crucial in the encoding phase (Fujita et al., 1993).

Conversely, there was no difference in the ability to estimate angle across visual input condition in both virtual environment and real world. These results suggested that visual input provided at either time point did not affect the ability in estimating angle in the real world. It is necessary to note that angle estimation was least accurate at full‐blinded baseline, compared to

(20)

all three visual input conditions, and were complemented by large effect size. This result supported the findings that visual input is crucial in the formation of cognitive map and is subsequently utilized in the spatial navigation process (Linkenauger et al., 2011). On the other hand, angle estimation in a virtual environment was less accurate when visual input is provided solely during the encoding phase, though with a relatively small difference. In addition, no difference in angle estimation was found between baseline and when visual input is provided solely during the encoding phase. These findings indicate that despite being small in magnitude, some meaningful differences may be present when visual input was provided solely during the encoding phase as compared to the other two visual input conditions in the virtual environment. A two‐stage process is proposed to explain the findings in the virtual environment when visual input is provided solely during the encoding phase: 1) visual input provided during the encoding phase in the virtual environment altered the cognitive map formed compared to that of the real world (Linkenauger et al., 2011); 2) Lack of visual cues during the responding phase forced the subjects to rely solely on the altered cognitive map. In contrast, by providing visual input during both encoding + responding phases in the virtual environment, subjects are able to rely on the visual cues despite being altered. On the other hand, visual cues were not used in the formation of cognitive map when visual input is provided solely during the responding phase. Subjects may instead rely solely on the proprioceptive information (i.e. physical displacement) (Klatzky et al., 1997), without relying on visual cue provided during the responding phase. Results from the real world condition further suggested that visual input enhanced the alteration caused by the virtual environment; little differences were observed across visual input conditions in the real world. Therefore, visual input provided solely during the encoding phase in a virtual environment had greatest influence on spatial navigation process through the alternation for cognitive map formation.

Lastly, the results revealed that the difference between virtual environment and the real world and different visual input conditions influenced the accuracy of angle estimation independently. This suggestion is further supported by the medium effect size, therefore, a larger sample size is needed to confirm whether these variables affect the angle estimation independently or collectively. It is necessary to note that the first hypothesis is partially supported by the results; while the accuracy of angle estimation differs between virtual environment and the real world, no difference was found across visual input conditions.

The second hypothesis expected that, when visual input is provided during the encoding phase, distance estimation will be less accurate during triangle completion task in a virtual environment compared to real world. Accuracy in estimating distance from both triangle completion and distance reproduction tasks will be compared and contrasted in addressing the

(21)

second hypothesis. Similar to angle estimation, the results revealed that during the triangle completion task, distance estimation was less accurate in a virtual environment when visual input is provided during both encoding and responding phases, and solely during the encoding phase. The effect size illustrating the difference between the real world and the virtual environment did not differ between the two visual input conditions. It is interesting to note that this trend is similar to that of angle estimation, which was also assessed in the triangle completion task; suggesting possibility of suggest similar mechanism underlying the results. Distance estimation during the triangle completion task was less accurate when visual input is provided at either one of the time points compared to when visual input is provided during both encoding and responding phases in the real world. This result is contrasted with angle estimation in the real world despite being assessed under the same task. This suggested that in the real world condition, accuracy in distance estimation was more prone to the lack of visual input compared to angle estimation. On the other hand, the results from the virtual environment revealed that distance estimation was less accurate when visual input is provided solely during the encoding phase as compared to the other two visual input conditions. No difference in distance estimation was found when comparing baseline with visual input provided solely during the encoding phase. This again emphasized the importance of sensory inputs during the encoding phase of cognitive map (Fujita et al., 1993). As the results from both angle and distance estimations demonstrated similar trend, discussion of these results from triangle completion task as a whole is thus necessary.

Similar to angle estimation that was addressed in the first hypothesis, the difference between virtual environment and the real world and different visual input conditions influenced the distance estimation independently during the triangle completion task. However, the results in angle estimation suggested the possibility of different results with larger sample size. This in turn suggests that the effect of different visual input conditions in the virtual environment on distance estimation is not as pronounced as demonstrated in angle estimation. This is possibly in relation to the difficulty level of the task (i.e. angle estimation is harder than distance estimation). At this juncture, it is necessary to discuss the results of distance estimation in distance reproduction task, as well as the similarities and differences between triangle completion task and distance reproduction task. In contrast to the distance estimation in triangle completion task, distance estimation in distance reproduction task did not differ between virtual environment and the real world; as well as between different visual input conditions. These results contradicted the findings that spatial dimensions of virtual environments are found to be consistently underestimated compared to the real world (Arthur et al., 1997; Sahm et al., 2005; Ziemer et al., 2009). The differences in results

(22)

between the tasks could be attributed to the difference in difficulty level of the tasks; triangle completion task involves formulating cognitive map of both angle and distance, while distance reproduction task involves formulating of cognitive map of only distance estimation (Adamo et al., 2012; Garcia Popov et al., 2013). Alternative explanation proposed that the results observed in distance estimation during the triangle completion task was at least partially contributed by angle estimation regardless of the visual input condition. This is in line with the cognitive load theory that greater cognitive load leads to reduced performance (Sweller, 1994). During the triangle completion task, subjects need to uphold the information for both angle and distance in working memory, in contrast to the distance reproduction task in which subjects are only required to uphold the information for distance. To investigate whether this theory has a role in the difference between the tasks, distance estimation during the triangle completion task should be analysed by statistically controlling the angle estimation. The theory can be supported if the results of distance estimation from both tasks demonstrate similar trend. This also suggests that utilizing triangle completion task, as compared to distance reproduction task, in conjunction with virtual environment + visual input provided solely during the encoding phase presents greater potential in studies utilizing virtual reality in spatial navigation behaviour and clinical implementation. Several strengths have provided this study with added advantages over the past studies on spatial navigation. Firstly, this study utilized two similar tasks in examining spatial navigation of healthy subjects; both tasks involve distance estimation, with triangle completion task being more complicated in which subjects are required to simultaneously perform angle estimation. This allowed comparison between the ability to estimate angle and distance; as well as the ability to estimate distance between the two different tasks; particularly between tasks with different cognitive loads. Secondly, this study employed different visual input conditions by exposing the subjects to visual inputs at different time points, either in the real world or virtual environment. This allowed the identification of the critical time point in which both visual input and the environment have the greatest impact on the spatial navigation process, particularly in the formation of the cognitive map. The validity of the findings is further complemented with inclusion of a fully‐blindfolded baseline.

Lastly, this study required the subjects to walk physically compared to past studies that required the subjects to navigate the virtual environment with joystick. This approach was realised with Vive, which is also the latest virtual reality device. This allowed simultaneous examination of the effects of proprioceptive and visual stimuli on spatial navigation process.

(23)

Several methodological limitations should also be addressed. Firstly, only one triangle was used in the triangle completion task in which the turning direction was alternated. It has been found that the pattern and degree of errors in angle and distance estimations vary with the size of triangle (Adamo et al., 2012). Therefore, future studies should utilize, for example, a smaller and a larger triangle to investigate the effect of triangle size across the types of environment (i.e. virtual environment and real world) and different visual input condition on spatial navigation behaviour, represented by angle and distance estimation. In addition, this study did not investigate the subjective perception of the participants. These subjective parameters may include subjective distance perception (e.g. shorter distance walked perceived in the virtual environment as compared to the real world), as well as perceived task difficulties and satisfaction levels. This can be further examined by correlating objective metric errors and subjective perception. Both objective and subjective findings are crucial in clinical setting; objective effectiveness of a rehabilitation programme is rendered useless if patients considered the tasks as uncomfortable or difficult which subsequently demotivates their participation (Maclean et al., 2000). As discussed earlier, utilizing the triangle completion task in conjunction with virtual environment + visual input provided solely during the encoding phase, presents great potential in clinical implementation. Thus, it is important to evaluate the subjective perceptions prior clinical implementation of the triangle completion task. This can be done by deriving a questionnaire with a set of items to quantify the subject’s subjective experiences. In conclusion, the present study compared the spatial navigation process between virtual environment and the real world with varying degree of visual inputs. The results revealed that angle and distance estimations were less accurate in the virtual environment when visual input is provided solely during the encoding phase only during the triangle completion task; indicating the role of visual input in a virtual environment on the formation of cognitive map during the encoding phase. However, such a trend was not found in the distance reproduction task, which is possibly attributed to its relative task simplicity. While the current results suggest that triangle completion task offers potential for future studies, further research is needed into the context in which triangles with different sizes are to be utilized to investigate whether the pattern and degree of errors in angle and distance estimations vary with the size of the triangle. In addition, the subjects’ objective and subjective experiences across different conditions of the tasks needs to be investigated prior to deployment into clinical setting.

(24)

References

Adamo, D. E., Briceño, E. M., Sindone, J. A., Alexander, N. B., & Moffat, S. D. (2012). Age differences in virtual environment and real world path integration. Frontiers in aging neuroscience, 4(25), 34‐42. Alldredge, B. K., l Corelli, R., Ernst, M. E., Guglielmo, B. J., Jacobson, P. A., Koda‐Kimble, M. A., . . . Young, L. Y. (2013). Applied therapeutics: the clinical use of drugs: Wolters Kluwer/Lipincot William & Wilkins. Alme, C. B., Miao, C., Jezek, K., Treves, A., Moser, E. I., & Moser, M.‐B. (2014). Place cells in the hippocampus: Eleven maps for eleven rooms. Proceedings of the National Academy of Sciences, 111(52), 18428‐18435. Alomari, M. A., Khabour, O. F., Alzoubi, K. H., & Alzubi, M. A. (2013). Forced and voluntary exercises equally improve spatial learning and memory and hippocampal BDNF levels. Behavioural brain research, 247, 34‐39. Arthur, E., Hancock, P., & Chrysler, S. (1997). The perception of spatial layout in real and virtual worlds. Ergonomics, 40(1), 69‐77. Bannerman, D. M., Sprengel, R., Sanderson, D. J., McHugh, S. B., Rawlins, J. N. P., Monyer, H., & Seeburg, P. H. (2014). Hippocampal synaptic plasticity, spatial memory and anxiety. Nature reviews neuroscience., 15(3), 181‐192. Barrett, A. M., & Craver‐Lemley, C. E. (2008). Is it what you see, or how you say it? Spatial bias in young and aged subjects. Journal of the International Neuropsychological Society, 14(04), 562‐570. Blasier, R., Carpenter, J., & Huston, L. (1994). Shoulder proprioception. Effect of joint laxity, joint position, and direction of motion. Orthopaedic review, 23(1), 45‐50. Bricken, M. (1991). Virtual reality learning environments: potentials and challenges. ACM SIGGRAPH Computer Graphics, 25(3), 178‐184. Chersi, F., & Burgess, N. (2015). The cognitive architecture of spatial navigation: Hippocampal and striatal contributions. Neuron, 88(1), 64‐77.

Crison, F., Lecuyer, A., d'Huart, D. M., Burkhardt, J.‐M., Michel, G., & Dautin, J.‐L. (2005). Virtual technical trainer:

Learning how to use milling machines with multi‐sensory feedback in virtual reality. Paper presented at the IEEE Proceedings. VR 2005. Virtual Reality, 2005. Fujita, N., Klatzky, R. L., Loomis, J. M., & Golledge, R. G. (1993). The encoding‐error model of pathway completion without vision. Geographical Analysis, 25(4), 295‐314. Garcia Popov, A., Paquet, N., & Lajoie, Y. (2013). Influence of gymnastic background on triangle completion performance in single and dual‐task conditions. The Open Sports Science Journal, 6, 15‐21.

Goodwin, G. M., McCloskey, D. I., & Matthews, P. B. (1972). Proprioceptive illusions induced by muscle vibration: contribution by muscle spindles to perception? Science, 175(4028), 1382‐1384.

Gould, N. F., M Kathleen Holmes, B., Fantie, B. D., Luckenbaugh, D. A., Pine, D. S., Gould, T. D., . . . Zarate Jr, C. A. (2007). Performance on a virtual reality spatial memory navigation task in depressed patients. American Journal of

Psychiatry, 164(3), 516‐519.

Groen, E. L., & Bos, J. E. (2008). Simulator sickness depends on frequency of the simulator motion mismatch: An observation. Presence, 17(6), 584‐593. Hebert, A. E., & Dash, P. K. (2004). Nonredundant roles for hippocampal and entorhinal cortical plasticity in spatial memory storage. Pharmacology Biochemistry and Behavior, 79(1), 143‐153. Henderson, A., Korner‐Bitensky, N., & Levin, M. (2007). Virtual reality in stroke rehabilitation: a systematic review of its effectiveness for upper limb motor recovery. Topics in stroke rehabilitation, 14(2), 52‐61. Iphofen, R. (2016). Ethical decision making in social research: A practical guide: Springer. Kearns, M. J., Warren, W. H., Duchon, A. P., & Tarr, M. J. (2002). Path integration from optic flow and body senses in a homing task. Perception, 31(3), 349‐374.

(25)

Klatzky, R. L., Loomis, J. M., & Golledge, R. G. (1997). Encoding spatial representations through nonvisually guided locomotion: Tests of human path integration. Psychology of Learning and Motivation, 37, 41‐84. Laakso, M. P., Lehtovirta, M., Partanen, K., Riekkinen, P. J., & Soininen, H. (2000). Hippocampus in Alzheimer’s disease: a 3‐year follow‐up MRI study. Biological psychiatry, 47(6), 557‐561. Laver, K., George, S., Thomas, S., Deutsch, J. E., & Crotty, M. (2012). Virtual reality for stroke rehabilitation. Stroke, 43(2), e20‐e21. Li, W. L., Cai, H. H., Wang, B., Chen, L., Zhou, Q. G., Luo, C. X., . . . Zhu, D. Y. (2009). Chronic fluoxetine treatment improves ischemia‐induced spatial cognitive deficits through increasing hippocampal neurogenesis after stroke.

Journal of neuroscience research, 87(1), 112‐122.

Linkenauger, S. A., Mohler, B. J., & Proffitt, D. R. (2011). Body‐Based Perceptual Rescaling Revealed through the Size— Weight Illusion. Perception, 40(10), 1251‐1253.

Livingstone, S. A., & Skelton, R. W. (2007). Virtual environment navigation tasks and the assessment of cognitive deficits in individuals with brain injury. Behavioural brain research, 185(1), 21‐31.

Luo, C. X., Jiang, J., Zhou, Q. G., Zhu, X. J., Wang, W., Zhang, Z. J., . . . Zhu, D. Y. (2007). Voluntary exercise‐induced neurogenesis in the postischemic dentate gyrus is associated with spatial memory recovery from stroke.

Journal of neuroscience research, 85(8), 1637‐1646.

Maclean, N., Pound, P., Wolfe, C., & Rudd, A. (2000). Qualitative analysis of stroke patients' motivation for rehabilitation.

Bmj, 321(7268), 1051‐1054.

Messing, R., & Durgin, F. H. (2005). Distance perception and the visual horizon in head‐mounted displays. ACM

Transactions on Applied Perception (TAP), 2(3), 234‐250.

Milford, M., & Schulz, R. (2014). Principles of goal‐directed spatial robot navigation in biomimetic models. Philosophical

Transactions of The Royal Society B, 369(1655), 20130484.

Muir, G. M., & Bilkey, D. K. (2001). Instability in the place field location of hippocampal place cells after lesions centered on the perirhinal cortex. The Journal of Neuroscience, 21(11), 4016‐4025.

Nedelska, Z., Andel, R., Laczó, J., Vlcek, K., Horinek, D., Lisy, J., . . . Hort, J. (2012). Spatial navigation impairment is proportional to right hippocampal volume. Proceedings of the National Academy of Sciences, 109(7), 2590‐ 2594. Nico, D., Piccardi, L., Iaria, G., Bianchini, F., Zompanti, L., & Guariglia, C. (2008). Landmark based navigation in brain‐ damaged patients with neglect. Neuropsychologia, 46(7), 1898‐1907. Rankin, K., Mucke, L., Miller, B., & Gorno‐Tempini, M. (2007). Spatial cognition and the human navigation network in AD and MCI. Neurology, 69(10), 986‐997.

Richardson, A. R., & Waller, D. (2007). Interaction with an immersive virtual environment corrects users' distance estimates. Human Factors: The Journal of the Human Factors and Ergonomics Society, 49(3), 507‐517. Riecke, B. E., Van Veen, H. A., & Bülthoff, H. H. (2002). Visual homing is possible without landmarks: A path integration study in virtual reality. Presence: Teleoperators and Virtual Environments, 11(5), 443‐473. Sahm, C. S., Creem‐Regehr, S. H., Thompson, W. B., & Willemsen, P. (2005). Throwing versus walking as indicators of distance perception in similar real and virtual environments. ACM Transactions on Applied Perception (TAP), 2(1), 35‐45. Saposnik, G., Teasell, R., Mamdani, M., Hall, J., McIlroy, W., Cheung, D., . . . Group, S. O. R. C. W. (2010). Effectiveness of virtual reality using Wii gaming technology in stroke rehabilitation a pilot randomized clinical trial and proof of principle. Stroke, 41(7), 1477‐1484.

Schinazi, V. R., Thrash, T., & Chebat, D. R. (2016). Spatial navigation by congenitally blind individuals. Wiley

(26)

So, R. H., Lo, W., & Ho, A. T. (2001). Effects of navigation speed on motion sickness caused by an immersive virtual environment. Human Factors: The Journal of the Human Factors and Ergonomics Society, 43(3), 452‐461. Stark, M. (1996). Impairment of an egocentric map of locations: Implications for perception and action. Cognitive

Neuropsychology, 13(4), 481‐524. Sweller, J. (1994). Cognitive load theory, learning difficulty, and instructional design. Learning and instruction, 4(4), 295‐312. Taube, J. S. (2007). The head direction signal: origins and sensory‐motor integration. Annu. Rev. Neurosci., 30, 181‐207. Tiantai, G., & Yintong, W. (2003). Virtual reality and augmented reality. Machinery, 41(466), 7‐9. Tolman, E. C. (1948). Cognitive maps in rats and men. Psychological review, 55(4), 189.

Tsirlin, I., Dupierrix, E., Chokron, S., Coquillart, S., & Ohlmann, T. (2009). Uses of virtual reality for diagnosis, rehabilitation and study of unilateral spatial neglect: review and analysis. Cyberpsychology & behavior, 12(2), 175‐181.

Vasconcellos, A., Tabajara, A., Ferrari, C., Rocha, E., & Dalmaz, C. (2003). Effect of chronic stress on spatial memory in rats is attenuated by lithium treatment. Physiology & behavior, 79(2), 143‐149.

Wang, J., Ho, L., Chen, L., Zhao, Z., Zhao, W., Qian, X., . . . Rosendorff, C. (2007). Valsartan lowers brain β‐amyloid protein levels and improves spatial learning in a mouse model of Alzheimer disease. The Journal of clinical

investigation, 117(11), 3393‐3402. Wolbers, T., & Hegarty, M. (2010). What determines our navigational abilities? Trends in cognitive sciences, 14(3), 138‐ 146. Ziemer, C. J., Plumert, J. M., Cremer, J. F., & Kearney, J. K. (2009). Estimating distance in real and virtual environments: Does order make a difference? Attention, Perception, & Psychophysics, 71(5), 1095‐1106.

Referenties

GERELATEERDE DOCUMENTEN

In the second method, referred to as Hot Plate (HP), the membrane is placed directly onto a hot surface. Two model membranes were selected to explore the heating methods:

Hence, perceptions of a leader’s fairness in allocating rewards and benefits seem to significantly influence trust, both through the character- and relationship-based mechanisms

Zich beroemen op het verleden komt als een trek van de oude man goed uit de verf, terwijl mooi belicht wordt hoe hij zijn tong nog kan roeren, een bekwaam- heid die ook Bade

brunneum Cb15-III for A&amp;K strategy against the major subterranean termite pest species (Microtermes sp.) sampled in cocoa agroforests (Chapter six) showed no

De vervanging van voedetfosfaat door fytase bij een gelijk berekend verteerbaar fosforgehalte heeft in deze proef niet geleid tot een verbete- ring van de opfokresultaten van gespeende

- Een driehoekshok voor vleesvarkens met 0,45 m2 quent aflaten van mest door toepassing van drie- dichte vloer per dier veroorzaakt meer hokbevuiling hoekshokken bij

Gedurende de jaren 1991 t/m 1994 is op twee maaiproeven op kleigrond de stikstofbenutting voor Engels raaigras weidetype, laat hooitype en vroeg hooitype onderzocht.. In dit

Ook tijdens het voeren leek het wegduwen bij de voerbak het meest voor te komen bij de kooien zonder schotten, waarschijnlijk omdat de hennen hier met een groot aantal bij elkaar