michael naimark
 
Copyright 1997 Society of Photo-Optical Instrumentation Engineers

This paper was published in SPIE Proceedings, Vol. 3012, San Jose, 1997, and is made available as an electronic reprint with permission of SPIE. Single print or electronic copies for personal use only are allowed. Systematic or multiple reproduction, or distribution to multiple locations through an electronic listserver or other electronic means, or duplication of any material in this paper for a fee or for commercial purposes is prohibited. By choosing to view or print this document, you agree to all the provisions of the copyright law protecting it.




 
A 3D Moviemap and a 3D Panorama
 
Michael Naimark
Interval Research Corporation
Palo Alto, California


ABSTRACT

Two immersive virtual environments produced as art installations investigate "sense of place" in different but complimentary ways. One is a stereoscopic moviemap, the other a stereoscopic panorama. Moviemaps are interactive systems which allow "travel" along pre-recorded routes with some control over speed and direction. Panoramas are 360 degree visual representations dating back to the late 18th century but which have recently experienced renewed interest due to "virtual reality" systems. Moviemaps allow "moving around" while panoramas allow "looking around," but to date there has been little or no attempt to produce either in stereo from camera-based material.

"See Banff!" (1993-4) is a stereoscopic moviemap about landscape, tourism, and growth in the Canadian Rocky Mountains. It was filmed with twin 16mm cameras and displayed as a single-user experience housed in a cabinet resembling a century-old kinetoscope, with a crank on the side for "moving through" the material.

"BE NOW HERE (Welcome to the Neighborhood)" (1995-6) is a stereoscopic panorama filmed in public gathering places around the world, based upon the UNESCO World Heritage "In Danger" list. It was filmed with twin 35mm motion picture cameras on a rotating tripod and displayed using a synchronized rotating floor.

Keywords
: moviemaps, panoramas, immersive virtual environments, art installations



1. CAMERA-BASED IMMERSION

"Immersion," in the context of media and virtual environments, is often defined as the feeling of "presence" or "being there," of being "inside" rather than "outside looking in." Various attempts have been made to taxonomize the elements required for presence,
1,2,3,4 but at the very least studies suggest that visual presence is directly related to field of view (FOV). 5,6 Stereopsis and orthoscopy (i.e., where the viewing FOV matches the recording FOV, thus maintaining proper scale) often enhance immersion. 7 A variety of special-venue film formats have been developed to deliver immersive experiences. 8 These experiences are camera-based and represent the physical world, but are linear and don't allow any interaction or navigability. Computer-based immersive virtual environments do allow interaction and navigability, but are restricted to whatever can be made into 3D computer models. Since such computer models are built from scratch, they often represent imaginary or fantasy environments. Making computer models of actual places has proven to be non-trivial, and even today's best models display unwanted artifacts.
The work described here represents something between filmic and computer graphic immersion. It is camera-based and the imagery is of the physical world (in the documentary or ethnographic tradition), but it also has elements of interactivity and navigability, be they constrained.


2. MOVIEMAPS

Moviemaps allow virtual travel through pre-recorded spaces. Routes are pre-determined and filmed with a stop-frame camera triggered by distance rather than by time (typically done via an encoder on a wheel). Distance-triggering maintains constant speeds during playback at constant frame rates, which is often not practical or possible during production with a conventional (time-triggered) movie camera. The result, in a very real sense, is the transfer of speed control from the producer to the end-user, who has control over frame rate through an input device like a joystick or trackball.
In addition to speed control, limited control of direction is possible by filming registered turns at intersections. By match-cutting between a straight sequence and a turn sequence, the user can "move" from one route to another. Care must be taken to minimize visual discontinuities such as sun position and transient objects (e.g., cars and people). The goal is to make the cuts appear as seamless as possible.


2.1 Past Moviemaps

The first interactive moviemap was produced at MIT in the late 1970s of Aspen, Colorado. A gyroscopic stabilizer with 16mm stop-frame cameras was mounted on top of a camera car and a fifth wheel with an encoder triggered the cameras every 10 feet. Filming took place daily between 10 AM and 2 PM to minimize lighting discrepancies. The camera car carefully drove down the center of the street for registered match-cuts. In addition to the basic "travel" footage, panoramic camera experiments, thousands of still frames, audio, and data were collected. The playback system required several laserdisc players, a computer, and a touch screen display. Very wide-angle lenses were used for filming, and some attempts at orthoscopic playback were made.
9

The author has since conceived and directed several moviemap productions, each with its own unique playback configuration. The "Paris VideoPlan" (1986) was commissioned by the RATP (Paris Metro) to map the Madeleine district of Paris from the point-of-view of walking down the sidewalk. It was filmed with a stop-frame 35mm camera mounted on an electric cart, filming one frame every 2 meters. An encoder was attached to one of the cart's axles. Rather than filming all the turn possibilities at each intersection, a mime was employed to stand in each intersection and simply point in the possible turn directions. The idea was to substitute the perceptual continuity of actual match-cuts with cinematic continuity. The playback system was built in a kiosk and exhibited in the Madeleine Metro Station.

The "Golden Gate Videodisc Exhibit" (1987) was produced for San Francisco's Exploratorium as an aerial moviemap over a 10 by 10 mile grid of the Bay Area. It was filmed with a gyro-stabilized 35mm motion picture camera on a helicopter, which flew at a constant ground speed and altitude along one-mile grid lines determined by LORAN satellite navigation technology, effectively filming one frame every 30 feet. The camera was always pointed at the center of the Golden Gate Bridge, hence no turn sequences were necessary since the images always matched at each intersection regardless of travel direction. The playback system used a trackball to control both speed and direction, with the feel of "tight linkage" to the laserdiscs. The result was the sensation of moving smoothly over the Bay Area at speeds much faster than normal.

"VBK: A Moviemap of Karlsruhe" was commissioned by the Zentrum fur Kunst und Medientechnologie (ZKM). Karlsruhe, Germany, has a well-known tramway system, with over 100 kilometers of track snaking from the downtown pedestrian area out into the Black Forest. A 16mm stop-frame camera was mounted in front of a tram car and interfaced to the tram's odometer. Triggering was programmed to be at 2, 4, or 8 meter increments per frame depending on location. Filming on a track resulted in virtually perfect spatial registration. The playback system consisted of a pedestal with a throttle for speed control and 3 pushbuttons for choosing direction at intersections. The camera had a very wide-angle lens (85 degree horizontal FOV) and playback employed a 16 foot wide video projection. The input pedestal was strategically placed in front of the screen to achieve orthoscopically correct viewing, resulting in a strong sense of visual immersion.


2.2 The "See Banff!" Kinetoscope

But the immersive experience with the Karlsruhe Moviemap was monoscopic. One might argue that binocular disparity is not an important factor for landscape imagery (e.g., compared to infinity focus or motion disparity), but no one had yet made a stereoscopic moviemap.

In 1992, the author was working on field recording studies in the "Art and Virtual Environments" program at the Banff Centre for the Arts, and after several attempts at making computer models from camera-based imagery, reverted back to using moviemap production techniques, but this time in stereo.
10 The concept was to film scenic routes in and around the Banff region of the Canadian Rocky Mountains.


2.2.1 Camera Design and Production

The camera rig had to be small, portable, and rugged. As attractive as gyro stabilization may have been, it would have been much too heavy to take down mountain trails, on glaciers, and over narrow bridges. Without such stabilization, the stability of the imagery would be at the mercy of the terrain and, to some extent, the skill of the operator.

The basis of the camera rig was a 3-wheeled "super jogger" baby carriage, reinforced for extra sturdiness and modified to hold a tripod (see Figure 1). An encoder was installed on one of the rear wheels, with electronics for triggering the cameras from 1 frame every centimeter on up. A custom mount was built to hold two 16mm stop-frame cameras in parallel so that they could be released for film loading but would mount back in precisely the same position.

The cameras were fitted with the 85 degree horizontal FOV wide-angle lenses, with the intention of making a wide-angle orthostereoscopic display system.

The cameras were always triggered in sync. The stop-frame motors rotated at 1/8 second. The shutters were variable and modified to lock at 30 degrees, resulting in a shutter speed of 1/96 second, enough to freeze most motion if the rig moved at walking speed.

Figure 1. The "See Banff!" camera rig.


After much theoretical debate about optimal interocular distance between cameras, the minimum practical distance the cameras could be mounted apart was about 8 inches due to the size of the stop-frame motors. After reviewing several hundred turn-of-the-century landscape stereograms, it became clear that the exaggerated depth resulting from abnormally large interocular distances was the rule more than the exception, so no sleep was lost over our rig design.

The intent of production was to film a wide variety of unconnected routes without any intersections. The playback system would require speed control and route selection but not control over turns. This decision broadened the range of possible filming, since the notion of completion (i.e., covering all possibilities, filming turns, completing grids) wasn't necessary.

As it turned out, capturing the beauty of the landscape became overshadowed by the presence of tourists everywhere. Busload after busload appeared even in very remote areas. Dozens of people of all ages and cultures, clad in bright colors and toting cameras, wandered through the landscape. It became clear that, as an artwork, there was strength in counterpointing the beauty of the landscape with the actualities of tourism. The presence of tourists also created a lively foreground, giving the imagery a greater sense of 3D.

Filming took place during September 1993 in a wide range of locations. Stability ranged from finding smooth paved wheelchair trails to carrying the rig over rocky terrain. Frame rates were determined on-the-spot as a function of stability of the surface and distance of nearest objects, and ranged from 1 frame every centimeter to 1 frame every meter. Usually the cameras were pointed forward in the direction of movement but sometimes were pointed off to the side (and occassionally were stationary in a clock-driven timelapse mode). Over 120 scenes were filmed.


2.2.2 Playback System

Shortly after production, the film was transferred to videotape, edited, and transferred to 2 laserdiscs, one for each eye. A trackball-based interactive system was produced, with simple optics and mirrors arranged in a Wheatstone configuration for single-user stereoscopic viewing.

The first system approximated true orthoscopic viewing, with an 85 degree horizontal FOV. Such a wide FOV is considerably larger than sitting in the front row of the grandest of movie theaters, and the video resolution was extremely coarse spread out over such a large area. We backed off to a FOV closer to 60 degrees. It still looked very large and still appeared orthoscopically correct. (One is tempted to speculate that the human perceptual system acts like Saul Steinberg's "New Yorker's Map of the U.S., " that anything bigger or farther than that with which we are familiar appears so nonlinear that beyond some point it doesn't matter.)

The next step was to package it into a traveling exhibit. In roughing out a basic design - a podium-like box to house the hardware, an eyehood for single-user wide-angle orthostereo viewing, a one-dimensional input device - it became clear that a strikingly similar device had already been built. In April 1894, the Edison kinetoscope made its public debut. This was at a turbulent moment in the history of cinema, when the camera had already been actualized but projection had not. It was now exactly 100 years later, and the temptation to suggest an analogy was overwhelming (see Figure 2).

The final exhibit, called "See Banff!" (irony intentional), was built of walnut and brass in an authentic but exaggerated kinetoscope design, with a lever for selecting one of several scenes on exhibit and a crank for "moving through" the material. (History buffs will note that kinetoscopes never had cranks, mutoscopes did, in part because Edison was peddling another of his inventions, electricity and electric motors.) The crank was equipped with a force-feedback brake which freezes movement when the user reaches the boundaries of each scene, simulating film mechanics. (Perhaps not surprisingly, it feels "not right" when the brake is disabled.)

Partly for practical and partly for aesthetic reasons, a single laserdisc was used with field-sequential stereoscopic video and 30 Hz LCD shutters built into the optics. The flicker is noticeable, but after all, it is a kinetoscope.


3. PANORAMAS

The word "panorama" was coined in 1792 in London to describe the first of what became a popular form of public entertainment: a large elevated cylindrical room entirely covered by a 360 degree painting.
11 From the beginning, a distinction was made between displaying a panoramic image all at once (circular or stationary panoramas) and over time (moving panoramas). This distinction has carried over to cinema as well.


3.1 "Moving Movies"

In 1977, the author began investigating what happens when a projected motion picture image physically moves the same way as the original camera movement. If the angular movement of the projector equals the angular movement of the camera, and if the FOVs are equal, spatial correspondence is maintained, and the result appears as natural as looking around a dark space with a flashlight.
12 A simple demonstration was made using a super-8 film camera and projector on a slowly rotating turntable. Later a more complex system which recorded the pan and tilt axes was built. Finally a series of art installations were realized using the simple turntable inside a space furnished to resemble a livingroom, whose entire contents were spray-painted white after filming to become a 3D relief projection screen for itself.13 This phenomena of motion picture projection physically moving around a playback space was named "moving movies."

Moving movies have no lateral and only angular movements, which must be around the camera's nodal point. As such, they are strictly 2D spatial representations. Computer-based moving panoramas such as Apple Computer's Quicktime VR, Microsoft's Surround Video, and Omniview's Photobubbles (to name a few) are similar insofar as they rely on 2D representations from a single point of view, produced by tiling multiple images or by using a single fisheye lens. Making any 2D visual representation from more than one point of view results in distortion, by definition. (The essence of 2D photographic representation and indeed, photorealism, is strictly from a single point of view. Alternative forms like Cubism and David Hockney's photo-collages are fine counter-examples.)

A 2D panorama represents a single point of view, but stereopsis requires at least two points of view. Making stereoscopic panoramas with two 2D panoramic images is problematic. For example, if two panoramic images are taken with stationary cameras placed apart at an interocular distance, the disparity will vary as the user looks around (and will become zero on the axis between the 2 cameras). One the other hand, if the cameras move laterally during exposure to keep disparity constant, each camera's image will no longer represent a single point of view, resulting in distortion (e.g., circles become ovals, image appear fragmented, perspective becomes discontinuous).


3.2 "Be Now Here"

In December 1994, the author was invited to produce an art installation for the Center for the Arts at the Yerba Buena Gardens in San Francisco to open in December 1995. The installation proposed was called "Be Now Here (Welcome to the Neighborhood)." Just as the Banff kinetoscope was an experiment in making a stereoscopic version of interactively moving around, "Be Now Here" was to compliment it by making a stereoscopic version of interactively looking around.

The concept was to assemble an experimental camera system to film stereoscopic panoramas, then to go to public gathering places (commons) and film throughout the course of a day from a single position. The experience would be analogous to standing in a single place, with both eyes open, and being able to look around but not move from the spot.

Site selection was based on the "In Danger" list issued by the UNESCO World Heritage Centre in Paris.
14 Of the 440 UNESCO-designated World Heritage Sites, 17 had been further designated "In Danger." Of these, 4 are cities: Jerusalem, Dubrovnik (Croatia), Timbuktu (Mali), and Angkor (Cambodia). With assistance of UNESCO, the plan would be to visit each site, to work with local collaborators and to determine the most representative public commons and a single spot in which to set up the camera system. (Partly for art and partly for research reasons, going into interesting but fragile environments to make an statement about "place" seemed appropriate.)


3.2.1 Camera Design and Production

The camera design was based on 2 cameras (for stereo), 60 degree horizontal FOV lenses (for immersion), and a slowly rotating tripod (for panoramics), rotating once per minute (1 rpm). This is a compromise since it takes a minute to capture an entire 360 degree scene, but using multiple camera pairs to capture the entire scene at once was not practical. Sunlight variation was assumed to be negligible during the course of a minute, so using multiple images from the same scene for projection or panoramic tiling would have artifacts only from moving objects. A 1 rpm closed-loop crystal synchronized motor was mounted on a tripod.

The question of how to arrange the cameras with respect to the axis of rotation resulted in lively debate. Mimicking mammal head rotation suggested placing both cameras symmetrically in front of the axis of rotation. A colleague, John Woodfill, had a novel suggestion: rotate the camera pair around the nodal point of one of the cameras. Such a configuration would result in a perfect 2D panorama from one camera and would place all of the disparity difference in the other camera. As a vision researcher interested in the footage, he felt this could be useful. The "Woodfill Configuration" was deployed, with careful manual determination of one of the camera's nodal points. (One might speculate that looking at a stereoscopic panning movie where one eye sees no disparity and the other eye sees all the disparity would be noticeable, but one could counter-speculate that it's possible to sit on a rotating stool with one eye directly over the axis of rotation and conclude that there's nothing special about it.)

After much deliberation, it was decided to use 35mm motion picture film. The resolution would be 4 times greater than 16mm film and much greater than video, particularly with respect to dynamic range. It is also well-known that 35mm motion picture cameras are simple yet durable and time-tested, with less likelihood of failure than video in the field. Arriflex cameras with Zeiss lenses were selected. As with Banff, the size of the cameras made it difficult to obtain normal human interocular distance, so an exaggerated interoccular distance of 8 inches was used.

It was further decided to film at a frame rate of 60 frames per second (fps). Such footage could be transferred to video with each film frame corresponding to a single video field (half-frame). The result would have the best qualities of both film and video: it maintains the high dynamic range of film while having the motion smoothness of video (which updates at 60 fps). A sync box made by Cinematography Electronics was used to synchronize both cameras to a single controller, which allowed syncing phase as well as frame rate. The shutters were closed down to 30 degrees, resulting in an exposure time of 1/720 second, fast enough to freeze most everything.

Color negative film daylight-balanced with an ASA of 50 was used. Since all filming was to take place during daylight, the low ASA coupled with fast shutter speed wouldn't be a problem, with most filming possible at apertures between F4 and F11.

Using stereoscopic 35mm cameras with high quality lenses and low-speed film, running at 60 fps, and with synchronized shutter and rotational speeds would result in unrivaled fidelity. It would, at the very least, have 5 times the resolution of theatrical 35mm film (twice the spatial and 2.5 times the temporal resolution).

The complete camera rig, including cases, weighed 500 pounds but was built for travel (see Figure 3). All production took place during October 1995, including filming the Yerba Buena Gardens in San Francisco for counterpoint. A pro-DAT audio recorder with a shotgun microphone was used to collect sounds at each site for later mixing into 4 channel rotating sound. Enough stock to film 5 panoramas (10 reels of 400' film) was taken to each site. Miraculously, production stayed on schedule and everything came out.
15

Figure 3. The "Be Now Here" camera rig on location in Timbuktu.



3.2.2 Playback System

After production, the film was transferred to videotape, edited and mixed with the audio, then transferred to 2 laserdiscs. A simple input pedestal was made to allow site selection. Three scenes from each site were selected and aligned with each other such that perfectly registered time-of-day changes could be experienced within the same location.

Unlike Banff, Be Now Here would retain realtime motion and sacrifice browsability, enabling the user to control place and time but not speed. People movement would appear natural (not the case with See Banff) and coupled audio would be possible. This decision was based on the difference between footage of moving along long pathways (where browsability is desirable) and footage which, literally, goes around in circles.

A 12 by 16 foot highly reflective front projection screen would be used in conjunction with dual polarized video projection. The input pedestal would be strategically placed approximately 14 feet in front of the screen to entice the viewers to stand at the orthoscopically correct spot. The audience would wear inexpensive polarized glasses.

The obvious way to achieve spatially correspondent playback would be to rotate the projected image around the viewing space, another moving movie. But this proved impractical given the need for wide-angle projection. The solution: rather than rotate the projection around the static audience, to rotate the audience inside a static projection.

A 16 foot diameter rotating platform was used, rotating at 1 rpm in sync with the imagery (see Figure 4). The audience, limited to 10 at a time, stands on it (standing seems more desirable for ambient rather than narrative media). A black tent-like cylindrical structure surrounded most of the viewing space to mask out the non-rotating world.

Figure 4. The "Be Now Here" installation.



The resulting effect is difficult to describe. Most viewers reported that after several seconds it felt like they were still and that the image was rotating around them. This effect is similar to the "moving train illusion," when a train sitting in the station pulls out and observers on the adjacent (non-moving) train believe their train is the moving one. Some viewers reported feeling the rotational force from the turntable, but most did not. (In a NASA design study on space colonies, it was determined that 1 rpm was the maximum rotational rate which would be undetectable by the general population.
16) Though one might argue that physchophysical ambiguity exists between such audio-visual and vestibular cues, one could equally argue that a conventional panning image viewed in a conventional (non-rotating) movie theater produces the same degree of ambiguity between audio-visual and vestibular senses.

Most everyone reported feeling a strong visceral sense of place. And that's what the installation was about: conveying the feeling of presence by connecting our eyes and ears with the ground.



ACKNOWLEDGMENTS


The author wishes to express thanks and gratitude to the many collaborators  
17,18 for See Banff and Be Now Here. See Banff was produced with the Banff Centre for the Arts. Be Now Here was produced for the Center for the Arts Yerba Buena Gardens in San Francisco with special thanks to the UNESCO World Heritage Centre in Paris. Both projects were entirely supported by Interval Research Corporation in Palo Alto.




REFERENCES


1. M. Naimark, "Elements of Realspace Imaging: a Proposed Taxonomy," SPIE Electronic Imaging Proceedings, vol. 1457, pp. 169-179, 1991.


2. T. B. Sheridan, "Musings on Telepresence and Virtual Presence," Presence, vol. 1, no. 1, pp. 120-126, 1992.


3. D. Zeltzer, "Autonomy, Interaction, and Presence," Presence, vol. 1, no. 1, pp. 127-132, 1992.


4. J. Steuer, "Defining Virtual Reality: Dimensions Determining Telepresence," Journal Of Communication, vol. 42, no. 4, 1992.


5. T. Hatada, H. Sakata, H. Kusaka, "Psychophysical Analysis of the 'Sensation of Reality' Induced by a Visual Wide-Field Display," SMPTE Journal, vol. 89, pp. 560-569, 1980.


6. C. Hendrix and W. Barfield, "Presence within Virtual Environments as a Function of Visual Display Parameters," Presence, vol. 5, no. 3, pp. 274-289, 1996.


7. E. M. Howlett, "Wide Angle Orthostereo," SPIE Electronic Imaging Proceedings, vol. 1256, 1990.


8. M. Naimark, "Expo '92 Seville," Presence, vol. 1, no. 3, pp. 364-369, 1992.


9. R. Mohl, Cognitive Space in the Interactive Movie Map: An Investigation of Spatial Learning in Virtual Environments, PhD dissertation, Education and Media Technology, M.I.T., 1981.


10. M. Naimark, "Field Recording Studies," Immersed in Technology: Art and Virtual Environments, M. A. Moser, ed., pp. 299-302, MIT Press, Cambridge, 1996.


11. A. Miller, "The Panorama, the Cinema, and the Emergence of the Spectacular," Wide Angle, vol. 18, no. 2, pp. 34-69, 1996.


12. M. Naimark, "Spatial Correspondence in Motion Picture Display," SPIE Proceedings, vol. 462, pp. 78-81, 1984.


13. M. Naimark, "Moving Movie," Aspen Center for the Arts, 1980; "Movie Room," Center for Advanced Visual Studies, M.I.T., 1980; "Displacements," San Francisco Museum of Modern Art, 1984.


14. See: http://www.unesco.org/whc/list.htm .


15. See: http://www.naimark.net/writing/trips/bnhtrip.html.


16. R. D. Johnson and C. Holbrow (ed.), Space Settlements: A Design Study, NASA SP-413, p. 22, 1977.


17. See: http://www.naimark.net/projects/banff.html .


18. See: http://www.naimark.net/projects/benowhere.html .