The Transparent Video Facetop is a novel user interface concept that supports not only single-user interactions with a PC, but also close pair collaborations, such as that found in collaborative Web browsing, in distributed pair programming and in distributed extreme programming (dPP/dXP). We recently demonstrated the Vis-a-Vid facetop prototype as a single-user GUI for manipulating the elements of a traditional WIMP desktop [21]. In this paper we show how the single-user transparent video facetop can be used for fingertip control of a Web browser, and how a dual-head facetop can be used for paired synchronous (collaborative) Web browsing. The facetop is not a new browser, per se, but rather a novel way to interact with a Web brower, as well as a novel means of providing visual interaction among users collaboratively browsing the Web. BASIC FACETOP CONCEPTS The transparent video facetop is a novel enhancement of the traditional WIMP user interface, so nearly ubiquitous on today’s computers. In the facetop, the user sees him/her self as a “ghostly” image apparently behind the desktop, looking back at the icons and windows from the back. Instead of a traditional desktop, we see a “face” top. This self-image is used for visual feedback and communications both to the user as well as to collaborators; it is also used for desktop/application control and manipulation via a fingertipdriven “virtual mouse”. Figure 1: Facetop physical setup, with iBot video camera Figure 1 shows the physical setup for a computer with a facetop being displayed on a monitor. Note the video camera sitting on top the LCD panel pointing back at the user; in our current work we use a $100 Sony iBot, giving us an image that is 640 x 480 pixels of 24-bit color, captured 30 frames per second. The facetop video window shows the PC user sitting at his/her workspace; we reverse the image horizontally so that when the user moves a hand, say, to the left, the image of the hand mirrors this movement on the screen. In software, and using a high-performance 3D-graphics video card, we make the video window semi-transparent and composite it with the desktop image itself. Once we have the full screen video with transparent image compositing we get the illusion of the user watching the desktop from behind. Mirroring means if the user physically points to an icon on the desktop, the facetop image points to the icon as well (with proper spatial calibration of the camera and user locations). Using image analysis techniques we then track the user’s fingertip in the backing window, and optionally drive the mouse from this tracker. The user can then manipulate the desktop of a projected computer from his seat while successfully communicating the areas of interest on the screen to others watching the projection. Projected displays vs. monitors The facetop as a concept works fine on a PC with any display technology -a monitor, a projector, an immersive device -but its unique aspects are most pronounced and most effective in a projected environment. The concept of background user video as visual cues for control and communication came about when our research group was discussing other work using a projected PC. We were all sitting in chairs viewing the projection wall, but constantly pointing at the desktop image 5 to 10 feet away. Determining where on the wall/screen to look was largely an exercise in visually interpolating along the line formed by a person’s pointing arm. With the facetop, the user is right there in the desktop image and gives an immediate visual cue as to where to look when pointing. Figure 6 shows the facetop projected.
[1]
Donald C. Wells.
Extreme Programming: A gentle introduction
,
2003
.
[2]
Jonathan Grudin,et al.
Integration of interpersonal space and shared workspace: ClearBoard design and experiments
,
1993,
TOIS.
[3]
Ronald P. Uhlig,et al.
The office of the future
,
1979
.
[4]
Ivan Laptev,et al.
Tracking of Multi-state Hand Models Using Particle Filtering and a Hierarchy of Multi-scale Image Features
,
2001,
Scale-Space.
[5]
J. Luis Guasch,et al.
Costs and benefits
,
2019,
Joint Audit 2019 – Enhancing Tax Co-operation and Improving Tax Certainty.
[6]
Robert E. Kraut,et al.
The VideoWindow system in informal communication
,
1990,
CSCW '90.
[7]
Greg Welch,et al.
Working in the Office of "Real Soon Now"
,
2000,
IEEE Computer Graphics and Applications.
[8]
P. David Stotts,et al.
Virtual Teaming: Experiments and Experiences with Distributed Pair Programming
,
2003,
XP/Agile Universe.
[9]
Jonathan Grudin,et al.
Integration of Inter-Personal Space and Shared Workspace: ClearBoard Design and Experiments
,
1992,
CSCW.
[10]
Robert,et al.
The VideoWindow System in Informal Communications
,
1990
.
[11]
P. David Stotts,et al.
Exploring the Efficacy of Distributed Pair Programming
,
2002,
XP/Agile Universe.
[12]
Maureen C. Stone,et al.
A taxonomy of see-through tools: the video
,
1995,
CHI 95 Conference Companion.
[13]
P. David Stotts,et al.
An orthogonal taxonomy for hyperlink anchor generation in video streams using OvalTine
,
2000,
HYPERTEXT '00.
[14]
Yoichi Sato,et al.
Interactive object registration and recognition for augmented desk interface
,
2001,
INTERACT.
[15]
Laurie Williams,et al.
The costs and benefits of pair programming
,
2001
.
[16]
P. David Stotts,et al.
Semi-automated Hyperlink markup for archived video
,
2002,
HYPERTEXT '02.
[17]
Hiroshi Ishii,et al.
Transparent layered user interfaces: an evaluation of a display design to enhance focused and divided attention
,
1995,
CHI '95.
[18]
David Stotts,et al.
The Vis-a-Vid Transparent Video Facetop
,
2003
.
[19]
K. Beck,et al.
Extreme Programming Explained
,
2002
.
[20]
Donald R. Johnson,et al.
The Office of the Future.
,
1985
.
[21]
Steve Benford,et al.
User embodiment in collaborative virtual environments
,
1995,
CHI '95.
[22]
Laurie Williams,et al.
Exploring Pair Programming in Distributed Object-Oriented Team Projects
,
2002
.
[23]
Alan W. Esenther.
Instant Co-Browsing: Lightweight Real-Time Collaborative Web Browsing
,
2002,
WWW 2002.
[24]
Lars Bretzner,et al.
Use Your Hand as a 3-D Mouse, or, Relative Orientation from Extended Sequences of Sparse Point and Line Correspondences Using the Affine Trifocal Tensor
,
1998,
ECCV.
[25]
David Stotts,et al.
2 CobWeb : Visual Design of Collaboration Protocols for Dynamic Group Web Browsing
,
2002
.