Sharing digital information and collaborating based on that digital information is becoming increasingly common. Input devices capture digital information (e.g., user input on a computing device, digital cameras, scanning devices, etc.). Output devices output digital information for consumption by a user or group of users. Output devices may include digital displays or digital projectors that display digital information onto a display screen or into a workspace.
The following description includes discussion of figures having illustrations given by way of example of implementations of embodiments of the invention. The drawings should be understood by way of example, not by way of limitation. As used herein, references to one or more “embodiments” are to be understood as describing a particular feature, structure, or characteristic included in at least one implementation of the invention. Thus, phrases such as “in one embodiment” or “in an alternate embodiment” appearing herein describe various embodiments and implementations of the invention, and do not necessarily all refer to the same embodiment. However, they are also not necessarily mutually exclusive.
Embodiments described herein relate to a projection-capture system. The combination of projection, capture and user input control in the same workspace at the same time facilitates mixed reality (i.e., physical and virtual) collaboration. For example, virtual objects could be files on a user's computer (e.g., electronic presentation slides, electronic documents, digital photos, etc). Physical objects could be two-dimensional objects (e.g., photos, documents, etc.) or three-dimensional objects (e.g., models, widgets, etc.). As described herein, physical and virtual objects are treated as visually interchangeable “layers.” As described herein, these layers represent logical layers and allow systems and devices described herein to control the workspace environment such that one layer (e.g., physical object or set of digital information) has the appearance of being on top of another layer. In a collaborative environment where users in different locations collaborate via separate projection-capture systems, each user is able see and interact with both the local and remote objects (physical and virtual) because the camera on each system sends a live (e.g., video) feed to the other location for projection.
System 100 includes a camera 110, a display 120, a layering module 130 and a controller 140. Camera 110, display 120 and layering module 130 are operatively connected to controller 140 to facilitate altering visual adjacency between layers in a workspace. In an example, display 120 displays a map onto a work-surface. Display 120 could be a display screen placed onto a work-surface or it could be a digital projector that displays information on the work-surface via digital projection. Additionally, camera 110 detects a physical object (e.g., a model of a house or building) placed on the work-surface. Camera 110 could be a visible-light camera (e.g., digital image camera, a digital video camera, etc.) or an infrared (IR) camera.
Layering module 130 associates the physical object with one visual layer and associates the projected map with another visual layer. By maintaining state information for the various layers, controller 140 is able to control the visual adjacency of the layers. In other words, one layer may initially appear to be “on top of” the other layer, but controller 140 may alter the visual adjacency such that the “bottom” layer visually becomes the “top” layer.
System 200 includes a top 210 and a base 220. Base 220 includes an infrared (IR) camera 224 and a projector 222, along with housing other components (e.g., processor, memory, layering module, controller, etc.). Projector 222 may be any suitable light projector including, but not limited to, LED (laser emitting diode) and DLP (digital light processing) projection. In various embodiments, projector 222 projects digital information toward top 210 where a mirror reflects the projection onto work-surface 230. IR camera 224 detects the position of physical objects positioned on work-surface 230.
Projector 222 projects a projection 250 of digital information (e.g., lines) onto work-surface 230. Object 260 is a physical object placed in workspace 230. IR camera 224 detects the position of the object 260. A layering module (e.g., layering module 130) associates projection 250 with one logical layer and object 260 with another logical layer. As shown in
System 200 also includes a user input device 240 that allows the user to interact with system 200. In various embodiments, user input (UI) device 240 includes an infrared digital stylus and/or infrared camera to detect the position of UI device 240 in workspace 230. Although any suitable UI device may be used, a digital stylus has the advantage of allowing input in three dimensions, including along the surface of workspace 230, without a sensing pad or other special surface. Thus, system 200 can be used on a variety of work-surfaces.
User input received from UI device 240 or other input mechanism may indicate a request to alter the visual adjacency of layers in workspace 230. For example, touching object 260 with UI device 240 may indicate a request to have object 260 be the foreground layer, with projection 250 as the background layer. Touching the surface of workspace 230 might indicate a request to have projection 250 be the foreground layer, with object 260 as the background layer. System 200 can use this input and state information from the layering module to alter the projection from projector 222 such that the visual order of the layers is altered (e.g., foreground and background layers are swapped).
In response to user input or other control signal to alter the visual adjacency of the layers, the system layering module changes the layer associations from those of state 410 to those of state 420. In state 420, the physical object becomes layer L2 and the projection of digital information becomes layer L1. To create the visual appearance of the physical object being “on top of” the projection, the system may simply stop projecting the digital information into the workspace.
State 412 includes a projection of digital information (e.g., a map, document, image, etc.) initially associated with layer L1. In other words, the projection is visually adjacent L0. A physical object (e.g., two-dimensional or three-dimensional) is initially associated with layer L2 and is visually adjacent layer L1. To achieve the visual appearance of physical object being “on top of” the projection, the system may remove the digital information (e.g., map or image data, etc.) corresponding to the coordinates of the physical object from the projection. For example, at the position where the physical object is located, the system may project a blank space (e.g., white or other suitable-colored light) while still projecting the original digital information (e.g., map or image data, etc.) at all other positions in the workspace. Alternatively, the system may capture a digital image of the physical object (e.g., from directly above the physical object) and project that image at the position where the physical object is located (in lieu of projecting a blank space). In either example, the effect is the appearance of the physical object being “on top of” of the projected digital information.
In response to user input or other control signal to alter the visual adjacency of the layers, the system layering module changes the layer associations from those of state 412 to those of state 422. In state 422, the physical object becomes layer L1 and the projection of digital information becomes layer L2. To create the visual appearance of the projected digital information being “on top of” the physical object, the system may simply project all of the digital information into the workspace including at the position where the physical object is located.
It should be noted that multiple projection-capture systems may be used in remote locations for virtual collaboration. The states depicted in
A system identifies 510 a physical object in a workspace. The physical object can be two-dimensional or three-dimensional. Objects may be identified by a camera (e.g., infrared, digital image capture, digital video, etc.). In various embodiments, identification includes determining a position (e.g., coordinates) of the object within the workspace. A layering module associates 520 the physical object with a first layer. The system also displays (e.g., projects) 530 digital information (e.g., an image) as a second layer into the workspace. Again, the layering module maintains the association between the projected information and the second layer.
The system, perhaps in response to user input, alters 540 the visual adjacency between the first and second layers. For example, the first layer might initially be visually adjacent (e.g., appears to be “on top of”) the surface of the workspace. The second layer might initially be visually adjacent (e.g., appears to be “on top of”) the first layer. In response to a user request or other control signal, the system switches the visual order of the layers. Thus, the second layer becomes visually adjacent the surface of the workspace while the first layer becomes visually adjacent the second layer.
A projection-capture system identifies 610 a physical object in a workspace. Again, the physical object can be two-dimensional or three-dimensional. Objects may be identified by a camera (e.g., infrared, digital image capture, digital video, etc.). A system layering module associates 620 the physical object with a first layer. The system also displays (e.g., projects) 630 digital information (e.g., an image) as a second layer into the workspace. Again, the layering module maintains the association between the projected information and the second layer.
The system detects 640 a change in the position of the physical object. For example, the system may include an infrared camera or a visible-light camera (e.g., digital still-image camera or a digital video camera) for detecting the position of the physical object. In response to detecting the change in position of the physical object, the system maintains 650 the visual adjacency between the first and second layers. For example, if a workspace is logically divided into four quadrants, a physical object might initially be detected as occupying the first quadrant. If the physical object (associated with one layer) is initially visually “on top of” projected digital information (associated with a different layer), then the system may abstain from projecting the digital information that would otherwise be located in the first quadrant, consistent with the visual order of the layers. However, in response to defecting movement of the physical object from the first quadrant to the second quadrant, the system may then project the digital information associated with the first quadrant while removing from the projection the digital information associated with the second quadrant, thereby maintaining the appearance that the physical object is “on top of” the projected digital information.
In examples described herein, methods and systems manage the visual adjacency of first and second layers. The methods and systems described herein are readily extendible to more than two layers. For example, additional layers of projected digital information could be managed by a projection capture system. Also, additional physical objects could each be associated with a distinct layer. The visual adjacency between these multiple physical and projected layers may be managed according to the embodiments described herein.
Various modifications may be made to the disclosed embodiments and implementations of the invention without departing from their scope. Therefore, the illustrations and examples herein should be construed in an illustrative, and not a restrictive sense.
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/US2011/045983 | 7/29/2011 | WO | 00 | 12/13/2013 |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2013/019190 | 2/7/2013 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
695460 | Jayne | Mar 1902 | A |
4986651 | Vanderwerf | Jan 1991 | A |
5511148 | Wellner | Apr 1996 | A |
6067112 | Wellner | May 2000 | A |
6431711 | Pinhanez | Aug 2002 | B1 |
6840627 | Olbrich | Jan 2005 | B2 |
6965460 | Gann et al. | Nov 2005 | B1 |
7023536 | Zhang et al. | Apr 2006 | B2 |
7038846 | Mandella et al. | May 2006 | B2 |
7088440 | Buermann et al. | Aug 2006 | B2 |
7110100 | Buermann et al. | Sep 2006 | B2 |
7113270 | Buermann et al. | Sep 2006 | B2 |
7161664 | Buermann et al. | Jan 2007 | B2 |
7203384 | Carl et al. | Apr 2007 | B2 |
7268956 | Mandella et al. | Sep 2007 | B2 |
7432917 | Wilson et al. | Oct 2008 | B2 |
7474809 | Carl et al. | Jan 2009 | B2 |
7589747 | Matsui et al. | Sep 2009 | B2 |
7626578 | Wilson et al. | Dec 2009 | B2 |
7729515 | Mandella et al. | Jun 2010 | B2 |
7743348 | Robbins | Jun 2010 | B2 |
7826641 | Mandella et al. | Nov 2010 | B2 |
7843470 | Bannai et al. | Nov 2010 | B2 |
7959300 | Hirahara et al. | Jun 2011 | B2 |
7961909 | Mandella et al. | Jun 2011 | B2 |
9069382 | Starner | Jun 2015 | B1 |
20020186221 | Bell | Dec 2002 | A1 |
20030035061 | Iwaki et al. | Feb 2003 | A1 |
20030092470 | Kurakane | May 2003 | A1 |
20040095562 | Moffatt | May 2004 | A1 |
20050168437 | Carl et al. | Aug 2005 | A1 |
20060010400 | Dehlin et al. | Jan 2006 | A1 |
20060126128 | Ahmed et al. | Jun 2006 | A1 |
20070008340 | Endo et al. | Jan 2007 | A1 |
20080030461 | Matsui et al. | Feb 2008 | A1 |
20080094588 | James et al. | Apr 2008 | A1 |
20080192017 | Hildebrandt et al. | Aug 2008 | A1 |
20090002342 | Terada et al. | Jan 2009 | A1 |
20090027337 | Hildreth | Jan 2009 | A1 |
20090115721 | Aull et al. | May 2009 | A1 |
20090262098 | Yamada | Oct 2009 | A1 |
20100017722 | Cohen | Jan 2010 | A1 |
20100157388 | Pishdadian et al. | Jun 2010 | A1 |
20100188474 | Robinson et al. | Jul 2010 | A1 |
20100259633 | Kii | Oct 2010 | A1 |
20100271394 | Howard | Oct 2010 | A1 |
20100299103 | Yoshikawa | Nov 2010 | A1 |
20100315825 | Kawamura | Dec 2010 | A1 |
20110025650 | Jocelyne | Feb 2011 | A1 |
20110063256 | Ye et al. | Mar 2011 | A1 |
20110169778 | Nungester et al. | Jul 2011 | A1 |
20110227915 | Mandella et al. | Sep 2011 | A1 |
20110234879 | Kashitani | Sep 2011 | A1 |
20120038549 | Mandella et al. | Feb 2012 | A1 |
20120054355 | Arrasvuori et al. | Mar 2012 | A1 |
20120069180 | Kawamura | Mar 2012 | A1 |
20130034484 | Zones et al. | Feb 2013 | A1 |
20130194418 | Gonzalez-Banos et al. | Aug 2013 | A1 |
20140139717 | Short | May 2014 | A1 |
20150042678 | Alt | Feb 2015 | A1 |
Number | Date | Country |
---|---|---|
101292516 | Oct 2008 | CN |
101810003 | Aug 2010 | CN |
10222436 | Aug 1998 | JP |
2003-131319 | May 2003 | JP |
2003-131319 | May 2003 | JP |
2004192623 | Jul 2004 | JP |
2004192623 | Jul 2004 | JP |
2006-189712 | Jul 2006 | JP |
2011081556 | Apr 2011 | JP |
1020110003705 | Jan 2011 | KR |
WO-2007006051 | Oct 2008 | WO |
WO-2009018161 | Feb 2009 | WO |
WO-2010137496 | Dec 2010 | WO |
Entry |
---|
Scott R. Klemmer, Katherine M. Everitt, James A. Landay, Integrating Physical and Digital Interactions on Walls for Fluid Design Collaboration, 2008, Human-Computer Interaction, 23(2):138-213. |
Bjorn Hartmann, Meredith R. Morris, Hrvoje Benko, Andrew D. Wilson, Augmenting Interactive Tables with Mice and Keyboards, 2009,Proceedings of the 22nd annual ACM symposium on User interface software and technology UIST '09, pp. 149-152. |
Tyler Denis, Superimpose a Car onto Another Image in Photoshop, 2011, retrieved from <<http://designinstruct.com/digital-art/photo-manipulation/superimpose-a-car-onto-another-image-in-photoshop/>> Accessed Apr. 2, 2015. |
Michael Wooldridge, Chapter 8: Working with Layers, 2009, Teach Yourself Visually Adobe Photoshop CS4, Chapter 8, pp. 168-199, ISBN: 9780470339473. |
Gao, Rui et al; Microsoft Research—Mobile Surface; Microsoft Research; 2010; http://research.microsoft.com/en-us/projects/mobilesurface/. |
Hand, Randall; Infinite Z Launches zSpace Virtual Holographic 3D Display for Designers; VizWorld.com; Dec. 13, 2011; http://www.vizworld.com/2011/12/infinite-launches-zspace-virtual-holographic-3d-display-designers/#sthash.j6Ys61PX.dpbs. |
Harrison, Beverly & Ziola, Ryder; Bringing Toys to Life: Intel Labs OASIS Project; Augmented Engineering; Jan. 26, 2011; http://augmentedengineering.wordpress.com/2011/01/26/bringing-toys-to-life-intel-labs-oasis-project/. |
Harrison, Chris et al; OmniTouch: Wearable Multitouch Interaction Everywhere; UIST'11; Oct. 16, 2011; Santa Barbara, California; http://research.microsoft.com/en-us/um/people/awilson/publications/HarrisonUIST2011/HarrisonUIST2011.html. |
Hartmann, Bjorn et al; Pictionaire: Supporting Collaborative Design Work by Integrating Physical and Digital Artifacts; CSCW 2010; Feb. 6, 2010; Savannah, Georgia; http://research.microsoft.com/en-us/um/people/awilson/publications/HartmannCSCW2010/HartmannCSCW2010.html. |
Hinckley, Ken et al; Pen + Touch = New Tools; UIST'10; Oct. 3, 2010; New York, New York; http://research.microsoft.com/en-us/um/people/awilson/publications/HinckleyUIST2010/HinckleyUIST2010.html. |
Izadi, Shahram et al; C-Slate: A Multi-Touch and Object Recognition System for Remote Collaboration Using Horizontal Surfaces; IEEE; 2007; http://research.microsoft.com/pubs/132551/cslate1.pdf. |
Kane, S.K. et al., Bonfire: a Nomadic System for Hybrid Laptop-tabletop interaction, (Research Paper), Oct. 4-7, 2009, pp. 129-138. |
Kane, Shaun K. et al; Bonfire: A Nomadic System for Hybrid Laptop-Tabletop Interaction; UIST'09; Oct. 4, 2009; Victoria, British Columbia, Canada; http://dub.washington.edu/djangosite/media/papers/uist09.pdf. |
Linder, Natan et al; LuminAR: Portable Robotic Augmented Reality Interface Design and Prototype; UIST'10, Oct. 3, 2010; New York, New York; http://fluid.media.mit.edu/sites/default/files/2010-10-03-luminar_uist10_demo.pdf. |
Melanson, Donald; Microsoft Research Working on Portable Surface; Mar. 2, 2010: http://www.engadget.com/2010/03/02/microsoft-research-working-on-portable-surface/. |
Melanson, Donald; Wiimote Repurposed for Multi-Point Interactive Whiteboard; Dec. 10, 2007; http://www.engadget.com/2007/12/10/wiimote-repurposed-for-multi-point-interactive-whiteboard/. |
Simonite, Tom; A Kitchen Countertop With a Brain; MIT Technology Review; Jul. 2, 2010; http://www.technologyreview.com/news/419639/a-kitchen-countertop-with-a-brain/. |
Wilson, Andrew D. et al; Combining Multiple Depth Cameras and Projectors for Interactions on, Above, and Between Surfaces; USTI'10; Oct. 3, 2010; New York, New York; http://research.rnicrosoft.com/en-us/um/people/awilson/publications/WilsonUIST2010/WilsonUIST2010.html. |
Wilson, Andrew D.; Using a Depth Camera as a Touch Sensor; ITS 2010: Devices & Algorithms; Nov. 7, 2010; Saarbrucken, Germany; http://research.microsoft.com/en-us/um/people/awilson/publications/WilsonITS2010/WilsonITS2010.html. |
Yuanyuan Ding et al; Catadioptric Projectors; http://www.eecis.udel.edu/˜jye/lab_research/09/cprojector.pdf. |
International Search Authority Search Report, dated Feb. 20, 2012. PCT Application No. PCT/US2011/045983. Filing Date Jul. 29, 2011. |
Anton Fuhrmann et al., “Occlusion in Collaborative Augmented Environments,” Computers & Graphics, vol. 23, No. 6, Dec. 1, 1999, pp. 809-819, Elsevier Science, Ltd. |
Daniel Kurz et al., “Mutual Occlusions on Table-top Displays in Mixed Reality Applications,” VRST '88, 1988, pp. 1-4, New York, New York, USA. |
Extended European Search Report, EP Application No. 11870359.4, dated Apr. 23, 2015, pp. 1-6. |
Ramesh Raskar et al., “The Office of the Future: A Unified Approach to Image-Based Modeling and Spatially Immersive Displays,” SIGGRAPH '98, Jul. 19-24, 1998, Orlando, Florida, USA, pp. 1-10. |
Sasa Junuzovic et al., Microsoft Research, IllumiShare, 2012, pp. 1-2, Microsoft. |
Number | Date | Country | |
---|---|---|---|
20140125704 A1 | May 2014 | US |