System for capturing and projecting images, use of the system and method for capturing, projecting and inserting images

Information

  • Patent Grant
  • 12101573
  • Patent Number
    12,101,573
  • Date Filed
    Friday, February 7, 2020
    4 years ago
  • Date Issued
    Tuesday, September 24, 2024
    2 months ago
  • Inventors
    • da Costa Marques Arrabal; Alexandre Thadeu
    • Barcelos Xavier Filho; Rodolpho
  • Original Assignees
  • Examiners
    • Gadomski; Stefan
    Agents
    • Miles; Craig R.
    • CR MILES P.C.
Abstract
The invention relates to a system and a method for capturing and projecting images for use in an integrated studio including a real location and panels which render these premises partially virtual, using images generated from outside to create an image that is part real and part virtual by showing the images.
Description
FIELD OF INVENTION

The invention refers to the field of television production, more specifically to image capture and projection systems and methods, and their use in a TV studio, and the extension of images through virtual projections.


STATE OF THE ART

Currently, the TV and cinema market uses the scenario extension technique, where part of the studio is physically assembled, and another part is created through graphics.


The two main ways to implement scene expansion today are through chrome-keying or through the use of projection screens.


Extending scenarios with Chroma-keying is one of the most used techniques today. In this case, the setting is made up of floors and walls of a homogeneous color, usually green or blue, and may or may not contain additional furniture. The process of replacing the image on the wall is done using a computer program. This way, everything that the camera sees in tone will be exchanged for an image, video, or virtual graphics.


U.S. Pat. No. 6,201,579 describes a virtual studio object position detection device, which comprises providing an object with an emitting device and using the device in conjunction with a patterned chroma-key plane or background to provide the exact position of a TV camera or alone to provide a position for another object or person.


Patent application WO1997028654 describes a video system that creates a virtual shadow of an object filmed in front of a blue screen with a main video camera. The virtual shadow is created using a second camera located as a virtual light source. This system uses the chroma-key color system.


The extension of scenarios with projection screens consists of a simpler and easier to apply technique, although quite restrictive. In this case, a panel or screen is placed as an image, video or virtual graphics, through the presenter who is inside the studio. This technique, however, limits the camera's movements and any wider movement can show the limits of the panel and harm the composition of the image. In this technique, use is restricted to cameras that are stationary or with minimal movement.


The present invention provides a system that allows a virtual image integrated with a real image, executed in real time.





FIGURES


FIG. 1 shows the scenario extension using LED panels.



FIG. 2 shows a perspective view of the scenographic elements such as the glass background and LED panels.



FIG. 3 shows a top view of the studio and the configuration of the panels and scenographic elements.



FIG. 4 shows infographics and an animated virtual character inserted into the video by graphics servers.



FIG. 5 shows the physical layout of the studio, with the anchor points of the panels.



FIG. 6 shows a configuration of the asymmetric frustrum projection technique.



FIG. 7 shows the asymmetric frustrum distortions.





DETAILED DESCRIPTION OF THE INVENTION

The present invention solves the aforementioned problems of studio size and camera movement by offering a more complete and robust approach, delivering a final image where the viewer is unable to perceive what is real or virtual and executing everything in real time, that is, permissible in live programs.


In this way, the solution presented expands the perception of space in the studio through the concept of scene extension, making use of LED screens, a glass background overlooking a real environment, cameras with movement tracking sensors and software responsible for controlling the images displayed on the screens.


LED Screens


LED screens are arranged laterally, perpendicular to the floor and rotated in relation to the background window, forming an angle of 120 degrees. This angle may vary, however it must be greater than 90 degrees.


LED technology was chosen due to its high resolution compared to other existing technologies. This resolution is greater not only in the number of pixels but also in the intensity of the colors. There is additional importance in the intensity of colors due to the incidence of this lighting on the real environment. This effect will contribute to the presenter's perception of insertion in the virtual environment.


Glass Background


The background is important to reinforce the effect of inserting the studio into a real environment.


Next to the glass back (1) and hidden behind the LED side panels (2, 3) are two static cameras (4, 5) that capture complementary images to those viewed through the glass back. These complementary images are sent to servers that distort them according to the studio camera's perspective and forward them for display on the side panels. When viewed through the studio camera, the images on the panels extend the view offered by the glass background. However, if viewed from any point of view other than the camera, they appear distorted and out of context.


Cameras with Motion Sensors


Two cameras with motion sensors (10, 11) are positioned inside the studio. They have the characteristics of not only tracking movements, but also promoting changes in the lens (zoom and focus, for example). This data is captured and sent to the graphics server that interprets it and, thus, deforms the images captured by the static cameras next to the glass background to be displayed on the screens according to the perspective of the cameras positioned inside the studio.


Sensors and tracking systems are commonly known in the art.


Software is used to generate and provide graphics for projections on LED screens.


One embodiment of the invention allows the composition of an entirely virtual studio, consisting of LED panels on the walls, ceiling and floor.


The term “Virtual insertions” can be understood as all computer graphics inserted on top of the image, such as: scoreboards, cards, objects, characters and graphic elements in general.


The physical studio (16) is made up of an entire front wall made of glass, and the other walls are opaque, behaving transparently. The effect sought would be to recreate a 180-degree studio (Panoramic) and, to achieve this, the side walls should be made of glass, or invisible.


The physical studio can have any desired size, and the LED panels are then positioned on each side, rotated approximately 30 degrees, with the panel being the hypotenuse of the triangle formed between it, the glass and the wall, with said angle formed between the wall and the panel.


The inclination of the panels avoids aberrations in the image due to the angle at which the LEDs light emitted, when observed by the camera, otherwise they could suffer a total reflection effect internally and appear dark, in addition to creating a hidden area to position between the panel and wall a static camera with a wide-angle lens (6, 7) (170 degrees) glued to the glass to capture the exterior in process (called stitching).


The floor made of LED boards allows the graphics to be changed, to display images or complete the current floor itself.


Application of filters and films (9) is carried out to reach luminance levels on the glass compatible with the equalization of the images generated in the LED and the images received through the glass on the front wall.


Also as an integral part of the studio's internal assembly, a sensorized crane was used for camera tracking purposes.


A graphical computing system (17) is used and is installed on servers that are located in the central art and receives the camera and lens tracking data by a connection, such as ethernet.


The process for the final effect consists of the following steps:

    • capture of images to complement the scenery through 1 or 2 or more cameras with wide-angle lens(es);
    • the images from these cameras go through a process called stitching, which consists of creating a single image by superimposing the common areas (12, 13) between the 2 cameras, resulting in a filmed panoramic image (14);
    • the panoramic image being sent to the computer graphics system (17), which projects it onto a virtual plane (15) that is moved away and scaled to overlap and artistically match the real image;
    • camera and lens tracking sensors are added, so that, in real time, the cameras used for transmission have their intrinsic and extrinsic data, which is informed to the computer graphics system;
    • for each transmission camera, a computer graphics system is required receiving all the information mentioned above;
    • each computer graphics system calculates the projective distortion of the panoramic image, which is projected onto the LED panels, always illustrating the current camera's point of view. This way the image of the LEDs merges with the real image of the window;
    • all video signals need to be adjusted by TV systems so that they have a calculated delay, so that the videos are displayed without delay and in the right order.


The computer graphics system is installed on servers located in the technical center and receives camera and lens tracking data (18) via a connection, such as ethernet.


The studio can be made up of two side LED panels, optionally an LED panel on the floor and a glass window at the back. The side panels are perpendicular to the floor and rotated in relation to the background window forming an angle of 120 degrees (for correct operation, the angle must be greater than 90 degrees).


Behind each side panel is one or more static cameras with one or more wide-angle lenses (6, 7) with a viewing angle approaching 180 degrees, pointing out of the studio through the glass window. The video captured by each of these cameras is fed into a processor and sent to the corresponding side panel. This processor, in addition to video from the rear camera on the side panel, also receives position and lens data from the studio's main camera (8).


In an alternative embodiment, only a static camera can be used to capture the external image.


In an alternative embodiment, the still camera may have one or more wide-angle lenses.


frustrum projection technique, available and implemented in all software and programming frameworks for real-time computer graphics already known in the art.


The distortion from frustrum to asymmetric is also implemented by commercially available software, represented by the function described below.


Once the system receives camera position data in relation to the studio, that is, geolocating the camera in the studio, a virtual camera is created with the same characteristics and changes in lens, camera and position occurring in real time, this data is duplicated and simultaneously feed two virtual systems. The first system processes everything related to distortions for the LED screens. The second system applies augmented reality elements.


The LED panel has a processing time to display the images, therefore, once system 2, responsible for virtual insertions, needs to be delayed so that system 1 projects the image complementary to the glass onto the LED, before System 2 Captures the studio image to make the virtual insertions. The delay, known as “delay”, is variable, being in the range of 3 frames in the worst case.


To enable the calculation of projective distortion and other resources, the servers in the technical center, where the systems are installed, are fed with tracking data from the camera and lens, for example via ethernet or any other compatible means, and with video, such as via video cables. The system is divided into modules: the projective distortion of the panoramic image is calculated by part of the system, which is commercially available software; and the 3D elements inserted in the studio are generated by systems that will not be described here.


The perception of a virtual window is due to the fact that there is always temporal coherence between the real image seen through the glass and the studio complemented by LED panels. The effect as described generates a perception of a real scenario, much larger than the physical space available with fluid graphic interventions.


From the foregoing, it will be seen that numerous modifications and variations can be made without departing from the true spirit and scope of the new concepts of the present invention. It should be understood that no limitations with respect to the specific embodiments illustrated are intended or should be inferred. The description is intended to cover all said modifications that fall within the scope of the invention.

Claims
  • 1. An image acquisition and projection system for a studio, comprising: a glass background window, Light Emitting Diode (LED) side panels disposed laterally with respect to the glass background window, perpendicular to the floor and rotated relative to the glass background window and forming an angle of greater than 90 degrees, the glass background window with a view to a real environment;cameras with motion tracking sensors positioned inside the studio and configured to capture images from within the studio;static cameras that are behind each LED side panel with one or more wide angle lenses with a view angle of 180 degrees, pointing outside the studio through the glass background window and configured to acquire complementary images to those viewed through the glass background window; anda computer system configured to generate and provide graphics from the complementary images and from tracking data from the motion tracking sensors for projections on the LED side panels.
  • 2. The image acquisition and projection system of claim 1, comprising side-walls, wherein the LED side panels are rotated 30 degrees with respect to the side-walls.
  • 3. The image acquisition and projection system of claim 1, wherein the computer system is configured to receive the complementary images from the static cameras and distort the complementary images according to a perspective of one of the cameras with motion tracking sensors and forward distorted complementary images for display on the LED side panels.
  • 4. The image acquisition and projection system of claim 1, wherein the static cameras are configured to capture images that complement the internal scene.
  • 5. The image acquisition and projection system of claim 1, wherein the complementary images of the static cameras comprise common overlapping zones suitable for image stitching to create a single image.
  • 6. The image acquisition and projection system of claim 1, wherein the static cameras are configured to create a panoramic image obtained by image stitching that is sent to the computer system that scales and projects the panoramic image onto a virtual plane to overlap and merge with the image viewed through the glass background window.
  • 7. The image acquisition and projection system of claim 1, wherein the cameras with motion tracking sensors are capable of tracking movements and promoting changes in the lens according to the movement and capturing the tracking data, wherein the computer system is configured to receive and interpret the tracking data.
  • 8. The image acquisition and projection system of claim 1, wherein the computer system is configured to receive the tracking data from the cameras with motion tracking sensors through a data connection.
  • 9. The image acquisition and projection system of claim 1, wherein the motion tracking sensors cameras further comprise camera and lens tracking sensors.
  • 10. An image acquisition and projection system according to claim 3, wherein each computer system calculates the projected distortion of the panoramic image obtained by stitching.
  • 11. The image acquisition and projection system according to claim 1, wherein videos displayed without delay and in the correct order are adjusted by TV systems having a calculated delay.
  • 12. A studio including the image acquisition and projection system according to claim 1, wherein the glass background window comprises filters and films to balance the light intensity of the view to the real environment transmitted through the glass background window and the projections on the LED side panels.
  • 13. A method of image acquisition, projection and insertion method using the system as defined in claim 1, comprising the steps of: capturing the complementary images to those viewed through the glass background window by the static cameras and creating a panoramic image by image stitching;sending the panoramic image obtained by image stitching to the computer system and scaling and projecting the panoramic image onto a virtual plane to overlap and merge with the image viewed through the glass background window;using the cameras with motion tracking sensors for real-time transmission having intrinsic and extrinsic data input to the computer system;calculating the projected distortion of the panoramic image by the computer system;setting video signals by TV systems, which calculate a delay; anddisplay of the video signals without delay calculated at previous step and in the correct order.
Priority Claims (1)
Number Date Country Kind
1020190064650 Mar 2019 BR national
PCT Information
Filing Document Filing Date Country Kind
PCT/BR2020/000006 2/7/2020 WO
Publishing Document Publishing Date Country Kind
WO2020/198823 10/8/2020 WO A
US Referenced Citations (27)
Number Name Date Kind
6191812 Tzidon Feb 2001 B1
6752498 Covannon Jun 2004 B2
7468778 Thomas et al. Dec 2008 B2
8269900 O'Connell Sep 2012 B2
9615054 McNelley Apr 2017 B1
10286309 Lee May 2019 B2
10898818 Karafin Jan 2021 B2
20030224333 Vastvedt Dec 2003 A1
20070009222 Koo Jan 2007 A1
20070098368 Carley May 2007 A1
20080178501 Crowell Jul 2008 A1
20080246759 Summers Oct 2008 A1
20100014053 Brentnall, III Jan 2010 A1
20110128555 Rotschild Jun 2011 A1
20110157297 O'Connell Jun 2011 A1
20120188637 Joseph Jul 2012 A1
20130229482 Vilcovsky Sep 2013 A1
20140118336 Smithwick May 2014 A1
20140300830 Wang Oct 2014 A1
20150029314 Reichow Jan 2015 A1
20150077592 Fahey Mar 2015 A1
20160343166 Inoko Nov 2016 A1
20170150108 Kong May 2017 A1
20170223312 McNelley Aug 2017 A1
20170251172 McNelley Aug 2017 A1
20180176506 McNelley Jun 2018 A1
20190141291 McNelley May 2019 A1
Foreign Referenced Citations (10)
Number Date Country
202608633 Dec 2012 CN
203658722 Jun 2014 CN
204889366 Dec 2015 CN
207082314 Mar 2018 CN
0956709 Apr 2002 EP
2408191 Jan 2012 EP
2006047610 Feb 2006 JP
20130142458 Dec 2013 KR
101433751 Aug 2014 KR
101593136 Feb 2016 KR
Non-Patent Literature Citations (1)
Entry
PCT International Patent Application No. PCT/BR20/00006, International Preliminary Report on Patentability and International Search Report (English Translation) mailed Apr. 14, 2020, 12 pages.
Related Publications (1)
Number Date Country
20220150386 A1 May 2022 US