The invention refers to the field of television production, more specifically to image capture and projection systems and methods, and their use in a TV studio, and the extension of images through virtual projections.
Currently, the TV and cinema market uses the scenario extension technique, where part of the studio is physically assembled, and another part is created through graphics.
The two main ways to implement scene expansion today are through chrome-keying or through the use of projection screens.
Extending scenarios with Chroma-keying is one of the most used techniques today. In this case, the setting is made up of floors and walls of a homogeneous color, usually green or blue, and may or may not contain additional furniture. The process of replacing the image on the wall is done using a computer program. This way, everything that the camera sees in tone will be exchanged for an image, video, or virtual graphics.
U.S. Pat. No. 6,201,579 describes a virtual studio object position detection device, which comprises providing an object with an emitting device and using the device in conjunction with a patterned chroma-key plane or background to provide the exact position of a TV camera or alone to provide a position for another object or person.
Patent application WO1997028654 describes a video system that creates a virtual shadow of an object filmed in front of a blue screen with a main video camera. The virtual shadow is created using a second camera located as a virtual light source. This system uses the chroma-key color system.
The extension of scenarios with projection screens consists of a simpler and easier to apply technique, although quite restrictive. In this case, a panel or screen is placed as an image, video or virtual graphics, through the presenter who is inside the studio. This technique, however, limits the camera's movements and any wider movement can show the limits of the panel and harm the composition of the image. In this technique, use is restricted to cameras that are stationary or with minimal movement.
The present invention provides a system that allows a virtual image integrated with a real image, executed in real time.
The present invention solves the aforementioned problems of studio size and camera movement by offering a more complete and robust approach, delivering a final image where the viewer is unable to perceive what is real or virtual and executing everything in real time, that is, permissible in live programs.
In this way, the solution presented expands the perception of space in the studio through the concept of scene extension, making use of LED screens, a glass background overlooking a real environment, cameras with movement tracking sensors and software responsible for controlling the images displayed on the screens.
LED Screens
LED screens are arranged laterally, perpendicular to the floor and rotated in relation to the background window, forming an angle of 120 degrees. This angle may vary, however it must be greater than 90 degrees.
LED technology was chosen due to its high resolution compared to other existing technologies. This resolution is greater not only in the number of pixels but also in the intensity of the colors. There is additional importance in the intensity of colors due to the incidence of this lighting on the real environment. This effect will contribute to the presenter's perception of insertion in the virtual environment.
Glass Background
The background is important to reinforce the effect of inserting the studio into a real environment.
Next to the glass back (1) and hidden behind the LED side panels (2, 3) are two static cameras (4, 5) that capture complementary images to those viewed through the glass back. These complementary images are sent to servers that distort them according to the studio camera's perspective and forward them for display on the side panels. When viewed through the studio camera, the images on the panels extend the view offered by the glass background. However, if viewed from any point of view other than the camera, they appear distorted and out of context.
Cameras with Motion Sensors
Two cameras with motion sensors (10, 11) are positioned inside the studio. They have the characteristics of not only tracking movements, but also promoting changes in the lens (zoom and focus, for example). This data is captured and sent to the graphics server that interprets it and, thus, deforms the images captured by the static cameras next to the glass background to be displayed on the screens according to the perspective of the cameras positioned inside the studio.
Sensors and tracking systems are commonly known in the art.
Software is used to generate and provide graphics for projections on LED screens.
One embodiment of the invention allows the composition of an entirely virtual studio, consisting of LED panels on the walls, ceiling and floor.
The term “Virtual insertions” can be understood as all computer graphics inserted on top of the image, such as: scoreboards, cards, objects, characters and graphic elements in general.
The physical studio (16) is made up of an entire front wall made of glass, and the other walls are opaque, behaving transparently. The effect sought would be to recreate a 180-degree studio (Panoramic) and, to achieve this, the side walls should be made of glass, or invisible.
The physical studio can have any desired size, and the LED panels are then positioned on each side, rotated approximately 30 degrees, with the panel being the hypotenuse of the triangle formed between it, the glass and the wall, with said angle formed between the wall and the panel.
The inclination of the panels avoids aberrations in the image due to the angle at which the LEDs light emitted, when observed by the camera, otherwise they could suffer a total reflection effect internally and appear dark, in addition to creating a hidden area to position between the panel and wall a static camera with a wide-angle lens (6, 7) (170 degrees) glued to the glass to capture the exterior in process (called stitching).
The floor made of LED boards allows the graphics to be changed, to display images or complete the current floor itself.
Application of filters and films (9) is carried out to reach luminance levels on the glass compatible with the equalization of the images generated in the LED and the images received through the glass on the front wall.
Also as an integral part of the studio's internal assembly, a sensorized crane was used for camera tracking purposes.
A graphical computing system (17) is used and is installed on servers that are located in the central art and receives the camera and lens tracking data by a connection, such as ethernet.
The process for the final effect consists of the following steps:
The computer graphics system is installed on servers located in the technical center and receives camera and lens tracking data (18) via a connection, such as ethernet.
The studio can be made up of two side LED panels, optionally an LED panel on the floor and a glass window at the back. The side panels are perpendicular to the floor and rotated in relation to the background window forming an angle of 120 degrees (for correct operation, the angle must be greater than 90 degrees).
Behind each side panel is one or more static cameras with one or more wide-angle lenses (6, 7) with a viewing angle approaching 180 degrees, pointing out of the studio through the glass window. The video captured by each of these cameras is fed into a processor and sent to the corresponding side panel. This processor, in addition to video from the rear camera on the side panel, also receives position and lens data from the studio's main camera (8).
In an alternative embodiment, only a static camera can be used to capture the external image.
In an alternative embodiment, the still camera may have one or more wide-angle lenses.
frustrum projection technique, available and implemented in all software and programming frameworks for real-time computer graphics already known in the art.
The distortion from frustrum to asymmetric is also implemented by commercially available software, represented by the function described below.
Once the system receives camera position data in relation to the studio, that is, geolocating the camera in the studio, a virtual camera is created with the same characteristics and changes in lens, camera and position occurring in real time, this data is duplicated and simultaneously feed two virtual systems. The first system processes everything related to distortions for the LED screens. The second system applies augmented reality elements.
The LED panel has a processing time to display the images, therefore, once system 2, responsible for virtual insertions, needs to be delayed so that system 1 projects the image complementary to the glass onto the LED, before System 2 Captures the studio image to make the virtual insertions. The delay, known as “delay”, is variable, being in the range of 3 frames in the worst case.
To enable the calculation of projective distortion and other resources, the servers in the technical center, where the systems are installed, are fed with tracking data from the camera and lens, for example via ethernet or any other compatible means, and with video, such as via video cables. The system is divided into modules: the projective distortion of the panoramic image is calculated by part of the system, which is commercially available software; and the 3D elements inserted in the studio are generated by systems that will not be described here.
The perception of a virtual window is due to the fact that there is always temporal coherence between the real image seen through the glass and the studio complemented by LED panels. The effect as described generates a perception of a real scenario, much larger than the physical space available with fluid graphic interventions.
From the foregoing, it will be seen that numerous modifications and variations can be made without departing from the true spirit and scope of the new concepts of the present invention. It should be understood that no limitations with respect to the specific embodiments illustrated are intended or should be inferred. The description is intended to cover all said modifications that fall within the scope of the invention.
Number | Date | Country | Kind |
---|---|---|---|
1020190064650 | Mar 2019 | BR | national |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/BR2020/000006 | 2/7/2020 | WO |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2020/198823 | 10/8/2020 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
6191812 | Tzidon | Feb 2001 | B1 |
6752498 | Covannon | Jun 2004 | B2 |
7468778 | Thomas et al. | Dec 2008 | B2 |
8269900 | O'Connell | Sep 2012 | B2 |
9615054 | McNelley | Apr 2017 | B1 |
10286309 | Lee | May 2019 | B2 |
10898818 | Karafin | Jan 2021 | B2 |
20030224333 | Vastvedt | Dec 2003 | A1 |
20070009222 | Koo | Jan 2007 | A1 |
20070098368 | Carley | May 2007 | A1 |
20080178501 | Crowell | Jul 2008 | A1 |
20080246759 | Summers | Oct 2008 | A1 |
20100014053 | Brentnall, III | Jan 2010 | A1 |
20110128555 | Rotschild | Jun 2011 | A1 |
20110157297 | O'Connell | Jun 2011 | A1 |
20120188637 | Joseph | Jul 2012 | A1 |
20130229482 | Vilcovsky | Sep 2013 | A1 |
20140118336 | Smithwick | May 2014 | A1 |
20140300830 | Wang | Oct 2014 | A1 |
20150029314 | Reichow | Jan 2015 | A1 |
20150077592 | Fahey | Mar 2015 | A1 |
20160343166 | Inoko | Nov 2016 | A1 |
20170150108 | Kong | May 2017 | A1 |
20170223312 | McNelley | Aug 2017 | A1 |
20170251172 | McNelley | Aug 2017 | A1 |
20180176506 | McNelley | Jun 2018 | A1 |
20190141291 | McNelley | May 2019 | A1 |
Number | Date | Country |
---|---|---|
202608633 | Dec 2012 | CN |
203658722 | Jun 2014 | CN |
204889366 | Dec 2015 | CN |
207082314 | Mar 2018 | CN |
0956709 | Apr 2002 | EP |
2408191 | Jan 2012 | EP |
2006047610 | Feb 2006 | JP |
20130142458 | Dec 2013 | KR |
101433751 | Aug 2014 | KR |
101593136 | Feb 2016 | KR |
Entry |
---|
PCT International Patent Application No. PCT/BR20/00006, International Preliminary Report on Patentability and International Search Report (English Translation) mailed Apr. 14, 2020, 12 pages. |
Number | Date | Country | |
---|---|---|---|
20220150386 A1 | May 2022 | US |