With current technology, users are able to view streaming video content in various different locations and on numerous different client devices, from television sets to laptop computers, tablet computers, smartphones, and other devices. Users have great flexibility in where and when they will view content. However, those users are not necessarily provided a significant amount of flexibility in the views that are displayed to them; those views are generally determined by a director or other content creator. To some extent, user control over the presented views is limited by bandwidth. Streaming multiple views of a video to a user and allowing the user to select which view to display would consume an excessive amount of bandwidth. Alternatively, if a user were to request different streams for display at different times, bandwidth consumption would be more limited, but the user may experience an unsatisfactory level of latency between the time when a new stream is requested by the user and the time when the new stream has been retrieved from a distant server and is displayed on the client device. It would be desirable to provide users with more flexible options for obtaining different views in a video and for providing a high level of quality for that video while making efficient use of limited available bandwidth.
Exemplary systems and methods are described to enable viewers of video presentations to select from among a plurality of available objects of interest in order to view, with minimal latency, zoomed versions of the selected object or objects of interest. In some embodiments, different zoom factors may be selected for the zoomed content in response to user input.
In exemplary embodiments, zoomable areas within a video are signaled by a video encoder or other content server to a client device. The client device enables the end customer to choose a zoom region and with little latency (e.g. within a single frame time) switches over to either a video stream based on local processed video in baseband or based on an incoming alternative video stream representative of the chosen zoomed stream.
In some embodiments, a client device is receiving streaming video along with metadata that identifies locations (e.g. bounding boxes) of objects of interest within the streaming video. The metadata may include information identifying one or more alternative streams associated with each object of interest. An alternative stream associated with an object of interest may be, for example, a video of a zoomed version of the respective object of interest.
In an exemplary embodiment, the client device receives from the user a selection of an object of interest to zoom into. In response, the client may retrieve the associated stream identified in the metadata (e.g. by issuing an HTTP request or by other means). The client decodes the alternative stream or streams representing the zoomed object of interest and causes display of the content to the user, e.g. by directly displaying the content on a screen of the client or by sending video data (e.g. uncompressed video) to an external display device.
In some embodiments, the display of the selected zoomed content may be rendered forward in time after the user selection at the zoomed resolution in a full screen display mode, such that the selected object of interest continues to be displayed in a zoomed format until, e.g, a user instructs otherwise or video of the object of interest is no longer available. In some embodiments, the selected object of interest remains substantially centered on the screen even if the object changes position with respect to the un-zoomed video.
In some embodiments, the zoomed content is displayed simultaneously with the original un-zoomed content. For example, the zoomed content may be displayed in a picture-in-picture relationship with the un-zoomed content (with either the zoomed or the un-zoomed content being displayed in the picture-in-picture inset), in a side-by-side presentation, in an alpha-blended combination, or otherwise displayed simultaneously to the viewer.
In exemplary embodiments, the zoomed content is retrieved as a separate stream, different from the stream encoding the un-zoomed content. In some embodiments, a manifest file, such as an MPEG DASH media presentation description (MPD) file is provided to the client. The manifest file may include information identifying one or more available video segments that encode the un-zoomed content and one or more available video segments that encode zoomed video of an associated object of interest. Available zoomed and un-zoomed streams may be identified using a URL or a template from which a URL may be generated. A manifest may further include data (e.g. in XML format) identifying an object of interest with which a stream is associated, or data indicating that the stream is an un-zoomed stream. Various different streams may be identified in a manifest file, including streams encoded at different bitrates to enable adaptive bitrate streaming.
In one exemplary method performed by a client device, the device receives an original video stream over a network and causes display of the original video. The client also receives, over the network, information identifying at least one object of interest and a spatial position of the object of interest within the original video stream. The device detects a user input selecting the object of interest and, in response, generates a first zoomed video of the object of interest from the received video stream and causes display of the first zoomed video.
In another exemplary method, a client receives a primary video stream. The client also receives tracking information for a plurality of tracked objects in the primary video stream, the tracking information indicating a time-varying position of each of the tracked objects (e.g. coordinates of the objects themselves or of bounding boxes corresponding to the objects). The client displays the primary video stream. While the primary stream is being displayed, the client receives a user input (e.g. on a touch screen) selecting (i) a spatial location in the primary video stream and (ii) a zoom factor. Based on the selected spatial location, the client identifies which tracked object is being selected. Based on the time-varying position and the selected zoom factor, the client crops and upscales the primary video stream to generate a first cropped and upscaled video stream that follows the selected tracked object at the selected zoom factor, and the client displays the first cropped and upscaled video stream.
In some such embodiments, while the first cropped and upscaled video stream is displayed, the client receives a subsequent user input to change the zoom factor to a subsequent zoom factor. In response to the subsequent user input, the client uses the time-varying position and the subsequent zoom factor to crop and upscale the primary video stream to generate a second cropped and upscaled video stream that follows the selected tracked object at the subsequent zoom factor. The client then switches to display of the second cropped and upscaled video stream.
In some embodiments, the cropping is performed so as to keep the tracked selected tracked object substantially centered in the cropped and upscaled video stream even as the selected tracked object moves with respect to the primary video stream.
In some embodiments, the client receives user input selecting two or more spatial locations in the primary video stream and identifies a tracked object for each spatial location. In such an embodiment, the cropping and upscaling is performed based on the time-varying position of the two or more selected objects to generate a cropped and upscaled video stream that follows the two or more selected objects at the selected zoom factor.
In some embodiments, the primary video stream includes a plurality of segments, and metadata regarding tracked objects in each segment is provided on a segment-by-segment basis. Where the tracked objects are players in a sporting event, the metadata may provide, for example, names and/or biographical information regarding the players.
Further described herein are various user features that may be used to enable user selection and/or control of video display options. In some embodiments, such options may include scaling, alpha blending, zooming and/or cropping in order to display the zoomed content according to the users commands and/or preferences.
Exemplary embodiments further include client devices comprising a network interface, a processor, and a non-transitory storage medium storing instructions operative when executed on the processor to perform methods described herein. Such a client device may include a touch screen for display and user input.
Various techniques described herein enable the use of high resolution, high frame-rate and/or high bit depth capture of content to be repurposed to provide useful features to end consumers of video. In some embodiments, those features may include generating additional or alternate views for the end user, such as zoomed views of particular objects or regions of interest.
Exemplary User Interface Display.
In some embodiments, the user interface is operative to enable a user to map zoom buttons to selected objects of interest (e.g., the user's favorite athletes). In some embodiments, information is provided to a video player that tracks the objects of interest (e.g. favorite athletes) and identify which athlete is on screen at a given time. Such information may be generated using various techniques. For example, RFID tags may be used to track individual athlete locations at the sporting venue, and such location information may be compared to tracked camera location and orientation information to determine whether or not an athlete is visible in a particular camera at different points in time. The information tracking the location of objects of interest may be accompanied by identification information (IDs) identifying the object of interest (e.g. identifying a particular athlete). In response to receipt of information indicating that an object of interest is currently on-screen, a user interface button or other user interface element corresponding to the user may be activated (e.g. the button may appear on the screen, or the button may change from an inactive state to an enabled state). The button or other user interface element may disappear or may otherwise be disabled in response to receipt of information indicating that the object of interest is no longer visible on the screen.
In some embodiments, there may be many objects or athletes with available zoomed content streams. In some such embodiments, user preferences or past user viewing history may be used to limit the zoom controls to the athletes or objects that are favorites of the user or are otherwise selected explicitly or implicitly by the user. Objects of interest may be selected, for example, in response to a determination that the user has explicitly indicated a preference to follow those objects, or in response to a determination that the user has followed or zoomed into in those objects of interest in past sessions.
In some embodiments, in response to selection by a user of an active zoom button, the next frame of video displayed after the selection is a frame of zoomed video that corresponds to the zoom-selected object.
Exemplary Pinch and Zoom Functionality.
Although in some embodiments, zooming in on a particular athlete or other object of interest is accomplished through activation of a button displayed on the screen, other embodiments use different techniques to detect a user request for zooming functionality. A technique that is employed in some embodiments is a pinch-and-zoom user input. One such embodiment is illustrated with respect to
A client display device 200 presenting a streaming video of a soccer game is illustrated in
As illustrated in
The information identifying the objects of interest and providing accompanying information may be periodically sent as an alternative very low bit rate stream from the server, e.g. as a sequence of JPEGs. Alternatively, the coordinates of the objects of interest or of regions enclosing respective objects of interest may be communicated in-stream in the private data section of the video. Other techniques may alternatively be used to communicate the location of objects of interest. The player then renders the highlights in response to an interactive mode being enabled by the user. The user may then be provided with a “pinch and zoom” or other input capability on a touch screen or other input device that will play the zoom stream corresponding to that player.
For example, in some embodiments, a client device is equipped with a touch-screen display. The client device continually monitors for touch input by a user. Display of a zoomed stream associated with an object of interest may be triggered in various ways. For example, the client device may detect that (i) a user has touched the screen substantially simultaneously at two different touch points, (ii) an object of interest (or at least a portion of such an object) is located between the two touch points, and (iii) the distance between the touch points is increasing. In response to detection of an appropriate touch input, the client device may begin display of the zoomed stream corresponding to the requested object of interest. As described in further detail below, the level of zoom (or zoom factor) of the zoomed stream may be a predetermined default level of zoom, or the level of zoom may be determined at least in part on the touch screen user input. For example, the level of zoom may be proportional to the change in distance between the touch points, such that the user can achieve a low level of zoom by moving his fingers slightly apart on the screen and can achieve a greater level of zoom by moving his fingers farther apart on the screen.
In response to user selection of a zoomed stream corresponding to one of the objects of interest, the client device display may appear as illustrated in
As illustrated in
As illustrated in
The manifest file provided in step 308 includes information identifying one or more tracked objects in the video. Information identifying the tracked objects may be displayed to the user in step 314. As the first video is displayed to the user client in step 316, the user in step 318 provides an input to select a tracked object, for example using a pinch and zoom operation. In response to receiving the pinch and zoom input (step 320), the video player crops and upscales the original un-zoomed stream in step 322 to produce a locally-generated version of the zoomed stream (e.g. cropping areas outside the region of interest and upscaling the cropped un-zoomed video by an appropriate zoom factor). The locally-generated zoomed video may not have a video quality as high as the video quality of the dedicated zoomed stream due to the upscaling. However, because the locally-generated zoomed video is generated from a video stream that is already being received by the client device, the locally-generated zoomed video may be displayed (step 324) with little or no perceptible delay following the selection by the user. For example, in some embodiments, the next displayed frame of video following user selection may be a frame of the locally-generated zoomed video.
In parallel with the generation of the locally-generated zoomed video, the video player in step 326 requests from the server the zoomed stream that corresponds to the selected object of interest. The zoomed stream may have a higher resolution (or a higher quality in other respects) than the corresponding locally-generated zoomed video. However, there is likely to be some latency in retrieval (step 328) of the zoomed stream from the server. In an exemplary embodiment, the client device continues playing the locally-generated zoomed video until the device determines that a sufficient amount of the zoomed stream has been retrieved to permit display of the zoomed stream. The client device then switches to display of the zoomed stream in step 330.
In some exemplary embodiments, the rendering of the content may be performed with the zoomed stream being played as full-screen content. In other embodiments, the zoomed content may be rendered as an insert (e.g. picture-in-picture) in the un-zoomed content. In some embodiments, users are provided with the ability to select among different display options (e.g. full-screen display versus picture-in-picture display or side-by-side display). The zoomed content may be cropped differently for different display configurations.
A method performed by a client device is illustrated in the flow chart of
In response to the selection, the device scales and crops the un-zoomed video in step 408 to generate a low-resolution zoomed view. An example of a low-resolution zoomed view is illustrated in box 410. The low-resolution zoomed view may be displayed with little to no perceptible latency. For example, the low-resolution zoomed view may be displayed in the next frame after selection by the user is received. In parallel, the client device in step 412 retrieves a high-resolution zoomed stream (e.g. from a content server over a network). When a sufficient amount of the high-resolution zoomed stream has been received, the client device performs a synchronized switchover in step 414 to display of the high-resolution zoomed stream. An example of a high-resolution zoomed view is illustrated in box 416.
Anticipatory Streaming of Zoomed Streams.
In some exemplary embodiments described above, separate zoomed streams are not requested until a user request for a zoomed stream is detected. In some embodiments, however, the client device may receive and decode low-bitrate versions of zoomed streams to be used in the event that the user selects an object of interest for zooming.
While client A is receiving, decoding, and displaying the main content, Client A may also be receiving and decoding relatively lower-bitrate channel-change streams, with each of the channel-change streams corresponding to a particular region of interest. In response to user selection of the region of interest, Client A may switch to display of the channel-change stream that corresponds to the selected region of interest. Because the channel-change stream was already being decoded, display of the zoomed content may begin with little or no perceptible delay, e.g. in the next frame after user selection. Using adaptive bitrate streaming techniques, Client A may subsequently retrieve the requested zoomed content at a higher bitrate as network conditions permit. A channel-change stream may also be available for the un-zoomed content. For example, Client B is receiving, decoding, and displaying a zoomed stream for region of interest ROI 1. Concurrently, Client B is receiving and decoding but not displaying a low-bitrate channel-change stream for the main content, allowing the user to switch back to display of the un-zoomed content with low latency.
Some embodiments are implemented using MPEG DASH. A simplified example of an MPD (a “pseudo MPD”) that may be used in some embodiments is described below. The exemplary MPD identifies different content representations for zoomed streams within a single DASH period and adaptation set. Other locations and methods for sharing this data, such as other types of manifest file, may alternatively be used.
A pseudo MPD describing a primary stream, a zoom stream and a fast channel change stream is described below. Descriptions of the streams referenced in the MPD are further provided below. Note that the exemplary channel-change stream has short segment length as well as low resolution and bitrate. The deriving of segment numbers for the individual duration and segment length may be performed through Segment Templates illustrated below. Parameters to segment template allow specification of short segments for the channel-change stream ‘cc1’.
This example includes three different representations for media of total length 300 seconds. For the primary view ‘primary’, a representation with 30 segments of length 10 second and video frame resolution of 1920×1080 and 6.8 Mbs is used. For a zoomed view ‘zoom1’ a representation with 30 segments of length 10 second and video frame resolution of 1920×1080 and 6.8 Mbs is used. For the channel change stream corresponding to zoom1, namely ‘cc1’, a representation with short segment size of 1 second and lower resolution and bitrate is used. The relevant representations may be identified as follows in a DASH MPD.
In the context of a DASH MPD, the above-listed elements may appear as follows.
An example of a client device that may be used in some embodiments is the client device illustrated in the functional block diagram of
Client Device for Locally-Generated Zoomed Video.
An example of another client device that may be used in some embodiments is the client device illustrated in the functional block diagram of
An embodiment according to
In some embodiments, coordinates of the regions of interest as illustrated in
In some embodiments, methods described above are combined to provide greater flexibility in the generation of zoomed video. For example, a content server may advertise the availability of a 2× zoomed stream and a 4× zoomed stream for a particular object of interest. In response to a user input indicating that the user wishes to view a 3× zoomed video of a selected object of interest, the client device may locally generate a 3× zoomed video of the object of interest and concurrently request the 2× zoomed stream. The client up-scales the 2× zoomed stream (by 1.5×) to generate a higher-quality 3× zoomed stream. The client switches to display of the higher-quality 3× zoomed stream once it is available. In some embodiments, the client retrieves the zoomed stream with the highest zoom factor that is less than or equal to the zoom factor requested by the user. In some embodiments, the client retrieves the zoomed stream with the lowest zoom factor that is greater than or equal to the zoom factor requested by the user. For example, in a case where the content server advertises the availability of a 2× zoomed stream and a 4× zoomed stream for the object of interest, the client device, in response to user input requesting a 3× zoomed video, may request the 4× zoomed stream and then down-scale the 4× zoomed stream to a 3× zoomed stream for display to the user.
Note that various hardware elements of one or more of the described embodiments are referred to as “modules” that carry out (i.e., perform, execute, and the like) various functions that are described herein in connection with the respective modules. As used herein, a module includes hardware (e.g., one or more processors, one or more microprocessors, one or more microcontrollers, one or more microchips, one or more application-specific integrated circuits (ASICs), one or more field programmable gate arrays (FPGAs), one or more memory devices) deemed suitable by those of skill in the relevant art for a given implementation. Each described module may also include instructions executable for carrying out the one or more functions described as being carried out by the respective module, and it is noted that those instructions could take the form of or include hardware (i.e., hardwired) instructions, firmware instructions, software instructions, and/or the like, and may be stored in any suitable non-transitory computer-readable medium or media, such as commonly referred to as RAM, ROM, etc.
Object-Tracked Single Stream.
In some exemplary embodiments, a single stream with an appropriate slice or tile structure may be used for the generation of a single stream of compressed video. In such embodiments, objects or regions of interest that have been identified at the headend may be automatically tracked, and the tracking information is sent down to the receiver. Using this tracking information, a receiver may either locally extract the zoom regions of interest or issue a request to the head end server for alternative ABR streams.
Constructing Zoomed Streams from High-Resolution Slices.
In some embodiments, a client device may operate to construct zoomed streams by requesting selected spatial regions (e.g. slices) from a high-resolution video made available by a server. One such embodiment is illustrated in
Display Modes.
In some embodiments, a client bases its selection of a download mechanism at least in part on the selected viewing mode. For example, if the zoomed view is in full screen mode, then the video to be processed may be the zoomed view alone (unless channel-change streams are also being processed). Also, in this case, the best resolution available is preferably presented for rendering. Hence, if the client chooses full screen view of the zoomed mode, then the specific zoomed stream will be requested from the server (e.g. specific slices may be requested from the server). The content received from the server is then decoded and rendered by the client. Until the zoomed stream is available, a local crop of the zoomed region may be upscaled and rendered to effect a low-latency (e.g. apparently instantaneous) switch.
Exemplary Embodiment with Default Initial Zoom Factor.
In one exemplary embodiment, a server advertises metadata with location and extent of each zoomable object. A video client displays indications/highlights of the zoomable objects. A user selects an object for zoom. The object (or, e.g., the corresponding bounding box), is highlighted to confirm its selection to the user. The object expands into a default zoomed view which is rendered forward in time. An initial zoom factor may be a default zoom factor and may be conveyed in metadata. As the user continues to watch the zoomed view of the selected object, the client may use subsequent metadata to keep the object of interest within the zoomed view. The user may adjust the extent of the zoom using a ‘pinch-to-zoom’ gesture on a touch screen or touch pad. Client rendering continues to center on the zoomed object, while adjusting to the new extent. The desired zoom factor at a given time may be realized either by local scaling of received content or by requesting and transitioning to a higher/lower resolution alternative stream, as applicable.
One such exemplary embodiment is illustrated in
In
In response to selection of the object of interest 716, using techniques described in greater detail above, the client device causes display of a zoomed video 718 substantially centered on the selected object of interest, as illustrated in
As illustrated with respect to
In a variation of the embodiment of
Exemplary Embodiment with User-Input of Initial Zoom Factor.
In some embodiments, a user input may operate to select both a particular object of interest and a zoom factor to be used for display of the object. In one such embodiment, illustrated in
In response to the input, as illustrated in
Exemplary Embodiment with Multi-Object Selection.
Some embodiments operate to enable aggregate selection of multiple objects for combined zoom view. In an exemplary embodiment illustrated with respect to
In this example, during the selection process, the client does not yet execute a zoom action. Instead, the client awaits further input. As illustrated in
In the embodiment of
As the aggregate zoomed video is displayed, the client device may monitor for client input changing the zoom factor. For example, the client may detect a pinch-to-zoom gesture to further adjust the zoom factor of the aggregate zoomed video.
In some embodiments, the client device automatically adjusts the zoomed view. One factor that may be used by the client device in adjusting the zoomed view is the changing position and spatial relationship of the selected objects of interest over time. For example, if the objects become farther apart in the un-zoomed video, the client may automatically reduce the zoom factor automatically (effectively “zooming out”) to maintain the full set of selected objects in view. Another factor that may be used by the client device in adjusting the zoomed view is detection that a selected object of interest is no longer displayed in the un-zoomed video. For example, if an object is no longer visible in the main view (and no longer available as an alternative zoomed stream), then the client may remove the object from the aggregate zoomed view. From that point forward, the view may be generated using a bounding box comprising the reduced set of selected objects, with the no-longer available object removed. The zoom factor may likewise be adjusted when an object is no longer visible, e.g. a previous “zoom out” action due to object divergence may be reversed if a divergent object disappears from the main view, e.g. travels beyond the edge of the main view video content.
In an alternative embodiment, detection by the client device of an expanding-pinch gesture does not immediately trigger an adjustment of the zoom level but rather selects one or multiple objects and establishes and displays a visible bounding rectangle defined by the gesture. In response to completion of the gesture, the client executes a zoom where the zoom factor is determined based on the selected objects and the extent of this user-defined bounding rectangle. This allows the multiple object selection and the initial setting of zoom factor using a single gesture.
Exemplary Client Device.
Exemplary embodiments disclosed herein are implemented using one or more wired and/or wireless network nodes, such as a wireless transmit/receive unit (WTRU) or other network entity.
The processor 1218 may be a general purpose processor, a special purpose processor, a conventional processor, a digital signal processor (DSP), a plurality of microprocessors, one or more microprocessors in association with a DSP core, a controller, a microcontroller, Application Specific Integrated Circuits (ASICs), Field Programmable Gate Array (FPGAs) circuits, any other type of integrated circuit (IC), a state machine, and the like. The processor 1218 may perform signal coding, data processing, power control, input/output processing, and/or any other functionality that enables the WTRU 1202 to operate in a wireless environment. The processor 1218 may be coupled to the transceiver 1220, which may be coupled to the transmit/receive element 1222. While
The transmit/receive element 1222 may be configured to transmit signals to, or receive signals from, a base station over the air interface 1216. For example, in one embodiment, the transmit/receive element 1222 may be an antenna configured to transmit and/or receive RF signals. In another embodiment, the transmit/receive element 1222 may be an emitter/detector configured to transmit and/or receive IR, UV, or visible light signals, as examples. In yet another embodiment, the transmit/receive element 1222 may be configured to transmit and receive both RF and light signals. It will be appreciated that the transmit/receive element 1222 may be configured to transmit and/or receive any combination of wireless signals.
In addition, although the transmit/receive element 1222 is depicted in
The transceiver 1220 may be configured to modulate the signals that are to be transmitted by the transmit/receive element 1222 and to demodulate the signals that are received by the transmit/receive element 1222. As noted above, the WTRU 1202 may have multi-mode capabilities. Thus, the transceiver 1220 may include multiple transceivers for enabling the WTRU 1202 to communicate via multiple RATs, such as UTRA and IEEE 802.11, as examples.
The processor 1218 of the WTRU 1202 may be coupled to, and may receive user input data from, the speaker/microphone 1224, the keypad 1226, and/or the display/touchpad 1228 (e.g., a liquid crystal display (LCD) display unit or organic light-emitting diode (OLED) display unit). The processor 1218 may also output user data to the speaker/microphone 1224, the keypad 1226, and/or the display/touchpad 1228. In addition, the processor 1218 may access information from, and store data in, any type of suitable memory, such as the non-removable memory 1230 and/or the removable memory 1232. The non-removable memory 1230 may include random-access memory (RAM), read-only memory (ROM), a hard disk, or any other type of memory storage device. The removable memory 1232 may include a subscriber identity module (SIM) card, a memory stick, a secure digital (SD) memory card, and the like. In other embodiments, the processor 1218 may access information from, and store data in, memory that is not physically located on the WTRU 1202, such as on a server or a home computer (not shown).
The processor 1218 may receive power from the power source 1234 and may be configured to distribute and/or control the power to the other components in the WTRU 1202. The power source 1234 may be any suitable device for powering the WTRU 1202. As examples, the power source 1234 may include one or more dry cell batteries (e.g., nickel-cadmium (NiCd), nickel-zinc (NiZn), nickel metal hydride (NiMH), lithium-ion (Li-ion), and the like), solar cells, fuel cells, and the like.
The processor 1218 may also be coupled to the GPS chipset 1236, which may be configured to provide location information (e.g., longitude and latitude) regarding the current location of the WTRU 1202. In addition to, or in lieu of, the information from the GPS chipset 1236, the WTRU 1202 may receive location information over the air interface 1216 from a base station and/or determine its location based on the timing of the signals being received from two or more nearby base stations. It will be appreciated that the WTRU 1202 may acquire location information by way of any suitable location-determination method while remaining consistent with an embodiment.
The processor 1218 may further be coupled to other peripherals 1238, which may include one or more software and/or hardware modules that provide additional features, functionality and/or wired or wireless connectivity. For example, the peripherals 1238 may include sensors such as an accelerometer, an e-compass, a satellite transceiver, a digital camera (for photographs or video), a universal serial bus (USB) port, a vibration device, a television transceiver, a hands free headset, a Bluetooth® module, a frequency modulated (FM) radio unit, a digital music player, a media player, a video game player module, an Internet browser, and the like.
Exemplary Server.
Communication interface 1392 may include one or more wired communication interfaces and/or one or more wireless-communication interfaces. With respect to wired communication, communication interface 1392 may include one or more interfaces such as Ethernet interfaces, as an example. With respect to wireless communication, communication interface 1392 may include components such as one or more antennae, one or more transceivers/chipsets designed and configured for one or more types of wireless (e.g., LTE) communication, and/or any other components deemed suitable by those of skill in the relevant art. And further with respect to wireless communication, communication interface 1392 may be equipped at a scale and with a configuration appropriate for acting on the network side—as opposed to the client side—of wireless communications (e.g., LTE communications, Wi-Fi communications, and the like). Thus, communication interface 1392 may include the appropriate equipment and circuitry (perhaps including multiple transceivers) for serving multiple mobile stations, UEs, or other access terminals in a coverage area.
Processor 1394 may include one or more processors of any type deemed suitable by those of skill in the relevant art, some examples including a general-purpose microprocessor and a dedicated DSP.
Data storage 1396 may take the form of any non-transitory computer-readable medium or combination of such media, some examples including flash memory, read-only memory (ROM), and random-access memory (RAM) to name but a few, as any one or more types of non-transitory data storage deemed suitable by those of skill in the relevant art could be used. As depicted in
Although features and elements are described above in particular combinations, one of ordinary skill in the art will appreciate that each feature or element can be used alone or in any combination with the other features and elements. In addition, the methods described herein may be implemented in a computer program, software, or firmware incorporated in a computer-readable medium for execution by a computer or processor. Examples of computer-readable storage media include, but are not limited to, a read only memory (ROM), a random access memory (RAM), a register, cache memory, semiconductor memory devices, magnetic media such as internal hard disks and removable disks, magneto-optical media, and optical media such as CD-ROM disks, and digital versatile disks (DVDs). A processor in association with software may be used to implement a radio frequency transceiver for use in a WTRU, UE, terminal, base station, RNC, or any host computer.
The present application is a national stage application under 35 U.S.C. 371 of International Application No. PCT/US2018/018540, entitled SYSTEMS AND METHODS FOR SELECTIVE OBJECT-OF-INTEREST ZOOMING IN STREAMING VIDEO, filed on PCT Feb. 17, 2018, which claims benefit under 35 U.S.C. § 119(e) from U.S. Provisional Patent Application Ser. No. 62/460,651, entitled “SYSTEMS AND METHODS FOR SELECTIVE OBJECT-OF-INTEREST ZOOMING IN STREAMING VIDEO,” filed Feb. 17, 2017, the entirety of which is incorporated herein by reference.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/US2018/018540 | 2/17/2018 | WO |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2018/152437 | 8/23/2018 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
5610653 | Abecassis | Mar 1997 | A |
5892554 | Dicicco | Apr 1999 | A |
6282713 | Kitsukawa | Aug 2001 | B1 |
6750919 | Rosser | Jun 2004 | B1 |
6973130 | Wee | Dec 2005 | B1 |
7079176 | Freeman | Jul 2006 | B1 |
7114174 | Brooks | Sep 2006 | B1 |
7117517 | Milazzo | Oct 2006 | B1 |
7343617 | Katcher | Mar 2008 | B1 |
7577980 | Kienzle | Aug 2009 | B2 |
7782344 | Whittaker | Aug 2010 | B2 |
7949032 | Frost | May 2011 | B1 |
8055785 | Liu | Nov 2011 | B2 |
8160077 | Traversat | Apr 2012 | B2 |
8243797 | Lin | Aug 2012 | B2 |
8331760 | Butcher | Dec 2012 | B2 |
8621000 | Adimatyam | Dec 2013 | B2 |
8705527 | Addepalli | Apr 2014 | B1 |
9047236 | Gigliotti | Jun 2015 | B2 |
9060187 | Wu | Jun 2015 | B2 |
9064313 | Seshadrinathan | Jun 2015 | B2 |
9137027 | Matthews | Sep 2015 | B2 |
9137558 | Gibbon | Sep 2015 | B2 |
9170707 | Laska | Oct 2015 | B1 |
9177225 | Cordova-Diba | Nov 2015 | B1 |
9218641 | Ellis | Dec 2015 | B1 |
9231838 | Ould-Brahim | Jan 2016 | B2 |
9288545 | Hill | Mar 2016 | B2 |
9317688 | Yegin | Apr 2016 | B2 |
9336060 | Nori | May 2016 | B2 |
9344888 | Starsinic | May 2016 | B2 |
9380623 | Kim | Jun 2016 | B2 |
9515372 | Chang | Dec 2016 | B2 |
9992553 | Bennett | Jun 2018 | B2 |
10152826 | Saito | Dec 2018 | B2 |
10530757 | Seed et al. | Jan 2020 | B2 |
11122027 | Seed et al. | Sep 2021 | B2 |
20020056136 | Wistendahl | May 2002 | A1 |
20020059588 | Huber | May 2002 | A1 |
20020065678 | Peliotis | May 2002 | A1 |
20020120931 | Huber | Aug 2002 | A1 |
20020120934 | Abrahams | Aug 2002 | A1 |
20020126990 | Rasmussen | Sep 2002 | A1 |
20020131511 | Zenoni | Sep 2002 | A1 |
20020133817 | Markel | Sep 2002 | A1 |
20020147987 | Reynolds | Oct 2002 | A1 |
20020174425 | Markel | Nov 2002 | A1 |
20030028873 | Lemmons | Feb 2003 | A1 |
20030149983 | Markel | Aug 2003 | A1 |
20040031062 | Lemmons | Feb 2004 | A1 |
20040032495 | Ortiz | Feb 2004 | A1 |
20040190779 | Sarachik | Sep 2004 | A1 |
20050093976 | Valleriano | May 2005 | A1 |
20050286466 | Tagg | Dec 2005 | A1 |
20060155842 | Yeung et al. | Jul 2006 | A1 |
20060193295 | White | Aug 2006 | A1 |
20070024706 | Brannon, Jr. | Feb 2007 | A1 |
20070042784 | Anderson | Feb 2007 | A1 |
20070077943 | Hamilla | Apr 2007 | A1 |
20070086669 | Berger | Apr 2007 | A1 |
20070101418 | Wood | May 2007 | A1 |
20070104369 | Weatherhead | May 2007 | A1 |
20070153739 | Zheng | Jul 2007 | A1 |
20070171921 | Wookey | Jul 2007 | A1 |
20070226761 | Zalewski | Sep 2007 | A1 |
20080077965 | Kamimaki | Mar 2008 | A1 |
20080127253 | Zhang | May 2008 | A1 |
20080134305 | Hinton | Jun 2008 | A1 |
20080225116 | Kang | Sep 2008 | A1 |
20080261711 | Tuxen | Oct 2008 | A1 |
20090002564 | Barnhoefer | Jan 2009 | A1 |
20090003718 | Liu | Jan 2009 | A1 |
20090007023 | Sundstrom | Jan 2009 | A1 |
20090046152 | Aman | Feb 2009 | A1 |
20090047902 | Nory | Feb 2009 | A1 |
20090201917 | Maes | Aug 2009 | A1 |
20090217339 | Kim | Aug 2009 | A1 |
20090220080 | Herne | Sep 2009 | A1 |
20090225164 | Renkis | Sep 2009 | A1 |
20090276805 | Andrews, II | Nov 2009 | A1 |
20090300692 | Mavlankar | Dec 2009 | A1 |
20090305708 | Matsui | Dec 2009 | A1 |
20090316795 | Chui | Dec 2009 | A1 |
20090322489 | Jones | Dec 2009 | A1 |
20100097221 | Kreiner | Apr 2010 | A1 |
20100188531 | Cordes | Jul 2010 | A1 |
20100232504 | Feng | Sep 2010 | A1 |
20100321389 | Gay | Dec 2010 | A1 |
20110013836 | Gefen | Jan 2011 | A1 |
20110016321 | Sundaram | Jan 2011 | A1 |
20110113480 | Ma | May 2011 | A1 |
20110145430 | Ha | Jun 2011 | A1 |
20110231653 | Bollay | Sep 2011 | A1 |
20110299832 | Butcher | Dec 2011 | A1 |
20110304772 | Dasher | Dec 2011 | A1 |
20120030637 | Dey | Feb 2012 | A1 |
20120047551 | Pattar | Feb 2012 | A1 |
20120062732 | Marman | Mar 2012 | A1 |
20120072952 | Vaysman | Mar 2012 | A1 |
20120078712 | Fontana | Mar 2012 | A1 |
20120117140 | Wang | May 2012 | A1 |
20120195574 | Wallace | Aug 2012 | A1 |
20130003576 | Segura | Jan 2013 | A1 |
20130016910 | Murata | Jan 2013 | A1 |
20130031582 | Tinsman | Jan 2013 | A1 |
20130036442 | Wingert | Feb 2013 | A1 |
20130050268 | Lohrenz | Feb 2013 | A1 |
20130061262 | Briggs | Mar 2013 | A1 |
20130065622 | Hwang | Mar 2013 | A1 |
20130091430 | Zhai | Apr 2013 | A1 |
20130091515 | Sakata | Apr 2013 | A1 |
20130142118 | Cherian | Jun 2013 | A1 |
20130155106 | Rolleston | Jun 2013 | A1 |
20130155847 | Li | Jun 2013 | A1 |
20130163507 | Hoshino | Jun 2013 | A1 |
20130167062 | Herring | Jun 2013 | A1 |
20130188503 | Anepu | Jul 2013 | A1 |
20130188515 | Pinheiro | Jul 2013 | A1 |
20130223339 | Nakahara | Aug 2013 | A1 |
20130246784 | Zhang | Sep 2013 | A1 |
20130266286 | Yu et al. | Oct 2013 | A1 |
20130312042 | Venson | Nov 2013 | A1 |
20130336305 | Yan | Dec 2013 | A1 |
20130342539 | Khan | Dec 2013 | A1 |
20140022074 | Balinski | Jan 2014 | A1 |
20140026052 | Thorwirth | Jan 2014 | A1 |
20140029462 | Stewart | Jan 2014 | A1 |
20140033074 | Thibaux | Jan 2014 | A1 |
20140045452 | Ma | Feb 2014 | A1 |
20140059457 | Min | Feb 2014 | A1 |
20140082054 | Denoual | Mar 2014 | A1 |
20140086217 | Park | Mar 2014 | A1 |
20140089990 | Van Deventer | Mar 2014 | A1 |
20140168056 | Swaminathan | Jun 2014 | A1 |
20140192204 | Glazer | Jul 2014 | A1 |
20140195918 | Friedlander | Jul 2014 | A1 |
20140204206 | Itoi | Jul 2014 | A1 |
20140204247 | Bilgen | Jul 2014 | A1 |
20140210714 | Kang | Jul 2014 | A1 |
20140215043 | Ryu | Jul 2014 | A1 |
20140245367 | Sasaki | Aug 2014 | A1 |
20140269581 | Song | Sep 2014 | A1 |
20140269932 | Su | Sep 2014 | A1 |
20140279068 | Systrom | Sep 2014 | A1 |
20140280781 | Gregotski | Sep 2014 | A1 |
20140282751 | Lee | Sep 2014 | A1 |
20140282909 | Cherian | Sep 2014 | A1 |
20150016504 | Auyeung | Jan 2015 | A1 |
20150032901 | Wang | Jan 2015 | A1 |
20150033312 | Seed | Jan 2015 | A1 |
20150062339 | Ostrom | Mar 2015 | A1 |
20150070587 | Emeott | Mar 2015 | A1 |
20150089023 | Phillips | Mar 2015 | A1 |
20150089072 | Phillips | Mar 2015 | A1 |
20150104155 | Bloch | Apr 2015 | A1 |
20150117229 | Avudainayagam | Apr 2015 | A1 |
20150145782 | Brown | May 2015 | A1 |
20150172775 | Yee | Jun 2015 | A1 |
20150179219 | Gao | Jun 2015 | A1 |
20150201222 | Mertens | Jul 2015 | A1 |
20150208103 | Guntur | Jul 2015 | A1 |
20150222815 | Wang | Aug 2015 | A1 |
20150264096 | Swaminathan | Sep 2015 | A1 |
20150268822 | Waggoner | Sep 2015 | A1 |
20150296250 | Casper | Oct 2015 | A1 |
20150341091 | Park | Nov 2015 | A1 |
20150348247 | McLaughlin | Dec 2015 | A1 |
20150350659 | Auyeung | Dec 2015 | A1 |
20150351084 | Werb | Dec 2015 | A1 |
20150382268 | Hampel | Dec 2015 | A1 |
20160042496 | Ichimi | Feb 2016 | A1 |
20160057494 | Hwang | Feb 2016 | A1 |
20160073106 | Su | Mar 2016 | A1 |
20160156949 | Hattori | Jun 2016 | A1 |
20160165309 | Van Brandenburg | Jun 2016 | A1 |
20160182593 | Denoual | Jun 2016 | A1 |
20160192433 | Deenoo | Jun 2016 | A1 |
20160217623 | Singh | Jul 2016 | A1 |
20160249269 | Niu | Aug 2016 | A1 |
20160249353 | Nakata | Aug 2016 | A1 |
20160253240 | Cocagne | Sep 2016 | A1 |
20160269771 | Bangma | Sep 2016 | A1 |
20160306036 | Johnson | Oct 2016 | A1 |
20160360172 | Tao | Dec 2016 | A1 |
20160360235 | Ramasubramonian | Dec 2016 | A1 |
20160364087 | Thompson | Dec 2016 | A1 |
20160366454 | Tatourian | Dec 2016 | A1 |
20170070735 | Ramasubramonian | Mar 2017 | A1 |
20170118540 | Thomas | Apr 2017 | A1 |
20170136948 | Sypitkowski | May 2017 | A1 |
20170163645 | Bradley et al. | Jun 2017 | A1 |
20170212583 | Krasadakis | Jul 2017 | A1 |
20170236288 | Sundaresan | Aug 2017 | A1 |
20170302719 | Chen | Oct 2017 | A1 |
20180098063 | Chen | Apr 2018 | A1 |
20180332632 | Vikberg | Nov 2018 | A1 |
20180343043 | Hakola | Nov 2018 | A1 |
20190036910 | Choyi | Jan 2019 | A1 |
20190320189 | Cooper | Oct 2019 | A1 |
20200014961 | Ramaswamy | Jan 2020 | A1 |
20210130629 | Suzuki | May 2021 | A1 |
Number | Date | Country |
---|---|---|
2936318 | Aug 2015 | CA |
101313578 | Nov 2008 | CN |
101331735 | Dec 2008 | CN |
101350923 | Jan 2009 | CN |
102238547 | Nov 2011 | CN |
102752877 | Oct 2012 | CN |
102907068 | Jan 2013 | CN |
102939573 | Feb 2013 | CN |
103190089 | Jul 2013 | CN |
103336663 | Oct 2013 | CN |
104012106 | Aug 2014 | CN |
104052788 | Sep 2014 | CN |
104092488 | Oct 2014 | CN |
104220975 | Dec 2014 | CN |
105103566 | Nov 2015 | CN |
105144727 | Dec 2015 | CN |
105264904 | Jan 2016 | CN |
105373331 | Mar 2016 | CN |
106023261 | Oct 2016 | CN |
106233745 | Dec 2016 | CN |
106416223 | Feb 2017 | CN |
1175785 | Jan 2002 | EP |
1443768 | Aug 2004 | EP |
2426645 | Mar 2012 | EP |
2530675 | Dec 2012 | EP |
2824883 | Jan 2015 | EP |
2824885 | Jan 2015 | EP |
2882110 | Jun 2015 | EP |
2919471 | Sep 2015 | EP |
3076672 | Oct 2016 | EP |
3110160 | Dec 2016 | EP |
2408196 | Jan 2017 | EP |
2004104351 | Apr 2004 | JP |
2007192988 | Aug 2007 | JP |
2009207114 | Sep 2009 | JP |
2013522965 | Jun 2013 | JP |
2014525191 | Sep 2014 | JP |
2014530534 | Nov 2014 | JP |
6313503 | Apr 2018 | JP |
20080041237 | May 2008 | KR |
20100065327 | Jun 2010 | KR |
20140043532 | Apr 2014 | KR |
20140098156 | Aug 2014 | KR |
20160021141 | Feb 2016 | KR |
20170128515 | Nov 2017 | KR |
2001018658 | Mar 2001 | WO |
2010021705 | Feb 2010 | WO |
2010021705 | Feb 2010 | WO |
2011082150 | Jul 2011 | WO |
2012021246 | Feb 2012 | WO |
2012114666 | Aug 2012 | WO |
2012141557 | Oct 2012 | WO |
2013142966 | Oct 2013 | WO |
2014057131 | Apr 2014 | WO |
2014067566 | May 2014 | WO |
2014111423 | Jul 2014 | WO |
2014183803 | Nov 2014 | WO |
2014190308 | Nov 2014 | WO |
2015013645 | Jan 2015 | WO |
2015013685 | Jan 2015 | WO |
2015014773 | Feb 2015 | WO |
2015059194 | Apr 2015 | WO |
2015122737 | Aug 2015 | WO |
2015130796 | Sep 2015 | WO |
2015197815 | Dec 2015 | WO |
2016040290 | Mar 2016 | WO |
2016040833 | Mar 2016 | WO |
2016098056 | Jun 2016 | WO |
2017061297 | Apr 2017 | WO |
2017196670 | Nov 2017 | WO |
Entry |
---|
Loungev Films, “HD 1080p—Nature Scenery Video”. YouTube link available at: www.youtube.com/watch?v=DQuhA5ZCV9M, published on Jan. 13, 2016, 3 pages. |
Flowplayer—“Bandwidth Detection”. Web Archived dated Dec. 24, 2016. Retrieved from http://web.archive.org/web/20161224034909/http://flash.flowplayer.org/plugins/streaming/bwcheck.html. |
International Preliminary Report on Patentability for PCT/US2018/020229 dated Sep. 10, 2019, 12 pages. |
International Organization for Standardization, “Information Technology—Dynamic Adaptive Streaming Over HTTP (DASH), Part 1: Media Presentation Description and Segment Formats/ Amendments 2: Spatial Relationship Description, Generalized URL Parameters and Other Extensions”. International Standard, ISO/IEC 23009-1:2015/FDAM 2:2015(E), Second Edition (Annex H), Feb. 25, 2015, 23 pages. |
IBC, “D-Zero—Low Latency ABR Streaming”. Sky Italia, available at: http://www.ibc.org/ibc-awards/ibc2016-innovation-award-entrants/ibc-innovation-award-for-content-delivery--2016-entrants, Sep. 2016. |
IBC, “True live OTT Solution Puts F1 in Pole Position”. Tata Communications, available at: http://www.ibc.org/ibc-awards/ibc2016-innovation-award-entrants/ibc-innovation-award-for-content-delivery--2016-entrants, Sep. 2016. |
SYE, “SYE”. Web Archive dated Sep. 5, 2016, available at: https://netinsight.net/start/technical-offer-sye/, 1 page. |
Sun, Lei, et. al., “The Dynamic Videobook: A Hierarchical Summarization for Surveillance Video”. IEEE International Conference on Image Processing, (2013), pp. 3963-3966. |
TATACOMMS . “Transforming media + entertainment (subtitles)”. Youtube link dated Sep. 10, 2015, available at: https://w,w.youtube.com/watch?v=KUxxPYocMiw, 19 pages. |
Tata Communications, “Media and Entertainment Services”. Tata Communications, Archived on Feb. 27, 2017. Retrieved from https://web.archive.org/web/20170217212843/https://www.tatacommunications.com/products-services/enterprises/media-entertainment-services, 1 page. |
International Search Report and Written Opinion of the International Searching Authority for PCT/US2018/018540 dated Apr. 6, 2018. |
Mavlankar, A. et. al., “An Interactive Region-Of-Interest Video Streaming System for Online Lecture Viewing”. Proceedings of 2010 IEEE 18th International Packet Video Workshop Dec. 13-14, 2010, pp. 64-68. |
Written Opinion of the International Preliminary Examining Authority for PCT/US2018/018540 dated Jan. 18, 2019, 7 pages. |
International Preliminary Report on Patentability for PCT/US2018/018540 completed Apr. 26, 2019, 7 pages. |
Mavlankar, A. et. al., “Video Streaming with Interactive Pan/Tilt/Zoom”, Oct. 2009. Available at: http://web.stanford.edu/˜bgirod/pdfs/Mavlankar_Girod_Chapter_Oct09.pdf. |
Mavlankar, A., et al., “An Interactive Region-of-Interest Video Streaming System for Online Lecture Viewing”. Proceedings of 2010 IEEE 18th International Packet Video Workshop, Dec. 13-14, 2010, pp. 64-71. |
International Search Report and The Written Opinion of the International Searching Authority for PCT/US2016/046317, dated Oct. 19, 2016, 11 pages. |
International Preliminary Report on Patentability for PCT/US2016/046317 dated Mar. 1, 2018. |
International Organization for Standardization, “Information Technology—Dynamic adaptive Streaming Over HTTP (DASH)—Part 3: Implementation Guidelines/AMD 1: Alignment with ISO/IEC 23009-1:2014 and Extensibility”. International Standard, ISO/IEC JTC 1/SC 29N, ISO/IEC TR 23009-3:2014/PDAM1, ISO/IEC JTC 1/SC 29/WG11, No. N15990, Mar. 2, 2016, 49 pages. |
International Search Report and Written Opinion of the International Searching Authority for PCT/US2017/040060 dated Sep. 11, 2017, 15 pages. |
Wang, X., et. al., “Multi-Stream Streaming in DASH (Multi-Stream DASH)”. Huawei Technology, LTD., Motion Picture Expert Group (MPEG) Meeting, No. m35885, Feb. 2015, 8 pages. |
International Preliminary Report on Patentability for PCT/US2017/040060 dated Jan. 1, 2019. |
International Search Report and Written Opinion of The International Searching Authority for PCT/US2017/026388 dated Jul. 24, 2017, 14 Pages. |
Written Opinion of the International Preliminary Examining Authority for PCT/US2017/026388 dated Mar. 5, 2018. |
International Preliminary Report on Patentability for PCT/US2017/026388 completed on Jul. 12, 2018. |
Yilmaz, A., et. al., “Object Tracking—A Survey”. ACM Computing Surveys, vol. 38, No. 4, Article 13, Dec. 2006, pp. 1-45. |
International Organization for Standardization, “Information Technology—Dynamic Adaptive Streaming Over HTTP (DASH), Part 1: Media Presentation Description and Segment Formats”. International Standard, ISO/IEC 23009-1, Second Edition, May 15, 2014, 152 pages. |
Lavigne, F., “Automatic Video Zooming for Sport Team Video Broadcasting on Smart Phones”. Available at: http://tcts.fpms.ac.be/publications/papers/2010/visapp2010_flfcxd.pdf, (2010), 7 pages. |
Written Opinion of the International Preliminary Examining Authority for PCT/US2016/053512 dated Aug. 28, 2017. |
International Preliminary Report on Patentability, Corrected Version, for PCT/US2016/053512 completed Jan. 15, 2018. |
International Preliminary Report on Patentability for PCT/US2016/053512 completed Jan. 15, 2018. |
Mavlankar, A., et. al., “Optimal Slice Size for Streaming Regions of High Resolution Video with Virtual Pan/Tilt/Zoom Functionality”, Jun. 2007. Available at: http://web.stanford.edu/˜bgirod/pdfs/MavlankarEUSIPCO07.pdf. |
International Search Report and Written Opinion of the International Searching Authority for PCT/US2016/053512, dated Nov. 28, 2016, 10 pages. |
Notification of Transmittal of the International Search Report and the Written Opinion of the International Searching Authority, or the Declaration for PCT/US2017/031335 dated Jul. 19, 2017, 16 pages. |
Giladi, A. et. al., “Use Cases for Conveying Source Information”. International Organization for Standardization Coding of Moving Pictures and Audio, ISO/IEC JTC1/SC29/WG11, m31594, Oct. 2013, 3 pages. |
Liu S., et al, “SVC Inter-Layer Prediction for SVC Bit-Depth Scalability”. Joint Video Team Of ISO/IEC JTC1/SC29/WG11 and ITU-T SG.1624, No. JVT-X075, Jun. 2007, 12 pages. |
International Search Report and Written Opinion of the International Searching Authority for PCT/US2017/041142 dated Oct. 11, 2017. |
Written Opinion of the International Preliminary Examining Authority for PCT/US2017/031335 dated Apr. 5, 2018. |
Written Opinion of the International Preliminary Examining Authority for PCT/US2017/041142 dated Jun. 20, 2018. |
International Preliminary Report on Patentability for PCT/US2017/031335 completed on Jun. 26, 2018. |
International Preliminary Report on Patentability for PCT/US2017/041142 completed on Sep. 19, 2018. |
Mantiuk, R., et. al., “Display Adaptive Tone Mapping”. ACM Transactions on Graphics (TOG), vol. 27, No. 3, Aug. 2008, available at: http://resources.mpi-inf.mpg.de/hdr/datmo/mantiuk08datm.pdf, 10 pages. |
Unifore Security, “What's ROI (Region of Interest) Video Surveillance Technology?”. Unifore.net, Oct. 23, 2015, 4 pages. |
Mavlankar, A., et. al., “Peer-To-Peer Multicast Live Video Streaming with Interactive Virtual Pan/Tilt/Zoom Functionality”. In 2008 15th IEEE International Conference on Image Processing, Oct. 2008, pp. 2296-2299. |
Chen, Y., et. al., “Intra-And-Inter-Constraint-Based Video Enhancement Based On Piecewise Tone Mapping”. IEEE Transactions on Circuits and Systems for Video Technology, vol. 23, No. 1, Jan. 2013, pp. 74-82. |
Vehkaperä, J., et. al., “Replacing Picture Regions in H. 264/AVC Bitstream By Utilizing Independent Slices”. Proceedings of 2010 IEEE 17th International Conference on Image Processing, Sep. 26-29, 2010, pp. 3397-3400. |
ClassXteam, “ClassX Mobile”. You Tube Link, published on May 8, 2011, available at: https://www.youtube.com/watch?v=Kul-oCw4hj8. |
International Telecommunication Union, “Advanced Video Coding for Generic Audiovisual Services”. Series H: Audiovisual and Multimedia Systems, ITU-T, H.264, Feb. 2014, 790 pages. |
Ubuntu, “Display Adaptive Tone Mapping”. Ubuntu Manuals, available at: <http://manpages.ubuntu.com/manpages/xenial/en/man1/pfstmo_mantiuk08.1.html, (2010), 7pages. |
Invitation to pay additional Fees for PCT/US2018/020229, issued by the International Searching Authority, and dated Jun. 8, 2018, 14 pages. |
International Search Report and Written Opinion of the International Searching Authority for PCT/US2018/020229 dated Jul. 31, 2018, 18 pages. |
Brandenburg, R., et., al., “Immersive Second-Screen Experiences Using Hybrid Media Synchronization”. Media Synchronization Workshop, (2013), 7 pages. |
HBB Next, “Immersive Second Screen Synchronization”. HBB Next, Next Generation Hybrid Media, Applications, available at: http://www.mediafutureweek.nl/wp-content/uploads/2014/05/poster_IBC_2013_applications.pdf, (2013), 1 page. |
Niamut, O.A., et. al., “Live Event Experiences—Interactive UHDTV On Mobile Devices”. IBC2014 Conference, Sep. 11-15, 2014, 8 pages. |
Lindgren, P., et., al., “How True, Synchronized Live OTT Can Change the Second Screen and Social TV Game”. IBC2016 Conference, Sep. 8-12, 2016, pp. 1-8. |
Richardson, Iain E., et. al., “The H.264 Advanced Video Compression Standard, Chapter 10: Extensions and Directions”. John Wiley, Second Edition, (2010), pp. 287-311. |
Shibata, F., et. al., “A View Management Method for Mobile Mixed Reality Systems”. In Virtual Environments 2008: 14th Eurographics Symposium on Virtual Env., May 29-30, 2008, 8 pages. |
International Search Report and Written Opinion of the International Searching Authority for PCT/US2017/043248 dated Oct. 5, 2017. |
Written Opinion of the International Preliminary Examining Authority for PCT/US2017/043248 dated Jun. 28, 2018. |
International Preliminary Report on Patentability for PCT/US2017/043248 completed Sep. 27, 2018. |
Taylor, T. “NFL using Zebra RFID Chips to Track Player Movements, Gather Data”. Sports Illustrated article, available at: http://web.archive.org/web20160703053654/http://www.si.com:80/, updated Mar. 6, 2015, 4 pages. |
Zebra, “Zebra and the NFL: Change the Game”. Zebra.com article, available at: http://web.archive.org/web/20160604013821/https://www.zebra.com/us/en/nfl.html, web archive dated Jun. 4, 2016, 4 pages. |
Sheikh Y., et. al., “Geodetic Alignment of Aerial Video Frames”. In Shah M., Kumar R. (eds) Video Registration, Chapter 7, (2003), pp. 141-177. |
“NASCAR Pointer Screen Capture”, retrieved from video available at: http://www.sportvision.com/media/pointers, on Jul. 8, 2016, 1 page. |
Aberdeen, “Case Study: FCC Challenges Video Editors to Make Room for Caption Placement”, Aug. 4, 2015, 3 pages. |
Intel Corporation, “Overview of new radio access technology requirements and designs”, 3GPP Tdoc R1-162379, 3GPP TSG RAN WG1 Meeting #84bis, Busan, Korea, Apr. 11-15, 2016, 4 pages. |
ETSI TS 102690 V1.1.1, Machine-to-Machine Communications (M2M), “Functional Architecture” Oct. 2011, 280 pages. |
European Telecommunications Standards Institute (ETSI / TS 102 690 V1.2.1; Machine-to-Machine Communications (M2M); Functional Architecture, Jun. 2013, 279 pages. |
Qualcomm incorporated S2-162339 “Solution for key issue 1 on network slicing : network slice and network functions selection based on evolved eDeCOR model” S2-162339, SA WG2 Meeting #115, May 23-27, 2016, Nanjing, P.R. China (7 pages). |
ZTE: “Considerations on channel access procedure for NR-U”, 3GPP Draft; R2-1811281 Considerations on Random Access Procedure for NR-U, 3rd Generation Partnership Project (3GPP), Mobile Competence Centre ; 650, Route Des Lucioles ; F-06921 Sophia-Antipolis Cedex ; France, vol. RAN WG2, No. Gothenburg, Sweden; Aug. 20, 2018-Aug. 24, 2018 Aug. 9, 2018, (Aug. 9, 2018), XP051520945. |
Huawei et al: “Numerology and wideband operation in NR unlicensed”, 3GPP Draft; R1-1803677, 3rd Generation Partnership Project (3GPP), Mobile Competence Centre ; 650, Route Des Lucioles ; F-06921 Sophia-Antipolis Cedex ; France, vol. RAN WG1, No. Sanya, China; Apr. 16, 2018-Apr. 20, 2018 Apr. 15, 2018 (Apr. 15, 2018), XP051425974. |
Nokia et al: “Feature Lead's Summary on Channel Access Procedures” 3GPP Draft; R1-1809801 NRCJ Channel Access Summary, 3rd Generation Partnership Project (3GPP), Mobile Competence Centre ; 650, Route Des Lucioles ; F-06921 Sophia-Antipolis Cedex ; France, vol. RAN WGI, No. Gothenburg, Sweden; Aug. 20, 2018-Aug. 20, 2018 Aug. 23, 2018 (Aug. 23, 2018),XP051517160. |
Huawei et al: “Initial access in NR unlicensed”, 3GPP Draft; R1-1808062, 3rd Generation Partnership Project (3GPP),Mobile Competence Centre ; 650, Route Des Lucioles F-06921 Sophia-Antipolis Cedex ; France, vol. RAN W1I, No. Gothenburg, Sweden; Aug. 20, 2018-Aug. 24, 2018 Aug. 10, 2018 (Aug. 10, 2018), XP051515466. |
European Telecommunications Standards Institute (ETSI), TS 102 690 V2.1.1, Technical Specification, Machine-to-Machine Communications (M2M); Functional Architecture, Oct. 2013, 332 pages. |
OneM2M Technical Specification, oneM2M-TS-0001 oneM2M Functional Architecture-V-0.0.7, oneM2M Functional Architecture, Aug. 26, 2013, 43 pages. |
Third Generation Partnership Project, “Technical Specification Group Radio Access Network; Feasibility study for Further Advancements for E-UTRA (LTE-Advanced) (Release 13),” 3GPP TR 36.912 V13.0.0, Dec. 2015 (62 pages). |
Third Generation Partnership Project, “Technical Specification Group Radio Access Network; Study on Elevation Beamforming / Full-Dimension (FD) Multiple Input Multiple Output (MIMO) for LTE” (Release 13), 3GPP TR 36.897 V13.0.0, Jun. 2015 (58 pages). |
Third Generation Partnership Project, Technical Specification Group Radio Access Network; Evolved Universal Terrestrial Radio Access (E-UTRA); Study on Latency Reduction Techniques for LTE (Release 14) 3GPP TR 36.881 V14.0.0, Jun. 2016 (99 pages). |
Third Generation Partnership Project, “Technical Specification Group Services and System Aspects; IP Multimedia Substystem (IMS)” Stage 2 (Release 12), 3GPP TS 23.228 V12.1.0, Jun. 2013, (296 pages). |
Third Generation Partnership Project, “Technical Specification Group Services and System Aspects; Feasibility Study on New Services and Markets Technology Enablers for Critical Communications” Stage 1 (Release 14), 3GPP TR 22.862 V14.1.0, Sep. 2016, (31 pages). |
Third Generation Partnership Project, “Technical Specification Group Services and System Aspects; Feasibility Study on New Services and Markets Technology Enablers—Network Operation” Stage 1 (Release 14), 3GPP TR 22.864 V14.1.0, Sep. 2016 (35 pages). |
Third Generation Partnership Project, “Technical Specification Group Services and System Aspects; Feasibility Study on New Services and Markets Technology Enablers—Enhanced Mobile Broadband” Stage 1 (Release 14), 3GPP TR 22.863 V14.1.0, Sep. 2016 (21 pages). |
IEEE P802.11, Wireless LANs, Proposed TGax draft specification, Comment Resolutions on UL MU Operation, Sep. 13, 2016, (30 pages). |
Third Generation Partnership Project, “Technical Specification Group GSM/EDGE Radio Access Network; Cellular system support for ultra-low complexity and low throughput Internet of Things (CloT)” (Release 13), 3GPP TR 45.820 V13.1.0, Nov. 2015 (495 pages). |
Third Generation Partnership Project, “Technical Specification Group Services and System Aspects; Feasibility Study on New Services and Markets Technology Enablers” Stage 1 (Release 14), 3GPP TR 22.891 V1.1.0, Nov. 2015 (95 pages). |
Third Generation Partnership Project, “Technical Specification Group Services and System Aspects; Study on architecture enhancements for Cellular Internet of Things” (Release 13), 3GPP TR 23.720 V13.0.0, Mar. 2016 (94 pages). |
ITU-R, “IMT Vision—Framework and Overall Objectives of the Future Development of IMT for 2020 and Beyond”, Recommendation ITU-R M.2083-0, M Series, Mobile, Radiodetermination, Amateur and Related Satellite Services, Sep. 2015 (21 pages). |
Third Generation Partnership Project, “Technical Specification Group Radio Access Network; Evolved Universal Terrestrial Radio Access (E-UTRA); Medium Access Control (MAC) protocol specification” (Release 13), 3GPP TS 36.321 V13.0.0, Dec. 2015 (82 pages). |
Third Generation Partnership Project, “Technical Specification Group Radio Access Network; Study on Scenarios and Requirements for Next Generation Access Technologies” (Release 14), 3GPP TR 38.913 V0.3.0, Jun. 2017 (30 pages). |
Third Generation Partnership Project, “Technical Specification Group Radio Access Network; Evolved Universal Terrestrial Radio Access (E-UTRA); Radio Resource Control (RRC); Protocol specification” (Release 13), 3GPP TS 36.331 V13.0.0, Dec. 2015 (507 pages). |
Third Generation Partnership Project, “Technical Specification Group Radio Access Network; Evolved Universal Terrestrial Radio Access (E-UTRA); Physical layer procedures (Release 12),” 3GPP TS 36.213 V13.0.0, Dec. 2015 (6 pages). |
Third Generation Partnership Project, “Way Forward on Frame Structure”, 3GPP TSG RAN WG1 #85, R1-165669, Qualcomm and etc., Nanjing, China, May 23-27, 2016, Agenda Item 7.1.4 (2 pages). |
Third Generation Partnership Project, “Technical Specification Group Radio Access Network, Study on Scenarios and Requirements for Next Generation Access Technologies, (Release 14)”, (3GPP), TR 38.913 V0.2.0, Feb. 2016 (19 pages). |
Third Generation Partnership Project, “Technical Specification Group Radio Access Network; Evolved Universal Terrestrial Radio Access (E-UTRA); Physical Channels and Modulation (Release 13)”, 3GPP TS 36.211 V13.1.0, Mar. 2016 (155 pages). |
Third Generation Partnership Project, 3GPP TSG RAN WG1 #85, R1-164013, “Framework for Beamformed Access” May 2016 (4 pages). |
Hattachi, R,E, , et al., “White Paper” NGMN 5G Initiative, v1.0, Feb. 17, 2015 (125 pages). |
Third Generation Partnership Project, “Frame Structure for NR”, 3GPP TSG-RAN WG1 #85, R1-164628, Ericsson, Nanjing, China, May 23-27, 2016 (3 pages). |
Third Generation Partnership Project, “Introduction of Virtual Cell” 3GPP TSG RAN WG2 Meeting #93bis, R2-162571, Apr. 2016 (3 pages). |
Third Generation Partnership Project, “Further enhancements on signaling reduction to enable light connection for LTE”, 3GPP RP-160425 TSG RAN Meeting #71, Intel Corporation, Gothenburg, Sweden, Mar. 7-10, 2016 (7 pages). |
Huawei, “Network slice selection” RAN WG3 Meeting #92, R3-16113, Agenda Item 10.2.3, Nanjing, China, May 23-27, 2016 (4 pages). |
Third Generation Partnership Project, “Update the network slicing use case in Smarter”, 3GPP S1-152395 Revision of S1-152074, ZTE Corporation et al., ZTE Smarter Update the Network Slicing Use case REV3, vol. SA WG1, No. Belgrade Serbia, Aug. 24, 2015 (3 pages). |
Chu, D., “Polyphase Codes With Good Periodic Correlation Properties”, IEEE Transactions On Information Theory, Jul. 1972, pp. 531-532 (2 pages). |
Third Generation Partnership Project, “Technical Specification Group Services and Systems Aspects, Feasibility Study on New Services and Markets Technology Enablers for Massive Internet of Things, Stage 1” (Release 14), 3GPP TR 22.861 V14.1.0, Sep. 2016, (28 pages). |
European Telecommunications Standards Institute (ETSI TS) 124 302 V13.5.0, Universal Mobile Telecommunications System (UMTS); LTE; Access to the 3GPP Evolved Packet Core (EPC) via non-3GPP access networks; Stage 3 (Release 13), 3GPP TS 24.302 V13.5.0, Apr. 2016 (128 pages). |
European Telecommunications Standards Institute (ETSI TS) 136 304 V13.0.0, “LTE; Evolved Universal Terrestrial Radio Access (E-UTRA); User Equipment (UE) procedures in idle mode” (Release 13), 3GPP TS 36.304 V13.0.0, Feb. 2016 (44 pages). |
Damour, N., List of Common Services Functions—Call for Contributions, M2M Functional Architecture Technical Specification, Doc# oneM2M-ARC-2013-0304R04—List of oneM2M CSFs—Call for contributions—2.DOC Input Contribution, Jul. 11, 2013, 1-5 pages. |
Third Generation Partnership Project, (3GPP) TSG RAN WG1 #85, R1-164014, Discussion on RS for Beamformed Access, Samsung, Nanjing, China, May 23-27, 2016 (3 pages). |
Third Generation Partnership Project, (3GPP) S2-161324 SA WG2 Meeting #113, Solution to Key Issue on Mobility Framework, Sophia Antipolis, FR, Feb. 23-26, 2016 (3 pages). |
European Telecommunications Standards Institute (ETSI TS) Digital cellular telecommunications system (Phase 2+) (GSM); Universal Mobile Telecommunications System (UMTS); General Packet Radio Service (GPRS); Service Description; Stage 2 (3GPP TS 23.060 version 13.6.0 (Release 13) Apr. 2016 (364 pages). |
European Telecommunications Standards Institute (ETSI TS) LTE; General Packet Radio Service (GPRS) enhancements for Evolved Universal Terrestrial Radio Access Network (E-UTRAN) access (Release 13), 3GPP TS 23.401 V13.6.1, May 2016 (367 pages). |
European Telecommunications Standards Institute (ETSI TS) Technical Specification LTE; Evolved Universal Terrestrial Radio Access (E-UTRA); Requirements for support of radio resource management (Release 14), 3GPP TS 36.133 V14.7.0, Apr. 2018 (2999 pages). |
OMA Open Mobile Alliance, “Lightwieght Machine to Machine Technical Specification, Approved Version 1.0”, OMA-TS-LightweightM2M-V1_0-20170208-A, Feb. 8, 2017 (138 pages). |
A-Yong et al., “Research on Tower Crane Automatic Tracking and Zooming Video Monitoring System based on Improved Camshift algorithm”, 2011 International Conference on Electric Information and Control Engineering, Wuhan, 2011, 4 pages. |
Third Generation Partnership Project, “Technical Specification Group Radio Access Network Evolved Universal Terrestrial Radio Access (E-UTRA)”; Multiplexing and channel coding (Release 13), 3GPP TS 36.212 V13.2.0 (Jun. 2016), 140 pages. |
European Telecommunications Standards Institute, Machine-to-Machine communications (M2M); Functional architecture, ETSI TS 102 690 V1.1.1 (Oct. 2011), 280 pages. |
Number | Date | Country | |
---|---|---|---|
20200014961 A1 | Jan 2020 | US |
Number | Date | Country | |
---|---|---|---|
62460651 | Feb 2017 | US |