The subject matter described herein relates to streaming of video and audio data, for example, relating to real-time game play and commentary from multiple sources to multiple viewers simultaneously.
Video game streamers typically engage with their audience by showing live footage of their own gameplay. Sometimes their own commentary is provided during the gameplay. While this is an effective means of displaying what the streamer is seeing on his or her game screen, there is no effective means for a streamer to display and represent what is happening to other players in separate parts of a game or eSports competition.
In an aspect, user action data characterizing action by a player in a game environment executing at a user client is received at a server. The game environment is created by the user client separate from the server. Data characterizing a selected viewing position is received. The selected viewing position is different than a player viewing position. The selected viewing position characterizes a viewing location within the game environment. A recreated game environment is generated from the user action data at the server. A video stream of the recreated game environment is generated. The video stream includes video from a perspective of the selected viewing position. The video stream is transmitted to a viewing client.
One or more of the following features can be included in any feasible combination. For example, the user action data can include data characterizing user interface actions by the player with the game environment. Generating the recreated game environment can include executing a game engine using game state data and the user activity data. The server can include the game engine that receives the selected viewing position and the user action data and renders the recreated game environment from the perspective of the selected viewing position.
The user activity data can be retrieved from a database and the user activity data can be associated with a historical game play. The user activity data can be received from the user client or a game hosting server. The user activity data can be from an ongoing executing game environment. Generating the video stream can include capturing a rendered recreated game environment. An audio stream can be received from a commentator client. The audio stream can be combined with the video stream prior to transmission of the video stream. Statistics on the player can be generated using user action data. The statistics can be transmitted to the viewing client.
As eSports grow in popularity, there is a need for a toolkit that facilitates streamers, as individuals, to perform the same level of commentary and gameplay analytics that is performed by analyst teams in the parallel industry of physical sports.
The presently described subject matter can provide an automated broadcast studio for a streamer to cover an eSports tournament being played in a plurality of different games. The presently described subject matter can capture video data, or action data, from all tournament participants. A video queue interface can be delivered for streamers to view, select, and play videos or clips of events occurring within the tournament. In some variations, views as seen by players in a tournament can be included in the video queue. In other variations, the view point can be selected by the commentator, or streamer, of the streaming content. The selected view point need not be a player view point.
The presently described subject matter can break down video into highlight reels and key segments based on an analysis of event data. The presently described subject matter can facilitate the provision of game analytics, and assemble a suggested broadcast for the streamer.
Implementations of the current subject matter can include, but are not limited to, methods consistent with the descriptions provided herein as well as articles that comprise a tangibly embodied machine-readable medium operable to cause one or more machines (e.g., computers, etc.) to result in operations implementing one or more of the described features. Similarly, computer systems are also described that may include one or more processors and one or more memories coupled to the one or more processors. A memory, which can include a computer-readable storage medium, may include, encode, store, or the like one or more programs that cause one or more processors to perform one or more of the operations described herein. Computer implemented methods consistent with one or more implementations of the current subject matter can be implemented by one or more data processors residing in a single computing system or multiple computing systems. Such multiple computing systems can be connected and can exchange data and/or commands or other instructions or the like via one or more connections, including but not limited to a connection over a network (e.g. the Internet, a wireless wide area network, a local area network, a wide area network, a wired network, or the like), via a direct connection between one or more of the multiple computing systems, etc.
The details of one or more variations of the subject matter described herein are set forth in the accompanying drawings and the description below. Other features and advantages of the subject matter described herein will be apparent from the description and drawings, and from the claims. It should be readily understood that features described herein are not intended to be limiting. The claims that follow this disclosure are intended to define the scope of the protected subject matter.
The accompanying drawings, which are incorporated in and constitute a part of this specification, show certain aspects of the subject matter disclosed herein and, together with the description, help explain some of the principles associated with the disclosed implementations. In the drawings,
A tool kit is provided to effectively and efficiently stream game play video and audio data and commentary data to viewers of streaming content during an eSports competition. The tool kit can provide an automated broadcast studio for a streamer to cover an eSports tournament occurring in a game. The toolkit can facilitate video and audio capture from all tournament participants. The toolkit can facilitate delivery of a video queue interface for streamers to view, select, and play videos or clips. The toolkit can be configured to break down game play into highlight reels and key segments, provide game analytics, and assemble a suggested broadcast for the streamer based on game play events. In some implementations, a game environment can be recreated for the commentator to explore and/or generate new video content from any desired viewing position, angle, and/or point in time. This approach can reduce the amount of data that needs to be transmitted between game participant devices, streaming servers, and game observer devices.
Server 102, client device 104, commentator device 106, external resources 108, and/or other components of system 100, can be in electronic communication through a network 110. The network can be a local area network, wide area network, the Internet, a cloud network, or the like.
The server 102 can be configured to maintain a streamers toolkit. The streamers toolkit can be configured to facilitate streaming of game play and commentary to one or more client devices 104. In some variations, facilitating streaming of game play can include receiving audio and video data from a device associated with a game participant. The audio and video data can be representative of the view and sounds presented by the game to the game participant. In other variations, facilitating the streaming of game play can include receiving player action information representative of a player's actions being taken within the game by a game participant.
In some variations, where data representative of a player's actions is obtained, the streaming toolkit can be configured to reconstruct the game play on client devices 104 to which the game play has been streamed. In this way, a greatly reduced amount of data can be transmitted between game participant devices and game observer devices.
Game participants and game observers may be connected to the system 100 through similar devices, such as client devices 104.
In some variations, game play data can be streamed directly from a game participant device to a game observer device without the game play data being routed through a server 102, or through a commentator device 106. The commentator data, being data generated by the commentator device 106, can be transmitted to game observer devices such that the commentary and game play data can be synchronized.
The game play data, whether video and audio streams, or game participant action data, can be first provided to the commentator device 106. The commentator, through the commentator device 106 can curate the stream of game play that is provided to the game observers, together with any commentary or animations that the commentator desires to add.
The system 100 can facilitate receipt, by game observers, of streaming data provided by the commentator.
The toolkit can be configured to facilitate capture of video replay of all game participants who participate in a tournament. The toolkit, or streaming management system can be configured to facilitate providing access, by a streamer, to a list of all competitors in the event about which he or she is commentating. The streamer, or commentator, can have access to each participant's associated replay(s).
For each participant, the streamer has multiple actions that they can take using the captured video replay of the participant. For example, the streamer can view a preview of the game play for a particular participant. The streamer can privately view the replay of a participant without sharing that video to the live stream. This tool gives the streamer flexibility to review plays before having to speak about them, seek out a specific moment of the replay that he or she wants to show, or otherwise preview the replay before showing it to his or her audience.
The streamer can add video replay sections for participants in the game to a queue. The streamer can annotate the video replay, add statistics for display from the annotations feed, and the like.
The video queue graphical user interface 500 can allow the streamer to interact with and control what is being broadcasted to his or her stream. In some exemplary implementations, the graphical user interface 500 can include multiple areas, such as for example, a now playing stream area, a queue area, and the like.
The now playing stream area can be configured to show the streamer what replay is being currently broadcast to their stream. This is a persistent space that allows the streamer to monitor what they are currently sharing even while using the window to preview streams or adjust replays in a way that will not be shared with the audience.
The queue area can be configured to facilitate control of the order that replays are to be broadcast on the streamer's stream. The graphical user interface 500 can facilitate adding of individual replays that the streamer would like to show next, change the order of videos are to be broadcast, or the like. In some implementations, the toolkit can be configured to automatically stream video of game play in the event that the streamer chooses not to add any content to the queue, or if the queue becomes empty. Preset criteria can be provided to dictate what is shown when a content queue becomes empty. For example, video content showing the game play of the top three competitors can be presented, content of major actions or events within the tournament can be presented, or the like.
In some variations, user action information can be transmitted to the commentator device 106. The system 100 can be configured to recreate the game environment in which the tournament is being played for the commentator or streamer. The streamer can select a location and a viewing angle at which to stream from, this can be referred to, for example, as the streamer's viewpoint. The system 100 can be configured to facilitate moving of the streamer's viewpoint by the streamer. Game play footage can be streamed to game observers from the viewpoint of the streamer.
The system 100 can be configured to analyze game play and present interesting game play footage to streamers for streamers to select and add to their stream queues, in the graphical user interface 500. The system 100 can be configured to analyze game play and select interesting game play footage based on actions taken by game play participants, based on events experienced by game play participants, game play events selected by other game streamers, or the like.
The system 100 can analyze the game play to provide statistics to the streamer. The system 100 can facilitate the streamer to work as a one-person team and have interesting footage readily available to stream without having to take breaks from the live stream to look up past footage or statistics on players and/or non-player characters.
In some variations, automated analysis of the game play can include an automatic capture of a full history of all of a player's game records and details of the player's in-game play. The system 100 can be configured to compute a player's skill level based on a player's play history across all games, the current game, the current tournament within the game, or the like. The player information can be used to automatically provide a stream of interesting statistics on a particular player that a streamer can use to enhance their broadcast. For example, statistics can include information such as Player A beats Player B 57% of the time in general, but only 47% of the time on the particular level being broadcast, and only 43% of the time given his current position in this replay.
The system 100 can be configured to cause replays and player lists to appear in an order selected by the streamer. For example the order can be their current rank in the tournament, their current skill level in a particular game, their current rank among all players in a common game platform that hosts multiple games, or the like. When the order is based on players' current rank, a presumption can be made that the most interesting events involve participation of the most highly ranked players. Alternatively, a metric of how exciting an event is likely to be can be based on a relative difference in ranking of the participants in a multi-player game. In other words, players who are closely watched in skill level can be assumed to have a higher likelihood of participating in a more interesting game. Consequently, this configuration can make it easy to find the more interesting game replay. Prior to watching the video playback of the game play for a user, the streamer will know what level of player has been selected and what the outcome of their game will likely be.
The system 100 can be configured to provide historical game replays. In some variations, links can be provided to historical replays of the same game that occurred during previous tournaments, actions taken by the player in similar but different games, or the like. This can facilitate the streamer to provide their favorite footage for easy reference. Example uses include showing the best game ever recorded, keeping a reference of themselves playing the game, using footage with a specifically interesting or relevant play or strategy, showing a matchup between the players in a previous game, or the like.
The graphical user interface 500 can be configured to provide a notification to the streamer, or commentator, that a change in the leaderboard has occurred. In some variations, this can be limited to a certain number of players at the top of the leaderboard, or the bottom of the leaderboard. The graphical user interface 500 can be configured to notify the streamer when a game participant moves up or down an uncharacteristically large number of places on the leaderboard. The analytics provided by system 100 can be configured to detect segments where a change in a score (or other success metric) per unit time is abnormally high or low. This can facilitate easy detection of key plays within the game, where a player secures victory or focus on areas where an otherwise good player struggled and may have lost the game.
The graphical user interfaces 400 and 500 can be configured to facilitate provision of real-time score updates to the streamer and indicate whether or not the player is currently on track to beat their previous score or secure a top spot in the leaderboard. This takes the current concept of “splits”, as seen in the Olympics, and makes it work automatically for any eSport, regardless of type of scoring.
A plurality of players 210i (1=1, 2, . . . , N) operate respective player clients 220i. Each player client 220i, includes a third party game application 230i. The game application 230i, can be any online digital game (e.g., video game). In some implementations, multiple players 210, can compete against one another online. Games can be consistent across game applications 230, (e.g., if the players 210, are playing chess, each game application 230, is an instance of an electronic chess game). Each game application 230 can be in communication with and receiving game data from a game server 250. The game server 250 provides game data necessary to operate the game. The game server 250 can be a third party game server, e.g., separate from the tournament server 260 and with limited accessibility by the tournament server 260 (e.g., the tournament server 260 has either no communication or limited API call access to the game server). Player clients 220 may include mobile devices (e.g., smartphones, tablets, and the like), and the player clients 220, game server 250, and tournament server 260 can implement the tournament over a network, such as the internet.
Each game application 230 includes a peer-tournament module 240i. Peer-tournament module 240 integrates into game application 230 and enables players 210 to enroll and participate in an online game competition. Peer-tournament module 240 communicates with and works in tandem with the tournament server 260. While a game application 230 is running a match (e.g., a game instance), the game application 230 has an internal game state that changes over time as the associated player 210 interacts with (e.g., plays) the game. Peer-tournament module 240 can capture the game state data periodically, and/or continuously and transmit the captured game state to the tournament server 260. The game state data can include user action data including a characterization of user input into an interface such as key-presses, mouse/joystick/direction-pad inputs, and the like. The tournament server 260 can receive the game state from multiple peer-tournament modules 240. The matches may be separate in that they do not share a game state. For example, each game application 230 can be an asynchronous single player game. In some implementations, the matches share a game state (e.g., multiple players interact within one game environment).
In some implementations, the peer-tournament module 240 and the tournament syntax server 260 does not provide game data to the game application 230i. In other words, in some implementations, they do not contribute to implementing a game instance but rather observe the game, capture a game state, and enable a separate layer of multi-match tournament functionality.
Reconstruction engine 605 can include an extended game engine that enables reconstruction of the game environment from user action data and/or from game state data. The reconstruction of the game environment can be performed in an ongoing manner to provide for commentator coverage of a live event. The reconstruction of the game environment can be performed for any past point in time enabling the commentator to analyze and generate video of past gameplay with a new view-point. This can be achieved by maintaining a database of historical user action data and/or game state data. Reconstruction engine 605 can be an extension of the game engine that is incorporated into the game application 230 that is executing on player clients 220. Game state data and/or user action data that originates at the player client 220 can be utilized by the reconstruction engine to recreate the gaming environment. Reconstruction engine 605 can utilize data files that are normally utilized by the gaming application such as maps, level specifications, and the like.
Reconstruction engine 605 can include game program logic 610, a rendering engine 615, audio engine 620, physics engine 625, artificial intelligence 630, and commentator interface 635. Game program logic 610 includes actual game logic describing game mechanics. Rendering engine 615 generates animated graphics (e.g., 2D or 3D) and utilizes one or more rendering application programming interfaces (APIs), such as Direct3D or OpenGL. Audio engine 620 processes audio data and can utilize APIs, such as OpenAL, SDL audio, XAudio 2, Web Audio, and the like. Physics engine 625 emulates the laws of physics for the gaming environment. Artificial intelligence 630 provides guidance for some game controlled elements, such as non-player characters (NPCs).
Commentator interface 635 can enable input from commentator 320. This input can include providing viewing parameters such as viewing position, viewing angle, field of view, visible or non-visible portions of the field of view, and the like. The commentator interface 635 can provide the commentator input data to the game program logic rendering of the game environment using the commentator viewing parameters.
At 710, user action data is received. The user action data describes or characterizes action by a player in an executing game environment. For example, the user action data can include user interface inputs made by the user such as key strokes, button presses, mouse movement, joystick movement, directional pad input, and the like. In some implementations, the user action data is created by the player client 220 (e.g., a user client) and can be received by media management server 310 from a player client 220. In some implementations, the user action data can be received from game server 250. Game state data can also be received. The game state data can include, for example, character position, character orientation, and any other attributes and/or data necessary to reconstruct the gaming environment at a particular point in time. The game state data can include the user action data.
At 720, a selected viewing position can be received. The selected viewing position can be received from commentator 320. The selected viewing position can be different than a player viewing position and can include a location for viewing the game environment (e.g., a field of view).
At 730, a recreation of the game environment can be generated at the media management server 310. The gaming environment can be recreated from the user action data and/or the game state data. In some implementations, recreating the game environment can include executing a game engine (e.g., reconstruction engine 605) using game state data, the user activity data, and the selected viewing position. The game engine can receive the viewing position, game state data, and the user action data and renders the game environment from the perspective of the viewing position.
At 740, a video stream of the recreated game environment is generated. The video stream can by captured from the recreated environment rendered by the game engine. The video stream may include view from a perspective of the selected viewing position. The video stream can be generated by the media management server 310.
At 750, the video stream can be provided. The video stream can be transmitted to a viewing client device. The video stream can be provided to a user interface for further editing and/or curation by the commentator, for example, by utilizing the interface illustrated in
In some implementations, different viewing parameters can be used, such as viewing angle, a specific field of view, viewing effects (e.g., black and white, adjustment of contrast/color/brightness, additional video overlay, and the like), and the like. In some implementations, an audio stream may be received from commentator 320 and/or a player client 220. The audio stream may be combined with the generated video stream prior to transmission to viewing clients 330 or can be transmitted with the video stream for simultaneous media consumption.
In general, video game streamers can engage with their audience by showing live footage of their own gameplay. But in an eSports tournament, mere streaming of a display does not convey what is happening to other players in separate parts of a game or eSports competition. As eSports grow in popularity, the media management server 310 solves a need for a toolkit that allows streamers, as individuals, to perform the same level of commentary and gameplay analytics that is performed by analyst teams in the parallel industry of physical sports.
Media management server 310 provides an automated broadcast studio for a streamer 320 to cover (e.g., comment on) an eSports tournament in any game. The media management server can have two types of end users. The first is the viewing clients 330. The viewing clients 330 do not have exposure to stream tools that are available as part of the media management server 310. Instead, the viewing clients 330 are able to see the content shared by the tournament commentator 320. An example of the content shared by the tournament commentator 320 is illustrated in
The second end type of user is the tournament commentator 320 or “streamer” who uses the media management server 310 to deliver a high quality and engaging video feed to their audience. The streamer is able to capture video from all tournament participants; preview the captured video to privately view the replay of a player without sharing that video to the live stream; interact with a video queue interface to view, select, and play videos or clips; and receive suggestions for automatically portioned and assembled video broken into highlight reels and/or key segments with game analytics.
Media management server 310 in combination with peer-tournament module 240, can automatically capture and record video replay of all players 210, who participate in a tournament. In addition, a video stream can be generated by recreating the game environment and allowing the commentator to select viewing parameters allowing a controllable commentator view-point, for example, as described above with reference to
Through the media management server 310, a streamer 320 hosting a tournament has access to a list of all competitors or players 210, in the event that he or she is hosting and each player's 210 associated video capture. For each player 210, the streamer 320 has at least two actions that they can take using the captured video replay. They can preview the captured video in which the streamer 320 privately views the captured video of a player 210 without sharing that video to the live stream. This tool gives the streamer 320 flexibility to review plays before having to speak about them, seek out a specific moment of the replay that he or she wants to show, or otherwise preview the video before showing it to his or her audience. The streamer 320 can also add the selected video to a video queue, which specifies the next video portions to be streamed on the live broadcast stream.
The now playing stream 510B shows the streamer 320 what replay is being currently broadcasted (e.g., the live video stream or feed). This is a persistent space that allows the streamer 320 to monitor what they are currently sharing even while using the window to preview streams or adjust replays in a way that will not be shared with the audience.
The queue 510A gives the streamer 320 the capability to control the order that video portions or replays will be shown. The streamer 320 may add individual video portions that they would like to show next, or change the order of what videos are up next to be broadcasted. If the streamer 320 chooses not to add anything to the queue, or if the queue becomes empty, the system can automatically stream video of one of players using the across-match comparison, for example, by streaming video of one of the top three performing players 210, the video of the player 210 having the greatest change in score over a one minute period of time, and the like.
Video control panel 500 includes a list of players 520 participating in tournament matches and for which video capture is available for broadcast to a wider audience. The list of players 520 includes options to preview 520A or add 520B their replay to the queue 510A. Video control panel 500 further includes a preview space 530 for viewing video and a list of historical game replays 540 with options to preview 540A or add 540B the historical game replay to the queue 510A.
In some implementations, media management server 310 automatically breaks down video captured at the player clients 220 into highlight reels and key segments, and assembles a suggested broadcast for the streamer 320. Thus media management server 310 enables automated functionality that acts as the streamer's 320 own “behind the scenes analyst team.” This allows the streamer 320 to work as a one person team and easily select the correct footage to play without having to take breaks from the live stream to look up past footage or stats. Automatic analytics and features include the rank ordering of replays that causes replays and player lists to appear in the queue 510A in order of their current rank in the tournament. This makes it easy to find the replay for any player based upon where they stand. Prior to watching the video the streamer can know what level of player they are selecting and what the outcome of their game might be.
Automated media management server 310 functionality includes providing a list of historical game replays. This can include links to historical replays of the same game that occurred during previous tournaments. This allows the streamer 320 to bring up their favorite footage for easy reference. For example, the streamer 320 can show the best game ever recorded, keep a reference of themselves playing the game, or use footage with a specifically interesting or relevant play or strategy. In some implementations, the historical replays can include game state data including user activity data that enables recreation of the gaming environment, allowing the commentator to recreate a past game and generate a new video of the gameplay, possibly from a new view point.
Automated media management server 310 functionality includes providing notifications when there is a change within top of leaderboard. The streamer 320 can be notified (in addition to the leaderboard changing) when there is a change in position amongst top players in a tournament.
Referring again to
Server 102 can include electronic storage 114. Electronic storage 114 can be configured to store information. Electronic storage 114 can be configured to store computer readable instructions which, when executed by processor 112, can cause processor 112 to perform one or more operations described in the present description.
Electronic storage 114 may comprise electronic storage media that electronically stores information. The electronic storage media of electronic storage 114 may include one or both of system storage that is provided integrally (i.e., substantially non-removable) with a computing device, such as server 102, client device 104, and/or removable storage that is removably connectable to server 102 via, for example, a port (e.g., a USB port, a firewire port, etc.) or a drive (e.g., a disk drive, etc.). Electronic storage 114 may include one or more of optically readable storage media (e.g., optical disks, etc.), magnetically readable storage media (e.g., magnetic tape, magnetic hard drive, floppy drive, etc.), electrical charge-based storage media (e.g., EEPROM, RAM, etc.), solid-state storage media (e.g., flash drive, etc.), and/or other electronically readable storage media. The electronic storage 114 may include one or more virtual storage resources (e.g., cloud storage, a virtual private network, and/or other virtual storage resources). Electronic storage 114 may store software algorithms, information determined by processor 112, information received from one or more computing devices, such as server 102, client computing devices 104 information that enables the one or more computing device to function, or the like.
Without in any way limiting the scope, interpretation, or application of the claims appearing herein, a technical effect of one or more of the example embodiments disclosed herein may include facilitating the provision of curated and commentated streams of video game play.
One or more aspects or features of the subject matter described herein can be realized in digital electronic circuitry, integrated circuitry, specially designed application specific integrated circuits (ASICs), field programmable gate arrays (FPGAs) computer hardware, firmware, software, and/or combinations thereof. These various aspects or features can include implementation in one or more computer programs that are executable and/or interpretable on a programmable system including at least one programmable processor, which can be special or general purpose, coupled to receive data and instructions from, and to transmit data and instructions to, a storage system, at least one input device, and at least one output device. The programmable system or computing system may include clients and servers. A client and server are generally remote from each other and typically interact through a communication network. The relationship of client and server arises by virtue of computer programs running on the respective computers and having a client-server relationship to each other.
These computer programs, which can also be referred to programs, software, software applications, applications, components, or code, include machine instructions for a programmable processor, and can be implemented in a high-level procedural language, an object-oriented programming language, a functional programming language, a logical programming language, and/or in assembly/machine language. As used herein, the term “machine-readable medium” refers to any computer program product, apparatus and/or device, such as for example magnetic discs, optical disks, memory, and Programmable Logic Devices (PLDs), used to provide machine instructions and/or data to a programmable processor, including a machine-readable medium that receives machine instructions as a machine-readable signal. The term “machine-readable signal” refers to any signal used to provide machine instructions and/or data to a programmable processor. The machine-readable medium can store such machine instructions non-transitorily, such as for example as would a non-transient solid-state memory or a magnetic hard drive or any equivalent storage medium. The machine-readable medium can alternatively or additionally store such machine instructions in a transient manner, such as for example as would a processor cache or other random access memory associated with one or more physical processor cores.
To provide for interaction with a user, one or more aspects or features of the subject matter described herein can be implemented on a computer having a display device, such as for example a cathode ray tube (CRT) or a liquid crystal display (LCD) or a light emitting diode (LED) monitor for displaying information to the user and a keyboard and a pointing device, such as for example a mouse or a trackball, by which the user may provide input to the computer. Other kinds of devices can be used to provide for interaction with a user as well. For example, feedback provided to the user can be any form of sensory feedback, such as for example visual feedback, auditory feedback, or tactile feedback; and input from the user may be received in any form, including, but not limited to, acoustic, speech, or tactile input. Other possible input devices include, but are not limited to, touch screens or other touch-sensitive devices such as single or multi-point resistive or capacitive trackpads, voice recognition hardware and software, optical scanners, optical pointers, digital image capture devices and associated interpretation software, and the like.
In the descriptions above and in the claims, phrases such as “at least one of” or “one or more of” may occur followed by a conjunctive list of elements or features. The term “and/or” may also occur in a list of two or more elements or features. Unless otherwise implicitly or explicitly contradicted by the context in which it used, such a phrase is intended to mean any of the listed elements or features individually or any of the recited elements or features in combination with any of the other recited elements or features. For example, the phrases “at least one of A and B;” “one or more of A and B;” and “A and/or B” are each intended to mean “A alone, B alone, or A and B together.” A similar interpretation is also intended for lists including three or more items. For example, the phrases “at least one of A, B, and C;” “one or more of A, B, and C;” and “A, B, and/or C” are each intended to mean “A alone, B alone, C alone, A and B together, A and C together, B and C together, or A and B and C together.” Use of the term “based on,” above and in the claims is intended to mean, “based at least in part on,” such that an unrecited feature or element is also permissible.
The subject matter described herein can be embodied in systems, apparatus, methods, and/or articles depending on the desired configuration. The implementations set forth in the foregoing description do not represent all implementations consistent with the subject matter described herein. Instead, they are merely some examples consistent with aspects related to the described subject matter. Although a few variations have been described in detail above, other modifications or additions are possible. In particular, further features and/or variations can be provided in addition to those set forth herein. For example, the implementations described above can be directed to various combinations and subcombinations of the disclosed features and/or combinations and subcombinations of several further features disclosed above. In addition, the logic flows depicted in the accompanying figures and/or described herein do not necessarily require the particular order shown, or sequential order, to achieve desirable results. Other implementations may be within the scope of the following claims.
This application is a continuation of U.S. application Ser. No. 16/536,956, filed Aug. 9, 2019, which is a continuation of U.S. application Ser. No. 16/001,259, filed Jun. 6, 2018 (now U.S. Pat. No. 10,421,011, issued Sep. 24, 2019), which is a continuation of U.S. application Ser. No. 15/459,313, filed Mar. 15, 2017 (now U.S. Pat. No. 10,016,674, issued Jul. 10, 2018), which claims the benefit of U.S. Provisional Application No. 62/309,327, filed Mar. 16, 2016, the entire contents of each of which are incorporated by reference herein.
Number | Name | Date | Kind |
---|---|---|---|
5411258 | Wilson et al. | May 1995 | A |
6443841 | Rossides | Sep 2002 | B1 |
6676517 | Beavers | Jan 2004 | B2 |
6699127 | Lobb et al. | Mar 2004 | B1 |
6856986 | Rossides | Feb 2005 | B1 |
8088000 | Ginsberg et al. | Jan 2012 | B2 |
8210926 | Asher et al. | Jul 2012 | B2 |
8313368 | Filipour et al. | Nov 2012 | B2 |
8360873 | Wickett et al. | Jan 2013 | B1 |
8414387 | Paradise et al. | Apr 2013 | B1 |
8485899 | Rom et al. | Jul 2013 | B1 |
8512129 | Ginsberg et al. | Aug 2013 | B2 |
8545330 | Wickett et al. | Oct 2013 | B2 |
8562422 | Lutnick | Oct 2013 | B2 |
8568222 | Gagner et al. | Oct 2013 | B2 |
8613662 | Morrow et al. | Dec 2013 | B2 |
8641511 | Ginsberg et al. | Feb 2014 | B2 |
8651948 | Asher et al. | Feb 2014 | B2 |
8683272 | Cadima et al. | Mar 2014 | B2 |
8715077 | Paradise et al. | May 2014 | B2 |
8790170 | Arnone et al. | Jul 2014 | B2 |
8882576 | Paradise et al. | Nov 2014 | B1 |
8900054 | Patel | Dec 2014 | B2 |
8926435 | Perry et al. | Jan 2015 | B2 |
9240101 | Paradise et al. | Jan 2016 | B2 |
9349246 | Paradise et al. | May 2016 | B1 |
9446315 | Paradise et al. | Sep 2016 | B2 |
9479602 | Paradise et al. | Oct 2016 | B1 |
9630097 | Paradise et al. | Apr 2017 | B2 |
9649564 | Paradise et al. | May 2017 | B2 |
9697680 | Paradise et al. | Jul 2017 | B2 |
9767644 | Paradise et al. | Sep 2017 | B2 |
10016674 | Paradise | Jul 2018 | B2 |
10421011 | Paradise | Sep 2019 | B2 |
10960306 | Paradise | Mar 2021 | B2 |
20010031663 | Johnson | Oct 2001 | A1 |
20010044339 | Cordero et al. | Nov 2001 | A1 |
20020037767 | Ebin | Mar 2002 | A1 |
20020073021 | Ginsberg et al. | Jun 2002 | A1 |
20020160824 | Goto et al. | Oct 2002 | A1 |
20030038805 | Wong et al. | Feb 2003 | A1 |
20040204217 | Herman | Oct 2004 | A1 |
20050043089 | Nguyen et al. | Feb 2005 | A1 |
20050090307 | Walker et al. | Apr 2005 | A1 |
20060058103 | Danieli et al. | Mar 2006 | A1 |
20060080175 | Rowe et al. | Apr 2006 | A1 |
20060098013 | Wong et al. | May 2006 | A1 |
20060189382 | Muir et al. | Aug 2006 | A1 |
20070004509 | Banton | Jan 2007 | A1 |
20070026934 | Herbrich et al. | Feb 2007 | A1 |
20070037623 | Romik | Feb 2007 | A1 |
20070072676 | Baluja | Mar 2007 | A1 |
20070136817 | Nguyen | Jun 2007 | A1 |
20070191101 | Coliz et al. | Aug 2007 | A1 |
20070265092 | Betteridge | Nov 2007 | A1 |
20080033734 | Carry | Feb 2008 | A1 |
20080108438 | Sugiyama et al. | May 2008 | A1 |
20080153588 | Muir et al. | Jun 2008 | A1 |
20080161113 | Hansen et al. | Jul 2008 | A1 |
20080200242 | Ginsberg et al. | Aug 2008 | A1 |
20080201159 | Gabrick et al. | Aug 2008 | A1 |
20080207327 | Van et al. | Aug 2008 | A1 |
20080214301 | Sandige et al. | Sep 2008 | A1 |
20080234047 | Nguyen | Sep 2008 | A1 |
20080268961 | Brook et al. | Oct 2008 | A1 |
20080287183 | Reeves | Nov 2008 | A1 |
20090048010 | Kroeckel et al. | Feb 2009 | A1 |
20090099924 | Lensch et al. | Apr 2009 | A1 |
20090131177 | Pearce | May 2009 | A1 |
20090170608 | Herrmann et al. | Jul 2009 | A1 |
20090197665 | Christensen | Aug 2009 | A1 |
20090208181 | Cottrell | Aug 2009 | A1 |
20090209350 | Kelly et al. | Aug 2009 | A1 |
20090215540 | Perlman et al. | Aug 2009 | A1 |
20090227313 | Minka et al. | Sep 2009 | A1 |
20090325709 | Shi | Dec 2009 | A1 |
20100022307 | Steuer et al. | Jan 2010 | A1 |
20100144426 | Winner et al. | Jun 2010 | A1 |
20100166062 | Perlman et al. | Jul 2010 | A1 |
20100216536 | Gagner et al. | Aug 2010 | A1 |
20100222124 | Goshen | Sep 2010 | A1 |
20100234101 | Morrow et al. | Sep 2010 | A1 |
20100241699 | Muthukumarasamy et al. | Sep 2010 | A1 |
20100304860 | Gault et al. | Dec 2010 | A1 |
20100311496 | Taylor et al. | Dec 2010 | A1 |
20110010386 | Zeinfeld | Jan 2011 | A1 |
20110072056 | Bakalash et al. | Mar 2011 | A1 |
20110130197 | Bytnar et al. | Jun 2011 | A1 |
20110212766 | Bowers et al. | Sep 2011 | A1 |
20110254223 | Valentine | Oct 2011 | A1 |
20110281645 | Wolfson et al. | Nov 2011 | A1 |
20110300949 | Filipour et al. | Dec 2011 | A1 |
20120004039 | Perry et al. | Jan 2012 | A1 |
20120013622 | Mahajan et al. | Jan 2012 | A1 |
20120028718 | Barclay et al. | Feb 2012 | A1 |
20120100918 | Ginsberg et al. | Apr 2012 | A1 |
20120122553 | Bunch et al. | May 2012 | A1 |
20120156668 | Zelin | Jun 2012 | A1 |
20120178514 | Schulzke et al. | Jul 2012 | A1 |
20120196687 | Graham et al. | Aug 2012 | A1 |
20120252579 | Sethi et al. | Oct 2012 | A1 |
20120281080 | Wang | Nov 2012 | A1 |
20130029760 | Wickett et al. | Jan 2013 | A1 |
20130079072 | Filipour et al. | Mar 2013 | A1 |
20130121614 | Intwala | May 2013 | A1 |
20130151614 | Beerse et al. | Jun 2013 | A1 |
20130165196 | Kroeckel et al. | Jun 2013 | A1 |
20130172086 | Ikenaga | Jul 2013 | A1 |
20130179798 | Korupolu et al. | Jul 2013 | A1 |
20130217482 | Kelly et al. | Aug 2013 | A1 |
20130226983 | Beining et al. | Aug 2013 | A1 |
20130273987 | Laycock et al. | Oct 2013 | A1 |
20130331177 | Arnone et al. | Dec 2013 | A1 |
20130339473 | McCaffrey et al. | Dec 2013 | A1 |
20130344960 | Perry et al. | Dec 2013 | A1 |
20130344963 | Gupta et al. | Dec 2013 | A1 |
20140024427 | Cole et al. | Jan 2014 | A1 |
20140024437 | Vann et al. | Jan 2014 | A1 |
20140031132 | Wickett et al. | Jan 2014 | A1 |
20140045589 | Paradise et al. | Feb 2014 | A1 |
20140094267 | Davis et al. | Apr 2014 | A1 |
20140100023 | Arnone et al. | Apr 2014 | A1 |
20140128147 | Yu Cheng et al. | May 2014 | A1 |
20140194188 | Kosta et al. | Jul 2014 | A1 |
20140200062 | Paradise et al. | Jul 2014 | A1 |
20140323199 | Dimichele et al. | Oct 2014 | A1 |
20140323204 | Paradise et al. | Oct 2014 | A1 |
20150024839 | Zahn et al. | Jan 2015 | A1 |
20150065226 | Paradise et al. | Mar 2015 | A1 |
20150121437 | Tan | Apr 2015 | A1 |
20150202529 | Paradise et al. | Jul 2015 | A1 |
20160055710 | Paradise et al. | Feb 2016 | A1 |
20160110960 | Paradise et al. | Apr 2016 | A1 |
20160253866 | Paradise et al. | Sep 2016 | A1 |
20170225075 | Paradise et al. | Aug 2017 | A1 |
20170249802 | Paradise et al. | Aug 2017 | A1 |
20170266549 | Paradise | Sep 2017 | A1 |
20170266552 | Paradise et al. | Sep 2017 | A1 |
20170270751 | Paradise | Sep 2017 | A1 |
20180280796 | Paradise | Oct 2018 | A1 |
20190358538 | Paradise | Nov 2019 | A1 |
Number | Date | Country |
---|---|---|
103052429 | Apr 2013 | CN |
103945912 | Jul 2014 | CN |
1684228 | Jul 2006 | EP |
1738810 | Jan 2007 | EP |
2695650 | Feb 2014 | EP |
2003144745 | May 2003 | JP |
2003164672 | Jun 2003 | JP |
2009506690 | Feb 2009 | JP |
2009247562 | Oct 2009 | JP |
2010211728 | Sep 2010 | JP |
2013533000 | Aug 2013 | JP |
2015013102 | Jan 2015 | JP |
1020000018932 | Apr 2000 | KR |
20130122523 | Nov 2013 | KR |
2011149564 | Dec 2011 | WO |
2017160932 | Sep 2017 | WO |
Entry |
---|
International Preliminary Report on Patentability issued in International Application No. PCT/US2014/035876, dated Nov. 12, 2015, 8 pages. |
International Search Report and Written Opinion for Application No. PCT/US2017/022423 dated Nov. 8, 2017. |
International Search Report and Written Opinion for Application No. PCT/US2017/022440 dated Jun. 6, 2017. |
International Search Report and Written Opinion issued in International Application No. PCT/US2017/022412, dated Jun. 16, 2017, 11 pages. |
International Search Report issued in International Application No. PCT/US2014/035876, dated Oct. 17, 2014, 5 pages. |
Dreef et al., “Measuring Skill in Games: Several Approaches Discussed”, Mathematical Methods of Operations Research, 2004, 59(3):375-391. |
Fiedler et al., “Quantifying Skill in Games—Theory and Empirical Evidence for Poker”, Gaming Law Review and Economics, 2009, 13(1):50-57. |
Number | Date | Country | |
---|---|---|---|
20210178259 A1 | Jun 2021 | US |
Number | Date | Country | |
---|---|---|---|
62309327 | Mar 2016 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 16536956 | Aug 2019 | US |
Child | 17187149 | US | |
Parent | 16001259 | Jun 2018 | US |
Child | 16536956 | US | |
Parent | 15459313 | Mar 2017 | US |
Child | 16001259 | US |