Methods and systems for determining decal projections intersecting spatial units in a frame of a game space

Information

  • Patent Grant
  • 12064688
  • Patent Number
    12,064,688
  • Date Filed
    Thursday, December 23, 2021
    3 years ago
  • Date Issued
    Tuesday, August 20, 2024
    4 months ago
  • Inventors
    • Szymczyk; Michal (Woodland Hills, CA, US)
    • Drobot; Michal Dominik (Venice, CA, US)
    • Vance; Michael Kenneth (Falmouth, ME, US)
  • Original Assignees
  • Examiners
    • Rowland; Steve
    Agents
    • Novel IP
Abstract
A method of shading spatial units while rendering a frame in a video game. Initially, the virtual camera field of view is divided into a group of spatial units and each spatial unit in the group of spatial units is tested to determine which decals intersect the spatial unit. To do so, an initial list of decals associated with a given game space is tested. Once a subset of decals for a given spatial unit is identified, that spatial unit is subdivided further into smaller spatial units. The smaller spatial units are tested again, however, instead of using the initial list of decals, only the smaller subset of decals that was identified is used. The iterative process of subdividing the spatial units into smaller and smaller spatial units and testing each spatial unit continues until a given (higher) resolution is reached for the game space.
Description
FIELD

The present specification is related generally to the field of video games and computer graphics. More specifically, the present specification is related to efficiently determining decal projections that intersect spatial units of a frame being rendered.


BACKGROUND

In video games, there are places in the virtual landscape in which a high-detail or patterned object or landscape may need to contact a low-detail object or landscape. In such scenarios, to enhance realistic visuals, it is desirable to address an abrupt visual transition from the low-detail object or landscape to the high-detail object or landscape without significantly increasing the time required to author objects and environments, without degrading the frame rate of the render, and without consuming significant computer memory and processing resources.


In computer graphics, decals are visual representations that are often used for their aesthetic value such as, for example, to transfer a pattern, a text, a glyph, or a symbol, from a 2D image to a 3D surface upon contact in game scenes. When decals are projected into a game scene, they interact with the scene's lighting and wrap around solid objects and meshes. Stated differently, a decal projection is a virtual shape (like a box) that is used to cover another part of a virtual landscape and direct a projection of some graphic onto that part of the virtual landscape. Among other functions, decal projections can serve to mask transitions from low-detail world geometry patches to high-detail 3D models.



FIG. 5A shows a game scene 500 without projected decals. A left portion 505 of the scene shows a low-detail world geometry patch transitioning abruptly into a second portion 510 having a high-detail landscape. On the other hand, FIG. 5B shows the same game scene 500 with decal projections, in accordance with some embodiments of the present specification. The transition from the low-detail world geometry patch of the left portion 505 to the high-detail landscape of the second portion 510 is hardly discernible after decals are applied.


However, a typical field of view of a virtual camera may have hundreds and thousands of decal projections. Evaluating so many decal projections for every spatial unit in order to determine what is being projected onto the spatial unit, and therefore how the spatial unit should be shaded during rendering of a frame, often have prohibitive computational costs and may lead to rendering performance degradation.


Accordingly, there is a need for systems and methods that enable computationally efficient determination of which decals intersect or are being projected onto which spatial unit(s) when rendering a frame of the game and during the shading of a spatial unit.


SUMMARY

The following embodiments and aspects thereof are described and illustrated in conjunction with systems, tools and methods, which are meant to be exemplary and illustrative, not limiting in scope. The present application discloses numerous embodiments.


In some embodiments, the present specification discloses a method of shading spatial units during rendering of a frame of a game space in a multiplayer video game, wherein the game space is implemented on a plurality of player client devices by at least one server in data communication with the plurality of player client devices over a network, the method being implemented by the at least one server executing a plurality of programmatic instructions and comprising: acquiring a number of pixels in a field of view of a virtual camera associated with the game space, wherein the field of view has an associated plurality of decal projections; generating one or more groups of spatial units by progressively dividing the field of view, wherein the one or more groups of spatial units range from a lowest resolution group of spatial units through one or more intermediate resolution groups of spatial units to a highest resolution group of spatial units, and wherein each group of the one or more groups of spatial units represents a higher resolution compared to an immediately previous group of spatial units, and wherein each spatial unit has a predefined number of pixels acquired from the field of view of the virtual camera associated with the game space; testing each spatial unit of a lowest resolution group of spatial units to determine a subset of decal projections, of the plurality of decal projections, that intersect each spatial unit of the lowest resolution group of spatial units; iteratively testing each spatial unit of a next higher resolution group of spatial units to determine a next subset of decal projections using the subset of decal projections determined for the immediate previously tested lower resolution group of spatial units, wherein the iterative testing determines which of the subset of decal projections intersect the spatial unit of the next higher resolution group of spatial units, and wherein the iterative testing ends with testing of each spatial unit of the highest resolution group of spatial units to determine a final subset of decal projections that intersect each spatial unit of the highest resolution group of spatial units; and shading spatial units using the final subset of decal projections per spatial unit.


Optionally, the field of view is progressively divided based on a predefined divisor, and wherein the divisor is a natural number greater than 1.


Optionally, a number of decal projections in the subset of decal projections is less than the plurality of decal projections. Still optionally, a number of decal projections in the next subset of decal projections is less than the subset of decal projections.


Optionally, the testing is performed using Separating Axis Theorem.


Optionally, at least the final subset of decal projections is sorted, in ascending order, based on an identification number associated with each decal projection.


Optionally, if a spatial unit, of the one or more groups of spatial units, is determined to be devoid of intersecting decals then no further tests are conducted for higher resolution spatial units encapsulated by the spatial unit.


Optionally, each spatial unit has a length comprising a first predefined number of pixels, a width comprising a second predefined number of pixels, and a depth defined in terms of world space units representing a distance/depth into a world space corresponding to the game space of the multiplayer video game.


In some embodiments the present specification discloses a system for determining shading of spatial units during rendering a frame of a game space, the system comprising: at least one server in data communication with a plurality of player client devices over a network, wherein the at least one server implements the game space on the plurality of player client devices and is configured to: acquiring a number of pixels in a field of view of a virtual camera associated with the game space, wherein the field of view has associated a plurality of decal projections; generating one or more groups of spatial units by progressively dividing the field of view, wherein the one or more groups of spatial units range from a lowest resolution group of spatial units through one or more intermediate resolution groups of spatial units to a highest resolution group of spatial units, and wherein each group of the one or more groups of spatial units represents a higher resolution compared to an immediately previous group of spatial units, and wherein each spatial unit has a predefined number of pixels acquired from the field of view of the virtual camera associated with the game space; testing each spatial unit of a lowest resolution group of spatial units to determine a subset of decal projections, of the plurality of decal projections, that intersect the spatial unit of the lowest resolution group of spatial units; iteratively testing each spatial unit of a next higher resolution group of spatial units to determine a next subset of decal projections, of the subset of decal projections determined for the immediate previously tested lower resolution group of spatial units, that intersect the spatial unit of the next higher resolution group of spatial units, wherein the iterative testing ends with testing of each spatial unit of the highest resolution group of spatial units to determine a final subset of decal projections that intersect the spatial unit of the highest resolution group of spatial units; and using the final subset of decal projections per spatial unit during the shading of spatial units.


Optionally, the field of view is progressively divided based on a predefined divisor, and wherein the divisor is a natural number greater than 1.


Optionally, a number of decal projections in the subset of decal projections is less than the plurality of decal projections. Still optionally, a number of decal projections in the next subset of decal projections is less than the subset of decal projections.


Optionally, the testing is performed using Separating Axis Theorem.


Optionally, at least the final subset of decal projections is sorted, in ascending order, based on an identification number associated with each decal projection.


Optionally, if a spatial unit in a group of spatial units, of the one or more groups of spatial units, is determined to be devoid of intersecting decals then no further tests are conducted for smaller spatial units, encapsulated by the spatial unit, in higher resolution groups of spatial units.


Optionally, each spatial unit has a length comprising a first predefined number of pixels, a width comprising a second predefined number of pixels, and a depth defined in terms of world space units representing a distance/depth into a world space corresponding to the game space.


In some embodiments, the present specification discloses a method of determining shading of spatial units while rendering a frame of a game space, wherein the game space has an associated virtual camera defining a field of view of a plurality of pixels, the method being implemented in a computer having a processor and a random access memory, wherein the processor is in data communication with a display and with a storage unit, the method comprising: using the field of view to generate one or more groups of spatial units, wherein the one or more groups of spatial units range from a lowest resolution group of spatial units through one or more intermediate resolution groups of spatial units to a highest resolution group of spatial units, and wherein each group of the one or more groups of spatial units represents a higher resolution compared to an immediately previous group of spatial units, and wherein each spatial unit has a predefined number of pixels; for each spatial unit of a lowest resolution group of spatial units, determining a subset of decal projections that intersect the spatial unit of the lowest resolution group of spatial units; for each spatial unit of a next higher resolution group of spatial units, iteratively determining a next subset of decal projections, of the subset of decal projections determined for the immediate previous lower resolution group of spatial units, that intersect the spatial unit of the next higher resolution group of spatial units, wherein the iterative determining ends when for each spatial unit of the highest resolution group of spatial units a final subset of decal projections is determined that intersect the spatial unit of the highest resolution group of spatial units; and using the final subset of decal projections during the shading of spatial units.


Optionally, the one or more groups of spatial units are generated by progressively dividing the field of view by a predefined divisor, and wherein the divisor is a natural number greater than 1.


Optionally, a number of decal projections in the subset of decal projections is less than the plurality of decal projections.


Still optionally, a number of decal projections in the next subset of decal projections is less than the subset of decal projections determined for the immediate previous lower resolution group of spatial units.


Optionally, a Separating Axis Theorem is used to determine an intersection of decal projections with a spatial unit.


Optionally, at least the final subset of decal projections is sorted, in ascending order, based on an identification number associated with each decal projection.


Optionally, if a spatial unit in a group of spatial units, of the one or more groups of spatial units, is determined to be devoid of intersecting decals then smaller spatial units, encapsulated by the spatial unit, in higher resolution groups of spatial units are not processed for determining intersecting decal projections.


Optionally, each spatial unit has a length comprising a first predefined number of pixels, a width comprising a second predefined number of pixels, and a depth defined in terms of world space units representing a distance/depth into a world space corresponding to the game space.


In some embodiments, the present specification discloses a system for determining shading of spatial units while rendering a frame of a game space, wherein the game space has an associated virtual camera defining a field of view of a plurality of pixels, the system comprising: a computer having a processor and a random access memory, wherein the processor is in data communication with a display and with a storage unit, the computer configured to: using the field of view to generate one or more groups of spatial units, wherein the one or more groups of spatial units range from a lowest resolution group of spatial units through one or more intermediate resolution group of spatial units to a highest resolution group of spatial units, and wherein each group of the one or more groups of spatial units represents a higher resolution compared to an immediately previous group of spatial units, and wherein each spatial unit has a predefined number of pixels defined by the camera's field of view; for each spatial unit of a lowest resolution group of spatial units, determining a subset of decal projections that intersect the spatial unit of the lowest resolution group of spatial units; for each spatial unit of a next higher resolution group of spatial units, iteratively determining a next subset of decal projections, of the subset of decal projections determined for the immediate previous lower resolution group of spatial units, that intersect the spatial unit of the next higher resolution group of spatial units, wherein the iterative determining ends when for each spatial unit of the highest resolution group of spatial units a final subset of decal projections is determined that intersect the spatial unit of the highest resolution group of spatial units; and using the final subset of decal projections during the shading of spatial units. Optionally, the one or more groups of spatial units are generated by progressively dividing the field of view by a predefined divisor, and wherein the divisor is a natural number greater than 1.


Optionally, a number of decal projections in the subset of decal projections is less than the plurality of decal projections. Still optionally, a number of decal projections in the next subset of decal projections is less than the subset of decal projections determined for the immediate previous lower resolution group of spatial units.


Optionally, an intersection of decal projections with a spatial unit is determined based on Separating Axis Theorem.


Optionally, at least the final subset of decal projections is sorted, in ascending order, based on an identification number associated with each decal projection.


Optionally, if a spatial unit in a group of spatial units, of the one or more groups of spatial units, is determined to be devoid of intersecting decals then smaller spatial units, encapsulated by the spatial unit, in higher resolution groups of spatial units are not processed for determining intersecting decal projections.


Optionally, each spatial unit has a length comprising a first predefined number of pixels, a width comprising a second predefined number of pixels, and a depth defined in terms of world space units representing a distance/depth into a world space corresponding to the game space.


In some embodiments, the present specification discloses a method of shading spatial units during rendering of a frame of a game space in a multiplayer video game, wherein the game space is implemented on a plurality of player client devices by at least one server in data communication with the plurality of player client devices over a network, the method being implemented by the at least one server executing a plurality of programmatic instructions and comprising: acquiring a field of view of a virtual camera associated with the game space, wherein the field of view has an associated plurality of decal projections; generating one or more groups of spatial units within the field of view by progressively dividing the field of view, wherein the one or more groups of spatial units range from a lowest resolution group of spatial units through one or more intermediate resolution group of spatial units to a highest resolution group of spatial units, and wherein each group of the one or more groups of spatial units represents a higher resolution compared to an immediately previous group of spatial units, and wherein each spatial unit comprises a predefined number of pixels; testing each spatial unit of a lowest resolution group of spatial units to determine a subset of decal projections, of the plurality of decal projections, that intersect each spatial unit of the lowest resolution group of spatial units; iteratively testing each spatial unit of a next higher resolution group of spatial units to determine a next subset of decal projections using the subset of decal projections determined for the immediate previously tested lower resolution group of spatial units, wherein the iterative testing determines which of the subset of decal projections intersect the spatial unit of the next higher resolution group of spatial units, and wherein the iterative testing ends with testing of each spatial unit of the highest resolution group of spatial units to determine a final subset of decal projections that intersect each spatial unit of the highest resolution group of spatial units; and shading spatial units using the final subset of decal projections per spatial unit.


Optionally, the field of view is progressively divided based on a predefined divisor, wherein the divisor is a natural number greater than 1.


Optionally, a number of decal projections in the subset of decal projections is less than the plurality of decal projections. Optionally, a number of decal projections in the next subset of decal projections is less than the subset of decal projections.


The testing may be performed using Separating Axis Theorem.


Optionally, at least the final subset of decal projections is sorted, in ascending order, based on an identification number associated with each decal projection.


Optionally, if a spatial unit, of the one or more groups of spatial units, is determined to be devoid of intersecting decals, then no further tests are conducted for higher resolution spatial units encapsulated by the spatial unit.


Optionally, each spatial unit has a length comprising a first predefined number of pixels, a width comprising a second predefined number of pixels, and a depth defined in terms of world space units representing a distance/depth into a world space corresponding to the game space of the multiplayer video game.


In some embodiments, the present specification also discloses a system for determining shading of spatial units during rendering a frame of a game space, the system comprising: at least one server in data communication with a plurality of player client devices over a network, wherein the at least one server implements the game space on the plurality of player client devices and is configured to: acquiring a field of view of a virtual camera associated with the game space, wherein the field of view has associated a plurality of decal projections; generating one or more groups of spatial units by progressively dividing the field of view, wherein the one or more groups of spatial units range from a highest resolution group of spatial units through one or more intermediate resolution group of spatial units to a lowest resolution group of spatial units, and wherein each group of the one or more groups of spatial units represents a lower resolution compared to an immediately previous group of spatial units, and wherein each spatial unit comprises a predefined number of pixels; testing each spatial unit of a lowest resolution group of spatial units to determine a subset of decal projections, of the plurality of decal projections, that intersect the spatial unit of the lowest resolution group of spatial units; iteratively testing each spatial unit of a next higher resolution group of spatial units to determine a next subset of decal projections, of the subset of decal projections determined for the immediate previously tested lower resolution group of spatial units, that intersect the spatial unit of the next higher resolution group of spatial units, wherein the iterative testing ends with testing of each spatial unit of the highest resolution group of spatial units to determine a final subset of decal projections that intersect the spatial unit of the highest resolution group of spatial units; and using the final subset of decal projections per spatial unit during the shading of spatial units.


Optionally, the field of view is progressively divided based on a predefined divisor, wherein the divisor is a natural number greater than 1.


Optionally, a number of decal projections in the subset of decal projections is less than the plurality of decal projections. Optionally, a number of decal projections in the next subset of decal projections is less than the subset of decal projections.


The testing may be performed using Separating Axis Theorem.


Optionally, at least the final subset of decal projections is sorted, in ascending order, based on an identification number associated with each decal projection.


Optionally, if a spatial unit in a group of spatial units, of the one or more groups of spatial units, is determined to be devoid of intersecting decals, then no further tests are conducted for smaller spatial units, encapsulated by the spatial unit, in higher resolution groups of spatial units.


Optionally, each spatial unit has a length comprising a first predefined number of pixels, a width comprising a second predefined number of pixels, and a depth defined in terms of world space units representing a distance/depth into a world space corresponding to the game space.


In some embodiments, the present specification also discloses a method of determining shading of spatial units while rendering a frame of a game space, wherein the game space has an associated virtual camera defining a field of view of a plurality of spatial units, the method being implemented in a computer having a processor and a random access memory, wherein the processor is in data communication with a display and with a storage unit, the method comprising: using the field of view to generate one or more groups of spatial units, wherein the one or more groups of spatial units range from a highest resolution group of spatial units through one or more intermediate resolution group of spatial units to a lowest resolution group of spatial units, and wherein each group of the one or more groups of spatial units represents a lower resolution compared to an immediately previous group of spatial units, and wherein each spatial unit comprise a predefined number of pixels; for each spatial unit of a lowest resolution group of spatial units, determining a subset of decal projections that intersect the spatial unit of the lowest resolution group of spatial units; for each spatial unit of a next higher resolution group of spatial units, iteratively determining a next subset of decal projections, of the subset of decal projections determined for the immediate previous lower resolution group of spatial units, that intersect the spatial unit of the next higher resolution group of spatial units, wherein the iterative determining ends when for each spatial unit of the highest resolution group of spatial units a final subset of decal projections is determined that intersect the spatial unit of the highest resolution group of spatial units; and using the final subset of decal projections during the shading of each spatial unit.


Optionally, the one or more groups of spatial units are generated by progressively dividing the field of view by a predefined divisor, wherein the divisor is a natural number greater than 1.


Optionally, a number of decal projections in the subset of decal projections is less than the plurality of decal projections. Optionally, a number of decal projections in the next subset of decal projections is less than the subset of decal projections determined for the immediate previous lower resolution group of spatial units.


The intersection of decal projections with a spatial unit may be determined based on Separating Axis Theorem.


Optionally, at least the final subset of decal projections is sorted, in ascending order, based on an identification number associated with each decal projection.


Optionally, if a spatial unit in a group of spatial units, of the one or more groups of spatial units, is determined to be devoid of intersecting decals, then smaller spatial units, encapsulated by the spatial unit, in higher resolution groups of spatial units are not processed for determining intersecting decal projections.


Optionally, each spatial unit has a length comprising a first predefined number of pixels, a width comprising a second predefined number of pixels, and a depth defined in terms of world space units representing a distance/depth into a world space corresponding to the game space.


In some embodiments, the present specification also discloses a system for determining shading of spatial units while rendering a frame of a game space, wherein the game space has an associated virtual camera defining a field of view of a plurality of spatial units, the system comprising: a computer having a processor and a random access memory, wherein the processor is in data communication with a display and with a storage unit, the computer configured to: using the field of view to generate one or more groups of spatial units, wherein the one or more groups of spatial units range from a highest resolution group of spatial units through one or more intermediate resolution group of spatial units to a lowest resolution group of spatial units, and wherein each group of the one or more groups of spatial units represents a lower resolution compared to an immediately previous group of spatial units, and wherein each spatial unit comprises a predefined number of pixels; for each spatial unit of a lowest resolution group of spatial units, determining a subset of decal projections that intersect the spatial unit of the lowest resolution group of spatial units; for each spatial unit of a next higher resolution group of spatial units, iteratively determining a next subset of decal projections, of the subset of decal projections determined for the immediate previous lower resolution group of spatial units, that intersect the spatial unit of the next higher resolution group of spatial units, wherein the iterative determining ends when for each spatial unit of the highest resolution group of spatial units a final subset of decal projections is determined that intersect the spatial unit of the highest resolution group of spatial units; and using the final subset of decal projections during the shading of each spatial unit.


Optionally, the one or more groups of spatial units are generated by progressively dividing the field of view by a predefined divisor, wherein the divisor is a natural number greater than 1.


Optionally, a number of decal projections in the subset of decal projections is less than the plurality of decal projections. Optionally, a number of decal projections in the next subset of decal projections is less than the subset of decal projections determined for the immediate previous lower resolution group of spatial units.


The intersection of decal projections with a spatial unit may be determined based on Separating Axis Theorem.


Optionally, at least the final subset of decal projections is sorted, in ascending order, based on an identification number associated with each decal projection.


Optionally, if a spatial unit in a group of spatial units, of the one or more groups of spatial units, is determined to be devoid of intersecting decals, then smaller spatial units, encapsulated by the spatial unit, in higher resolution groups of spatial units are not processed for determining intersecting decal projections.


Optionally, each spatial unit has a length comprising a first predefined number of pixels, a width comprising a second predefined number of pixels, and a depth defined in terms of world space units representing a distance/depth into a world space corresponding to the game space.


The aforementioned and other embodiments of the present specification shall be described in greater depth in the drawings and detailed description provided below.





BRIEF DESCRIPTION OF THE DRAWINGS

These and other features and advantages of the present specification will be appreciated, as they become better understood by reference to the following detailed description when considered in connection with the accompanying drawings, wherein:



FIG. 1A is a block diagram illustration of a multi-player online gaming system or environment, in accordance with some embodiments of the present specification;



FIG. 1B is a block diagram illustration of an offline gaming system or environment, in accordance with some embodiments of the present specification;



FIG. 2 is a flowchart of a plurality of exemplary steps of a method of determining one or more decal projections that intersect a spatial unit, in a game space, by generating a spatial acceleration grid, structure or groups of spatial subunits, in accordance with some embodiments of the present specification;



FIG. 3 illustrates a game space frame showing highlighted boxes and areas where decals have effect, in accordance with some embodiments of the present specification;



FIG. 4 illustrates the game space frame divided into a group of spatial units, in accordance with some embodiments of the present specification;



FIG. 5A shows a game scene without projected decals; and



FIG. 5B shows the same game scene with projected decals, in accordance with some embodiments of the present specification.





DETAILED DESCRIPTION

The present specification is directed towards multiple embodiments. The following disclosure is provided in order to enable a person having ordinary skill in the art to practice the invention. Language used in this specification should not be interpreted as a general disavowal of any one specific embodiment or used to limit the claims beyond the meaning of the terms used therein. The general principles defined herein may be applied to other embodiments and applications without departing from the spirit and scope of the invention. Also, the terminology and phraseology used is for the purpose of describing exemplary embodiments and should not be considered limiting. Thus, the present invention is to be accorded the widest scope encompassing numerous alternatives, modifications and equivalents consistent with the principles and features disclosed. For purpose of clarity, details relating to technical material that is known in the technical fields related to the invention have not been described in detail so as not to unnecessarily obscure the present invention.


The term “a multi-player online gaming environment” or “massively multiplayer online game” may be construed to mean a specific hardware architecture in which one or more servers electronically communicate with, and concurrently support game interactions with, a plurality of client devices, thereby enabling each of the client devices to simultaneously play in the same instance of the same game. Preferably the plurality of client devices number in the dozens, preferably hundreds, preferably thousands. In one embodiment, the number of concurrently supported client devices ranges from 10 to 5,000,000 and every whole number increment or range therein. Accordingly, a multi-player gaming environment or massively multi-player online game is a computer-related technology, a non-generic technological environment, and should not be abstractly considered a generic method of organizing human activity divorced from its specific technology environment.


In various embodiments, a computing device includes an input/output controller, at least one communications interface and system memory. The system memory includes at least one random access memory (RAM) and at least one read-only memory (ROM). These elements are in communication with a central processing unit (CPU) to enable operation of the computing device. In various embodiments, the computing device may be a conventional standalone computer or alternatively, the functions of the computing device may be distributed across multiple computer systems and architectures.


In some embodiments, execution of a plurality of sequences of programmatic instructions or code enable or cause the CPU of the computing device to perform various functions and processes. In alternate embodiments, hard-wired circuitry may be used in place of, or in combination with, software instructions for implementation of the processes of systems and methods described in this application. Thus, the systems and methods described are not limited to any specific combination of hardware and software. It should be appreciated that every method, step, process or action described herein is achieved by using the hardware, within the overall computing architecture, to execute one or more of the programmatic instructions.


The term “module”, “application” or “engine” used in this disclosure refers to computer logic utilized to provide a desired functionality, service or operation by programming or controlling a general purpose processor. Stated differently, in some embodiments, a module, application or engine implements a plurality of instructions or programmatic code to cause a general purpose processor to perform one or more functions. In various embodiments, a module, application or engine can be implemented in hardware, firmware, software or any combination thereof. The module, application or engine may be interchangeably used with unit, logic, logical block, component, or circuit, for example. The module, application or engine may be the minimum unit, or part thereof, which performs one or more particular functions.


The term “decal” used in this disclosure refers to a well-defined geometric shape, preferably a box, with 8 vertices and parallel faces. Edges of a decal can be of different lengths. A single decal volume is typically made of multiple images, such as a color map, a specular map, a normal map, a glossiness map, and/or an opacity map. Opacity map is a mask that allows artists to create smooth transition between decal and underlying world textures and, in particular, achieve a smooth fade out close to the boundaries of the decal volume's box which looks more appealing than hard-cut at the box's edge.


The term “Separating Axis Theorem (SAT)” used in this disclosure refers to a theorem which states that two convex geometries do not overlap or intersect if there exists a line (referred to as an axis) onto which the two geometries' projections do not overlap. Thus, SAT enables testing whether two convex geometries intersect or not.


The term “spatial unit” or “spatial subunit” refers to a volume of game space defined by a length comprising a first amount of pixels, a width comprising a second amount of pixels, and a depth extending into the virtual game space, where the first amount of pixels may equal the second amount of pixels or may be different than the second amount of pixels. The term “spatial unit” or “spatial subunit” therefore encompasses a set of pixels and may also be referred to as a cell.


The terms “low resolution”, “lower resolution” and/or “lowest resolution” used in this disclosure with reference to a grid or group of spatial units are indicative of a relatively large size of spatial units in the group compared to a given resolution of a display screen or that of a camera field of view. On the other hand, the terms “high resolution”, “higher resolution” or “highest resolution” used in this disclosure with reference to a group of spatial units may be indicative of a relatively small size of spatial units in the group.


The terms “grid” and “group” are used interchangeably in the present specification. In the description and claims of the application, each of the words “comprise” “include” and “have”, and forms thereof, are not necessarily limited to members in a list with which the words may be associated. It should be noted herein that any feature or component described in association with a specific embodiment may be used and implemented with any other embodiment unless clearly indicated otherwise.


As used herein, the indefinite articles “a” and “an” mean “at least one” or “one or more” unless the context clearly dictates otherwise.


Overview



FIG. 1A illustrates an embodiment of a multi-player online gaming or massively multiplayer online gaming system/environment 100 in which the systems and methods of the present specification may be implemented or executed. The system 100 comprises client-server architecture, where one or more game servers 105 are in communication with one or more client devices 110 over a network 115. Players and non-players, such as an administrator, may access the system 100 via the one or more client devices 110. The client devices 110 comprise computing devices such as, but not limited to, personal or desktop computers, laptops, Netbooks, handheld devices such as smartphones, tablets, and PDAs, gaming consoles and/or any other computing platform known to persons of ordinary skill in the art. Although four client devices 110 are illustrated in FIG. 1A, any number of client devices 110 can be in communication with the one or more game servers 105 over the network 115.


The one or more game servers 105 can be any computing device having one or more processors and one or more computer-readable storage media such as RAM, hard disk or any other optical or magnetic media. The one or more game servers 105 include a plurality of modules operating to provide or implement a plurality of functional, operational or service-oriented methods of the present specification. In some embodiments, the one or more game servers 105 include or are in communication with at least one database system 120. The database system 120 stores a plurality of game data associated with at least one game that is served or provided to the client devices 110 over the network 115. In some embodiments, the one or more game servers 105 may be implemented by a cloud of computing platforms operating together as game servers 105.


In accordance with aspects of the present specification, the one or more game servers 105 provide or implement a plurality of modules or engines such as, but not limited to, a master game module 130 and a spatial acceleration structure (SAS) module 132. The one or more client devices 110 are configured to implement or execute one or more of a plurality of client-side modules some of which are same as or similar to the modules of the one or more game servers 105. In some embodiments each of the player and non-player client devices 110 executes a client-side game module 130′ (also referred to as—client game module 130′) that may further integrate a client-side SAS module 132′. In some embodiments, however, the client-side SAS module 132′ may not be required and all processes related to the spatial acceleration structure functionality may be implemented at the one or more game servers 105.


In some embodiments, the at least one non-player client device 110g is used by an administrator to log into the one or more game servers 105 (via the client game module 130′) and execute the module 132 on the server to generate one or more GUIs that enable the administrator to customize and set one or more parameters, attributes or characteristics related to the spatial acceleration structure. For example, the one or more GUIs may be used to predefine a value of a divisor ‘d’ in order to generate the spatial acceleration structure. It should be appreciated that the administrator includes computer graphics designers or artists, members of visual effects teams, gameplay engineers and any other non-player personnel responsible for design and development of the game.


While various aspects of the present specification are being described with reference to functionalities or programming distributed across modules or engines 130 and 132, it should be appreciated that, in some embodiments, some or all of the functionalities or programming associated with these modules or engines may be integrated within fewer modules or in a single module—such as, for example, in the master game module 130 itself on the server side.


In embodiments, the master game module 130 is configured to execute an instance of an online game to facilitate interaction of the players with the game. In embodiments, the instance of the game executed may be synchronous, asynchronous, and/or semi-synchronous. The master game module 130 controls aspects of the game for all players and receives and processes each player's input in the game. In other words, the master game module 130 hosts the online game for all players, receives game data from the client devices 110 and transmits updates to all client devices 110 based on the received game data so that the game, on each of the client devices 110, represents the most updated or current status with reference to interactions of all players with the game. Thus, the master game module 130 transmits game data over the network 115 to the client devices 110 for use and rendering by the game module 130′ to provide local versions and current status of the game to the players.


On the client-side, each of the one or more player client devices 110 implements the game module 130′ that operates as a gaming application to provide a player with an interface between the player and the game. The game module 130′ generates the interface to render a virtual environment, virtual space, game space, map or virtual world associated with the game and enables the player to interact in the virtual environment to perform a plurality of game and other tasks and objectives. The game module 130′ accesses at least a portion of game data, received from the game server 105, to provide an accurate representation of the game to the player. The game module 130′ captures and processes player inputs and interactions within the virtual world or environment and provides at least a portion as updates to the game server 110 over the network 115.


The database system 120 described herein may be, include, or interface to, for example, an Oracle™ relational database sold commercially by Oracle Corporation. Other databases, such as Informix™, DB2 (Database 2) or other data storage, including file-based, or query formats, platforms, or resources such as OLAP (On Line Analytical Processing), SQL (Structured Query Language), a SAN (storage area network), Microsoft Access™ or others may also be used, incorporated, or accessed. The database system 120 may comprise one or more such databases that reside in one or more physical devices and in one or more physical locations.


While the embodiments of the present specification will be illustrated with reference to the multi-player online gaming or massively multiplayer online gaming system/environment 100 of FIG. 1A, the embodiments are equally applicable to offline gaming environments such as the system/environment 140 of FIG. 1B. As shown in FIG. 1B, the system 140 includes a display 150, graphics hardware 152, a computing platform 154, a user interface 156 and a storage device, unit or memory 158. The computing platform 154 includes a processor and random access memory as well as components for processing data in order to render the game space or virtual environment, such as a graphics system 160 and the SAS module 132.


The graphics system 160 may perform rendering operations based on the results of the SAS module 132. The rendering operations, performed by the graphics system 160, may include visibility, culling, tessellation and spatial unit shading operations, for example. Based on results of the rendering operations, commands (such as Open GL commands) may be issued by graphics system 160 to the graphics hardware 152 to cause a display of a rendered frame on display 150.


Decals and Decal Projections


Decal projections typically involve programmatically acquiring a two-dimensional (2D) texture (like worn down looking wood, for example) and projecting it onto a virtual three-dimensional (3D) surface (like a staircase, for example). A video game artist may define a decal projection by generating an appropriately sized 3D box in the virtual environment. The 3D box is defined by three axes (x, y and z). A 2D image is then projected along the z axis onto the 3D geometry. As an example, FIG. 3 illustrates a game space frame 300 showing highlighted boxes and areas 305 where decals have a visual effect, in accordance with some embodiments of the present specification.


When rendering a frame of the game and during the shading of a spatial unit, a world space position, based on a world coordinate system, of the spatial unit is reconstructed. In embodiments, a world space position is a position in a 3-dimensional coordinate space having three coordinates (X, Y, Z). In an embodiment, with reference to a video game, a camera position is considered as the origin of world space position and is designated the coordinates (0,0,0). Positions of all other objects in the game are specified/designated coordinates in the same 3D space with reference to the camera position coordinates.


The spatial unit has a world space position that may intersect a decal volume and shading of the spatial unit would need to be performed based on its position in the decal volume. Therefore, the world space position is used to find 3D position of the spatial unit within the decal's bounding box (volume). Positions within the 3D box has X, Y, and Z coordinates, however, the game textures are two dimensional. Hence, 3D position in decal volumes' frame of reference is finally converted to 2D texturing coordinates by dropping/ignoring the Z-coordinate. This process of dropping z-component of the 3D position is called plane projection, wherein all points within the decal volume/box are projected onto the XY plane.


The XY coordinates undergo couple transformations and are used to sample 2D texture. In embodiments, couple transformations comprise scaling, rotating and translating the obtained 2D texture coordinates to achieve a required aesthetic look. This is presented to game design artists in the form of three parameters, namely: X and Y scale, angle rotation around Z axis and X and Y translation. The couple transformations are applied to the obtained 2D texture coordinates in a 2D texture coordinate space.


In some embodiments, 2D textures are stored in a large 2D texture atlas—such as the atlas 310 visible on the left side of the frame 300 of FIG. 3.


As known to persons of ordinary skill in the art, pixel/unit shading is a method used for rendering various graphical features in a frame. As is known, shading involves executing a software program per fragment/pixel/vertex or other user defined data unit, which causes various graphical features to be rendered in a frame on a screen. Video games often utilize spatial unit shaders in conjunction with GPUs (Graphics Processing Units) to render these graphical features on a per pixel basis. A spatial unit shader, also known as a fragment shader, is a software program that dictates the color, brightness, contrast, and other graphical features or characteristics of a single spatial unit (fragment). For instance, if a spatial unit to be rendered is on a visible surface of a pyramid, factors affecting its appearance comprise:

    • The light sources that illuminate this fragment of the pyramid and the types of light sources. In addition, the color of the light is a factor.
    • The visual appearance of specular reflections of the pyramid's surface material and whether it reflects light sharply such as metal or whether it diffuses light, such as plastic.
    • Whether there is a texture mapped to the surface? If so, the corresponding spatial unit in the texture's bitmap may contribute to the color of the fragment.
    • Whether there is a bump associated with the surface of the pyramid? If so, a bitmap or mathematical procedure affects the brightness of the light to simulate a bumpy texture on the surface.
    • Whether the surface is partially transparent and whether there are partially transparent objects between the pyramid and the camera.
    • Whether any other objects cast shadows on the pyramid.


Thus, a spatial unit shader is a graphics function that determines visual effects on a per-spatial unit basis. Depending on resolution, millions of pixels in a plurality of spatial units may need to be rendered, lit, shaded, and colored for each frame, at, for example, 60 frames per second. This creates a tremendous computational load. Further, a typical camera field of view or frustum may have hundreds and thousands of decal projections. Testing all decal projections for every pixel, in order to determine what is being projected onto a pixel and therefore how the pixel or a corresponding spatial unit should be shaded, has prohibitive computational and performance costs. To address these issues, the systems 100, 140 of FIGS. 1A and 1B implement the SAS module 132, which is described in further detail below.


Spatial Acceleration Structure (SAS) Module 132


In accordance with some aspects of the present specification, the SAS module 132 implements a plurality of instructions or programmatic code to subdivide an entire camera frustum (region of space in the modeled world that may appear on a display screen and is a field of view of a perspective virtual camera) into groups of spatial units. In embodiments, each spatial unit has a length comprising a first predefined number of pixels, and a width comprising a second predefined number of pixels, wherein the second predefined number of pixels may or may not be the same as the first predefined number of pixels. Further, each spatial unit has a depth defined in terms of world space units which depends upon the spatial unit's position in the world space as described above, and represents a distance/depth into the world space. For example, in an exemplary embodiment, a spatial unit may have a length and width of 32 pixels each and a depth of 1 meter (depth into the game world) on the world coordinate system.


To begin with, the camera field of view (frustum) is divided into a group of large spatial units (representing a low resolution). Each large spatial unit comprises a plurality of smaller spatial units. In embodiments, each large spatial unit has a first length, width and depth into the world space, while each comprising spatial unit has a second smaller length, width and depth into the world space. For example, a scene may be divided into 8 (width)×5 (length)×4 (depth) spatial units, where each spatial unit is, for example, 32 pixels (width)×32 pixels (length)×1 meter (depth). In embodiments, each spatial unit is tested to determine which decal projections intersect the spatial unit. To do so, an initial list of decals associated with a given game space is tested using an approach such as, but not limited to, the Separating Axis Theorem (SAT). In an embodiment, an approximation method is used for testing the decals associated with a given game space. The method comprises approximating spatial unit and decal projection with a sphere and performing an intersection test of two such spheres. This comprises comparing distance between the centers of the two sphere by using the radii of said spheres. In another embodiment, spatial unit and decal projections are approximated by using axis-aligned bounding boxes, as this method provides, in some instances, an increased accuracy as compared to the method comprising intersection test of two such spheres. In embodiments, SAT is used for testing decals associated with a given game space, as SAT provides accurate results while avoiding any false positives. In an embodiment the SAT based tests are performed within the world space coordinates.


Once a subset or shortlist of decals for a given spatial unit is identified, that spatial unit is subdivided further into smaller spatial units (representing a higher resolution). The smaller spatial units are tested again, however, instead of using the initial larger list of decals, only the smaller subset or shortlist of decals that was identified is used. The iterative process of subdividing the spatial units into smaller and smaller spatial units and testing each spatial unit continues until a given resolution is reached for the game space. Thus, the SAS module 132 is configured to implement a spatial acceleration grid or structure of spatial units that enables testing fewer and fewer decals as the module subdivides from larger to smaller and smaller spatial units (that is, moving from lower resolution to higher resolution).



FIG. 2 is a flowchart of a plurality of exemplary steps of a method 200 of determining one or more decal projections that intersect a spatial unit, in a game space, by generating a spatial acceleration grid, structure or groups of spatial units, in accordance with some embodiments of the present specification. In embodiments, the method 200 is implemented by the SAS module 132 in any of the systems 110, 140 of FIGS. 1A and 1B.


At step 202, for a frame, a frustum or field of view of a perspective virtual camera is acquired. In some embodiments, the frustum defines the game space in a simulated virtual environment that may be displayed on a screen of a player's client device. In embodiments, the frustum or field of view has a plurality of decals associated therewith. In some embodiments, each of the plurality of decals has an associated identification number. Thus, in various embodiments, decals are placed in a game map with associated world position and size. When a frame is rendered, the decals in the game map are iterated to identify which decals are visible to (or lay in the field of view of) the virtual camera.


In various embodiments, the frustum or field of view encompasses a resolution or a number of pixels (hereinafter referred to as ‘total’ resolution or ‘total’ number of pixels) of the screen of the player's client device. In various embodiments, the total resolution of the screen is represented as m×n pixels, wherein ‘m’ denotes a number of pixels along a width dimension of the screen and ‘n’ denotes a number of pixels along a height dimension of the screen. For example, in some embodiments, a total resolution of a screen may be 720×480 pixels, 720×576 pixels, HD (High Definition) having 1280×720 pixels, Full HD having 1920×1080 pixels, Quad HD having 2560×1440 pixels, Ultra HD having 3840×2160 pixels or any other resolution known to persons of ordinary skill in the art.


At step 204, the frustum, field of view, total resolution or the total number of pixels is progressively divided into one or more successive groups of spatial units, wherein each successive groups of spatial units represents a smaller area (block of pixels) than the previous group of spatial units, such that each successive group or sub-division of spatial units represents a higher resolution compared to an immediately previous group of spatial units. In embodiments, the one or more successive groups of spatial units range from a largest (that is, lowest resolution) group of spatial units to a smallest, (that is highest resolution) group of spatial units, having increasing resolution with each iterative sub-division.


In various embodiments, each spatial unit of a group of spatial units includes a predefined number of pixels, such as, p×q pixels, wherein ‘p’ denotes a number of pixels along a width dimension of a spatial unit and ‘q’ denotes a number of pixels along a height dimension of the spatial unit. Further, each spatial unit is characterized by a depth (display screens are flat but a projection needs to extend into the screens). Consequently, each spatial unit of a group of spatial units includes p×q×r pixels, wherein ‘r’ denotes the depth in terms of a number of pixels. In some embodiments, p is equal to q is equal to r. For example, in some embodiments, each spatial unit of a group of spatial units includes 32×32×32 pixels (that is, p=q=r=32). In various embodiments, the one or more successive groups of spatial units are determined by progressively dividing the total number of pixels by multiples of a predefined divisor ‘d’. In some embodiments, the divisor is a natural number greater than 1, such as 2, 3, 4, 5 or any other natural number.


For example, a first group of spatial units, representative of a highest, or first resolution, has ‘m/p×n/q×r’ spatial units (‘m/p’ being the width, ‘n/q’ being the height and ‘r’ being the depth of the cluster). A successive second group of spatial units, representative of a second resolution (the second resolution being lower than the first resolution), has ‘m/(p*d)×n/(q*d)×r/d’ spatial units. A successive third group of spatial units, representative of a third resolution (the third resolution being lower than the second resolution), has ‘m/(p*d2)×n/(q*d2)×r/d2’ spatial units. Similarly, a successive fourth group of spatial units, representative of a fourth resolution (the fourth resolution being lower than the third resolution), has ‘m/(p*d3)×n/(q*d3)×r/d3’ spatial units, and so on. This example illustrates a case of moving from a highest resolution group of spatial units to a lowest resolution group of spatial units.


It should be appreciated that when generating the first, second, third and fourth group of spatial units, division of a number of pixels by the divisor ‘d’ may result in a non-integer number of pixels. For example, m/(p*d), n/(q*d) and/or r/d may be non-integer values. In such scenarios the resulting non-integer number of pixels is rounded up or down to an integer number.


It should be appreciated that it is not necessary to generate a specific number (for example, four in the present example) of groups of spatial units. In alternate embodiments, lesser or more number of groups of spatial units may be generated. In embodiments, the number of groups of spatial units generated depends on a screen resolution. The higher the screen resolution, the more groups are generated. For example, doubling the screen resolution corresponds to adding one more spatial unit group and, similarly, halving the resolution corresponds to removing one spatial unit group. In an embodiment, for a screen resolution of 1920×1080, an optimal number of spatial unit groups is determined as four. In various embodiments, depending on an actual screen resolution used, the determined optimal number of groups may be increased or decreased.


Choice of the predefined divisor depends at least upon the total resolution (the total number of pixels), the plurality of decals associated with the camera field of view, and the total number of groups of spatial units desired to be generated to optimize the time required to render authored objects and environments without consuming significant computer memory and processing resources.


The method 200 describes determining one or more decal projections that intersect a spatial unit, in a game space, beginning by testing the decals against a lowest resolution group of spatial units and then, subsequently testing against higher and higher resolution group of spatial units until a highest resolution group of spatial units is reached.


At step 206, each spatial unit in a first group of spatial units, representing the lowest resolution is tested, checked or analyzed to determine a subset of decal projections, of the plurality of decal projections, which intersect the spatial unit. In some embodiments, the decal projections are tested using the Separating Axis Theorem (SAT) in order to determine the subset of decal projections intersecting each spatial unit. It should be appreciated that the number of decal projections in the subset are less than the plurality of decal projections. In some embodiments, the subset of decal projections is sorted, in ascending order, based on the identification number associated with each decal projection.


Continuing with the example, each spatial unit in the first group of spatial units, representative of the first resolution (that is, the lowest resolution in the example), is tested to determine a first subset of decal projections, of the plurality of decal projections, that intersect the spatial unit. Assume, for illustration, that the camera field of view encompasses ‘N’ number of decals and that all decals are evenly distributed to every spatial unit in a group of spatial units. Therefore, the first group of spatial units requires m/(p*d3)×n/(q*d3)×r/d3×N tests. After performing these many tests, every spatial unit in the first group of spatial units has a first subset of decals ND1=N/(m/(p*d3)×n/(q*d3)×r/d3) decals. If the first subset is a non-integer number then it is rounded up or down. It should be appreciated that ND1<N. In embodiments, ND1 is an average number of decals per spatial unit and, since it is rarely a multiple of number of spatial units, after every step, ND1 is rounded up in order to avoid a situation where, after a few steps, the number of decal projections becomes zero.


At step 208, it is determined if all of the one or more groups of spatial units have been tested for intersection with decal projections. If not, then, at step 210, each spatial unit in a next group of spatial units, representing a higher resolution than the already tested group of spatial units, is tested to determine a next subset of decal projections, of the subset of decal projections determined for the immediate previously tested (lower resolution) group of spatial units, that intersect the spatial unit. In other words, the subset of decal projections, determined for the immediate previously tested (lower resolution) group of spatial units, is tested for intersection with each spatial unit of the next higher resolution group of spatial units. That is, the next group of spatial units is tested with the shortlist or subset of intersecting decal projections determined for the immediate previously tested group of spatial units. Thus, the next group of spatial units is being tested with fewer number of decal projections compared to the previous group of spatial units. In some embodiments, the next subset of decal projections is sorted, in ascending order, based on the identification number associated with each decal projection.


Thereafter, as part of an iterative process, the flow moves back to step 208. Continuing with the example, each spatial unit in the second group of spatial units, representative of the second resolution (that is, of higher resolution than that of the first group of spatial units), is tested to determine a second subset of decal projections, of the first subset of decal projections, that intersect the spatial unit.


Therefore, the second group of spatial units requires m/(p*d2)×n/(q*d2)×r/d2×ND1 tests. After performing these many tests, every spatial unit in the second group of spatial units has a second subset of decals ND2=ND1/d3 decals, as each spatial unit is subdivided ‘d’ times along each of the three axes X, Y, Z, and decals are assumed to be evenly distributed to all spatial units of the next group. Again, if the second subset is a non-integer number then it is rounded up or down. It should be appreciated that ND1<ND1.


The iterative process (using steps 208 and 210) of testing decals for intersection with spatial units continues with successive groups of spatial units, wherein each of the successive groups of spatial units represents a higher resolution than that of an immediate previously tested group of spatial units. Test at each group of spatial units results in a shortlist or subset of intersecting decal projections per spatial unit wherein the shortlist or subset of intersecting decal projections per spatial unit is sorted, in ascending order, based on the identification number associated with each decal projection. As a result, every spatial unit has an associated short list of decals (e.g., less than a dozen) that potentially intersects pixels that fall into the spatial unit.


The resulting shortlist or subset of intersecting decal projections per spatial unit for a group of spatial units is used as input list of decal projections to be tested for a next/subsequent group of spatial units, wherein the next/subsequent group of spatial units is representative of a higher resolution compared to the immediate previously tested group of spatial units. Consequently, fewer and fewer number of decals are tested moving from the lowest resolution group of spatial units, through intermediate resolution group of spatial units (if any), to the highest resolution group of spatial units.


The iterative process of testing decals for intersection with spatial units continues until the highest resolution group of spatial units is reached. The testing of decals for intersection with each spatial unit of the highest resolution group of spatial units results in a final shortlist or subset of decals per spatial unit. In some embodiments, the final shortlist or subset of intersecting decal projections per spatial unit, is sorted, in ascending order, based on the identification number associated with each decal projection. In some embodiments, all determined shortlists or subsets of intersecting decal projections per spatial unit are sorted in ascending order at each stage of testing.


Continuing with the example, each spatial unit in the third group of spatial units, representative of the third resolution (that is, of higher resolution than that of the second group of spatial units), is tested to determine a third subset of decal projections, of the second subset of decal projections, that intersect the spatial unit. Therefore, the third group of spatial units requires m/(p*d)×n/(q*d)×r/d×ND2 tests. After performing these many tests, every spatial unit in the third group of spatial units has a third subset of decals ND3. Again, if the third subset is a non-integer number then it is rounded up or down. It should be appreciated that ND3<ND2.


Finally, each spatial unit in the fourth group of spatial units, representative of the fourth resolution (that is, of highest resolution), is tested to determine a fourth subset of decal projections, of the third subset of decal projections, that intersect the spatial unit. Therefore, the fourth group of spatial units requires m/p×n/q×r×ND3 tests. After performing these many tests, every spatial unit in the fourth group of spatial units has a fourth subset of decals ND4. Again, if the fourth subset is a non-integer number then it is rounded up or down. It should be appreciated that ND4<ND3. In the example, the fourth subset is the final shortlist or subset of decal projections per spatial unit.


If, at step 208, it is determined that all of the one or more groups of spatial units have been tested for intersection with decal projections then, at step 212, the (sorted) final shortlist or subset of decal projections per spatial unit is used during spatial unit shading, for the frame, and the method 200 terminates. Continuing with the example, the fourth subset of decal projections per spatial unit ND4 is used during spatial unit shading.


As discussed, in some embodiments, the shortlist or subset of intersecting decal projections per spatial unit for each group of spatial units, is sorted in ascending order, based on the identification number associated with each decal projection. This enables faster spatial unit shading. In some embodiments, the output or resulting shortlist or subset of intersecting decal projections per spatial unit, for each group of spatial units, is stored in a flat array instead of a linked list. This also improves cache utilization, because during the iterative process (using steps 208 and 210) there is a high possibility that a next decal identification number (or index) is already in cache. Linked list tends to result in more random access patterns. Since decal projection counts per spatial unit vary greatly, for example, for a single frame, said counts may vary from zero to a couple of dozens, efficient implementation on graphics processing units might be very complex. Hence, in embodiments, the shortlist or subset of intersecting decal projections per spatial unit for each group of spatial units is sorted throughout during the processing process rather than at the end when a final list is obtained.


Persons of ordinary skill in the art would appreciate that, in practice, decals may not be evenly distributed across all spatial units. There may be spatial units with no decals at all and other spatial units with several decals. Therefore, in some embodiments, if a spatial unit in a group of spatial units is determined to be empty or devoid of intersecting decal projections (based on a test using the SAT) then no further tests are conducted for the smaller spatial units, encapsulated by the spatial unit. Thus, an entire hierarchy of tests is skipped for the spatial unit, thereby, further reducing the number of tests for determining intersecting decals.


In sum, the SAS module 132 implements a tile-rendered volume decal method 200 of FIG. 2, where: a) decals are applied to both object and world geometry during the spatial unit shading step of the render process; b) the decals are defined as planar projection volumes in the game space; c) the display screen or camera field of view is divided into at least one grid or group of spatial units (spatial acceleration structure); d) the decal volumes are rendered into the at least one grid during a preparatory render step to determine which decals are visible in which grid spatial units; and e) when rendering the object and world geometry within a spatial unit, the spatial unit shader considers only the decals visible in that spatial unit.


Stated differently, the SAS module 132 implements the method 200 of FIG. 2 in order to determine where and how decals intersect a spatial unit in in the virtual world space of the game. For this, the module 132 subdivides the virtual camera field of view into grids of spatial unit wherein every spatial unit intersects a number of world space pixels. Eventually, the module 132 is configured to generate a list of decals that intersect those spatial units. In various embodiments, as camera field of view changes, the module 132 builds the list of intersecting decals per frame by scratch.


An additional advantage is that the method 200 maps seamlessly to asynchronously compute on current generation GPUs. Cost of clustering (building a list of decal projections per spatial unit) is effectively free, because the computations related to clustering are scheduled at a time when the GPU has available free resources for computations.


Use Case Scenario


Described below is an exemplary use case scenario of the method 200 of FIG. 2. As explained earlier, in embodiments, each spatial unit has a length comprising a first predefined number of pixels, and a width comprising a second predefined number of pixels, and a depth defined in terms of world space units representing a distance/depth into the world space. For example, in an exemplary embodiment, a spatial unit may have a length and width of 32 pixels each and a depth of 1 meter (depth into the game world) on the world coordinate system. In a specific embodiment, a spatial unit may be defined in terms of a plurality of specific pixel dimensions. For example, a Full HD display screen of a player's client device is considered having a resolution of 1920×1080 pixels. That is, m=1920 and n=1080. This total number of pixels is divided into groups of spatial units such that each spatial unit has 32×32 pixels. That is, p=32 and q=32. To generate successive groups of spatial units, defined in terms of pixels, the divisor ‘d’ is chosen as 2. Consequently, the groups of spatial units are determined as follows:

    • First group of spatial units (representative of a lowest, or first resolution) has m/(p*d3)×n/(q*d3)×r/d3 spatial units=1920/(32*8)×1080/(32*8)×32/8 spatial units=7.5×4.2×4=8×5×4 spatial units (7.5 and 4.2 being rounded up to integer numbers).
    • Second group of spatial units (representative of a second resolution higher than the first resolution) has m/(p*d2)×n/(q*d2)×r/d2 pixels=1920/(32*4)×1080/(32*4)×32/4 pixels=15×8.4×8=15×9×8 pixels (8.4 being rounded up to an integer number).
    • Third group of spatial units (representative of a third resolution higher than the second resolution) has m/(p*d)×n/(q*d)×r/d pixels=1920/(32*2)×1080/(32*2)×32/2 pixels=30×16.85×16=30×17×16 pixels (16.85 being rounded up to an integer number).
    • Fourth group of spatial units (representative of a fourth resolution higher than the third resolution) has m/p×n/q×r pixels=1920/32×1080/32×32 spatial units=60×34×32 pixels.


It is assumed that N=1000 and, to simplify calculations, that all decals for a group of spatial units are evenly distributed to every spatial unit. Iterative process of testing intersection of decals with each spatial unit progresses as follows:


Step 1: Testing for intersection of the decals with each spatial unit begins with the lowest resolution group of spatial units—that is, with the first group of spatial units. Consequently, the first group of spatial units requires 8×5×4×1000=160000 tests. After the tests, every spatial unit has a first subset of 1000/160=6.25=approximately 7 decals.


Step 2: Each spatial unit in the second group of spatial units is now tested only against the first subset of 7 decals from the parent spatial unit. Every spatial unit of the first group of spatial units encapsulates 8 spatial units of the second group of spatial units. Consequently, the second group of spatial units requires 15×9×8×7=7560 tests. After the tests, every spatial unit has a second subset of ⅞=0.875=approximately 1 decal (assuming even distribution).


Step 3: Next, each spatial unit in the third group of spatial units is tested only against the second subset of 1 decal from the parent spatial unit. Every spatial unit of the second group of spatial units encapsulates 8 spatial units of the third group of spatial units. In embodiments, if d=2, then a spatial unit encapsulates 2*2*2 spatial units of the next group. Consequently, the third group of spatial units requires 30×17×16×1=8160 tests. After the tests, every spatial unit may have a third subset of decals. Let us assume every spatial unit will reference only one decal projection, although at this point, there will be many spatial units with no decal projections at all. For example, black spatial units 405 illustrated in a frame 400 of FIG. 4 are indicative of spatial units with no decal projections.


Step 4: Finally, each spatial unit in the fourth group of spatial units (that is, the highest resolution group of spatial units) is tested only against the third subset of one or more decals from the parent spatial unit. Every spatial unit of the third group of spatial units encapsulates 192 spatial units of the fourth group of spatial units. Consequently, the fourth group of spatial units requires 60×34×32×1=65280 tests (assuming every spatial unit will reference only one decal projection). After the tests, every spatial unit may have a fourth subset of decals. The fourth subset of decals per spatial unit determined at this last step is used for spatial unit shading.


The total number of tests amount to 160000+7560+8160+65280=241000 tests. In contrast, a conventional brute force method would test intersection between every spatial unit and every decal volume. Therefore, for 1000 decals in camera view this would require: 60×34×32×1000=65280000 tests=˜65M tests. This is a large number and approximately 300 times more than the number of tests required using the method 200 of the present specification. Thus, every spatial unit contains a shortlist or subset of decals that potentially intersect pixels that fall into the spatial unit. Processing the shortlist or subset of decals per spatial unit has a manageable computational cost.


Also, in practice, decals are not evenly distributed. There are spatial units with no decals at all and spatial units with many decals. Therefore, tests may be skipped entirely for spatial units that are empty. For instance, if at step 1 a spatial unit is determined to be empty (that is, has no intersecting decal projection), the entire hierarchy of tests—2×2×2, 4×4×4, 8×8×8—is skipped, further reducing the number of tests.


The above examples are merely illustrative of the many applications of the system of present specification. Although only a few embodiments of the present specification have been described herein, it should be understood that the present specification might be embodied in many other specific forms without departing from the spirit or scope of the specification. Therefore, the present examples and embodiments are to be considered as illustrative and not restrictive, and the specification may be modified within the scope of the appended claims.

Claims
  • 1. A method of shading spatial units during rendering of a frame of a game space in a multiplayer video game, wherein the game space is implemented on a plurality of player client devices by at least one server in data communication with the plurality of player client devices over a network, the method being implemented by the at least one server executing a plurality of programmatic instructions and comprising: acquiring a number of pixels in a field of view of a virtual camera associated with the game space, wherein the field of view has an associated plurality of decal projections;generating one or more groups of spatial units by progressively dividing the field of view, wherein the one or more groups of spatial units range from a lowest resolution group of spatial units through one or more intermediate resolution groups of spatial units to a highest resolution group of spatial units, and wherein each group of the one or more groups of spatial units represents a higher resolution compared to an immediately previous group of spatial units, and wherein each spatial unit has a predefined number of pixels acquired from the field of view of the virtual camera associated with the game space;testing each spatial unit of a lowest resolution group of spatial units to determine a subset of decal projections, of the plurality of decal projections, that intersect each spatial unit of the lowest resolution group of spatial units;iteratively testing each spatial unit of a next higher resolution group of spatial units to determine a next subset of decal projections using the subset of decal projections determined for the immediate previously tested lower resolution group of spatial units, wherein the iterative testing determines which of the subset of decal projections intersect the spatial unit of the next higher resolution group of spatial units, and wherein the iterative testing ends with testing of each spatial unit of the highest resolution group of spatial units to determine a final subset of decal projections that intersect each spatial unit of the highest resolution group of spatial units; andshading spatial units using the final subset of decal projections per spatial unit.
  • 2. The method of claim 1, wherein the field of view is progressively divided based on a predefined divisor, and wherein the divisor is a natural number greater than 1.
  • 3. The method of claim 1, wherein a number of decal projections in the subset of decal projections is less than the plurality of decal projections.
  • 4. The method of claim 3, wherein a number of decal projections in the next subset of decal projections is less than the subset of decal projections.
  • 5. The method of claim 1, wherein at least the final subset of decal projections is sorted, in ascending order, based on an identification number associated with each decal projection.
  • 6. The method of claim 1, wherein if a spatial unit, of the one or more groups of spatial units, is determined to be devoid of intersecting decals then no further tests are conducted for higher resolution spatial units encapsulated by the spatial unit.
  • 7. The method of claim 1 wherein each spatial unit has a length comprising a first predefined number of pixels, a width comprising a second predefined number of pixels, and a depth defined in terms of world space units representing a distance/depth into a world space corresponding to the game space of the multiplayer video game.
  • 8. A system for determining shading of spatial units during rendering a frame of a game space, the system comprising: at least one server in data communication with a plurality of player client devices over a network, wherein the at least one server implements the game space on the plurality of player client devices and is configured to: acquiring a number of pixels in a field of view of a virtual camera associated with the game space, wherein the field of view has associated a plurality of decal projections;generating one or more groups of spatial units by progressively dividing the field of view, wherein the one or more groups of spatial units range from a lowest resolution group of spatial units through one or more intermediate resolution groups of spatial units to a highest resolution group of spatial units, and wherein each group of the one or more groups of spatial units represents a higher resolution compared to an immediately previous group of spatial units, and wherein each spatial unit has a predefined number of pixels acquired from the field of view of the virtual camera associated with the game space;testing each spatial unit of a lowest resolution group of spatial units to determine a subset of decal projections, of the plurality of decal projections, that intersect the spatial unit of the lowest resolution group of spatial units;iteratively testing each spatial unit of a next higher resolution group of spatial units to determine a next subset of decal projections, of the subset of decal projections determined for the immediate previously tested lower resolution group of spatial units, that intersect the spatial unit of the next higher resolution group of spatial units, wherein the iterative testing ends with testing of each spatial unit of the highest resolution group of spatial units to determine a final subset of decal projections that intersect the spatial unit of the highest resolution group of spatial units; andusing the final subset of decal projections per spatial unit during the shading of spatial units.
  • 9. The system of claim 8, wherein the field of view is progressively divided based on a predefined divisor, and wherein the divisor is a natural number greater than 1.
  • 10. The system of claim 8, wherein a number of decal projections in the subset of decal projections is less than the plurality of decal projections.
  • 11. The system of claim 10, wherein a number of decal projections in the next subset of decal projections is less than the subset of decal projections.
  • 12. The system of claim 8, wherein at least the final subset of decal projections is sorted, in ascending order, based on an identification number associated with each decal projection.
  • 13. The system of claim 8, wherein if a spatial unit in a group of spatial units, of the one or more groups of spatial units, is determined to be devoid of intersecting decals then no further tests are conducted for smaller spatial units, encapsulated by the spatial unit, in higher resolution groups of spatial units.
  • 14. The system of claim 8 wherein each spatial unit has a length comprising a first predefined number of pixels, a width comprising a second predefined number of pixels, and a depth defined in terms of world space units representing a distance/depth into a world space corresponding to the game space.
  • 15. A method of determining shading of spatial units while rendering a frame of a game space, wherein the game space has an associated virtual camera defining a field of view of a plurality of pixels, the method being implemented in a computer having a processor and a random access memory, wherein the processor is in data communication with a display and with a storage unit, the method comprising: using the field of view to generate one or more groups of spatial units, wherein the one or more groups of spatial units range from a lowest resolution group of spatial units through one or more intermediate resolution groups of spatial units to a highest resolution group of spatial units, and wherein each group of the one or more groups of spatial units represents a higher resolution compared to an immediately previous group of spatial units, and wherein each spatial unit has a predefined number of pixels;for each spatial unit of a lowest resolution group of spatial units, determining a subset of decal projections that intersect the spatial unit of the lowest resolution group of spatial units;for each spatial unit of a next higher resolution group of spatial units, iteratively determining a next subset of decal projections, of the subset of decal projections determined for the immediate previous lower resolution group of spatial units, that intersect the spatial unit of the next higher resolution group of spatial units, wherein the iterative determining ends when for each spatial unit of the highest resolution group of spatial units a final subset of decal projections is determined that intersect the spatial unit of the highest resolution group of spatial units; andusing the final subset of decal projections during the shading of spatial units.
  • 16. The method of claim 15, wherein the one or more groups of spatial units are generated by progressively dividing the field of view by a predefined divisor, and wherein the divisor is a natural number greater than 1.
  • 17. The method of claim 15, wherein a number of decal projections in the subset of decal projections is less than the plurality of decal projections and wherein a number of decal projections in the next subset of decal projections is less than the subset of decal projections determined for the immediate previous lower resolution group of spatial units.
  • 18. The method of claim 15, wherein at least the final subset of decal projections is sorted, in ascending order, based on an identification number associated with each decal projection.
  • 19. The method of claim 15, wherein if a spatial unit in a group of spatial units, of the one or more groups of spatial units, is determined to be devoid of intersecting decals then smaller spatial units, encapsulated by the spatial unit, in higher resolution groups of spatial units are not processed for determining intersecting decal projections.
  • 20. The method of claim 15 wherein each spatial unit has a length comprising a first predefined number of pixels, a width comprising a second predefined number of pixels, and a depth defined in terms of world space units representing a distance/depth into a world space corresponding to the game space.
CROSS-REFERENCE

The present application relies on, for priority, U.S. Patent Provisional Application No. 63/131,966, titled “Methods and Systems for Determining Decal Projections Intersecting Pixels in a Frame of a Game Space” and filed on Dec. 30, 2020. The above-referenced application is herein incorporated by reference in its entirety.

US Referenced Citations (291)
Number Name Date Kind
5530796 Wang Jun 1996 A
5561736 Moore Oct 1996 A
5563946 Cooper Oct 1996 A
5685775 Bakoglu Nov 1997 A
5706507 Schloss Jan 1998 A
5708764 Borrel Jan 1998 A
5736985 Lection Apr 1998 A
5737416 Cooper Apr 1998 A
5745678 Herzberg Apr 1998 A
5768511 Galvin Jun 1998 A
5825877 Dan Oct 1998 A
5835692 Cragun Nov 1998 A
5878233 Schloss Mar 1999 A
5883628 Mullaly Mar 1999 A
5900879 Berry May 1999 A
5903266 Berstis May 1999 A
5903271 Bardon May 1999 A
5911045 Leyba Jun 1999 A
5920325 Morgan Jul 1999 A
5923324 Berry Jul 1999 A
5969724 Berry Oct 1999 A
5977979 Clough Nov 1999 A
5990888 Blades Nov 1999 A
6014145 Bardon Jan 2000 A
6025839 Schell Feb 2000 A
6059842 Dumarot May 2000 A
6069632 Mullaly May 2000 A
6081270 Berry Jun 2000 A
6081271 Bardon Jun 2000 A
6091410 Lection Jul 2000 A
6094196 Berry Jul 2000 A
6098056 Rusnak Aug 2000 A
6104406 Berry Aug 2000 A
6111581 Berry Aug 2000 A
6134588 Guenthner Oct 2000 A
6144381 Lection Nov 2000 A
6148328 Cuomo Nov 2000 A
6184888 Yuasa Feb 2001 B1
6185614 Cuomo Feb 2001 B1
6201881 Masuda Mar 2001 B1
6222551 Schneider Apr 2001 B1
6271842 Bardon Aug 2001 B1
6271843 Lection Aug 2001 B1
6282547 Hirsch Aug 2001 B1
6311206 Malkin Oct 2001 B1
6334141 Varma Dec 2001 B1
6336134 Varma Jan 2002 B1
6337700 Kinoe Jan 2002 B1
6353449 Gregg Mar 2002 B1
6356297 Cheng Mar 2002 B1
6411312 Sheppard Jun 2002 B1
6426757 Smith Jul 2002 B1
6445389 Bossen Sep 2002 B1
6452593 Challener Sep 2002 B1
6462760 Cox, Jr. Oct 2002 B1
6469712 Hilpert, Jr. Oct 2002 B1
6473085 Brock Oct 2002 B1
6499053 Marquette Dec 2002 B1
6505208 Kanevsky Jan 2003 B1
6525731 Suits Feb 2003 B1
6549933 Barrett Apr 2003 B1
6567109 Todd May 2003 B1
6618751 Challenger Sep 2003 B1
RE38375 Herzberg Dec 2003 E
6657617 Paolini Dec 2003 B2
6657642 Bardon Dec 2003 B1
6684255 Martin Jan 2004 B1
6717600 Dutta Apr 2004 B2
6734884 Berry May 2004 B1
6765596 Lection Jul 2004 B2
6781607 Benham Aug 2004 B1
6819669 Rooney Nov 2004 B2
6832239 Kraft Dec 2004 B1
6836480 Basso Dec 2004 B2
6886026 Hanson Apr 2005 B1
6948168 Kuprionas Sep 2005 B1
RE38865 Dumarot Nov 2005 E
6993596 Hinton Jan 2006 B2
7028296 Irfan Apr 2006 B2
7062533 Brown Jun 2006 B2
7143409 Herrero Nov 2006 B2
7209137 Brokenshire Apr 2007 B2
7230616 Taubin Jun 2007 B2
7249123 Elder Jul 2007 B2
7263511 Bodin Aug 2007 B2
7287053 Bodin Oct 2007 B2
7305438 Christensen Dec 2007 B2
7308476 Mannaru Dec 2007 B2
7404149 Fox Jul 2008 B2
7426538 Bodin Sep 2008 B2
7427980 Partridge Sep 2008 B1
7428588 Berstis Sep 2008 B2
7429987 Leah Sep 2008 B2
7436407 Doi Oct 2008 B2
7439975 Hsu Oct 2008 B2
7443393 Shen Oct 2008 B2
7447996 Cox Nov 2008 B1
7467181 McGowan Dec 2008 B2
7475354 Guido Jan 2009 B2
7478127 Creamer Jan 2009 B2
7484012 Hinton Jan 2009 B2
7503007 Goodman Mar 2009 B2
7506264 Polan Mar 2009 B2
7515136 Kanevsky Apr 2009 B1
7525964 Astley Apr 2009 B2
7552177 Kessen Jun 2009 B2
7565650 Bhogal Jul 2009 B2
7571224 Childress Aug 2009 B2
7571389 Broussard Aug 2009 B2
7580888 Ur Aug 2009 B2
7596596 Chen Sep 2009 B2
7640587 Fox Dec 2009 B2
7667701 Leah Feb 2010 B2
7698656 Srivastava Apr 2010 B2
7702784 Berstis Apr 2010 B2
7714867 Doi May 2010 B2
7719532 Schardt May 2010 B2
7719535 Tadokoro May 2010 B2
7734691 Creamer Jun 2010 B2
7737969 Shen Jun 2010 B2
7743095 Goldberg Jun 2010 B2
7747679 Galvin Jun 2010 B2
7765478 Reed Jul 2010 B2
7768514 Pagan Aug 2010 B2
7773087 Fowler Aug 2010 B2
7774407 Daly Aug 2010 B2
7782318 Shearer Aug 2010 B2
7792263 D Amora Sep 2010 B2
7792801 Hamilton, II Sep 2010 B2
7796128 Radzikowski Sep 2010 B2
7808500 Shearer Oct 2010 B2
7814152 McGowan Oct 2010 B2
7827318 Hinton Nov 2010 B2
7843471 Doan Nov 2010 B2
7844663 Boutboul Nov 2010 B2
7847799 Taubin Dec 2010 B2
7856469 Chen Dec 2010 B2
7873485 Castelli Jan 2011 B2
7882222 Dolbier Feb 2011 B2
7882243 Ivory Feb 2011 B2
7884819 Kuesel Feb 2011 B2
7886045 Bates Feb 2011 B2
7890623 Bates Feb 2011 B2
7893936 Shearer Feb 2011 B2
7904829 Fox Mar 2011 B2
7921128 Hamilton, II Apr 2011 B2
7940265 Brown May 2011 B2
7945620 Bou-Ghannam May 2011 B2
7945802 Hamilton, II May 2011 B2
7970837 Lyle Jun 2011 B2
7970840 Cannon Jun 2011 B2
7985138 Acharya Jul 2011 B2
7990387 Hamilton, II Aug 2011 B2
7996164 Hamilton, II Aug 2011 B2
8001161 George Aug 2011 B2
8004518 Fowler Aug 2011 B2
8005025 Bodin Aug 2011 B2
8006182 Bates Aug 2011 B2
8013861 Hamilton, II Sep 2011 B2
8018453 Fowler Sep 2011 B2
8018462 Bhogal Sep 2011 B2
8019797 Hamilton, II Sep 2011 B2
8019858 Bauchot Sep 2011 B2
8022948 Garbow Sep 2011 B2
8022950 Brown Sep 2011 B2
8026913 Garbow Sep 2011 B2
8028021 Reisinger Sep 2011 B2
8028022 Brownholtz Sep 2011 B2
8037416 Bates Oct 2011 B2
8041614 Bhogal Oct 2011 B2
8046700 Bates Oct 2011 B2
8051462 Hamilton, II Nov 2011 B2
8055656 Cradick Nov 2011 B2
8056121 Hamilton, II Nov 2011 B2
8057307 Berstis Nov 2011 B2
8062130 Smith Nov 2011 B2
8063905 Brown Nov 2011 B2
8070601 Acharya Dec 2011 B2
8082245 Bates Dec 2011 B2
8085267 Brown Dec 2011 B2
8089481 Shearer Jan 2012 B2
8092288 Theis Jan 2012 B2
8095881 Reisinger Jan 2012 B2
8099338 Betzler Jan 2012 B2
8099668 Garbow Jan 2012 B2
8102334 Brown Jan 2012 B2
8103640 Lo Jan 2012 B2
8103959 Cannon Jan 2012 B2
8105165 Karstens Jan 2012 B2
8108774 Finn Jan 2012 B2
8113959 De Judicibus Feb 2012 B2
8117551 Cheng Feb 2012 B2
8125485 Brown Feb 2012 B2
8127235 Haggar Feb 2012 B2
8127236 Hamilton, II Feb 2012 B2
8128487 Hamilton, II Mar 2012 B2
8131740 Cradick Mar 2012 B2
8132235 Bussani Mar 2012 B2
8134560 Bates Mar 2012 B2
8139060 Brown Mar 2012 B2
8139780 Shearer Mar 2012 B2
8140340 Bhogal Mar 2012 B2
8140620 Creamer Mar 2012 B2
8140978 Betzler Mar 2012 B2
8140982 Hamilton, II Mar 2012 B2
8145676 Bhogal Mar 2012 B2
8145725 Dawson Mar 2012 B2
8149241 Do Apr 2012 B2
8151191 Nicol, II Apr 2012 B2
8156184 Kurata Apr 2012 B2
8165350 Fuhrmann Apr 2012 B2
8171407 Huang May 2012 B2
8171408 Dawson May 2012 B2
8171559 Hamilton, II May 2012 B2
8174541 Greene May 2012 B2
8176421 Dawson May 2012 B2
8176422 Bergman May 2012 B2
8184092 Cox May 2012 B2
8184116 Finn May 2012 B2
8185450 McVey May 2012 B2
8185829 Cannon May 2012 B2
8187067 Hamilton, II May 2012 B2
8199145 Hamilton, II Jun 2012 B2
8203561 Carter Jun 2012 B2
8214335 Hamilton, II Jul 2012 B2
8214433 Dawson Jul 2012 B2
8214750 Hamilton, II Jul 2012 B2
8214751 Dawson Jul 2012 B2
8217953 Comparan Jul 2012 B2
8219616 Dawson Jul 2012 B2
8230045 Kawachiya Jul 2012 B2
8230338 Dugan Jul 2012 B2
8233005 Finn Jul 2012 B2
8234234 Shearer Jul 2012 B2
8234579 Do Jul 2012 B2
8239775 Beverland Aug 2012 B2
8241131 Bhogal Aug 2012 B2
8245241 Hamilton, II Aug 2012 B2
8245283 Dawson Aug 2012 B2
8265253 D Amora Sep 2012 B2
8310497 Comparan Nov 2012 B2
8334871 Hamilton, II Dec 2012 B2
8360886 Karstens Jan 2013 B2
8364804 Childress Jan 2013 B2
8425326 Chudley Apr 2013 B2
8442946 Hamilton, II May 2013 B2
8506372 Chudley Aug 2013 B2
8514249 Hamilton, II Aug 2013 B2
8554841 Kurata Oct 2013 B2
8607142 Bergman Dec 2013 B2
8607356 Hamilton, II Dec 2013 B2
8624903 Hamilton, II Jan 2014 B2
8626836 Dawson Jan 2014 B2
8692835 Hamilton, II Apr 2014 B2
8721412 Chudley May 2014 B2
8827816 Bhogal Sep 2014 B2
8838640 Bates Sep 2014 B2
8849917 Dawson Sep 2014 B2
8911296 Chudley Dec 2014 B2
8992316 Smith Mar 2015 B2
9007379 Shafer Apr 2015 B1
9083654 Dawson Jul 2015 B2
9152914 Haggar Oct 2015 B2
9205328 Bansi Dec 2015 B2
9286731 Hamilton, II Mar 2016 B2
9299080 Dawson Mar 2016 B2
9364746 Chudley Jun 2016 B2
9525746 Bates Dec 2016 B2
9583109 Kurata Feb 2017 B2
9682324 Bansi Jun 2017 B2
9764244 Bansi Sep 2017 B2
9789406 Marr Oct 2017 B2
9808722 Kawachiya Nov 2017 B2
20050266921 Hayashida Dec 2005 A1
20070270215 Miyamoto Nov 2007 A1
20080143722 Pagan Jun 2008 A1
20090113448 Smith Apr 2009 A1
20140038708 Davison Feb 2014 A1
20140344725 Bates Nov 2014 A1
20150258439 Prosin Sep 2015 A1
20150273333 Ciszewski Oct 2015 A1
20160191671 Dawson Jun 2016 A1
20170301135 Jagnow Oct 2017 A1
20180160049 Aizawa Jun 2018 A1
20190005712 Nevraev Jan 2019 A1
20190073747 Burch Mar 2019 A1
20190080495 Andronikos Mar 2019 A1
20200293176 Yoganandan Sep 2020 A1
20200357163 Wang Nov 2020 A1
20210117070 Muta Apr 2021 A1
20220261953 Li Aug 2022 A1
Foreign Referenced Citations (71)
Number Date Country
768367 Mar 2004 AU
2005215048 Oct 2011 AU
2143874 Jun 2000 CA
2292678 Jul 2005 CA
2552135 Jul 2013 CA
1334650 Feb 2002 CN
1202652 Oct 2002 CN
1141641 Mar 2004 CN
1494679 May 2004 CN
1219384 Sep 2005 CN
1307544 Mar 2007 CN
100407675 Jul 2008 CN
100423016 Oct 2008 CN
100557637 Nov 2009 CN
101001678 May 2010 CN
101436242 Dec 2010 CN
101801482 Dec 2014 CN
668583 Aug 1995 EP
0627728 Sep 2000 EP
0717337 Aug 2001 EP
0679977 Oct 2002 EP
0679978 Mar 2003 EP
0890924 Sep 2003 EP
1377902 Aug 2004 EP
0813132 Jan 2005 EP
1380133 Mar 2005 EP
1021021 Sep 2005 EP
0930584 Oct 2005 EP
0883087 Aug 2007 EP
1176828 Oct 2007 EP
2076888 Jul 2015 EP
2339938 Oct 2002 GB
2352154 Jul 2003 GB
3033956 Apr 2000 JP
3124916 Jan 2001 JP
3177221 Jun 2001 JP
3199231 Aug 2001 JP
3210558 Sep 2001 JP
3275935 Feb 2002 JP
3361745 Jan 2003 JP
3368188 Jan 2003 JP
3470955 Sep 2003 JP
3503774 Dec 2003 JP
3575598 Jul 2004 JP
3579823 Jul 2004 JP
3579154 Oct 2004 JP
3701773 Oct 2005 JP
3777161 Mar 2006 JP
3914430 Feb 2007 JP
3942090 Apr 2007 JP
3962361 May 2007 JP
4009235 Sep 2007 JP
4225376 Dec 2008 JP
4653075 Dec 2010 JP
5063698 Aug 2012 JP
5159375 Mar 2013 JP
5352200 Nov 2013 JP
5734566 Jun 2015 JP
117864 Aug 2004 MY
55396 Dec 1998 SG
2002073457 Sep 2002 WO
20020087156 Oct 2002 WO
2004086212 Oct 2004 WO
2005079538 Sep 2005 WO
2007101785 Sep 2007 WO
2008037599 Apr 2008 WO
2008074627 Jun 2008 WO
2008095767 Aug 2008 WO
2009037257 Mar 2009 WO
2009104564 Aug 2009 WO
2010096738 Aug 2010 WO
Non-Patent Literature Citations (1)
Entry
Martin Evans, Drawing Stuff On Other Stuff With Deferred Screenspace Decals, 2015, martindevans.me, pp. 1-8, at https://martindevans.me/game-development/2015/02/27/Drawing-Stuff-On-Other-Stuff-With-Deferred-Screenspace-Decals (last visited Dec. 12, 2023). (Year: 2015).
Related Publications (1)
Number Date Country
20220203231 A1 Jun 2022 US
Provisional Applications (1)
Number Date Country
63131966 Dec 2020 US