This application claims the benefit under 35 USC § 119(a) of Korean Patent Application No. 10-2022-0159547, filed on Nov. 24, 2022 in the Korean Intellectual Property Office, the entire disclosure of which is incorporated herein by reference for all purposes.
The following description relates to a device and method with scene component information estimation.
A goal of computer graphics may include a photorealistic synthesis of an image and a video. Methods and representations may mimic an imaging model of a real camera, including complex materials and global illumination processing. These methods may simulate light propagation from a light source to a virtual camera for synthesis based on the law of physics. For this, all physical parameters of a scene may need to be determined in a rendering process. For example, the parameters may include information about scene geometry and material properties such as reflectivity or opacity. When such information is provided, the latest ray tracing techniques may generate photorealistic images. In addition to the physics-based rendering methods, there are techniques, based on mathematical approximation, close to real imaging models.
This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.
In one or more general aspects, an electronic device includes: one or more processors configured to: extract, using an implicit neural representation (INR) model, a global geometry feature and information indicating whether a point is on a surface from a viewpoint and a view direction corresponding to an image pixel corresponding to a two-dimensional (2D) scene at the viewpoint within a field of view (FOV); determine an object surface position corresponding to the viewpoint and the view direction and normal information of the object surface position based on the information indicating whether the point is on the surface; estimate, using an albedo estimation model, albedo information independent of the view direction from the global geometry feature, the object surface position, and the normal information; and estimate, using a specular estimation model, specular information dependent on the view direction from the global geometry feature, the object surface position, the normal information, and the view direction.
The one or more processors may be configured to determine a pixel value of the image pixel based on scene component information including any one or any combination of any two or more of visibility information, indirect light information, and direct light information together with the albedo information and the specular information.
The one or more processors may be configured to estimate, using a machine learning model, the scene component information from the global geometry feature, the object surface position, and the normal information, for ray directions departing from the object surface position.
For the determining of the pixel value, the one or more processors may be configured to individually estimate visibility information using a visibility estimation model, indirect light information using an indirect light estimation model, and direct light information using a direct light estimation model from the global geometry feature, the object surface position and the normal information, for ray directions departing from the object surface position.
The one or more processors may be configured to: estimate scene component information including albedo information and specular information for view directions corresponding to image pixels corresponding to the 2D scene from the viewpoint; and generate a 2D image by determining pixel values of the image pixels using scene component information estimated for the image pixels.
The one or more processors may be configured to: adjust any one or any combination of any two or more of scene component of visibility information, indirect light information, direct light information, the albedo information, and the specular information, based on a user input; and determine a pixel value of a pixel of the 2D image corresponding to the viewpoint and the view direction based on the adjusted scene component and an estimated scene component.
The one or more processors may be configured to obtain the object surface position by repeatedly performing a ray marching based on the information indicating whether the point is on the surface, in the view direction from the viewpoint.
The one or more processors may be configured to: determine a point spaced apart from the viewpoint in the view direction; and generate, using the INR model, a global geometry feature corresponding to the determined point and distance information on a distance between the determined point and an object surface.
The one or more processors may be configured to determine normal information of the determined point by analyzing the viewpoint, the view direction, and information indicating whether the determined point is on a surface.
The INR model may be trained based on an output of a neural renderer, and the neural renderer may be configured to estimate a pixel value of an image pixel from the global geometry feature, the object surface position, the normal information, and the view direction.
The one or more processors may be configured to estimate visibility information using a visibility estimation model trained using visible distances between the object surface position and arrival points determined based on ray marching for ray directions departing from the object surface position.
The one or more processors may be configured to estimate indirect light information using an indirect light estimation model trained using color information of arrival points of rays departing from the object surface position viewed from the object surface position.
The color information of the arrival points for training of the indirect light estimation model may be estimated using the INR model and a neural renderer which are completely trained.
The albedo estimation model and the specular estimation model may be trained based on an objective function value between a ground truth (GT) 2D image and a temporary 2D image reconstructed based on albedo information output from the albedo estimation model, specular information output from the specular estimation model, and other scene component information output from a machine learning model.
The temporary 2D image may be reconstructed using an approximation that is based on a split of a rendering operation that determines an image pixel value based on scene component information into a reflection component and an illumination component.
In one or more general aspects, a processor-implemented method includes: extracting, using an implicit neural representation (INR) model, a global geometry feature and information indicating whether a point is on a surface from a viewpoint and a view direction corresponding to an image pixel corresponding to a two-dimensional (2D) scene at the viewpoint within a field of view (FOV); determining an object surface position corresponding to the viewpoint and the view direction and normal information of the object surface position based on the information indicating whether the point is on the surface; estimating, using an albedo estimation model, albedo information independent of the view direction from the global geometry feature, the object surface position, and the normal information; and estimating, using a specular estimation model, specular information dependent on the view direction from the global geometry feature, the object surface position, the normal information, and the view direction.
The method may include determining a pixel value of the image pixel based on scene component information including any one or any combination of any two or more of visibility information, indirect light information, and direct light information together with the albedo information and the specular information.
The determining of the pixel value may include individually estimating visibility information using a visibility estimation model, indirect light information using an indirect light estimation model, and direct light information using a direct light estimation model, from the global geometry feature, the object surface position and the normal information, for ray directions departing from the object surface position.
The method may include: adjusting any one or any combination of any two or more scene component of visibility information, indirect light information, direct light information, the albedo information, and the specular information, based on a user input; and determining a pixel value of a pixel of the 2D image corresponding to the viewpoint and the view direction based on the adjusted scene component and an estimated scene component.
In one or more general aspects, a non-transitory computer-readable storage medium stores instructions that, when executed by a processor, configure the processor to perform any one, any combination, or all of operations and/or methods described herein.
In one or more general aspects, an electronic device includes: one or more processors configured to: train, using an output of a neural renderer, an implicit neural representation (INR) model to output a global geometry feature of a point, the neural renderer being configured to estimate a pixel value of an image pixel corresponding to a two-dimensional (2D) scene based on the global geometry feature; train, using the global geometry feature output from the trained INR model, an albedo estimation model to output albedo information of the pixel value independent of a view direction corresponding to the image pixel; and train, using the global geometry feature output from the trained INR model, a specular estimation model to output specular information of the pixel value dependent on the view direction.
Other features and aspects will be apparent from the following detailed description, the drawings, and the claims.
Throughout the drawings and the detailed description, unless otherwise described or provided, the same drawing reference numerals will be understood to refer to the same elements, features, and structures. The drawings may not be to scale, and the relative size, proportions, and depiction of elements in the drawings may be exaggerated for clarity, illustration, and convenience.
The following detailed description is provided to assist the reader in gaining a comprehensive understanding of the methods, apparatuses, and/or systems described herein. However, various changes, modifications, and equivalents of the methods, apparatuses, and/or systems described herein will be apparent after an understanding of the disclosure of this application. For example, the sequences of operations described herein are merely examples, and are not limited to those set forth herein, but may be changed as will be apparent after an understanding of the disclosure of this application, with the exception of operations necessarily occurring in a certain order. Also, descriptions of features that are known after an understanding of the disclosure of this application may be omitted for increased clarity and conciseness.
Although terms, such as “first,” “second,” and “third”, or A, B, (a), (b), and the like may be used herein to describe various members, components, regions, layers, or sections, these members, components, regions, layers, or sections are not to be limited by these terms. Each of these terminologies is not used to define an essence, order, or sequence of corresponding members, components, regions, layers, or sections, for example, but used merely to distinguish the corresponding members, components, regions, layers, or sections from other members, components, regions, layers, or sections. Thus, a first member, component, region, layer, or section referred to in the examples described herein may also be referred to as a second member, component, region, layer, or section without departing from the teachings of the examples.
Throughout the specification, when a component or element is described as being “on”, “connected to,” “coupled to,” or “joined to” another component, element, or layer it may be directly (e.g., in contact with the other component or element) “on”, “connected to,” “coupled to,” or “joined to” the other component, element, or layer or there may reasonably be one or more other components, elements, layers intervening therebetween. When a component or element is described as being “directly on”, “directly connected to,” “directly coupled to,” or “directly joined” to another component or element, there can be no other elements intervening therebetween. Likewise, expressions, for example, “between” and “immediately between” and “adjacent to” and “immediately adjacent to” may also be construed as described in the foregoing.
The terminology used herein is for describing various examples only and is not to be used to limit the disclosure. The articles “a,” “an,” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. As non-limiting examples, terms “comprise” or “comprises,” “include” or “includes,” and “have” or “has” specify the presence of stated features, numbers, operations, members, elements, and/or combinations thereof, but do not preclude the presence or addition of one or more other features, numbers, operations, members, elements, and/or combinations thereof.
As used herein, the term “and/or” includes any one and any combination of any two or more of the associated listed items. The phrases “at least one of A, B, and C”, “at least one of A, B, or C”, and the like are intended to have disjunctive meanings, and these phrases “at least one of A, B, and C”, “at least one of A, B, or C”, and the like also include examples where there may be one or more of each of A, B, and/or C (e.g., any combination of one or more of each of A, B, and C), unless the corresponding description and embodiment necessitates such listings (e.g., “at least one of A, B, and C”) to be interpreted to have a conjunctive meaning.
Unless otherwise defined, all terms, including technical and scientific terms, used herein have the same meaning as commonly understood by one of ordinary skill in the art to which the present disclosure pertains and based on an understanding of the disclosure of the present application. Terms, such as those defined in commonly used dictionaries, are to be interpreted as having a meaning that is consistent with their meaning in the context of the relevant art and the disclosure of the present application, and are not to be interpreted in an idealized or overly formal sense unless expressly so defined herein.
The features described herein may be embodied in different forms, and are not to be construed as being limited to the examples described herein. Rather, the examples described herein have been provided merely to illustrate some of the many possible ways of implementing the methods, apparatuses, and/or systems described herein that will be apparent after an understanding of the disclosure of this application. The use of the term “may” herein with respect to an example or embodiment, e.g., as to what an example or embodiment may include or implement, means that at least one example or embodiment exists where such a feature is included or implemented, while all examples are not limited thereto.
Hereinafter, the examples will be described in detail with reference to the accompanying drawings. When describing the examples with reference to the accompanying drawings, like reference numerals refer to like components and a repeated description related thereto will be omitted.
According to an example, an electronic device (e.g., an electronic device 1000 of
The electronic device may generate an image corresponding to the 2D scene 120 when an object 191 in the 3D space 190 is viewed in an arbitrary view direction and FOV. For example, the electronic device may determine an image pixel corresponding to the viewpoint 110 and the view direction. A view direction from the viewpoint 110 may pass through an image pixel included in an image (e.g., an image of the 2D scene 120) corresponding to the 2D scene 120. The view direction passing through the image pixel may reach a point (e.g., a point on the object 191) in the 3D space 190. The electronic device may determine image information (e.g., a color value) on the image representing the point on the object 191 that the view direction reaches at the viewpoint 110 as a pixel value of a corresponding image pixel. As will be described below, the electronic device may determine a pixel value of an image pixel based on physical properties (e.g., scene component information) of a point. In the present specification, c may represent the viewpoint 110 (e.g., a position of a virtual camera), d may represent the view direction, to may represent a distance from the viewpoint 110 to a 3D point first encountered in the view direction, and p may represent the 3D point first encountered in the view direction. Information indicating the 3D point p may be referred to as point information and may include 3D coordinates. The 3D point p may also be referred to as an object surface position.
The electronic device of one or more embodiments may sequentially train each model in a network for an estimation of scene components. Accordingly, the electronic device of one or more embodiments may increase stability of training by the sequential training, compared to a typical electronic device which trains using integrated learning. Scene component information may be information indicating physical properties of each point of a scene. The electronic device may distinguish and estimate physical properties. The scene component information may include physical properties that will be described below. The physical properties may include at least one of albedo information, specular information, visibility information, indirect light information, or direct light information. An example of estimating the above five physical properties will be mainly described herein. Albedo information associated with an arbitrary point may be an albedo value (e.g., a value indicating a degree of diffuse reflection when a corresponding point receives light), and may include, for example, an albedo parameter of a bidirectional reflectance distribution function (BRDF) model (e.g., a Cook-Torrance BRDF model). The specular information may be a specular value (e.g., a value indicating a degree of specular reflection when a corresponding point receives light) of a corresponding point, and may include, for example, a specular parameter of a BRDF model. The visibility information may be information indicating visibility at a corresponding point for each of ray directions, and may include a visibility map including a value indicating a presence or an absence of a point at which rays traveling for each of ray directions departing from the corresponding point arrives. For example, when the point is present, the value may be “0”, and when the point is absent, the value may be “1”. The indirect light information may be information about indirect light for each of ray directions that have an influence on a corresponding point, and may indicate a color (or radiance) of indirect light corresponding to each of the ray directions from the corresponding point. The direct light information may indicate a color (e.g., a mixed color of light by a plurality of light sources) of a light source that has a direct influence on a corresponding point.
The electronic device of one or more embodiments may estimate albedo information and specular information using separate models (e.g., an albedo estimation model and a specular estimation model) depending on whether the albedo information and specular information depend on a view direction. Therefore, the electronic device of one or more embodiments may accurately distinguish and estimate physical properties of points in the 3D space 190 viewed in the 2D scene 120.
As will be described below, the electronic device of one or more embodiments may use ray marching with a smaller amount of computational operation than that of ray tracing. The electronic device of one or more embodiments may estimate physical properties of a point with a smaller amount of computational operation than that of physics-based rendering. Thus, the electronic device of one or more embodiments may be implemented in a mobile device as well as a graphics device of a personal computer (PC) to perform high-quality scene rendering in real time. The electronic device of one or more embodiments may perform realistic rendering on a mobile device that includes a graphics processing unit (GPU) and a neural processing unit (NPU).
According to an example, an image-based rendering (IBR) scheme of estimating a geometry of a 3D scene based on captured images and performing rendering may be combined with deep learning. For example, the electronic device may estimate scene property information (e.g., physical properties) corresponding to a 2D scene 120 obtained by observing the 3D space 190 from a new camera view (e.g., a view that has not been previously captured), based on a neural structure (e.g., a neural network) trained using images (e.g., color images based on an RGB color space) captured or rendered from various camera views (e.g., various viewpoints 110 and view directions), and may perform image rendering. The electronic device may perform realistic rendering in various dynamic changes of a physical environment (e.g., lighting, etc.).
The electronic device of one or more embodiments may render a high-quality 3D scene under various conditions (e.g., a view, a material, and lighting). The electronic device may render a realistic image representing the 2D scene 120 for visualization of a 3D game application or a 3D scene of a metaverse. When the electronic device is implemented as a mobile device, the electronic device may provide a user with a realistic experience without space constraints.
Hereinafter, an example of a network structure (e.g., a scene component estimation network) for an estimation of scene component information and an example of an inference (e.g., an estimation) of scene component information using the scene component estimation network are described with reference to
An electronic device (e.g., the electronic device 1000 of
In the present specification, a neural network may include layers, and each of the layers may include a node. The node may have a node value determined based on an activation function. A node of an arbitrary layer may be connected to a node (e.g., another node) of another layer through a link (e.g., a connection edge) having a connection weight. The node value of the node may be propagated to other nodes via links. For example, in an inference operation of the neural network, node values may be propagated forward from a previous layer to a next layer. A node value of a corresponding node may be propagated (e.g., forward propagated) to a node (e.g., a next node) of a next layer connected to the node through a connection edge. For example, a node may receive a value weighted by a connection weight from previous nodes (e.g., a plurality of nodes) connected through a connection edge. The node value of the node may be determined based on a sum (e.g., a weighted sum) of weighted values received from the previous nodes, to which an activation function is applied. Parameters of the neural network may include, for example, the connection weight described above. The parameters of the neural network may be updated such that an objective function value that will be described below may be changed in a targeted direction, for example, to minimize a loss. A non-limiting example of an objective function will be described below with reference to
According to an example, the electronic device may determine a viewpoint 211 and view directions 212 corresponding to image pixels of a 2D scene obtained by observing a 3D scene including an object at the viewpoint 211 in a FOV. A pixel position corresponding to each image pixel of an image (e.g., a 2D scene image) corresponding to a 2D scene may be determined based on a resolution of a 2D image to be reconstructed. As described above with reference to
For example, in operation 201, the electronic device may extract the global geometry feature 218 and information indicating whether a point is on a surface (e.g., the surface identification information 219), based on the INR model 210, from the viewpoint 211 and the view direction 212 corresponding to an image pixel corresponding to the 2D scene at the viewpoint 211 within the FOV. The electronic device may determine the point information 213 from the viewpoint 211 and the view direction 212 described above. The point information 213 may be information (e.g., coordinates of a point) about a point first encountered in the view direction 212 departing from the viewpoint 211 within a 3D space. The electronic device may extract the global geometry feature 218 and the surface identification information 219 from the point information 213 based on the INR model 210.
The global geometry feature 218 may be information (e.g., feature data or a feature vector) representing a feature that abstracts (or encodes) a geometry of a point indicated by the point information 213, and may be extracted from the point information 213 based on the INR model 210 as shown in
The surface identification information 219 may be information indicating whether the point indicated by the point information 213 is on a surface of an arbitrary object, and may also be referred to as information indicating whether a point is on a surface. In the present specification, the surface identification information 219 may have or include a value of a signed distance function (SDF). The SDF may have a value of “0” when the point indicated by the point information 213 is on a surface. When the point indicated by the point information 213 is inside a volume defined by the surface, the SDF may have a value indicating a shortest distance from the point to the surface using a negative sign (e.g., −). When the point indicated by the point information 213 is outside the volume defined by the surface, the SDF may have a value indicating the shortest distance from the point to the surface using a positive sign (e.g., +). However, in the present specification, while the surface identification information 219 is expressed as an SDF value for convenience of description, examples are not limited thereto.
According to an example, the electronic device may repeatedly or iteratively perform operation 201 on a point to which a ray is advanced in the view direction 212 by a distance determined based on ray marching until the surface identification information 219 indicates a surface (e.g., until the SDF has a value of “0”). For example, when the surface identification information 219 includes an SDF value (e.g., an SDF value less than or greater than “0”), the electronic device may perform an operation on the point information 213 indicating a position of a point to which a ray is advanced in the view direction 212 by the SDF value, based on the INR model 210. The electronic device may repeatedly or iteratively perform ray matching and estimation based on the INR model 210 until the SDF value becomes zero. A non-limiting example of the ray matching will be described with reference to
In operation 202, the electronic device may calculate (e.g., determine) the object surface position 221 corresponding to the viewpoint 211 and the view direction 212 and the normal information 222 of the object surface position 221, based on the information indicating whether the point is on the surface (e.g., the surface identification information 219). The electronic device may calculate the object surface position 221 and the normal information 222 from the surface identification information 219, the viewpoint 211, and the view direction 212. For example, the electronic device may use an analysis algorithm 220 to calculate the object surface position 221 and the normal information 222. The normal information 222 may include a normal value (e.g., a normal vector) at the calculated object surface position 221. The analysis algorithm 220 may include an analytic function having a gradient calculation based on the surface identification information 219, the viewpoint 211, and the view direction 212.
In operation 203, the electronic device may estimate the albedo information 231 independent of the view direction 212, based on the albedo estimation model 230, from the global geometry feature 218, the object surface position 221, and the normal information 222. For estimating the albedo information 231 using the albedo estimation model 230, the electronic device may exclude the view direction 212, and may perform an albedo estimation model-based operation (e.g., a neural network operation) on the global geometry feature 218, the object surface position 221, and the normal information 222.
In operation 204, the electronic device may estimate the specular information 241 dependent on the view direction 212, based on the specular estimation model 240, from the global geometry feature 218, the object surface position 221, the normal information 222, and the view direction 212. For estimating the specular information 241 using the specular estimation model 240, the electronic device may perform an operation using the specular estimation model 240 on the global geometry feature 218, the object surface position 221, and the normal information 222 together with the view direction 212.
According to an example, the electronic device may estimate the scene component information 251 based on the machine learning model 250 from the global geometry feature 218, the object surface position 221, and the normal information 222, for ray directions 205 departing from the object surface position 221. The ray directions 205 may be all directions from the object surface position 221 (e.g., a direction toward points on a virtual hemisphere with the object surface position 221 as the center of the base of the hemisphere). A non-limiting example of the ray directions 205 will be described with reference to
According to an example, the electronic device may determine a pixel value 290 of an image pixel based on scene component information that includes at least one of visibility information, indirect light information, or direct light information, together with the albedo information 231 and the specular information 241. For example, the electronic device may calculate the pixel value 290 from the albedo information 231, the specular information 241, the visibility information, the indirect light information, and the direct light information based on Equation 16 or 17 shown below, for example. The electronic device may estimate scene component information including the albedo information 231 and the specular information 241 for view directions 212 corresponding to image pixels corresponding to a 2D scene from the viewpoint 211. The electronic device may generate a 2D image by determining pixel values 290 of image pixels using the scene component information estimated for the image pixels. The electronic device may obtain a 2D image by determining pixel values of all image pixels corresponding to a 2D scene.
In addition, the electronic device may adjust at least one scene component of the visibility information, the indirect light information, the direct light information, the albedo information 231, or the specular information 241, based on a user input. For example, when a user determines to generate an image under different lighting conditions, the electronic device may adjust at least one of the indirect light information or the direct light information based on a user input. The electronic device may determine a pixel value 290 of a pixel of a 2D image corresponding to the viewpoint 211 and the view direction 212 based on the adjusted scene component and an estimated scene component. For example, as described above, the electronic device may obtain a 2D image by determining pixel values of all image pixels corresponding to a 2D scene.
Hereinafter, an example of training a scene component estimation network through three operations is described with reference to
Although an electronic device (e.g., the electronic device 1000 of
A training dataset 301 may include an image obtained by capturing a 3D space, and a viewpoint and a view direction in which the image is captured. The viewpoint and the view direction may be training inputs, and a pixel value of the image may be a ground truth (GT) as a training output. An image pixel value of the training dataset 301 may be referred to as a “GT pixel value”. The electronic device may perform training of the scene component estimation network using a plurality of training datasets 301. For example, each of the plurality of training datasets 301 may include images obtained by capturing the same 3D space in various viewpoints and various view directions.
According to an example, the electronic device may sequentially perform a first training operation 381, a second training operation 382, and a third training operation 383. In the second training operation 382, parameters of an INR model 310 and a neural renderer 315 may be fixed, and parameters of a visibility estimation model 350 and an indirect light estimation model 360 may be updated. In the third training operation 383, parameters of an albedo estimation model 330, a specular estimation model 340, and a direct light estimation model 370 may be updated, and parameters of the INR model 310, the visibility estimation model 350, and the indirect light estimation model 360 may be fixed. For reference, the visibility estimation model 350, the indirect light estimation model 360, the albedo estimation model 330, the specular estimation model 340, and the direct light estimation model 370 may be branched from the INR model 310 as branches distinguished from each other.
In the first training operation 381, the electronic device may train the INR model 310 and the neural renderer 315. The neural renderer 315 may be a model (e.g., a neural network) configured and trained to output image pixel values (e.g., color values) corresponding to outputs of the above-described INR model 310. Parameters (e.g., connection weights between nodes) of the INR model 310 and the neural renderer 315 may be updated based on an objective function value (e.g. A loss function value) calculated between a GT pixel value of the training dataset 301 and a temporary pixel value 395 output by the neural renderer 315 in the first training operation 381. The temporary pixel value 395 may be a color value (e.g., an RGB value) inferred for a 3D point corresponding to a viewpoint and a view direction. The GT pixel value may be a color value (e.g., an RGB value as a GT color value) of a pixel position corresponding to a viewpoint and a view direction in an image of the training dataset 301. The objective function value may be an error between the inferred color value and the GT color value, and the parameters of the INR model 310 and the neural renderer 315 may be updated to reduce the error. A non-limiting example of the first training operation 381 will be described with reference to
In the second training operation 382, the visibility estimation model 350 and the indirect light estimation model 360 may be trained. For example, in the second training operation 382, the trained INR model 310 obtained by the above-described first training operation 381 may be used. The electronic device may train the visibility estimation model 350 using outputs of the completely trained INR model 310. For example, the electronic device may calculate temporary visibility information based on the visibility estimation model 350 from light source directions together with the outputs of the completely trained INR model 310. A GT visibility value may be calculated, a non-limiting example of which will be described below with reference to
In the third training operation 383, the albedo estimation model 330, the specular estimation model 340, and the direct light estimation model 370 may be trained. When the third training operation 383 is completed, the electronic device may extract all physical properties (e.g., scene component information) used to reconstruct an image pixel value for an arbitrary view direction at an arbitrary viewpoint through inference of each model. During runtime, the electronic device may estimate scene component information for an arbitrary viewpoint and view direction based on the INR model 310, the visibility estimation model 350, the indirect light estimation model 360, the albedo estimation model 330, the specular estimation model 340, and the direct light estimation model 370, and may render a scene based on the estimated scene component information, a non-limiting example of which will be described below with reference to
In the third training operation 383, the albedo estimation model 330, the specular estimation model 340, and the direct light estimation model 370 may be additionally trained. The albedo estimation model 330 and the specular estimation model 340 may be models to estimate different types of physical material properties, respectively.
In an example, according to a rendering operation of a physical light reflection model, the physical material properties may be presented as a BRDF indicating that light energy incident in an arbitrary direction at an arbitrary 3D point is reflected in an arbitrary direction. Light reflection may be divided into diffuse reflection and specular reflection. The diffuse reflection may indicate that light energy coming in an arbitrary direction is absorbed into a surface and reflected with the same energy in all directions. The diffuse reflection may be observed as the same light reflection at all camera angles. The specular reflection may be a reflection in which incident light energy tends to be reflected from a surface in a limited direction (e.g., in one direction). In the specular reflection, observed light energy may vary depending on an angle of a camera facing a surface. In other words, the diffuse reflection may be independent of an angle of a camera (e.g., a view direction), but the specular reflection may be dependent on the view direction. As described above, the electronic device may estimate scene component information using a machine learning model of a branch separated according to a dependency on the view direction. The electronic device may estimate BRDF coefficients including a diffuse reflection parameter and a specular reflection parameter.
Albedo information output by the albedo estimation model 330 and specular information output by the specular estimation model 340 may be information associated with physical material properties (e.g., diffuse reflection and specular reflection) of a corresponding point (e.g., a 3D surface point where rays traveling in a view direction from an arbitrary viewpoint are encountered). The albedo information may include a value indicating a degree of diffuse reflection, and the specular information may include a value indicating a degree of specular reflection. Direct light information output by the direct light estimation model 370 may be information associated with a lighting condition and may indicate a color of a light source. For training of the albedo estimation model 330, the specular estimation model 340, and the direct light estimation model 370, the electronic device may extract visibility information and indirect light information based on the INR model 310, the visibility estimation model 350, and the indirect light estimation model 360 which are completely trained. The electronic device may additionally extract temporary albedo information, temporary specular information, and temporary direct light information, and may calculate the temporary pixel value 395 using the extracted scene component information.
The albedo information may include, for example, an albedo value related to diffuse reflection among BRDF coefficients, and the albedo value may be expressed as a color value (e.g., an RGB value) at a 3D point. The specular information may include a specular map representing a reflectivity of specular reflection for each of channels (e.g., RGB channels). The specular map may be a value field indicating a degree of reflection observed from camera views corresponding to a 3D point.
The direct light information may include color values of light coming to a corresponding 3D point in all directions.
According to an example, the albedo estimation model 330 and the specular estimation model 340 may be trained based on an objective function value between a GT 2D image and a temporary 2D image reconstructed based on the albedo information output from the albedo estimation model 330, the specular information output from the specular estimation model 340, and other scene component information output from other machine learning models. It may be difficult to substantially calculate GT values for the albedo information and the specular information. Values other than the albedo information and the specular information may be obtained by machine learning models that are completely trained in the first training operation 381 and the second training operation 382. The electronic device may apply the albedo information and the specular information to a rendering operation together with the other values obtained by the completely trained machine learning models. A pixel value for a corresponding 3D point may be calculated by the rendering operation. The electronic device may calculate a difference between the calculated pixel value and a GT pixel value of a corresponding pixel position in an image of the training dataset 301 as an objective function value. Since it is difficult to directly calculate GT values for the albedo information and the specular information, as described above, the electronic device may perform indirect supervised learning using the rendering operation together with other component values.
In the field of computer graphics for physics-based rendering, a typical electronic device may calculate a color value (e.g., an RGB value) at a corresponding 3D point by the rendering operation, where the rendering operation may have a high complexity. The high complexity may lead to a decrease in a probability of stable convergence in training. According to an example, in contrast to the typical electronic device, the electronic device of one or more embodiments may calculate a temporary color value from scene property information estimated using a rendering operation that is approximated and modified from the rendering operation. For example, the temporary 2D image may be reconstructed using an approximation (e.g., a split-sum approximation) that is based on a split of a rendering operation that determines an image pixel value based on scene component information into a reflection component and an illumination component. A non-limiting example of a rendering operation based on a BRDF will be described with reference to Equations 1 to 16 shown below, and a non-limiting example of the approximation will be described with reference to Equation 17 below.
The rendering operation based on the BRDF described above may be, for example, an operation that approximates reflection and refraction properties of a material based on the microfacet theory, and may be represented as shown in Equation 1 below, for example.
L
o(p,ωo)=∫Ωfr(p,ωi,ωo)L(p,ωi)(ωi·n)dωi Equation 1:
In Equation 1, L0 denotes a sum of light energy corresponding to light Li incident from an arbitrary 3D point p in all directions Q and output in an arbitrary reflection direction wo by a material reflectivity (e.g. A diffuse reflection coefficient
and a specular reflection coefficient
of a BRDF). wi denotes a direction of an arbitrary ray departing from the 3D point p and may be one of light vectors corresponding to all the directions Ω. wo may be an outgoing vector and may be the same as a view vector viewed by a camera. When d is a view direction of the camera, “wo=−d” may be satisfied. In addition, n denotes a normal vector at the 3D point p, which is the same as n(p). Here, L0 may be interpreted as a pixel value (e.g., an RGB color value) of a corresponding image pixel viewed at a viewpoint. fr may be expressed as in Equation 2 below, for example.
f
r
=k
d
f
lambert
+k
s
f
cook-torrance Equation 2:
Lambertian diffuse in Equation 2 described above may be represented as a constant factor as shown in Equation 3 below, for example.
Equation 3:
In Equation 3, ρ denotes an albedo value. A factor fcookTorrance representing a Cook-Torrance BRDF in Equation 2 described above may be expressed as in Equation 4 below, for example.
Equation 4 described above may be an equation that approximates a portion of reflection properties of a surface. In Equation 4, D denotes a normal distribution function, G denotes a geometry function, and F denotes a Fresnel equation. The normal distribution function D may be a distribution function of normal vectors and may be a probability value indicating a distribution of normal vectors of a microfacet at an arbitrary point p in the concept of the microfacet. D, G, and F are respectively described with reference to Equations 5 to 10 below, for example.
Equation 5 described above may represent the normal distribution function D and indicate a value obtained by statistically approximating an amount of alignment of microfacets to a half vector h affected by a roughness of a surface. The half vector h is described with reference to Equation 6. The half vector h may be a unit vector pointing in a center direction of a view vector and a light vector, and may be a vector pointing in a center direction of wo (e.g., a vector corresponding to −d) and wi. Also, n denotes a normal vector n(p).
In Equation 7 described above, the geometry function G may represent a value describing self-shadowing properties of a microfacet. Equation 8 may represent a sub-function of the geometry function G. k in Equation 8 is described with reference to Equation 9.
Equation 10 represents the Fresnel equation F, which is a ratio of reflected light to refracted light, and may be approximated using a Fresnel-Schlick approximation. In Equation 11, F0 may represent a level of base reflectivity when looking straight at a surface of a material.
L(p, wi) in Equation 1 described above may be expressed as in Equation 12 below, for example.
L(p,ωi)=Ldirect(p,ωi)·v(p,ωi)+Lindirect(p,ωi) Equation 12:
In Equation 12, a notation of Ldirect(p, wi) denoting direct light may be changed to Li(p, wi), and a notation of Lindirect(p, wi) denoting indirect light may be changed to e(p, wi), which may be expressed as in Equation 13 below, for example.
L(p,wi)=Li(p,wi)·v(p,wi)+e(p,wi) Equation 13:
In Equation 13, L(p, wi) may represent all light coming in all directions to an arbitrary point p, and may be classified as direct light information Li(p, wi), visibility information v(p, wi), and indirect light information e(p, wi). L(p, wi) may be a component obtained by adding the indirect light information e(p, wi) to a product of the direct light information Li(p, wi) and the visibility information v(p, wi).
In addition, fr(p, wi, wo) in Equation 1 described above may be expressed as in Equation 14 below, for example.
Equation 14 may be divided into a diffuse BRDF and a specular BRDF. A notation of kdρ in the diffuse BRDF is expressed as in Equation 15 below using albedo information (ρ(p)), and a notation of DFG in the specular BRDF is expressed as in Equation 15 below, for example, using the specular information f(p, d).
Therefore, an expression of the rendering operation according to Equation 1 may be summarized as shown in Equation 16 below, for example.
Equation 1 described above may be summarized as Equation 16 according to Equations 2 through 15. Equation 16 described above may be expressed as Equation 17 shown below, for example, based on the split-sum approximation.
In Equation 17, ρ(p)/π denotes a diffuse reflection component, and
denotes a specular reflection component. Therefore, in Equation 17, the BRDF (e.g., f (p, d)) and light properties (e.g., Li, v, and e) may be individually bundled as independent integral terms by the split-sum approximation, and a dependency between components may be further reduced. This is because a reflection component and an illumination component are separated.
For reference, Equations 1 to 16 are used to represent a process of deriving Equation 17. The electronic device may use outputs of the scene component estimation network that will be described below as values corresponding to each variable or a combination of variables in Equation 17. For example, Vi(p, wi) in Equation 17 may be a visible distance from an object surface position p in a ray direction wi as visibility information, and may be estimated by a visibility estimation model. Li(p, wi), which is direct light information, may indicate a color value of direct light received from the object surface position p in the ray direction wi and may be estimated by a direct light estimation model. Among diffuse reflection components, ρ(p) may correspond to albedo information estimated by an albedo estimation model that will be described below. Among specular reflection components, f(p,d) may correspond to specular information estimated by a specular estimation model.
According to an example, the electronic device may perform a calculation by converting a continuous integral term of Equation 17 described above into a discrete sigma sum based on Monte-Carlo estimation.
In the third training operation 383, the electronic device may calculate the temporary pixel value 395 based on Equation 17 from scene component information corresponding to temporary outputs 390 of each machine learning model. The electronic device may calculate a loss function value (e.g., a reconstruction loss) corresponding to a difference between the temporary pixel value 395 and the GT pixel value. The electronic device may update parameters (e.g., connection weights between nodes) of the albedo estimation model 330, the specular estimation model 340, and the direct light estimation model 370 so that the loss function value may be reduced. Therefore, a stability of training in the third training operation 383 may be further enhanced.
According to an example, in a first training, an electronic device (e.g., the electronic device 1000 of
The electronic device may determine a point spaced apart from a viewpoint c in a view direction d. As described above, the view direction d may correspond to an image pixel at an arbitrary pixel position within a 2D scene. Coordinates of the point spaced apart from the viewpoint c in the view direction d may be expressed as c+t0d. Each of the viewpoint c and the view direction d may be expressed as a vector according to a 3D coordinate system. to may represent a distance away from the viewpoint c.
The electronic device may generate a global geometry feature z corresponding to the determined point and distance information SDF on a distance between the determined point and an object surface, based on the INR model 410. As described above with reference to
The electronic device may calculate normal information n(p) of the determined point by analyzing information indicating whether the determined point is on a surface, the viewpoint c, and the view direction d. For example, the electronic device may calculate the object surface position p and normal information n(p) from the distance information SDF, the viewpoint c, and the view direction d, through an analysis algorithm 420. Based on the neural renderer 415, the electronic device may determine a pixel value (e.g., a color value) (RGB(p, d)) of a corresponding image pixel from the global geometry feature z, the object surface position p, and the normal information n(p).
In the above operation, the distance t0 may also be determined as a distance between the viewpoint c and the object surface. In other words, the object surface position p and the coordinates of the point spaced apart from the viewpoint c may ideally be the same. However, before the first training is completed, it may be difficult to guarantee the object surface position p that is based on an output of the INR model 410 as an object surface, and difficult to determine the distance t0. Accordingly, when the first training is initiated, the electronic device may set the distance t0 to an arbitrary value. At a point in time in which the first training is initiated, only the viewpoint c and the view direction d may be given. The electronic device may temporarily determine the coordinates c+t0d of the point spaced apart from the viewpoint c using an arbitrarily set distance t0. The electronic device may temporarily estimate the distance information SDF and the global geometry feature z by inputting the temporarily determined coordinates c+t0d to the INR model 410. The electronic device may adjust a temporary distance t0, which is temporarily away from the viewpoint c, until temporary distance information SDF becomes zero. Here, the temporary distance t0 may be adjusted based on ray marching that will be described below with reference to
The electronic device may estimate a temporary object surface position p from the viewpoint c, the view direction d, and the temporary distance information SDF through the analysis algorithm 420. The electronic device may estimate a temporary pixel value through the neural renderer 415 using the temporary object surface position p. The electronic device may estimate temporary pixel values of image pixels included in a 2D scene, as described above. The electronic device may update parameters (e.g., connection weights between nodes) of the INR model 410 and the neural renderer 415 based on a loss between the temporary pixel values and GT pixel values of a training dataset 490. The electronic device may update the parameters of the INR model 410 and the neural renderer 415 based on a backpropagation so that a calculated loss may be reduced.
For reference, when the first training has not been completed yet, it may be difficult to guarantee the temporary distance t0 to be a distance from the viewpoint c to the object surface, even though the above-described temporary distance information SDF is “0”. In other words, the temporary object surface position p may not be a position on the object surface. The electronic device may update the temporary distance t0 as well as the parameters of the INR model 410 and the neural renderer 415 so that the loss may be reduced during the above-described backpropagation process.
Therefore, the INR model 410 may be trained based on an output of the neural renderer 415 that estimates a pixel value of an image pixel from the global geometry feature z, the object surface position p, the normal information n(p), and the view direction d. The completely trained INR model 410 may be used in other training operations (e.g., second and third training operations) and an inference operation along with the analysis algorithm 420 for calculating the normal information n(p).
According to an example, an electronic device (e.g., the electronic device 1000 of
In a second training, the electronic device may determine GT visible distances for training of the visibility estimation model 550. According to an example, the visibility estimation model 550 may be trained using visible distances between the object surface position p and arrival points determined based on ray marching 570 for the ray directions wi departing from the object surface position p. For example, the electronic device may determine the object surface position p based on the completely trained INR model 510 and the analysis algorithm 520. The electronic device may perform the ray marching 570 from the object surface position p in the ray directions wi. As a result of the ray marching 570, the electronic device may determine arrival points p′ for each ray direction. The electronic device may determine a visible distance map including visible distances from the object surface position p to the arrival points p′ as a GT visible distance map.
The electronic device may update parameters (e.g., connection weights between nodes) of the visibility estimation model 550 based on an objective function value (e.g., a loss function value) between a GT visible distance map and temporary visibility information (e.g., a temporary visible distance map) (v(p, wi)) output based on the visibility estimation model 550. The electronic device may repeatedly or iteratively update the parameters of the visibility estimation model 550 based on a backpropagation so that a loss may be reduced.
For reference, each space in a visible distance map as visibility information may correspond to an arbitrary ray direction departing from the object surface position p, and a value assigned to a corresponding space may indicate a visible distance. In
Although a considerable amount of time may be used to perform an operation of a visible distance map based on the ray marching 570, an amount of time used to estimate visibility information may be reduced through the visibility estimation model 550 that is completely trained.
According to an example, ray marching may be an algorithm (e.g., sphere tracing) that iteratively performs an operation of advancing rays, by a minimum distance among distances from an arbitrary point to surrounding object surfaces, from a corresponding point in a target direction 619. For example, referring to
Referring to the example described above with reference to
In addition, the electronic device may perform ray marching in an inference operation, a non-limiting example of which will be described with reference to
According to an example, an electronic device (e.g., the electronic device 1000 of
The electronic device may obtain GT indirect light information based on ray marching and a scene rendering model 719. For example, the indirect light estimation model 760 may be trained based on color information of arrival points p′ at which rays departing from the object surface position p arrive, viewed from the object surface position p. The color information of the arrival points p′ for training of the indirect light estimation model 760 may be estimated using the INR model 710 and a neural renderer which are completely trained. For example, the electronic device may calculate the arrival points p′ based on ray marching from the object surface position p and ray directions based on the completely trained INR model 710 and the analysis algorithm 720. For reference, in the ray marching, both coordinates of the arrival points p′ and distances to the arrival points p′ may be calculated. Accordingly, the electronic device may also separate the object surface position p and a corresponding ray direction wi from the coordinates of the arrival points p′. The electronic device may calculate the GT indirect light information based on the scene rendering model 719 that is completely trained as described above with reference to
An output of the indirect light estimation model 760 that is not completely trained in the second training may also be referred to as temporary indirect light information. The electronic device may update parameters (e.g., connection weights between nodes) of the indirect light estimation model 760 based on an objective function value (e.g., a loss function value) between the temporary indirect light information and the GT indirect light information. The electronic device may repeatedly or iteratively update the parameters of the indirect light estimation model 760 based on a backpropagation so that a loss may be reduced.
In a third training, an electronic device (e.g., the electronic device 1000 of
When the third training is completed, the electronic device may reconstruct a scene image based on the completely trained scene component estimation network. The completely trained scene component estimation network may be interpreted as a network that may represent an arbitrary 3D space.
For example, the electronic device may determine a viewpoint c and a view direction d determined by a user, based on a user input. The electronic device may estimate a global geometry feature z, an object surface position p, and normal information n(p) based on an INR model 910 and an analysis algorithm 920, from the viewpoint c and the view direction d. For reference, as described above, when a distance from the viewpoint c to an object surface is not given in scene estimation, the electronic device may determine a distance t0 to the object surface by repeatedly or iteratively performing ray marching. For example, the electronic device may obtain an SDF value by inputting a temporary point on the view direction d determined using an arbitrary initial distance into the INR model 910. When the SDF value is not zero, the electronic device may update a point on the view direction d by performing ray marching as much as the obtained SDF value. The electronic device may determine the object surface position p by estimating the SDF value based on the INR model 910 and repeatedly or iteratively performing ray marching based on the estimated SDF value.
The electronic device may estimate the albedo information p(p) based on the albedo estimation model 930 from the global geometry feature z, the object surface position p, and the normal information n(p). The electronic device may estimate the visibility information v(p, wi), the indirect light information e(p, wi), the specular information f(p, d), and the direct light information L(p, wi) based on the visibility estimation model 950, the indirect light estimation model 960, the specular estimation model 940, and the direct light estimation model 970 from ray directions wi, in addition to the global geometry feature z, the object surface position p, and the normal information n(p). The electronic device may reconstruct an image pixel value RGB(p, d) from the above estimated scene component information. For example, the electronic device may calculate the image pixel value RGB(p, d) based on Equation 17 from the estimated scene component information. The electronic device may generate a 2D scene image by calculating all image pixel values RGB(p, d) corresponding to a 2D scene.
In addition, as described above, the electronic device may adjust a portion of the scene component information and reconstruct the image pixel value RGB(p, d) based on the adjusted scene component information. Thus, the electronic device may also reconstruct an image to a scene of an environment different from the training dataset.
Referring to
The input interface 1010, the processor 1020, the memory 1030, the display 1040, and the communication circuit 1050 may communicate with each another via a communication bus.
The input interface 1010 may receive an input from a user through a haptic, video, audio, or touch input. The input interface 1010 may include, for example, a keyboard, a mouse, a touch screen, a microphone, or any other device that detects an input from the user and transmits the detected input.
The processor 1020 may execute functions and instructions to be executed in the electronic device 1000 (e.g., a computing device). For example, the processor 1020 may process instructions stored in the memory 1030. The processor 1020 may extract a global geometry feature and information indicating whether a point is on a surface, based on an INR model, from a viewpoint and a view direction corresponding to an image pixel corresponding to a 2D scene at the viewpoint within a FOV. The processor 1020 may calculate an object surface position corresponding to the viewpoint and the view direction and normal information of the object surface position based on the information indicating whether the point is on the surface. The processor 1020 may estimate albedo information independent of the view direction based on an albedo estimation model from the global geometry feature, the object surface position, and the normal information. The processor 1020 may estimate specular information dependent on the view direction based on a specular estimation model from the global geometry feature, the object surface position, the normal information, and the view direction. The processor 1020 may perform the at least one operation described above with reference to
The processor 1020 may perform training using a loss function that is based on a split-sum approximation-based rendering operation in which a material integral term and a light integral term are separated to stabilize complex inverse rendering training in a training process. In addition, the processor 1020 of one or more embodiments may utilize global geometry features learned in an initial stage as main information to extract BRDF and light properties in a next training stage based on sequential learning rather than a typical processor which trains uses integrated learning.
The memory 1030 may store information or data used for execution of the processor 1020. For example, the memory 1030 may include the above-described machine learning models. The memory 1030 may store an INR model, an analysis algorithm, a visibility estimation model, an indirect light estimation model, an albedo estimation model, a specular estimation model, and a direct light estimation model. In addition, the memory 1030 may store a neural renderer and a training dataset for training. As described above, the albedo estimation model and the specular estimation model may be separated according to a view dependency. The memory 1030 may include a non-transitory computer-readable storage medium or a non-transitory computer-readable storage device. The memory 1030 may store instructions to be executed by the processor 1020 and may store related information while software and/or an application is executed by the electronic device 1000. For example, the memory 1030 may store instructions that, when executed by the processor 1020, configure the processor 1020 to perform any one, any combination, or all of the operations and methods described herein with reference to
The display 1040 may visually provide an output of the electronic device 1000 to a user. For example, the display 1040 may display a 2D scene image reconstructed by the processor 1020.
The communication circuit 1050 may communicate with an external device through a wired or wireless network. The communication circuit 1050 may receive the above-described machine learning models and parameters (e.g., completely trained parameters) of each of the machine learning models from an external device (e.g., a server).
The electronic devices, input interfaces, processors, memories, displays, communication circuits, electronic device 1000, input interface 1010, processor 1020, memory 1030, display 1040, communication circuit 1050, and other apparatuses, devices, and components described and disclosed herein with respect to
The methods illustrated in
Instructions or software to control computing hardware, for example, one or more processors or computers, to implement the hardware components and perform the methods as described above may be written as computer programs, code segments, instructions or any combination thereof, for individually or collectively instructing or configuring the one or more processors or computers to operate as a machine or special-purpose computer to perform the operations that are performed by the hardware components and the methods as described above. In one example, the instructions or software include machine code that is directly executed by the one or more processors or computers, such as machine code produced by a compiler. In another example, the instructions or software includes higher-level code that is executed by the one or more processors or computer using an interpreter. The instructions or software may be written using any programming language based on the block diagrams and the flow charts illustrated in the drawings and the corresponding descriptions herein, which disclose algorithms for performing the operations that are performed by the hardware components and the methods as described.
The instructions or software to control computing hardware, for example, one or more processors or computers, to implement the hardware components and perform the methods as described above, and any associated data, data files, and data structures, may be recorded, stored, or fixed in or on one or more non-transitory computer-readable storage media. Examples of a non-transitory computer-readable storage medium include read-only memory (ROM), random-access programmable read only memory (PROM), electrically erasable programmable read-only memory (EEPROM), random-access memory (RAM), dynamic random access memory (DRAM), static random access memory (SRAM), flash memory, non-volatile memory, CD-ROMs, CD-Rs, CD+Rs, CD-RWs, CD+RWs, DVD-ROMs, DVD-Rs, DVD+Rs, DVD-RWs, DVD+RWs, DVD-RAMs, BD-ROMs, BD-Rs, BD-R LTHs, BD-REs, blue-ray or optical disk storage, hard disk drive (HDD), solid state drive (SSD), flash memory, a card type memory such as multimedia card micro or a card (for example, secure digital (SD) or extreme digital (XD)), magnetic tapes, floppy disks, magneto-optical data storage devices, optical data storage devices, hard disks, solid-state disks, and any other device that is configured to store the instructions or software and any associated data, data files, and data structures in a non-transitory manner and provide the instructions or software and any associated data, data files, and data structures to one or more processors or computers so that the one or more processors or computers can execute the instructions. In one example, the instructions or software and any associated data, data files, and data structures are distributed over network-coupled computer systems so that the instructions and software and any associated data, data files, and data structures are stored, accessed, and executed in a distributed fashion by the one or more processors or computers.
While this disclosure includes specific examples, it will be apparent after an understanding of the disclosure of this application that various changes in form and details may be made in these examples without departing from the spirit and scope of the claims and their equivalents. The examples described herein are to be considered in a descriptive sense only, and not for purposes of limitation. Descriptions of features or aspects in each example are to be considered as being applicable to similar features or aspects in other examples. Suitable results may be achieved if the described techniques are performed in a different order, and/or if components in a described system, architecture, device, or circuit are combined in a different manner, and/or replaced or supplemented by other components or their equivalents.
Therefore, in addition to the above and all drawing disclosures, the scope of the disclosure is also inclusive of the claims and their equivalents, i.e., all variations within the scope of the claims and their equivalents are to be construed as being included in the disclosure.
| Number | Date | Country | Kind |
|---|---|---|---|
| 10-2022-0159547 | Nov 2022 | KR | national |