This application is a U.S. National Phase Application under 35 U.S.C. 371 of International Application No. PCT/KR2021/000217, filed on Jan. 7, 2021, which claims the benefit of Korean Patent Application No. 10-2020-0004128, filed on Jan. 13, 2020, and Korean Patent Application No. 10-2020-0182812, filed on Dec. 24, 2020. The entire disclosures of the above applications are herein incorporated by reference in their entireties. The present patent application claims priority to other applications to be filed in other countries, the disclosures of which are also incorporated by reference herein in their entireties.
The present invention provides a dental medical record device for creating a dental medical record and a dental medical record method thereof.
During dental treatment, to create a dental medical record, a doctor looks at the panoramic photo captured around a patient's teeth and performs an overall basic charting. Further, the doctor adds the data obtained during the oral examination by the patient to the dental medical record. After evaluating the patient's teeth, the doctor marks the dental medical record with text or symbols.
In this case, the doctor must evaluate each of the multiple teeth. Due to possible infection issues between the time of examination and the time of recording, instead of looking at the patient and filling out the dental medical record, the doctor dictates to an assistant doctor to allow the assistant doctor to record, or an assistant dictates and the doctor transcribes the dental medical record. Further, in order to review the treatment history for each of multiple teeth, the doctor should separately search for the medical record and receipt history for each tooth.
Currently, it thus takes a long time to create a dental medical record during dental treatment, and errors may occur while the assistant doctor transcribes when writing a dental medical record. Further, there is a problem in that it is difficult to dictate and record immediately when charting in a hospital or private clinic without a specialist.
Embodiments provide a dental medical record device and a dental medical record method capable of reducing the time and costs required for creating a dental medical record.
Embodiments also provide a dental medical record device and a dental medical record method capable of enhancing the accuracy of prognosis and reducing the possibility of cross-infection of patients.
Embodiments also provide a dental medical record device and a dental medical record method capable of easily creating, storing, and viewing dental medical records.
However, the objects of the embodiments are not limited thereto, and other objects may also be present.
The disclosure provides a dental medical record device and a dental medical record method that receive an image, such as a panoramic photo, a scan image, and a camera image, for a patient's oral cavity through artificial intelligence, perform charting using artificial intelligence, and if the treatment site is clicked in the corresponding image, allows the medical records for the treatment site in association with the chart.
In an aspect, a dental medical record method according to an embodiment comprises reading an image for a patient's oral cavity to mark the patient's oral cavity condition on a prognosis chart, using a text format, symbols, or both text and symbols and displaying the image and the prognosis chart.
In another aspect, a dental medical record device according to another embodiment comprises an input unit receiving an image for a patient's oral cavity, a controller reading the image for the patient's oral cavity to determine the patient's oral cavity condition and marking the patient's oral cavity condition on a prognosis chart, using a text format, symbols, or both text and symbols and an output unit displaying the image and the prognosis chart.
In another aspect, a non-transitory computer-readable storage medium according to another embodiment stores instructions, when executed by one or more processors, instructions that, when executed by one or more processors, enable the one or more processors to perform a dental medical record method.
When executed by the one or more processors, the instructions enable the computer device to read an image for a patient's oral cavity to mark the patient's oral cavity condition on a prognosis chart, using a text format, symbols, or both text and symbols and display the image and the prognosis chart.
By the dental medical record device and the dental medical record method according to embodiments, it is possible to reduce the time and costs required for creating a dental medical record.
By the dental medical record device and the dental medical record method according to embodiments, it is also possible to enhance the accuracy of prognosis and reduce the possibility of cross-infection of patients.
By the dental medical record device and the dental medical record method according to embodiments, a dental medical record device and a dental medical record method capable of easily creating, storing, and viewing dental medical records are provided.
Hereinafter, exemplary embodiments of the inventive concept will be described in detail with reference to the accompanying drawings. The inventive concept, however, may be modified in various different ways, and should not be construed as limited to the embodiments set forth herein. Like reference denotations may be used to refer to the same or similar elements throughout the specification and the drawings. However, the present invention may be implemented in other various forms and is not limited to the embodiments set forth herein. For clarity of the disclosure, irrelevant parts are removed from the drawings, and similar reference denotations are used to refer to similar elements throughout the specification.
In embodiments of the present invention, when an element is “connected” with another element, the element may be “directly connected” with the other element, or the element may be “electrically connected” with the other element via an intervening element. When an element “comprises” or “includes” another element, the element may further include, but rather than excluding, the other element, and the terms “comprise” and “include” should be appreciated as not excluding the possibility of presence or adding one or more features, numbers, steps, operations, elements, parts, or combinations thereof.
When the measurement of an element is modified by the term “about” or “substantially,” if a production or material tolerance is provided for the element, the term “about” or “substantially” is used to indicate that the element has the same or a close value to the measurement and is used for a better understanding of the present invention or for preventing any unscrupulous infringement of the disclosure where the exact or absolute numbers are mentioned. As used herein, “step of” A or “step A-ing” does not necessarily mean that the step is one for A.
As used herein, the term “part” may mean a unit or device implemented in hardware, software, or a combination thereof. One unit may be implemented with two or more hardware devices or components, or two or more units may be implemented in a single hardware device or component.
As used herein, some of the operations or functions described to be performed by a terminal or device may be, instead of the terminal or device, performed by a server connected with the terminal or device. Likewise, some of the operations or functions described to be performed by a server may be performed by a terminal or device connected with the server, instead of the server.
As used herein, some of the operations or functions described to be mapped or matched with a terminal may be interpreted as mapping or matching the unique number of the terminal, which is identification information about the terminal, or personal identification information.
Hereinafter, embodiments of the disclosure are described in detail with reference to the accompanying drawings.
Referring to
The image may be one of a panoramic photo, a scan image, or a camera image obtained by one or two or more of radiography, oral camera, scanning device, and occlusion check data for the patient's oral cavity, as is described below with reference to
Further, the patient's oral cavity condition may be at least one of the dental formula, dental conditions, or surrounding structures of the teeth.
In the step S110 of marking the prognosis chart, artificial intelligence is continuously trained with continuously collected training images and accumulated chart data so that artificial intelligence reads the images, as described below with reference to
Artificial intelligence may automatically learn the features for input values by being trained with a large amount of data from a deep neural network composed of a multi-layered network, and train the multi-layered network to minimize errors in prediction accuracy and read images.
As described below with reference to
Referring to
According to an embodiment, the dental medical record method 100 may read images, such as panoramic photos, scan images or camera images for the oral cavity, and move the results to an electronic medical record (EMR) chart, thereby obtaining a prognosis effect.
Referring to
Further, the dental medical record device may display the content read through the above-described images, in the format of text, symbols, or in a combination of text and symbols.
Referring to
For example, the text chart 140 may represent information indicating that tooth #16 has secondary caries, tooth #44 has a periapical lesion, teeth #41, 42, and 43 have attrition, and teeth #47, 46, and 37 have Old AM.
The symbol chart 142 is a chart in which the content read through the images is depicted through various symbols. For example, the symbol chart may depict tooth loss, periodontal disease, dental caries, restoration, and the like, based on standard symbol standards.
The dental medical record method 100 according to an embodiment may read the images through artificial intelligence using a machine learning model, a deep learning model, etc., as is described below with reference to
In this case, the artificial intelligence may set a dividing criterion while inputting video data for normal structures and lesions in the oral cavity during the learning process.
As an example, when the alveolar bone is positioned up to 1 mm below the area where the crown and the root meet, the alveolar bone may be recognized as normal. As another example, if there is no tooth in the alveolar bone, it may be recognized as tooth loss.
Meanwhile, the diagnosis of lesions using artificial intelligence may be performed based on factors, such as tooth color, tooth loss, and radiography information in the images (e.g., houndsfield scale).
The dental medical record method 100 according to an embodiment may continuously train the artificial intelligence with continuously collected training images and accumulated chart data, thereby increasing the accuracy for the artificial intelligence to read the images.
Further, the dental medical record method 100 according to an embodiment may, after reading the images through the artificial intelligence, chart the same and associate the images with the patient's treatment and receipt content, in the step S110 of marking the prognosis chart. For example, if the doctor clicks on the image, a pop-up 144 indicating information about the treatment date for the clicked area and the receipt memo window may be generated. Further, if the doctor clicks the receipt memo window on the pop-up 144, information about the chart associated with the corresponding memo window may be accessed.
Referring to
The roots of the teeth are positioned inside the bone (#11 to #48), the crown is positioned outside the bone. The teeth are nearly perpendicular to the bone.
The dental medical record method 100 according to an embodiment may determine that it is normal if the alveolar bone is positioned up to 1 mm in the root direction from the site (arrow) where the crown and the root of the tooth meet and that it is an alveolar bone loss if the alveolar bone is positioned therebelow. Alveolar bone loss includes horizontal loss and vertical loss. Further, the dental medical record method 100 according to an embodiment may identify the state of the furcation, which is the site where the root of the molar is split.
The dental medical record method 100 according to an embodiment determines the patient's periodontal condition of alveolar bone loss after reading the image shown in
As shown in
The dental medical record method 100 according to an embodiment determines the patient's periodontal condition of dental caries after reading the image shown in
As shown in
The dental medical record method 100 according to an embodiment reads the image shown in
The panoramic image capture device positions a capture subject, e.g., a patient, on a cylindrical body equipped with an X-ray source, and then rotates the X-ray source and panoramic image detector around the patient, thereby obtaining a plurality of images. The panoramic image capture device naturally connects these images to obtain a panoramic image as shown in
The artificial intelligence may be trained with the training images including guide chips.
Since the artificial intelligence is trained with the guide sample-containing training images, the patient's periodontal condition, such as dental caries, is more accurately determined after reading the images through the artificial intelligence in the step S110 of marking the prognosis chart.
In the step S110 of marking the prognosis chart, when the image is read through the trained artificial intelligence, the difference in dental caries and the case in which the transmission of radiations has been reduced, such as enamel, dentin, or pulp, may be more accurately read as compared with the normal tooth tissue in the panoramic image. Further, even when the radiodensity differs in the panoramic image depending on, e.g., the anatomical structure of each patient, the size of the patient, and the posture, it is possible to more accurately read the periodontal condition of dental caries.
In the above-described example, in the step S110 of marking the prognosis chart, the dental caries is determined after reading the image including the guide chip shown in
When capturing with a camera or scanner, an image including a guide sample is obtained as shown in
The artificial intelligence may be trained with the training images including guide samples. In the image, the tooth color is affected by lighting and shadow, the capture camera. Accordingly, since the artificial intelligence is trained with the guide sample-containing training images, the patient's tooth color may be more accurately determined after reading the images through the artificial intelligence in the step S110 of marking the prognosis chart.
Therefore, in the step S120 of displaying the image and the prognosis chart, if a specific tooth is clicked on the image, the color of the specific tooth, e.g., an optical tooth color, presented by the trained artificial intelligence is indicated with the color number (e.g., N3) and/or color (e.g., yellow) of the guide sample, on the image. Thus, it is possible to suggest the color of the restoration or prosthesis that best matches the color of the patient's teeth, providing an effective, aesthetic treatment method.
Referring to
Referring to
Further, if the receiving memo window is clicked, a chart linked to the receiving memo window may be output.
Referring to
The dental medical record method 100 according to an embodiment may enhance the accuracy of creating the prognosis chart by continuously learning individually input information or accumulated chart records through artificial intelligence using a machine learning model or a deep learning model in the step S110 of marking the prognosis chart.
In the step S120 of displaying the image and the chart, the dental medical record method 100 according to an embodiment may generate a treatment popup 144 including information, such as treatment title, date, and receipt memo window, if the doctor clicks the treatment site on the image (e.g., panoramic photo) for the patient's oral cavity and output a chart linked to the receipt memo window if the doctor clicks the receipt memo window.
By the dental medical record method 100 according to an embodiment, even a hospital or private clinic without a specialist may enhance convenience and accuracy when recording dental medical records. Further, it allows the specialist to focus on studying specialized care rather than dictation. Further, it is possible to increase the accuracy of determining the patient's oral condition based on the accumulated image information and charting data.
Referring to
The above-described deep learning model 221 may be a convolutional neural network (CNN), deep hierarchical network (DHN), convolutional deep belief network (CDBN), deconvolutional deep network (DDN), recurrent neural network (RNN), or generative adversarial network (GAN), but the present invention is not limited thereto, and may use various deep learning models that may be used currently or in the future.
The above-described deep learning model 221 may be implemented through a deep learning framework. The deep learning framework plays a role to provide a library of functions commonly used when developing the deep learning model 221 and to support the system software or hardware platform to be properly used. In this embodiment, the deep learning model 221 may be implemented using any deep learning framework that has been currently disclosed or will be disclosed in the future.
Referring back to
Convolution creates a feature map using a plurality of filters for each area of the medical image in the convolution layer. Subsampling or pooling reduces the size of the feature map in the subsampling layer to extract features for the image that is invariant to a change in position or rotation.
The feature extraction part 222 may repeat convolution and/or subsampling to extract various levels of features from low-level features, such as dots, lines, or planes, to complex, meaningful high-level features, from the medical image.
The deep learning model, e.g., a CNN-based deep learning model, targets optimally training the parameters present in each individual layer in the feature extraction part 222 and the output part 224. In the deep learning model, the order of data determines the initial parameter value.
The deep learning model 221 may apply random sampling (random in data order) and a regulation technique. Random sampling means that the order of the training data learned from the training data set is different.
The regulation technique is a technique that reduces overfitting, in which a deep learning model over-trained with training data including even noise deteriorates test or diagnosis accuracy. The regulation technique may be, e.g., a drop-out technique or a drop-connected technique.
The drop-out technique is a method for performing learning with the probable parameter value designated as 0 for a specific node. The drop-connected technique is a method for performing learning with the connections between nodes dropped. Although the drop-out technique, as the regulation technique, is described below as an example, it may be any current or future technique or algorithm for reducing overfitting.
The deep learning model 221 uses highly flexible nonlinear algorithms. Accordingly, the result values of the deep learning model 221 may exhibit a large deviation. The deviation between the result values of the deep learning models, i.e., output results, may be reduced by ensembling the output results of the deep learning models 221 based on one or more of a majority vote-based ensemble, an unanimity-based ensemble, and an uncertainty-based ensemble.
In other words, in each of the deep learning modules 221, the internal parameters are trained differently depending on the training scheme, e.g., the sampling order and the randomness of the dropout. Even when trained with the same data and the same deep learning models, each deep learning model may exhibit different results. Therefore, use of one deep learning model may lead to a risk of misjudgment. Therefore, the present embodiment may generate various deep learning models and minimize the risk of misjudgment through an ensemble technique.
Referring to
The image may be one of a panoramic photo, a scan image, or a camera image obtained by one or two or more of radiography, oral camera, scanning device, and occlusion check data for the patient's oral cavity.
The patient's oral cavity condition may be at least one of the dental formula, dental conditions, or surrounding structures of the teeth.
The controller 220 may continuously train artificial intelligence with continuously collected training images and accumulated chart data, allowing the artificial intelligence to read the image.
Artificial intelligence may automatically learn the features for input values by being trained with a large amount of data from a deep neural network composed of a multi-layered network, and train the multi-layered network to minimize errors in prediction accuracy and read images.
If a specific treatment site of the image is selected, the display unit 230 displays information about the corresponding treatment site, and if the image of the corresponding treatment site is selected, displays a treatment summary memo window, and if the treatment summary memo window is selected, outputs the chart for the corresponding treatment site.
For example, the controller 220 reads the image shown in
After reading the image shown in
After reading the image shown in
As the information about the treatment site, the display unit 230 displays “Jan. 2, 2020 #44, 45 extractions, 44 immediate implants, receipt: 200,000 won” as shown in
As shown in
The controller 220 may enhance the accuracy of creating the prognosis chart by continuously learning individually input information or accumulated chart records through artificial intelligence using a machine learning model or a deep learning model.
As described above with reference to
As described above with reference to
The above-described dental medical record method 100 may be implemented by a computing device including at least some of a processor, a memory, a user input device, and a presentation device. The memory is a medium that stores computer-readable software, applications, program modules, routines, instructions, and/or data, coded to perform specific tasks when executed by a processor. The processor may read and execute the computer-readable software, applications, program modules, routines, instructions, and/or data stored in the memory. The user input device may be a means for allowing the user to input a command to the processor to execute a specific task or to input data required for the execution of the specific task. The user input device may include a physical or virtual keyboard or keypad, key button, mouse, joystick, trackball, touch-sensitive input means, or a microphone. The presentation device may include, e.g., a display, a printer, a speaker, or a vibrator.
The computing device may include various devices, such as smartphones, tablets, laptops, desktops, servers, clients, and the like. The computing device may be a single stand-alone device and may include a plurality of computing devices operating in a distributed environment composed of a plurality of computing devices cooperating with each other through a communication network.
Further, the above-described dental medical record method 100 may be executed by a computing device that includes a processor and a memory storing computer readable software, applications, program modules, routines, instructions, and/or data structures, coded to perform an image diagnosis method utilizing a deep learning model when executed by the processor.
The present embodiments described above may be implemented through various means. For example, the present embodiments may be implemented by various means, e.g., hardware, firmware, software, or a combination thereof.
When implemented in hardware, the dental medical record method 100 using a deep learning model according to the present embodiments may be implemented by, e.g., one or more application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field programmable gate arrays (FPGAs), processors, controllers, micro-controllers, or micro-processors.
For example, the dental medical record method 100 according to embodiments may be implemented by an artificial intelligence semiconductor device in which neurons and synapses of the deep neural network are implemented with semiconductor devices. In this case, the semiconductor devices may be currently available semiconductor devices, e.g., SRAM, DRAM, or NAND or may be next-generation semiconductor devices, such as RRAM, STT MRAM, or PRAM, or may be combinations thereof.
When the dental medical record method 100 according to embodiments is implemented using an artificial intelligence semiconductor device, the results (weights) of training the deep learning model with software may be transferred to synaptic mimic devices disposed in an array, or learning may be performed in the artificial intelligence semiconductor device.
When implemented in firmware or hardware, the dental medical record method 100 according to the present embodiments may be implemented in the form of a device, procedure, or function performing the above-described functions or operations. The software code may be stored in a memory unit and driven by a processor. The memory unit may be positioned inside or outside the processor to exchange data with the processor by various known means.
The above-described terms, such as “system,” “processor,” “controller,” “component,” “module,” “interface,” “model,” or “unit,” described above may generally refer to computer-related entity hardware, a combination of hardware and software, software, or software being executed. For example, the above-described components may be, but are not limited to, processes driven by a processor, processors, controllers, control processors, entities, execution threads, programs, and/or computers. For example, both an application being executed by a controller or a processor and the controller or the processor may be the components. One or more components may reside within a process and/or thread of execution, and the components may be positioned in one device (e.g., a system, a computing device, etc.) or distributed in two or more devices.
Meanwhile, another embodiment provides a computer program stored in a computer recording medium for performing the above-described dental medical record method 100. Further, another embodiment provides a computer-readable recording medium storing a program for realizing the above-described dental medical record method.
The program recorded on the recording medium may be read, installed, and executed by a computer to execute the above-described steps.
As such, for the computer to read the program recorded on the recording medium and execute the implemented functions with the program, the above-described program may include code coded in a computer language, such as C, C++, JAVA, or machine language, which the processor (CPU) of the computer may read through a computer device interface.
Such code may include a function code related to a function defining the above-described functions or may include an execution procedure-related control code necessary for the processor of the computer to execute the above-described functions according to a predetermined procedure.
Further, the code may further include additional information necessary for the processor of the computer to execute the above-described functions or memory reference-related code as to the position (or address) in the internal or external memory of the computer the media should reference.
Further, when the processor of the computer needs to communicate with, e.g., another computer or a server at a remote site to execute the above-described functions, the code may further include communication-related code as to how the processor of the computer should communicate with the remote computer or server using the communication module of the computer and what information or media should be transmitted/received upon communication.
The above-described computer-readable recording medium may include, e.g., ROMs, RAMs, CD-ROMs, magnetic tapes, floppy disks, or optical data storage devices, or may also include carrier wave-type implementations (e.g., transmissions through the Internet).
Further, the computer-readable recording medium may be distributed to computer systems connected via a network, and computer-readable codes may be stored and executed in a distributed manner.
The functional programs for implementing the present invention and code and code segments related thereto may easily be inferred or changed by programmers of the technical field to which the present invention pertains, considering, e.g., the system environments of the computer reading and executing the program.
The dental medical record method 100 may be implemented in the form of recording media including computer-executable instructions, such as application or program modules. The computer-readable medium may be an available medium that is accessible by a computer. The computer-readable storage medium may include a volatile medium, a non-volatile medium, a separable medium, and/or an inseparable medium. The computer-readable medium may include a computer storage medium. The computer storage medium may include a volatile medium, a non-volatile medium, a separable medium, and/or an inseparable medium that is implemented in any method or scheme to store computer-readable commands, data architecture, program modules, or other data or information.
In a non-transitory computer-readable storage medium storing instructions that, when executed by one or more processors, enable the one or more processors to perform a dental medical record method, the instructions are executed by the one or more processors to enable the computer device to read an image for a patient's oral cavity to mark the patient's oral cavity condition on a prognosis chart, using a text format, symbols, or both text and symbols and display the image and the prognosis chart.
The above-described dental medical record method 100 may be executed by an application installed on a terminal, including a platform equipped in the terminal or a program included in the operating system of the terminal), or may be executed by an application (or program) installed by the user on a master terminal via an application providing server, such as a web server, associated with the service or method, an application, or an application store server. In such a sense, the above-described glaucoma surgery result diagnosis method may be implemented in an application or program installed as default on the terminal or installed directly by the user and may be recorded in a recording medium or storage medium readable by a terminal or computer. Although embodiments of the present invention have been described with reference to the accompanying drawings, It will be appreciated by one of ordinary skill in the art that the present disclosure may be implemented in other various specific forms without changing the essence or technical spirit of the present disclosure. Thus, it should be noted that the above-described embodiments are provided as examples and should not be interpreted as limiting. Each of the components may be separated into two or more units or modules to perform its function(s) or operation(s), and two or more of the components may be integrated into a single unit or module to perform their functions or operations.
It should be noted that the scope of the present invention is defined by the appended claims rather than the described description of the embodiments and include all modifications or changes made to the claims or equivalents of the claims.
The instant patent application claims priority under 35 U.S.C. 119(a) to Korean Patent Application Nos. 10-2020-0004128 and 10-2020-0182812, filed on Jan. 13, 2020 and Dec. 24, 2020, respectively, in the Korean Intellectual Property Office, the disclosures of which are herein incorporated by reference in their entireties. The present patent application claims priority to other applications to be filed in other countries, the disclosures of which are also incorporated by reference herein in their entireties.
Number | Date | Country | Kind |
---|---|---|---|
10-2020-0004128 | Jan 2020 | KR | national |
10-2020-0182812 | Dec 2020 | KR | national |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/KR2021/000217 | 1/7/2021 | WO |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2021/145607 | 7/22/2021 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
6099314 | Kopelman | Aug 2000 | A |
6334772 | Taub | Jan 2002 | B1 |
6334853 | Kopelman | Jan 2002 | B1 |
6463344 | Pavloskaia | Oct 2002 | B1 |
6542249 | Kofman | Apr 2003 | B1 |
6633789 | Nikolskiy | Oct 2003 | B1 |
6664986 | Kopelman | Dec 2003 | B1 |
6697164 | Babayoff | Feb 2004 | B1 |
6845175 | Kopelman | Jan 2005 | B2 |
6979196 | Nikolskiy | Dec 2005 | B2 |
7030383 | Babayoff | Apr 2006 | B2 |
7202466 | Babayoff | Apr 2007 | B2 |
7255558 | Babayoff | Aug 2007 | B2 |
7286954 | Kopelman | Oct 2007 | B2 |
7319529 | Babayoff | Jan 2008 | B2 |
7373286 | Nikolskiy | May 2008 | B2 |
7507088 | Taub | Mar 2009 | B2 |
7545372 | Kopelman | Jun 2009 | B2 |
7698068 | Babayoff | Apr 2010 | B2 |
7916911 | Kaza | Mar 2011 | B2 |
8108189 | Chelnokov | Jan 2012 | B2 |
8165213 | Ha | Apr 2012 | B2 |
8244028 | Kuo | Aug 2012 | B2 |
8582878 | Tzur | Nov 2013 | B1 |
8587582 | Matov | Nov 2013 | B2 |
8948482 | Levin | Feb 2015 | B2 |
D742518 | Barak | Nov 2015 | S |
9192305 | Levin | Nov 2015 | B2 |
9261356 | Lampert | Feb 2016 | B2 |
9261358 | Atiya | Feb 2016 | B2 |
9299192 | Kopelman | Mar 2016 | B2 |
D760901 | Barak | Jul 2016 | S |
9393087 | Moalem | Jul 2016 | B2 |
9408679 | Kopelman | Aug 2016 | B2 |
9431887 | Boltanski | Aug 2016 | B2 |
9439568 | Atiya | Sep 2016 | B2 |
9451873 | Kopelman | Sep 2016 | B1 |
D768861 | Barak | Oct 2016 | S |
D771817 | Barak | Nov 2016 | S |
9482939 | Moon | Nov 2016 | B2 |
9491863 | Boltanski | Nov 2016 | B2 |
D774193 | Makmel | Dec 2016 | S |
9510757 | Kopelman | Dec 2016 | B2 |
9660418 | Atiya | May 2017 | B2 |
9668829 | Kopelman | Jun 2017 | B2 |
9675430 | Verker | Jun 2017 | B2 |
9693839 | Atiya | Jul 2017 | B2 |
9717402 | Lampert | Aug 2017 | B2 |
9724177 | Levin | Aug 2017 | B2 |
9844426 | Atiya | Dec 2017 | B2 |
10076389 | Wu | Sep 2018 | B2 |
10098714 | Kuo | Oct 2018 | B2 |
10108269 | Sabina | Oct 2018 | B2 |
10111581 | Makmel | Oct 2018 | B2 |
10111714 | Kopelman | Oct 2018 | B2 |
10123706 | Elbaz | Nov 2018 | B2 |
10136972 | Sabina | Nov 2018 | B2 |
10251732 | Lee | Apr 2019 | B2 |
10380212 | Elbaz | Aug 2019 | B2 |
10390913 | Sabina | Aug 2019 | B2 |
10453269 | Furst | Oct 2019 | B2 |
10456043 | Atiya | Oct 2019 | B2 |
10499793 | Ozerov | Dec 2019 | B2 |
10504386 | Levin | Dec 2019 | B2 |
10507087 | Elbaz | Dec 2019 | B2 |
10517482 | Sato | Dec 2019 | B2 |
10695150 | Kopelman | Jun 2020 | B2 |
10708574 | Furst | Jul 2020 | B2 |
10726301 | Joyce | Jul 2020 | B2 |
10772506 | Atiya | Sep 2020 | B2 |
10806766 | Lee | Oct 2020 | B2 |
10813727 | Sabina | Oct 2020 | B2 |
10888399 | Kopelman | Jan 2021 | B2 |
10952816 | Kopelman | Mar 2021 | B2 |
10980613 | Shanjani | Apr 2021 | B2 |
11013581 | Sabina | May 2021 | B2 |
11096765 | Atiya | Aug 2021 | B2 |
11151421 | Joyce | Oct 2021 | B2 |
11154311 | Lee | Oct 2021 | B2 |
11494899 | Joyce | Nov 2022 | B2 |
11565513 | Lee | Jan 2023 | B2 |
11752650 | Joyce | Sep 2023 | B2 |
11759295 | Sabina | Sep 2023 | B2 |
11819223 | Lee | Nov 2023 | B2 |
11865007 | Lee | Jan 2024 | B2 |
20030065523 | Pruche | Apr 2003 | A1 |
20060239526 | Jonusauskas | Oct 2006 | A1 |
20100141931 | Ramirez Mancilla | Jun 2010 | A1 |
20110069149 | Park | Mar 2011 | A1 |
20120015316 | Sachdeva | Jan 2012 | A1 |
20140051040 | Kilcher | Feb 2014 | A1 |
20140267593 | Kim | Sep 2014 | A1 |
20140314291 | Souza | Oct 2014 | A1 |
20150057622 | Hyde | Feb 2015 | A1 |
20160331465 | Kim | Nov 2016 | A1 |
20170007374 | Lee | Jan 2017 | A1 |
20170035601 | Lee | Feb 2017 | A1 |
20170151039 | Lee | Jun 2017 | A1 |
20170157858 | Lee | Jun 2017 | A1 |
20190029784 | Moalem | Jan 2019 | A1 |
20190105842 | Dau | Apr 2019 | A1 |
20190175314 | Lagardere | Jun 2019 | A1 |
20190254787 | Lee | Aug 2019 | A1 |
20190313963 | Hillen | Oct 2019 | A1 |
20190388103 | Lee | Dec 2019 | A1 |
20190388193 | Saphier | Dec 2019 | A1 |
20200281700 | Kopelman | Sep 2020 | A1 |
20200281702 | Kopelman | Sep 2020 | A1 |
20200306142 | Mo | Oct 2020 | A1 |
20200315434 | Kopelman | Oct 2020 | A1 |
20200349698 | Minchenkov | Nov 2020 | A1 |
20200349705 | Minchenkov | Nov 2020 | A1 |
20200372677 | Yoon | Nov 2020 | A1 |
20200404243 | Saphier | Dec 2020 | A1 |
20210030503 | Shalev | Feb 2021 | A1 |
20210059796 | Weiss | Mar 2021 | A1 |
20210068773 | Moshe | Mar 2021 | A1 |
20220142782 | Lee | May 2022 | A1 |
20230039451 | Lee | Feb 2023 | A1 |
20230211612 | Yoon | Jul 2023 | A1 |
20230250224 | Park | Aug 2023 | A1 |
20230268482 | Lee | Aug 2023 | A1 |
Number | Date | Country |
---|---|---|
2009226096 | Oct 2009 | JP |
1020190053446 | May 2019 | KR |
102015224 | Oct 2019 | KR |
1020190129247 | Nov 2019 | KR |
1020190142687 | Dec 2019 | KR |
Entry |
---|
International Search Report (English and Korean) and Written Opinion of the ISA (Korean) issued in PCT/KR2021/000217, mailed Apr. 26, 2021; ISA/KR. |
Korean Office Action issued in KR1020200182812, dated Sep. 28, 2021. |
Number | Date | Country | |
---|---|---|---|
20230039451 A1 | Feb 2023 | US |