Brain-controlled interfaces or augmented reality devices, each including a user interface device, may be utilized by a user to transmit the user's intent (i.e., a desire or plan or accomplish something) or thoughts. The user's intent or thoughts may be received, processed, and mapped to an output corresponding to the user's intent. This output may include an audio, video and the like. However, this output is displayed only to the user wearing the device.
In situations when a user desires to communicate with another person but faces challenges in expressing himself or herself, it may be desirable for the user to display his/her thoughts or intent to an observer.
Therefore, there is a need to provide a system for bi-directional communication that helps the user to communicate effectively and express his/her thoughts clearly and accurately.
The disclosure includes a method for configuring a user interface device with dual-sided display. The method involves steps including reading the user's intent using a sensor, sending a user's intent to a processing unit, processing, analyzing and mapping the user's intent using the processing unit, projecting an output corresponding to the user's intent using the user interface device, displaying the output on a user's side of the user interface device and displaying a content corresponding to the output on an observer's side of the user interface device.
The disclosure provides a system of a user interface device with a dual-sided display that may include a processor and a memory storing instructions. When these instructions are executed by the processor, the system receives the user's intent from a sensor, sends the user's intent to a processing unit. The processing unit processes, analyzes and maps the user's intent into an output on the user interface device, projects the output using the user interface device, displays the output on a user's side of the user interface device and displays a content corresponding to the output on an observer's side of the user interface device.
To easily identify the discussion of any particular element or act, the most significant digit or digits in a reference number refer to the figure number in which that element is first introduced.
The disclosure relates to configuring a transparent or semi-transparent user interface device to simultaneously display asymmetric content based on the viewing angle. The content is determined based on a user's intent as a data input. The data input may include the user's intent or thoughts, and may be entered directly through a user interface device.
Many “actions” by a source user, (sensor, touch, voice, sound, push-message, Brain-Computer Interface (BCI), etc.) are mapped via the intent and message database to a communication (e.g., “hello”) that is dual-displayed, optionally language translated, or graphicly enhanced (i.e., not only raw text) that is observed by both the source user and an outside observer independently.
A transparent or semi-transparent user interface device may be used to simultaneously communicate asymmetric content to an observer and a user. The user interface device may be a dual-sided display on which an image is projected, but through which an observer can see through. A user's intent is mapped to an output that may be displayed to the user on the user's side of the transparent user interface device, while a content corresponding to the user's intent may be displayed to the observer on the observer's side of the transparent user interface device. The content corresponding to the user's intent may include text, images, video or audio and may be asymmetric to what is displayed to the user. The user interface device may include an Augmented Reality (AR) headset, a transparent or semi-transparent screen, an optical projection, or a medium through which a user and observer would still be able to see one another.
In an embodiment, an intent via a database is mapped to an action that may be displayed to the dual-display but may also be mapped to additional actions. As an exemplary example, speaking the words “Open sesame” (sensor input), if mapped to do so, may open a garage door, a front door to a house, or turn on a light, whatever series of actuations that are mapped to that intent and dual-displayed. In various embodiments, the intent may be a triggering eye-gaze in an AR headset, eye-tracking on to a button on a tablet device, or BCI selection of a steady state visually evoked potential (SSVEP) object, to initiate a control and/or communication. As non-limiting examples in the specification below, various embodiments shown in
A method of operating a context aware dual display system involves receiving a signal that represents a user's intent through a sensor or an input device on a user interface device, wherein the user initiates the signal by an activity. The signal may then be sent to an interpretation unit to determine an intent and message from the signal. The intent and message may then be sent to a user interface controller. The user interface controller may then display the intent and message through a user interface device.
The interpretation unit may include a converter, an intent database, a message database, and a comparator. The converter may convert the signal corresponding to the user's intent to a digital signal. The intent database may include intents associated with curated signals. The message database may include messages corresponding to curated intents. The comparator may generate at least one of an intent and a message corresponding to the intent by comparing the digital signal to curated signals in the intent database and by comparing the intent to curated intents in the message database.
In some configurations, the user initiates the signal by the activity that may include at least one of a physical activity, a mental activity, an environmental stimulus, an independently received digital message from a third party (e.g, Apple™ push message), and combinations thereof.
In some configurations, the user's intent includes biosignals comprising at least one of EEG (Electroencephalography), ECG (Electrocardiography), EMG (Electromyography), EOG (Electrooculography), steady state visually evoked potentials, audio evoked potentials, motion evoked potentials, motion based detection, and combinations thereof.
In some configurations, the user interface device comprises at least one of an augmented reality headset, a transparent or semi-transparent screen, an optical projection, a medium through which the user and an observer would see one another and combinations thereof.
In some configurations, the user interface controller may select to display at least one of the intent, the message and combinations thereof.
In some configurations, the sensor is at least one of a biometric sensor attached to a user's body, an accelerometer, and a camera tracking a user's eye movements.
In some configurations, the context aware dual display system may display the intent to a contextually related display. A “Contextually related display” refers to a display device with a set of physical and cognitive factors that determine the meaning of an otherwise ambiguous element(s) associated with those factors. The contextually related display may be a portion of a semi-transparent screen on the user interface device.
In some configurations, the signal may be a request to display a virtual keyboard on the user interface device. The inputs from the virtual keyboard may initiate a command to communicate to an observer.
In some configurations, the method may involve receiving a communication from a third party and displaying the communication to a contextually related display.
In some configurations, the method may involve determining if the intent database and the message database include a desired intent or a desired message, wherein at least one of the intent database and the message database is located on the user interface device. In some instances, at least one of the intent database and the message database may be located on a cloud server or local edge computing device (i.e., data is processed by the device itself or by a local computer or server, rather than being transmitted to a data center or cloud server). In some instances, the method may further involve updating at least one of the intent database and the message database with the desired intent or the desired message.
A computing apparatus may include a processor and memory storing instructions that, when executed by the processor, configure the apparatus to perform certain actions. The instructions may configure the apparatus to receive a signal that represents a user's intent through a sensor on a user interface device, wherein the user initiates the signal by an activity. The instructions may configure the apparatus to send the signal to an interpretation unit to determine an intent and message from the signal. The instructions may configure the apparatus to send the intent and the message to a user interface controller to display the intent and the message. The instructions may configure the apparatus to display the intent and the message on the user interface device.
The interpretation unit may include a converter, an intent database, a message database, and a comparator. The converter may convert the signal corresponding to the user's intent to a digital signal. The intent database may include intents associated with curated signals. The message database may include messages corresponding to curated intents. The comparator may generate at least one of an intent and a message corresponding to the intent by comparing the digital signal to curated signals in the intent database and by comparing the intent to curated intents in the message database.
In some configurations, the user interface device may include at least one of an augmented reality headset, a transparent or semi-transparent screen, a medium through which the user and an observer would see one another and combinations thereof. In some configurations, the transparent or semi-transparent screen may be a transparent organic light emitting diode (TOLED) screen or projection on a semi-reflective medium. In some configurations, the transparent or semi-transparent screen may be replaced by two non-transparent displays, wherein the two non-transparent displays show contextually related information.
In some configurations, the user interface controller may select to display at least one of the intent, the message and combinations thereof.
In some configurations, the user may initiate the signal by the activity that may include at least one of a physical activity, a mental activity, and environmental stimulus, an independently received digital message from a third party (e.g, APPLE™ push message), and combinations thereof.
Computer software, Hardware, and networks may be utilized in a variety of different system environments, including standalone, networked, remote-access (aka, remote desktop), virtualized, and/or cloud-based environments, among others.
The user's intent may include a biosignal such as EEG (Electroencephalography), ECG (Electrocardiography), EMG (Electromyography), EOG (Electrooculography), steady state visually evoked potentials, steady state audio evoked potentials, motion evoked potentials, motion detection such as derived from a gyroscope, accelerometer or magnetometer input, and the like.
In one embodiment, when a user sees a flashing light, the system processes what the user sees and gives a corresponding output. For example, if the flashing light is “green” in color and it is associated with “happy”, then the display on the user's side may show the text “happy” and the display on the observer's side may show a 3D image that corresponds to “happy.” In one embodiment, the visual word may appear correctly to observer 122, e.g., “Hello” and as a mirror image, e.g., “olleH” to the human user 114.
In one embodiment, a user generates a greeting expression of “Hello” and the user preference is set for informal communication style. The message may be displayed to the user in their native language (e.g., English=“Hi”)(translation 130). The observer may see the expression in their native language (e.g., Portuguese=“Oi”)(translation 128). The translation may be performed by a processing layer 132 that captures activating spoken and visual words 112 for translation and displays the translation 128 on the observer's side of the display 124.
In an embodiment as seen in
A human user 404 interacts with the environment, the sensor 406 reads the user's intent and triggers the operating system. The analog to digital converter 408 receives the sensor 406 output (e.g., user's intent). The analog to digital converter 408 transforms the sensor output into a digital signal which is sent to a processing unit 410. The signal is then processed using, for example, the system described in
The interpretation unit 508 may include a converter 516, a comparator 518, an intent database 510 and a message database 512.
In an embodiment, when the user interacts with a physical surface, physical document, virtual document and the like in an environment, his mental or physical activity may generate brain signals that trigger the sensor 506. These analog signals may be received by the sensor 506 and sent to the interpretation unit 508. The interpretation unit 508 converts the received signal to a digital signal using an analog to digital converter 516. The comparator 518 generates an intent and a message, by comparing the digital signal to curated signals in the intent database 510 and by comparing the intent to curated intents in the message database 512. The intent and the message may then be sent to a user interface controller, such as a graphical user interface controller 514. The user interface controller may provide display options to the user such as to display the intent and the message on one side, other side or both the sides of the user interface device. Depending upon the selected option, the intent, the message or both may be displayed on the user interface device 504. For example, the user may initiate a signal by eye movements, viewing certain colors or shapes and the like.
Examples of a user interface device 504 may include an augmented reality headset, a transparent or semi-transparent screen, an optical projection, a medium through which the user and the observer would see one another and the like.
In accordance with one embodiment,
According to embodiments of the disclosure,
The processing units 906 may in some cases comprise programmable devices such as bespoke processing units optimized for a particular function, such as AR related functions. The augmented reality device logic 900 may comprise other components that are not shown, such as dedicated depth sensors, additional interfaces, etc.
Some or all of the components in
The memory 920 comprises logic 922 to be applied to the processing units 906 to execute. In some cases, different parts of the logic 922 may be executed by different components of the processing units 906. The logic 922 typically comprises code of an operating system, as well as code of one or more applications configured to run on the operating system to carry out aspects of the processes disclosed herein.
Terms used herein should be accorded their ordinary meaning in the relevant arts, or the meaning indicated by their use in context, but if an express definition is provided, that meaning controls.
Herein, references to “one embodiment” or “an embodiment” do not necessarily refer to the same embodiment, although they may. Unless the context clearly requires otherwise, throughout the description and the claims, the words “comprise,” “comprising,” and the like are to be construed in an inclusive sense as opposed to an exclusive or exhaustive sense; that is to say, in the sense of “including, but not limited to.” Words using the singular or plural number also include the plural or singular number respectively, unless expressly limited to a single one or multiple ones. Additionally, the words “herein,” “above,” “below” and words of similar import, when used in this application, refer to this application as a whole and not to any particular portions of this application. When the claims use the word “or” in reference to a list of two or more items, that word covers all of the following interpretations of the word: any of the items in the list, all of the items in the list and any combination of the items in the list, unless expressly limited to one or the other. Any terms not expressly defined herein have their conventional meaning as commonly understood by those having skill in the relevant art(s).
It is to be understood that the disclosed subject matter is not limited in its application to the details of construction and to the arrangements of the components set forth in the following description or illustrated in the drawings. The disclosed subject matter is capable of other embodiments and of being practiced and carried out in various ways. Also, it is to be understood that the phraseology and terminology employed herein are for the purpose of description and should not be regarded as limiting.
As such, those skilled in the art will appreciate that the conception, upon which this disclosure is based, may readily be utilized as a basis for the designing of other structures, systems, methods and media for carrying out the several purposes of the disclosed subject matter. It is important, therefore, that the claims be regarded as including such equivalent constructions insofar as they do not depart from the spirit and scope of the disclosed subject matter.
Although the disclosed subject matter has been described and illustrated in the foregoing exemplary embodiments, it is understood that the present disclosure has been made only by way of example, and that numerous changes in the details of implementation of the disclosed subject matter may be made without departing from the spirit and scope of the disclosed subject matter, which is limited only by the claims which follow.
The user 1106 may greet the user 1108 by saying “HELLO,” and this message 1114 may be shown to the user 1106 on an occupant-facing partition 1110 of the contextually related display 1102. On the opposite side of the contextually related display 1102, the user 1108 may be shown the translated message 1116 (e.g., “Ni Hao”) on the outward-facing partition 1112 of the contextually related display 1102.
The term “network” as used herein and depicted in the drawings refers not only to systems in which remote storage devices are coupled together via one or more communication paths, but also to stand-alone devices that may be coupled, from time to time, to such systems that have storage capability. Consequently, the term “network” includes not only a “physical network” but also a “content network,” which is comprised of the data—attributable to a single entity—which resides across all physical networks.
The components may include data server 1410, web server 1406, and client computer 1404, laptop 1402. Data server 1410 provides overall access, control and administration of databases and control software for performing one or more illustrative aspects described herein. Data server data server 1410 may be connected to web server 1406 through which users interact with and obtain data as requested. Alternatively, data server 1410 may act as a web server itself and be directly connected to the internet. Data server 1410 may be connected to web server 1406 through the network 1408 (e.g., the internet), via direct or indirect connection, or via some other network. Users may interact with the data server 1410 using remote computer 1404, laptop 1402, e.g., using a web browser to connect to the data server 1410 via one or more externally exposed web sites hosted by web server 1406. Client computer 1404, laptop 1402 may be used in concert with data server 1410 to access data stored therein, or may be used for other purposes. For example, from client computer 1404, a user may access web server 1406 using an internet browser, as is known in the art, or by executing a software application that communicates with web server 1406 and/or data server 1410 over a computer network (such as the internet).
Servers and applications may be combined on the same physical machines, and retain separate virtual or logical addresses, or may reside on separate physical machines.
Each component data server 1410, web server 1406, computer 1404, laptop 1402 may be any type of known computer, server, or data processing device. Data server 1410, e.g., may include a processor 1412 controlling overall operation of the data server 1410. Data server 1410 may further include RAM 1416, ROM 1418, network interface 1414, input/output interfaces 1420 (e.g., keyboard, mouse, display, printer, etc.), and memory 1422. Input/output interfaces 1420 may include a variety of interface units and drives for reading, writing, displaying, and/or printing data or files. Memory 1422 may further store operating system software 1424 for controlling overall operation of the data server 1410, control logic 1426 for instructing data server 1410 to perform aspects described herein, and other application software 1428 providing secondary, support, and/or other functionality which may or may not be used in conjunction with aspects described herein. The control logic may also be referred to herein as the data server software control logic 1426. Functionality of the data server software may refer to operations or decisions made automatically based on rules coded into the control logic, made manually by a user providing input into the system, and/or a combination of automatic processing based on user input (e.g., queries, data updates, etc.).
memory 1422 may also store data used in performance of one or more aspects described herein, including a first database 1432 and a second database 1430. In some embodiments, the first database may include the second database (e.g., as a separate table, report, etc.). That is, the information can be stored in a single database, or separated into different logical, virtual, or physical databases, depending on system design. Web server 1406, computer 1404, laptop 1402 may have similar or different architecture as described with respect to data server 1410. Those of skill in the art will appreciate that the functionality of data server 1410 (or web server 1406, computer 1404, laptop 1402) as described herein may be spread across multiple data processing devices, for example, to distribute processing load across multiple computers, to segregate transactions based on geographic location, user access level, quality of service (QoS), etc.
One or more aspects may be embodied in computer-usable or readable data and/or computer-executable instructions, such as in one or more program modules, executed by one or more computers or other devices as described herein. Generally, program modules include routines, programs, objects, components, data structures, etc. that perform particular tasks or implement particular abstract data types when executed by a processor in a computer or other device. The modules may be written in a source code programming language that is subsequently compiled for execution, or may be written in a scripting language such as (but not limited to) HTML or XML. The computer executable instructions may be stored on a computer readable medium such as a nonvolatile storage device. Any suitable computer readable storage media may be utilized, including hard disks, CD-ROMs, optical storage devices, magnetic storage devices, and/or any combination thereof. In addition, various transmission (non-storage) media representing data or events as described herein may be transferred between a source and a destination in the form of electromagnetic waves traveling through signal-conducting media such as metal wires, optical fibers, and/or wireless transmission media (e.g., air and/or space). various aspects described herein may be embodied as a method, a data processing system, or a computer program product. Therefore, various functionalities may be embodied in whole or in part in Software, firmware and/or hardware or hardware equivalents such as integrated circuits, field programmable gate arrays (FPGA), and the like. Particular data structures may be used to more effectively implement one or more aspects described herein, and such data structures are contemplated within the scope of computer executable instructions and computer-usable data described herein.
Various functional operations described herein may be implemented in logic that is referred to using a noun or noun phrase reflecting said operation or function. For example, an association operation may be carried out by an “associator” or “correlator”. Likewise, switching may be carried out by a “switch”, selection by a “selector”, and so on.
Within this disclosure, different entities (which may variously be referred to as “units,” “circuits,” other components, etc.) may be described or claimed as “configured” to perform one or more tasks or operations. This formulation—[entity] configured to [perform one or more tasks]—is used herein to refer to structure (i.e., something physical, such as an electronic circuit). More specifically, this formulation is used to indicate that this structure is arranged to perform the one or more tasks during operation. A structure can be said to be “configured to” perform some task even if the structure is not currently being operated. A “credit distribution circuit configured to distribute credits to a plurality of processor cores” is intended to cover, for example, an integrated circuit that has Circuitry that performs this function during operation, even if the integrated circuit in question is not currently being used (e.g., a power supply is not connected to it). Thus, an entity described or recited as “configured to” perform some task refers to something physical, such as a device, circuit, memory storing program instructions executable to implement the task, etc. This phrase is not used herein to refer to something intangible.
The term “configured to” is not intended to mean “configurable to.” An unprogrammed FPGA, for example, would not be considered to be “configured to” perform some specific function, although it may be “configurable to” perform that function after programming.
Reciting in the appended claims that a structure is “configured to” perform one or more tasks is expressly intended not to invoke 35 U.S.C. § 112(f) for that claim element. Accordingly, claims in this application that do not otherwise include the “means for” [performing a function] construct should not be interpreted under 35 U.S.C § 112(f).
As used herein, the term “based on” is used to describe one or more factors that affect a determination. This term does not foreclose the possibility that additional factors may affect the determination. That is, a determination may be solely based on specified factors or based on the specified factors as well as other, unspecified factors. Consider the phrase “determine A based on B.” This phrase specifies that B is a factor that is used to determine A or that affects the determination of A. This phrase does not foreclose that the determination of A may also be based on some other factor, such as C. This phrase is also intended to cover an embodiment in which A is determined based solely on B. As used herein, the phrase “based on” is synonymous with the phrase “based at least in part on.”
As used herein, the phrase “in response to” describes one or more factors that trigger an effect. This phrase does not foreclose the possibility that additional factors may affect or otherwise trigger the effect. That is, an effect may be solely in response to those factors, or may be in response to the specified factors as well as other, unspecified factors. Consider the phrase “perform A in response to B.” This phrase specifies that B is a factor that triggers the performance of A. This phrase does not foreclose that performing A may also be in response to some other factor, such as C. This phrase is also intended to cover an embodiment in which A is performed solely in response to B.
As used herein, the terms “first,” “second,” etc. are used as labels for nouns that they precede, and do not imply any type of ordering (e.g., spatial, temporal, logical, etc.), unless stated otherwise. For example, in a register file having eight registers, the terms “first register” and “second register” can be used to refer to any two of the eight registers, and not, for example, just logical registers 0 and 1.
When used in the claims, the term “or” is used as an inclusive or and not as an exclusive or. For example, the phrase “at least one of x, y, or z” means any one of x, y, and z, as well as any combination thereof.
Having thus described illustrative embodiments in detail, it will be apparent that modifications and variations are possible without departing from the scope of the invention as claimed. The scope of inventive subject matter is not limited to the depicted embodiments but is rather set forth in the following Claims.
Terms used herein should be accorded their ordinary meaning in the relevant arts, or the meaning indicated by their use in context, but if an express definition is provided, that meaning controls.
Herein, references to “one embodiment” or “an embodiment” do not necessarily refer to the same embodiment, although they may. Unless the context clearly requires otherwise, throughout the description and the claims, the words “comprise,” “comprising,” and the like are to be construed in an inclusive sense as opposed to an exclusive or exhaustive sense; that is to say, in the sense of “including, but not limited to.” Words using the singular or plural number also include the plural or singular number respectively, unless expressly limited to a single one or multiple ones. Additionally, the words “herein,” “above,” “below” and words of similar import, when used in this application, refer to this application as a whole and not to any particular portions of this application. When the claims use the word “or” in reference to a list of two or more items, that word covers all of the following interpretations of the word: any of the items in the list, all of the items in the list and any combination of the items in the list, unless expressly limited to one or the other. Any terms not expressly defined herein have their conventional meaning as commonly understood by those having skill in the relevant art(s).
It is to be understood that the disclosed subject matter is not limited in its application to the details of construction and to the arrangements of the components set forth in the following description or illustrated in the drawings. The disclosed subject matter is capable of other embodiments and of being practiced and carried out in various ways. Also, it is to be understood that the phraseology and terminology employed herein are for the purpose of description and should not be regarded as limiting.
As such, those skilled in the art will appreciate that the conception, upon which this disclosure is based, may readily be utilized as a basis for the designing of other structures, systems, methods and media for carrying out the several purposes of the disclosed subject matter. It is important, therefore, that the claims be regarded as including such equivalent constructions insofar as they do not depart from the spirit and scope of the disclosed subject matter.
This application claims the benefit of U.S. provisional patent application Ser. No. 62/704,048 filed on Jan. 22, 2019, the contents of which are incorporated herein by reference in their entirety.
Number | Date | Country | |
---|---|---|---|
62704048 | Jan 2019 | US |