It has become increasingly important to monitor the health and activities of individuals. Accordingly, there is a need for improved devices that make monitoring these aspects of an individual easier and more convenient to use. Additionally, since such devices may comprise one or more computer processors and/or sensors, may require power to operate, and may be powered by one or more rechargeable batteries, there is also a need for improved systems and methods for charging such devices. Various embodiments of the present systems and methods recognize and address the foregoing considerations, and others, of prior art devices.
A computerized temple, according to various embodiments, comprises an elongated body having (a) a first end defining a first end wall, (b) a second end defining a second end wall, (c) a first side wall that extends between the first end and the second end, (d) a second side wall that extends between the first end and the second end, and (e) a blind bore formed in the first end wall that extends from the first end toward the second end. In various embodiments, the computerized temple further comprises at least one processor, one or more sensors operatively coupled to the at least one processor, and a power source operatively coupled to the at least one processor and the one or more sensors. In some embodiments, the temple body first end is configured to be releasably attached to an end of a frame, and the at least one processor and the power source are configured to be received in the blind bore.
In various embodiments, the computerized eyewear temple further comprises a closure that is coupled to the elongated body first end such that the closure at least partially covers a portion of the blind bore formed in the first end wall. In some embodiments, the closure forms a portion of a hinge that is used to rotationally couple the elongated body to the frame of the eyewear. In particular embodiments, an opening is formed through a portion of the hinge and a light is positioned in the opening. In some embodiments, the light is operatively coupled to the at least one processor.
In particular embodiments, the one or more sensors are configured to be received in the blind bore with the at least one processor and the power source. In other embodiments, the one or more sensors are configured to be coupled to the frame of the eyewear.
In various embodiments, the computerized eyewear temple further comprises a charging port connection (e.g., a USB port, a micro USB port, a magnetic charging port, etc.) that is operatively coupled to at least one of the at least one processor or the power source, where the charging port connection is accessible through the closure to allow a user to charge the power source. In some embodiments, the charging port connection is hidden when the computerized eyewear temple is coupled to the frame and positioned perpendicular to the frame. In various embodiments, the computerized eyewear temple further comprises a first half-lock and a second half-lock spaced apart from the first half-lock. In particular embodiments, at least one of the at least one processor or the power source is received intermediate the first and second half-locks such that when the first and second half-locks are placed into the elongated body blind bore, the first and second half-locks securely maintain the at least one of the at least one processor or the power source in the elongated body blind bore and prevent it from sliding or rotating in the elongated body blind bore.
Computerized eyewear according to certain embodiments comprises (1) a frame having (a) a first end, (b) a second end, (c) a front surface that extends between the first frame end and the second frame end, (d) a rear surface that extends between the first frame end and the second frame end, and (2) a computerized temple comprising an elongated body having (a) a first end defining a first end wall, wherein the elongated body first end is pivotally coupled to the frame first end, (b) a second end defining a second end wall, wherein the second end is configured to rest on an ear of a wearer of the computerized eyewear, (c) a first side wall and a second side wall that extend between the first end and the second end, and (d) a blind bore formed in the first end wall that extends from the first end toward the second end. In various embodiments, at least one processor and a power source is received in the blind bore formed in the first end wall of the elongated body. In some embodiments, the power source is operatively coupled to the at least one processor. In particular embodiments, at least one sensor is operatively coupled to the at least one processor and is also received in the elongated body blind bore.
In particular embodiments, the computerized eyewear further comprises a charging socket that is positioned at least partially over the blind bore formed in the first end of the elongated body, wherein the charging socket is operatively couple to the power source. In some embodiments, the charging socket further comprises a paramagnetic plate.
In certain embodiments, a computerized temple for use with a frame to form computerized eyewear comprises (1) an elongated body having (a) a first end defining a first end wall, wherein the elongated body first end is configured to be pivotally coupled to a frame first end, (b) a second end defining a second end wall, wherein the second end is configured to rest on an ear of a wearer, (c) a first side wall and a second side wall that extend between the first end and the second end, and (d) a blind bore formed in the first end wall that extends from the first end toward the second end intermediate the first and the second side walls, (2) electronic components received in the blind bore formed in the first end wall of the elongated body, and (3) a closure at least partially received on the elongated body first end. In various embodiments, the closure pivotally couples the elongated body first end to the frame. In some embodiments, the closure further comprises a portion of a hinge that is used to pivotally couple the elongated body first end to the frame.
In particular embodiments, the computerized temple further comprises a first half-lock, and a second half-lock that are spaced apart from one another so that a portion of the electronic components is positioned intermediate the first half-lock and the second half-lock. In various embodiments, the first and second half-locks are used to secure the electronic components within the blind bore. In some embodiments, a portion of a charging port may be positioned intermediate the first and second half locks in addition to one or more other electronic components. In this way, the sandwich of the charging port between the first and second half-locks aid in forming a water tight assembly since the charging port and first and second half-locks are received at the open end of the blind bore. In other embodiments, one or more fasteners pass through a portion of the closure and the first side wall. The one or more fasteners are secured to at least one of the first and the second half-locks so that the electronic components do not slide within the blind bore.
Various embodiments of a computerized temple for assessing a user's health and activities are described below. In the course of this description, reference will be made to the accompanying drawings, which are not necessarily drawn to scale and wherein:
Various embodiments will now be described more fully hereinafter with reference to the accompanying drawings. It should be understood that the invention may be embodied in many different forms and should not be construed as limited to the embodiments set forth herein. Rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the invention to those skilled in the art. Like numbers refer to like elements throughout.
Eyewear
In various embodiments, eyewear may be converted to computerized eyewear by attaching a computerized temple to the standard eyewear frame. The computerized temple may be formed from an elongated body having a first end and a second end, where the elongated body extends between the first and second ends. A blind bore is formed in a wall of the first end so that the blind bore extends from the wall of the temple first end to a point intermediate the temple first and second ends. One or more electronic components that are operatively coupled to one another may be placed into the blind bore. The electronic components may comprise one or more processors, one or more power sources and/or one or more sensors. A hinge portion having a closure may be seated on the temple first end to close off the blind bore and seal the electronic components in the blind bore. One or more O-rings may be used to form a water tight seal to protect the electronic components from water and sweat when the electronic eyewear is being worn by the user or being cleaned.
Eyewear
As shown in
Eyewear Frame
Still referring to
The eyewear frame 108 includes a first nose pad 122 and a second nose pad 124, which may be configured to maintain the eyewear 100 adjacent the front of a wearer's face such that the lenses 118, 120 are positioned substantially in front of the wearer's eyes while the wearer is wearing the eyewear 100. In particular embodiments, the nose pads 122, 124 may comprise a material that is configured to be comfortable when worn by the wearer (e.g., rubber, polymer, etc.). In other embodiments, the nose pads 122, 124 may include any other suitable material (e.g., plastic, metal, etc.). In still other embodiments, the nose pads 122, 124 may be integrally formed with the frame 108 and made from the same material as the eyewear frame 108.
The eyewear frame 108 includes a first connection receiving end 112 that attaches the computerized temple 102 to the frame first end 110, and a second connection receiving end 128 that attaches the second temple 114 to the frame second end 116. In various embodiments, the connection receiving ends 112, 128 may be releasably coupled to the computerized temple 102 and the second temple 114, respectively, by any suitable connection (e.g., tongue and groove, ball and socket, spring hinge, friction fit, screw, spring loaded ball and catch, spring loaded pin and catch, spring tab and catch, hinge and screw, etc.). In particular embodiments, the first and second connection receiving ends 112, 128 may be welded to, or integrally formed with, the eyewear frame 108. In some embodiments, the computerized temple 102 is pivotally coupled to the frame first end 110 by a hinge as described in more detail in
Computerized Temple
As shown in
Referring to
Still referring to
As a further example, the at least one processor 132 and the power source 134 may be embedded (e.g., received in a cavity formed in the temple 102) into the computerized temple 102. In some such embodiments, at least one of the one or more sensors 130 may be embedded or coupled to the computerized temple 102, another of the one or more sensors 130 may be coupled to the frame 108, and still another of the one or more sensors 130 may be operatively coupled to the nose piece 122 (
In particular embodiments, the computerized temple 102 may further include one or more user interfaces for communicating with a wearer of the computerized temple 102. For example, the computerized temple 102 may include one or more speakers, microphones, displays, haptic output devices and/or other user interface devices (e.g., lights, touch buttons, etc.) that are operatively coupled to facilitate the transfer of information between the wearer of the temple and the temple's one or more processors (e.g., while the wearer is wearing the temple). The computerized temple 102 may further include one or more wireless communications devices (e.g., a Bluetooth chip, a near field communications chip, or a cellular communications chip) for facilitating communication between the computerized temple 102 and one or more remote computing devices (e.g., a central server or the wearer's handheld computing device, laptop computer, etc.).
In particular embodiments, a haptic device (e.g., a device that applies forces, vibrations or motions to the user) may provide output signals to the wearer in response to input commands. For example, if the computerized eyewear is wireless linked to the wearer's mobile computing device, a haptic feedback device may notify a wearer when a call is received, a text message is received or when any other notification is being provided. In some embodiments, instead of, or in addition to, providing a notification from a handheld device linked to the computerized eyewear, the computerized eyewear itself may provide notifications to the wearer for actions or commands directly provided by the computerized eyewear. In some embodiments, the computerized eyewear may also contain haptic input devices that incorporate tactile sensors that measure forces exerted by the wearer on the haptic input device. Thus, the device can provide tactile output as well as receiving tactile inputs from the wearer.
In various embodiments, the one or more sensors 130, the at least one processor 132, and the power source 134 may be formed in any shape. In particular embodiments, the one or more sensors 130, the at least one processor 132, and the power source 134 may be formed on one or more circuit boards that are received in the temple 102 as described in greater detail herein.
Sensors
Still referring to
The one or more sensors 130 may include, for example: (1) one or more heart rate monitors; (2) one or more electrocardiograms (EKG); (3) one or more electroencephalograms (EEG); (4) one or more pedometers; (5) one or more thermometers; (6) one or more transdermal sensors; (7) one or more front-facing cameras; (8) one or more eye-facing cameras; (9) one or more microphones; (10) one or more accelerometers; (11) one or more blood pressure sensors; (12) one or more pulse oximeters; (13) one or more respiratory rate sensors; (14) one or more blood alcohol concentration (BAC) sensors; (15) one or more near-field communication sensors; (16) one or more motion sensors; (17) one or more gyroscopes; (18) one or more geomagnetic sensors; (19) one or more global positioning system (GPS) sensors; (20) one or more impact sensors; (21) one or more wireless communication sensors (e.g., a Bluetooth chip); (22) one or more tear sensors; (23) one or more olfactory sensors; and/or (24) any other suitable one or more sensors. In particular embodiments, the one or more sensors comprise a pulse oximeter, a front-facing camera, an eye-facing camera, an accelerometer and a gyroscope.
In particular embodiments, the system is configured to receive input from a user (e.g., a wearer of the eyewear) via one or more gestures, for example, using at least one of the sensors described immediately above. In various embodiments, the system may, for example, be configured to: (1) identify a gesture performed by the user; and (2) at least partially in response to identifying the gesture, perform a function associated with the gesture. In particular embodiments, the system may be configured to perform a particular function in response to identifying a particular gesture, where the particular gesture is associated with the particular function. In particular embodiments, the system may be configured to enable the user to provide one or more gestures for performing a particular function. In such embodiments, the system may, for example: (1) receive a selection of a particular function from the user; (2) receive input of one or more gestures from the user; and (3) associate the particular function with the one or more gestures.
Second Temple
In various embodiments, the second temple 114 substantially mirrors the shape of the computerized temple 102. Thus, for purposes of ease of understanding and clarity, only certain parts will be discussed to highlight the differences in the structure and operation of the embodiment shown in
In some embodiments, the second temple 114 may be different in form and shape from that of the computerized temple 102. For example, the second temple may be thinner in cross section along its length or made from a different material. In still other embodiments, the second temple may be shorter or longer in length as compared to the computerized temple 102. Thus, depending on the application of the computerized temple, the first and second temples may be substantially the same or they may be different. For example, in some embodiments, the computerized temple may be a replacement temple that is attached to a prefabricated standard eyewear frame. That is, a user may purchase the computerized temple and then attach the computerized temple 102 to an already existing pair of prescription frames thereby converting the preexisting prescription eyewear into computerized eyewear. In other embodiments, the computerized temple 102 may come pre-attached to a standard pair of non-computerized frames where the first and second temples are identical in form and color.
Computerized Temple Construction
Referring to
In various embodiments, the electronic components that are received in the cavity 300 of the computerized temple 102 are formed from one or more circuit boards 308 and 310 and the power source 134 that are operatively coupled to one another. The electronic components may be mounted in cavity 300 in any suitable manner, as further described herein. Referring particularly to
In various embodiments, the hinge portion 106 further comprises a partial closure 106a that is received over the first temple end 102a. The partial closure 106a defines an opening 106b that covers the temple first end 102a and also houses a charging port. In various embodiments, the charging port comprises a paramagnetic plate, as described in greater detail herein. The charging port is operatively coupled to one or more of the processor and the power source and allows a user to recharge the power source.
The hinge portion/closure 106 further defines an opening 324 (
In other embodiments, some or all of the electronic components may be partially or fully encased in a polymer material. For example, some or all of the electronic components may be encased in a polymer that is shaped to fit in the blind bore 300. In other embodiments, the electronic components may be placed intermediate one or more first and second half-locks so that the sandwiched electronic components are moveable with respect to each other pair of half-locks. These alternative configurations may allow for the electronic components to fit more snuggly in the length of the blind bore 300.
Recharging Port
In various embodiments and as described above, the computerized temple 102 contains a power source 134 (
Referring to
In particular embodiments, the charging port 334 contains a plate 336 formed from a magnetic, ferromagnetic or paramagnetic material. In embodiments where a magnetic or paramagnetic material is used for the plate 336, the charging connector may be formed from a magnetic material. In this way, when the charging cable comes into close proximity to the charging port 334, the magnetic male portion of the cable is attracted to the plate 336 so that the charging connector engages with and stays coupled to the charging port 334 until a sufficient force is applied to separate the charging connector from the charging port 334.
In various embodiments, the plate 336 may be made from a polymer material that forms a front wall of the blind bore 300 (
Method of Assembling the Computerized Temple
Assembly of the computerized temple is illustrated in
Referring to
Exemplary Computing Device Architecture
An exemplary computer device 820 includes a processing device 802, a main memory 804 (e.g., read-only memory (ROM), flash memory, dynamic random access memory (DRAM) such as synchronous DRAM (SDRAM) or Rambus DRAM (RDRAM), etc.), a static memory 806 (e.g., flash memory, static random access memory (SRAM), etc.), and a data storage device 818, which communicate with each other via a bus 823.
The processing device 802 represents one or more general-purpose or specific processing devices such as a microprocessor, a central processing unit (CPU), or the like. More particularly, the processing device 802 may be a complex instruction set computing (CISC) microprocessor, reduced instruction set computing (RISC) microprocessor, very long instruction word (VLIW) microprocessor, or processor implementing other instruction sets, or processors implementing a combination of instruction sets. The processing device 802 may also be one or more special-purpose processing devices such as an application specific integrated circuit (ASIC), a field programmable gate array (FPGA), a digital signal processor (DSP), network processor, or the like. The processing device 802 may be configured to execute processing logic 826 for performing various operations and steps discussed herein.
The computing device 820 may further include a network interface device 808. The computing device 820 may also include a video display unit 610 (e.g., a liquid crystal display (LCD) or a cathode ray tube (CRT)), an alpha-numeric input device 812 (e.g., a keyboard), a cursor control device 814 (e.g., a mouse), and a signal generation device 816 (e.g., a speaker).
The data storage device 818 may include a non-transitory computing device-accessible storage medium 830 (also known as a non-transitory computing device-readable storage medium, a non-transitory computing device-readable medium, or a non-transitory computer-readable medium) on which is stored one or more sets of instructions (e.g., software 822) embodying any one or more of the methodologies or functions described herein. The one or more sets of instructions may also reside, completely or at least partially, within the main memory 804 and/or within the processing device 802 during execution thereof by the computing device 820—the main memory 804 and the processing device 802 also constituting computing device-accessible storage media. The one or more sets of instructions may further be transmitted or received over a network 815 via a network interface device 808.
While the computing device-accessible storage medium 830 is shown in an exemplary embodiment to be a single medium, the term “computing device-accessible storage medium” should be understood to include a single medium or multiple media (e.g., a centralized or distributed database, and/or associated caches and servers) that store the one or more sets of instructions. The term “computing device-accessible storage medium” should also be understood to include any medium that is capable of storing, encoding, or carrying a set of instructions for execution by the computing device and that causes the computing device to include any one or more of the methodologies of the present invention. The term “computing device-accessible storage medium” should accordingly be understood to include, but not be limited to, solid-state memories, optical and magnetic media, etc.
Alternative Embodiment Having a Second Computerized Temple with Sensors
The eyewear frame 108 may be a standard eyewear frame already owned by the wearer. Similar to the eyewear 100 shown in
In various embodiments, the second computerized temple 114′ may be operatively coupled to the first computerized temple 102 via a wireless connection (e.g., WiFi, Bluetooth, near-field communications, etc.). In other embodiments, the second computerized temple 114′ may be operatively coupled to the first computerized temple 102 via a hardwire connection. For example, the first end of the first computerized temple 102a may plug into a connector (not shown) built into the frame front 108 that coupled to a second connector formed at the second end 116′ of the frame front 108 so that the second connector plugs into the charging port located at the first end 114a′ of the second computerized temple 114′. In this way, the first and second computerized temples 102 and 114′ are in a position in which they are perpendicular to the back of the frame front 108, each connector is plugged into the charging port thereby forming an electrical connection between the first and the second computerized temples 102 and 114′. It should be understood that an electrical connection between the first computerized temple 102 and the second computerized temple 114′ may be carried out in any other suitable method, for example, by a wire coupled between the first and second computerized temples.
Similar to the eyewear 100 of
Similar to the eyewear found in
Computerized Eyewear Having Voice Controlled Intelligent Personal Assistant Functionality
In various embodiments, the computerized eyewear discussed above further comprises functionality to provide voice-controlled intelligent personal assistant services. That is, the computerized eyewear can be placed into a default mode where the computerized eyewear continuously monitors all speech for a wake word to be spoken by the wearer. Once the wake word is spoken, the system can either prompt the user for a question or respond to the next question or command received by the wearer. In other modes, the wearer may manually wake the voice-controlled intelligent personal assistant services or use a voice-activated remote control. In particular embodiments, the computerized eyewear further comprises a digital media processor, a wireless and/or Bluetooth chip, a programmable LED driver, a low power stereo audio analog-to-digital converter and an integrated power-management integrated circuit, one or more microphones and one or more speakers.
In particular embodiments, the computerized eyewear is configured to wirelessly connect to a cloud based server system such that as the computerized eyewear receives voice recordings of the wear, the recording is sent to the remote service so that the recording can be analyzed to determine an appropriate answer or response. Examples of cloud based server systems includes, but are not limited to, Amazon Voices Services, Google Assistant, Google Now, Cortana intelligent personal assistant, and Siri intelligent personal assistant. In other embodiments, the questions or commands that are received by the computerized eyewear can be analyzed locally by the processor and other electronics in the computerized eyewear and an appropriate answer or response may be determined in real-time by the computerized eyewear as discussed in greater detail herein.
Exemplary Technical Platforms
As will be appreciated by one skilled in the relevant field, the present systems and methods may be, for example, embodied as a computer system, a method, or a computer program product. Accordingly, various embodiments may be entirely hardware or a combination of hardware and software. Furthermore, particular embodiments may take the form of a computer program product stored on a computer-readable storage medium having computer-readable instructions (e.g., software) embodied in the storage medium. Various embodiments may also take the form of Internet-implemented computer software. Any suitable computer-readable storage medium may be utilized including, for example, hard disks, compact disks, DVDs, optical storage devices, and/or magnetic storage devices.
Various embodiments are described below with reference to block diagram and flowchart illustrations of methods, apparatuses, (e.g., systems), and computer program products. It should be understood that each block of the block diagrams and flowchart illustrations, and combinations of blocks in the block diagrams and flowchart illustrations, respectively, can be implemented by a computer executing computer program instructions. These computer program instructions may be loaded onto a general purpose computer, a special purpose computer, or other programmable data processing apparatus that can direct a computer or other programmable data processing apparatus to function in a particular manner such that the instructions stored in the computer-readable memory produce an article of manufacture that is configured for implementing the functions specified in the flowchart block or blocks.
The computer instructions may execute entirely on the computerized eyewear, partly on the computerized eyewear, as a stand-alone software package, partly on the computerized eyewear and partly on a remote computer, or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the computerized eyewear through any type of network, including but not limited to: (1) a local area network (LAN); (2) a wide area network (WAN); (3) a cellular network; (4) the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider); or over a Bluetooth connection.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner such that the instructions stored in the computer-readable memory produce an article of manufacture that is configured for implementing the function specified in the flowchart block or blocks. The computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer-implemented process (e.g., method) such that the instructions that execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart block or blocks.
Example System Architecture
The one or more networks 1015 may include any of a variety of types of wired or wireless computer networks such as the Internet, a private intranet, a mesh network, a public switch telephone network (PSTN), or any other type of network (e.g., a network that uses Bluetooth or near field communications to facilitate communication between computing devices). The communication link between the Computerized Eyewear Devices 1050 and the Voice Command Server 1020 may be, for example, implemented via a Local Area Network (LAN) or via the Internet.
Exemplary System Platform
As noted above, a system, according to various embodiments, is adapted to receive voice commands from a wearer of the one or more computerized eyewear devices 1050, decipher the voice commands, determine one or more actions to take based on the voice commands and then carry out the one or more actions in response to the voice commands. Various aspects of the system's functionality may be executed by certain system modules, including the Voice Command Module 1100. The Voice Command Module 1100 is discussed in greater detail below.
Voice Command Module
The system begins, in various embodiments, at step 1105 when the computerized eyewear is activated either automatically when it recognizes the a “wake” word spoken by the wearer or manually when the wearer activates the voice command service (e.g., by pressing a button on the computerized eyewear, via a button on a remote control operatively coupled to the computerized eyewear, etc.). Once the voice command service has been awoken, the computerized eyewear monitors the wearer's speech for one or more commands (e.g., requests to take action, questions, etc.). In various embodiments, the computerized eyewear may capture the wearer's words in real-time as they are spoken. In other embodiments, the wearer's spoken words may be captured by recording them in memory.
At step 1110, the computerized eyewear converts the captured analog speech into one or more digital signals using a analog-to-digital converter. Once the captured speech is converted into a digital signal, at step 1115 the digital signal is put through a speech recognition and natural language understanding engines. That is, the one or more processors on the computerized eyewear utilizes algorithms through acoustic and language modeling to translate the spoken words into text. The computerized eyewear, at step 1120 interprets the one or more commands by evaluating the text extracted from the spoken one or more commands. The interpreted commands can include any one or more commands for using the computerized eyewear, questions to be answered through online web searches or entries into one or more applications associated with the computerized eyewear. At step 1125, the computerized eyewear executes the one or more interpreted commands.
For example, if the command is “what is the nearest grocery store?”, the computerized eyewear may carry out a Google® search to determine the closest grocery store and then provide a reply to the wearer via synthesized speech via a speaker operatively coupled to the computerized eyewear. If for example, the command is to record the medicine being taken by the wearer, the computerized eyewear could capture images of the medicine bottle being held by the wearer and then capture images of the medicine in the hand of the wearer. In this way, the system can analyze the captured images to determine the type and name of the medicine the wearer is taking and the quantity of medicine consumed by the wearer. The information may then be used by the computerized eyewear to track the user's compliance with a medicine regime.
In another example, the wearer can provide one or more commands to tell the computerized eyewear to capture data associated with exercise that is being performed by the wearer. For example, the wearer can tell the computerized eyewear to start collecting movement data, heartrate data and other physiological metrics associated with the wearer because the wearer is beginning to run. Thus, while the wearer is running, the computerized eyewear would use the various sensors contained in the computerized eyewear to capture physiological data associated with the wearer. Once the wear is finished exercising, the wearer can provide one or more commands to tell the computerized eyewear that the wearer has stopped the exercise. Based on the commands, the computerized eyewear can stop capturing the metrics or execute any other commands that are associate with the end of exercise, such as synching the physiological metrics captured while exercising with data collected by an application for tracking the information.
In yet other embodiments, the computerized eyewear can be operatively coupled to one or more other computer devices 1010a, 1010b, which may be a home automation computer, an electronic door lock computer system, an automobile computer system, etc. In various embodiments, the computerized eyewear system can validate the identity of the wearer via voice recognition or through some other identity recognition technology (e.g., biometrics, etc.). Once verified, the wearer can use voice commands that are captured, interpreted and executed by the one or more other computer devices in order to carry out functionality controlled by the one or more other computer devices 1010a, 1010b. For example, in the case where the computerized eyewear is operatively coupled to an automobile computer system, the user can use voice commands that are captured by the computerized eyewear, interpreted, and the commands may then be passed to the automobile computer system. Thus, the wearer may use one or more voice commands to unlock their automobile doors, start the car, change the radio station, turn on the air conditioner or carry out any other activity using voice commands that are captured by the computerized eyewear.
In the case of a home automation system, the wearer of the computerized eyewear may control, through the capture of voice commands by the computerized eyewear, lights, climate control, alarm systems or any other automated systems in the house. Additionally, the user may unlock electronic doors via their computerized eyewear. In still other embodiments, the computerized eyewear may execute a voice command such as “search for restaurants near me” and play an audio list of the search results. In other embodiments, the computerized eyewear may carry out the search and then ask whether the wearer would like to transfer the results to a mobile computing device that is operatively coupled to the computerized eyewear such as a smartwatch, a smartphone, a tablet or any other handheld computing device having a display. Thus, it should be understood that the computerized eyewear may have the processing ability to execute the interpreted commands, including searching the internet via a cellular connection, a WiFi connection or via any other network connection and then pass the results to another computing device. In the alternative, the computerized eyewear may also act as an input device for entering data on a handheld computing device that performs the search.
Many modifications and other embodiments of the invention will come to mind to one skilled in the art to which this invention pertains having the benefit of the teachings presented in the foregoing descriptions and the associated drawings. For example, as will be understood by one skilled in the relevant field in light of this disclosure, the invention may take form in a variety of different mechanical and operational configurations. For example, the computerized eyewear described in these embodiments may include any other suitable eyewear, such as, for example, ski or swim goggles, sunglasses, safety goggles or glasses, etc. Moreover, in various embodiments, instead of interpreting the voice commands locally on the computerized eyewear, the voice commands may be transmitted to a cloud service that interprets the voice commands and then returns commands to be locally executed by the computerized eyewear. Therefore, it is to be understood that the invention is not to be limited to the specific embodiments disclosed herein, and that the modifications and other embodiments are intended to be included within the scope of the appended exemplary concepts. Although specific terms are employed herein, they are used in a generic and descriptive sense only and not for the purposes of limitation.
Number | Name | Date | Kind |
---|---|---|---|
3505879 | Vanderberg | Apr 1970 | A |
3548663 | Radin | Dec 1970 | A |
3972038 | Nasa | Jul 1976 | A |
4100401 | Tutt et al. | Jul 1978 | A |
4186609 | Baermann | Feb 1980 | A |
4195642 | Price et al. | Apr 1980 | A |
4281663 | Pringle | Aug 1981 | A |
4407295 | Steuer et al. | Oct 1983 | A |
4434801 | Jiminez et al. | Mar 1984 | A |
4855942 | Bianco | Aug 1989 | A |
4878749 | McGee | Nov 1989 | A |
4919530 | Hyman | Apr 1990 | A |
5422816 | Sprague et al. | Jun 1995 | A |
5452480 | Ryden | Sep 1995 | A |
5497143 | Matsuo et al. | Mar 1996 | A |
5585871 | Linden | Dec 1996 | A |
5670872 | Van De Walle et al. | Sep 1997 | A |
5746501 | Chien et al. | May 1998 | A |
5891042 | Sham et al. | Apr 1999 | A |
5931764 | Freeman et al. | Aug 1999 | A |
5966680 | Butnaru | Oct 1999 | A |
5976083 | Richardson et al. | Nov 1999 | A |
6013007 | Root et al. | Jan 2000 | A |
6183425 | Whalen et al. | Feb 2001 | B1 |
6218958 | Eichstaedt et al. | Apr 2001 | B1 |
6241684 | Amano et al. | Jun 2001 | B1 |
6325507 | Jannard et al. | Dec 2001 | B1 |
6381482 | Jayaraman et al. | Apr 2002 | B1 |
6431705 | Linden et al. | Aug 2002 | B1 |
6439067 | Goldman et al. | Aug 2002 | B1 |
6513532 | Mault et al. | Feb 2003 | B2 |
6532298 | Cambier et al. | Mar 2003 | B1 |
6736759 | Stubbs et al. | May 2004 | B1 |
6769767 | Swab et al. | Aug 2004 | B2 |
6783501 | Takahashi et al. | Aug 2004 | B2 |
6790178 | Mault et al. | Sep 2004 | B1 |
6812845 | Yuzuki et al. | Nov 2004 | B2 |
7187960 | Abreu | Mar 2007 | B2 |
7192136 | Howell et al. | Mar 2007 | B2 |
7255437 | Howell et al. | Aug 2007 | B2 |
7376238 | Rivas et al. | May 2008 | B1 |
7380936 | Howell et al. | Jun 2008 | B2 |
7400257 | Rivas | Jul 2008 | B2 |
7401918 | Howell et al. | Jul 2008 | B2 |
7438410 | Howell et al. | Oct 2008 | B1 |
7454002 | Gardner et al. | Nov 2008 | B1 |
7457434 | Azar | Nov 2008 | B2 |
7481531 | Howell et al. | Jan 2009 | B2 |
7488294 | Torch | Feb 2009 | B2 |
7500746 | Howell et al. | Mar 2009 | B1 |
7500747 | Howell et al. | Mar 2009 | B2 |
7515054 | Torch | Apr 2009 | B2 |
7543934 | Howell et al. | Jun 2009 | B2 |
7581833 | Howell et al. | Sep 2009 | B2 |
7621634 | Howell et al. | Nov 2009 | B2 |
7634379 | Noble | Dec 2009 | B2 |
7648463 | Elhag et al. | Jan 2010 | B1 |
7677723 | Howell et al. | Mar 2010 | B2 |
7771046 | Howell et al. | Aug 2010 | B2 |
7792552 | Thomas et al. | Sep 2010 | B2 |
7793361 | Ishihara et al. | Sep 2010 | B2 |
7857772 | Bouvier et al. | Sep 2010 | B2 |
7806525 | Howell et al. | Oct 2010 | B2 |
7922321 | Howell et al. | Apr 2011 | B2 |
7987070 | Kahn et al. | Jul 2011 | B2 |
8007450 | Williams | Aug 2011 | B2 |
8081082 | Malik et al. | Dec 2011 | B2 |
8109629 | Howell et al. | Feb 2012 | B2 |
8157730 | Leboeuf et al. | Apr 2012 | B2 |
8188868 | Case | May 2012 | B2 |
8202148 | Young | Jun 2012 | B2 |
8294581 | Kamen | Oct 2012 | B2 |
8303311 | Forest | Nov 2012 | B2 |
8337013 | Howell et al. | Dec 2012 | B2 |
8384617 | Braun et al. | Feb 2013 | B2 |
8430507 | Howell et al. | Apr 2013 | B2 |
8448846 | Needham et al. | May 2013 | B2 |
8449471 | Tran | May 2013 | B2 |
8465151 | Howell et al. | Jun 2013 | B2 |
8494507 | Tedesco et al. | Jul 2013 | B1 |
8500271 | Howell et al. | Aug 2013 | B2 |
8510166 | Neven | Aug 2013 | B2 |
8540583 | Leech | Sep 2013 | B2 |
8568313 | Sadhu | Oct 2013 | B2 |
8594971 | Keal et al. | Nov 2013 | B2 |
8620600 | Vock et al. | Dec 2013 | B2 |
8630633 | Tedesco et al. | Jan 2014 | B1 |
8634701 | Kang et al. | Jan 2014 | B2 |
8647270 | Leboeuf et al. | Feb 2014 | B2 |
8690750 | Krueger | Apr 2014 | B2 |
8696113 | Lewis | Apr 2014 | B2 |
8733928 | Lewis | May 2014 | B1 |
8750971 | Tran | Jun 2014 | B2 |
8764651 | Tran | Jul 2014 | B2 |
8849610 | Molettiere et al. | Sep 2014 | B2 |
8892401 | Yuen et al. | Nov 2014 | B2 |
8905542 | Howell et al. | Dec 2014 | B2 |
8911087 | Publicover et al. | Dec 2014 | B2 |
8920332 | Hong et al. | Dec 2014 | B2 |
8931896 | Blum et al. | Jan 2015 | B2 |
8941560 | Wong et al. | Jan 2015 | B2 |
8961415 | Leboeuf et al. | Feb 2015 | B2 |
8964298 | Haddick et al. | Feb 2015 | B2 |
8979295 | Waters | Mar 2015 | B2 |
9001427 | Jacobs et al. | Apr 2015 | B2 |
9005129 | Venkatraman et al. | Apr 2015 | B2 |
9007220 | Johns et al. | Apr 2015 | B2 |
9028405 | Tran | May 2015 | B2 |
9031812 | Roberts et al. | May 2015 | B2 |
9033493 | Howell et al. | May 2015 | B2 |
9035970 | Lamb et al. | May 2015 | B2 |
9050033 | Yoneyama et al. | Jun 2015 | B2 |
9064342 | Yuen et al. | Jun 2015 | B2 |
9113794 | Hong et al. | Aug 2015 | B2 |
9113795 | Hong et al. | Aug 2015 | B2 |
9144405 | Kim et al. | Sep 2015 | B2 |
9215290 | Yuen et al. | Dec 2015 | B2 |
9235064 | Lewis | Jan 2016 | B2 |
9239473 | Lewis | Jan 2016 | B2 |
9241635 | Yuen et al. | Jan 2016 | B2 |
9244293 | Lewis | Jan 2016 | B2 |
9254100 | Beck et al. | Feb 2016 | B2 |
9304331 | Carrara | Apr 2016 | B2 |
9520638 | Baringer et al. | Dec 2016 | B2 |
9529197 | Olsson et al. | Dec 2016 | B2 |
9579060 | Lisy et al. | Feb 2017 | B1 |
20010031031 | Ogawa et al. | Oct 2001 | A1 |
20020151810 | Wong et al. | Oct 2002 | A1 |
20030195398 | Suzuki et al. | Oct 2003 | A1 |
20050033200 | Soehren et al. | Feb 2005 | A1 |
20050036103 | Bloch | Feb 2005 | A1 |
20050054942 | Melker et al. | Mar 2005 | A1 |
20060115130 | Kozlay | Jun 2006 | A1 |
20070273611 | Torch | Nov 2007 | A1 |
20080137916 | Lauper et al. | Jun 2008 | A1 |
20090030350 | Yang et al. | Jan 2009 | A1 |
20090195747 | Insua | Aug 2009 | A1 |
20090227853 | Wijesiriwardana | Sep 2009 | A1 |
20090267805 | Jin et al. | Oct 2009 | A1 |
20100042430 | Bartfeld | Feb 2010 | A1 |
20100045928 | Levy | Feb 2010 | A1 |
20100110368 | Chaum | May 2010 | A1 |
20100136508 | Zekhtser | Jun 2010 | A1 |
20100271587 | Pavlopoulos | Oct 2010 | A1 |
20100280336 | Giftakis et al. | Nov 2010 | A1 |
20100308999 | Chornenky | Dec 2010 | A1 |
20100332571 | Healey et al. | Dec 2010 | A1 |
20110169932 | Mula et al. | Jul 2011 | A1 |
20110224505 | Sadhu | Sep 2011 | A1 |
20120021806 | Maltz | Jan 2012 | A1 |
20120029367 | Hobeika | Feb 2012 | A1 |
20120127423 | Blum et al. | May 2012 | A1 |
20120133885 | Howell et al. | May 2012 | A1 |
20120135384 | Nakao | May 2012 | A1 |
20120142443 | Savarese et al. | Jun 2012 | A1 |
20120169990 | Burnstein | Jul 2012 | A1 |
20120191016 | Jastram | Jul 2012 | A1 |
20120203310 | Pugh et al. | Aug 2012 | A1 |
20120206485 | Osterhout et al. | Aug 2012 | A1 |
20120310442 | Doutaz et al. | Dec 2012 | A1 |
20130009907 | Rosenberg et al. | Jan 2013 | A1 |
20130024022 | Bowers | Jan 2013 | A1 |
20130024211 | Monteforte et al. | Jan 2013 | A1 |
20130041590 | Bunch et al. | Feb 2013 | A1 |
20130050258 | Liu et al. | Feb 2013 | A1 |
20130096397 | Kiso et al. | Apr 2013 | A1 |
20130138413 | Finch et al. | May 2013 | A1 |
20130157232 | Ehrenkranz | Jun 2013 | A1 |
20130242262 | Lewis | Sep 2013 | A1 |
20130274587 | Coza et al. | Oct 2013 | A1 |
20130274904 | Coza et al. | Oct 2013 | A1 |
20130307670 | Ramaci | Nov 2013 | A1 |
20130329183 | Blum et al. | Dec 2013 | A1 |
20130345168 | Kim et al. | Dec 2013 | A1 |
20140028456 | Sadhu | Jan 2014 | A1 |
20140031703 | Rayner et al. | Jan 2014 | A1 |
20140063242 | Hanina et al. | Mar 2014 | A1 |
20140073081 | Wang | Mar 2014 | A1 |
20140078049 | Parshionikar | Mar 2014 | A1 |
20140085190 | Erinjippurath et al. | Mar 2014 | A1 |
20140135593 | Jayalth et al. | May 2014 | A1 |
20140142459 | Jayalth et al. | May 2014 | A1 |
20140159862 | Yang et al. | Jun 2014 | A1 |
20140204334 | Stoll | Jul 2014 | A1 |
20140207264 | Quy | Jul 2014 | A1 |
20140218281 | Amayeh et al. | Aug 2014 | A1 |
20140228649 | Rayner et al. | Aug 2014 | A1 |
20140229220 | Yuen et al. | Aug 2014 | A1 |
20140247145 | Proud | Sep 2014 | A1 |
20140266988 | Fisher et al. | Sep 2014 | A1 |
20140276096 | Bonutti | Sep 2014 | A1 |
20140340221 | Yuen et al. | Nov 2014 | A1 |
20140346158 | Matthews | Nov 2014 | A1 |
20140375452 | Yuen et al. | Dec 2014 | A1 |
20140375470 | Malveaux | Dec 2014 | A1 |
20140378872 | Hong et al. | Dec 2014 | A1 |
20150057512 | Kapoor | Feb 2015 | A1 |
20150085245 | Howell et al. | Mar 2015 | A1 |
20150088464 | Yuen et al. | Mar 2015 | A1 |
20150173631 | Richards et al. | Jun 2015 | A1 |
20150179050 | Katingari et al. | Jun 2015 | A1 |
20150185506 | Lewis | Jul 2015 | A1 |
20150212329 | Sugihara | Jul 2015 | A1 |
20150223805 | Whitman et al. | Aug 2015 | A1 |
20150244910 | Marston et al. | Aug 2015 | A1 |
20150281879 | Saadi | Oct 2015 | A1 |
20150287338 | Wells et al. | Oct 2015 | A1 |
20150332149 | Kolb et al. | Nov 2015 | A1 |
20150342482 | Carrara | Dec 2015 | A1 |
20150366518 | Sampson | Dec 2015 | A1 |
20160007849 | Krueger | Jan 2016 | A1 |
20160034042 | Joo | Feb 2016 | A1 |
20160041404 | Palermo et al. | Feb 2016 | A1 |
20160041613 | Klanner et al. | Feb 2016 | A1 |
20160117937 | Penders et al. | Apr 2016 | A1 |
20160314468 | Smith et al. | Oct 2016 | A1 |
20170071528 | Chen | Mar 2017 | A1 |
Number | Date | Country |
---|---|---|
2005015163 | Feb 2005 | WO |
2005094667 | Oct 2005 | WO |
2008073806 | Jun 2008 | WO |
2010006370 | Jan 2010 | WO |
2010062479 | Jun 2010 | WO |
2010062481 | Jun 2010 | WO |
2011086466 | Jul 2011 | WO |
2012041485 | Apr 2012 | WO |
2013188343 | Dec 2013 | WO |
2014108481 | Jul 2014 | WO |
2014144918 | Sep 2014 | WO |
2014144940 | Sep 2014 | WO |
2014170280 | Oct 2014 | WO |
2014188244 | Nov 2014 | WO |
2015015025 | Feb 2015 | WO |
2015081299 | Jun 2015 | WO |
2015095924 | Jul 2015 | WO |
2015127143 | Aug 2015 | WO |
2015127441 | Aug 2015 | WO |
2016017997 | Feb 2016 | WO |
Entry |
---|
Final Office Action, dated Dec. 15, 2016, from corresponding U.S. Appl. No. 14/506,249. |
Final Office Action, dated Mar. 29, 2017, from corresponding U.S. Appl. No. 14/562,454. |
Final Office Action, dated Sep. 26, 2016, from corresponding U.S. Appl. No. 14/610,628. |
International Preliminary Report on Patentability, dated Mar. 16, 2017, from corresponding International Application No. PCT/US2015/048612. |
International Preliminary Report on Patentability, dated Mar. 16, 2017, from corresponding International Application No. PCT/US2015/048656. |
International Preliminary Report on Patentability, dated Mar. 16, 2017, from corresponding International Application No. PCT/US2015/048662. |
International Search Report, dated Dec. 18, 2015, from corresponding International Application No. PCT/US2015/048662. |
International Search Report, dated Jan. 21, 2016, from corresponding International Application No. PCT/US2015/048612. |
International Search Report, dated Jan. 21, 2016, from corresponding International Application No. PCT/US2015/048656. |
International Search Report, dated Jun. 2, 2016, from corresponding International Application No. PCT/US2016/015705. |
Invitation to Pay Additional Search Fees, dated Apr. 1, 2016, from corresponding International Application Serial No. PCT/US2016/015705. |
Invitation to Pay Additional Search Fees, dated Nov. 4, 2015, from corresponding International Application Serial No. PCT/US2015/048612. |
Invitation to Pay Additional Search Fees, dated Nov. 4, 2015, from corresponding International Application Serial No. PCT/US2015/048656. |
Maria S. Redin, “Marathon Man”, Article Jun. 15, 1998, MIT Media Laboratory. |
Michael Franco, Tzoa wearable turns you into a walking air-quality sensor, Dec. 9, 2014, CNET, https://www.cnet.com/news/tzoa-wearable-turns-you-into-a-walking-air-quality-sensor/. |
Notice of Allowance, dated Feb. 28, 2017, from corresponding U.S. Appl. No. 14/588,122. |
Office Action, dated Aug. 19, 2016, from corresponding U.S. Appl. No. 14/578,039. |
Office Action, dated Dec. 29, 2016, from corresponding U.S. Appl. No. 14/610,589. |
Office Action, dated Feb. 10, 2017, from corresponding U.S. Appl. No. 14/846,401. |
Office Action, dated Jul. 1, 2016, from corresponding U.S. Appl. No. 14/562,454. |
Office Action, dated Jul. 22, 2016, from corresponding U.S. Appl. No. 14/506,249. |
Office Action, dated Mar. 3, 2017, from corresponding U.S. Appl. No. 14/610,628. |
Office Action, dated Mar. 8, 2016, from corresponding U.S. Appl. No. 14/610,628. |
Office Action, dated Sep. 2, 2016, from corresponding U.S. Appl. No. 14/588,122. |
Phend, Crystal, “Calorie Intake Rises as Sleep Time Drops,” Medpage Today, Medpage Today, LLC, Mar. 15, 2012, Web Dec. 19, 2016, http://www.medpagetoday/com/cardiology/prevention/31663. |
Restriction Requirement, dated Nov. 10, 2016, from corresponding U.S. Appl. No. 14/846,401. |
Richard M. Satava, et al., “The Physiologic Cipher at Altitude: Telemedicine and Real-Time Monitoring of Climbers on Mount Everest”, Telemedicine Journal and e-Health, vol. 6, No. 3, 2000, Mary Ann Liebert, Inc. |
Ted Burnham, Wearable Air Quality Sensor: Tzoa, Jan. 5, 2015, Postscapes, http://www.postscapes.com/wearable-air-quality-sensor-tzoa/. |
Written Opinion of the International Searching Authority, dated Dec. 18, 2015, from corresponding International Application No. PCT/US2015/048662. |
Written Opinion of the International Searching Authority, dated Jan. 21, 2016, from corresponding International Application No. PCT/US2015/048612. |
Written Opinion of the International Searching Authority, dated Jan. 21, 2016, from corresponding International Application No. PCT/US2015/048656. |
Written Opinion of the International Searching Authority, dated Jun. 2, 2016, from corresponding International Application No. PCT/US2016/015705. |
Final Office Action, dated Jun. 30, 2017, from corresponding U.S. Appl. No. 14/610,589. |
Shankland, Stephen, “Google's electronic eyewear get ‘OK Glass’ voice commands”, Feb. 20, 2013, Cnet.com, https://www.cnet.com/news/googles-electronic-eyewear-gets-ok-glass-voice-commands/. |
Final Office Action, dated Jul. 10, 2017, from corresponding U.S. Appl. No. 14/846,401. |
Final Office Action, dated May 23, 2017, from corresponding U.S. Appl. No. 14/578,039. |
Notice of Allowance, dated Jun. 21, 2017, from corresponding U.S. Appl. No. 14/562,454. |
Office Action, dated Jun. 27, 2017, from corresponding U.S. Appl. No. 15/060,333. |
Tolentino, Mellisa, Udderly Clever Wearable Tech Solutions, http://siliconangle.com/blog/2014/03/25/udderly-clever-wearable-tech-solutions/, Mar. 25, 2014. |
Torres, Juan Carlos, ODG R-7 Smart Glasses Carries Its Own Android Inside, http://androidcommunity.com/odg-r-7-smart-glasses-carries-its-own-android-inside-20140919/, Sep. 19, 2014. |
Office Action, dated Sep. 29, 2017, from corresponding U.S. Appl. No. 14/506,249. |
Restriction Requirement, dated Oct. 4, 2017, from corresponding U.S. Appl. No. 14/610,439. |
Office Action, dated Sep. 26, 2017, from corresponding U.S. Appl. No. 14/846,401. |
Restriction Requirement, dated Sep. 13, 2017, from corresponding U.S. Appl. No. 14/550,406. |
Final Office Action, dated Nov. 16, 2017, from corresponding U.S. Appl. No. 14/610,628. |
Office Action, dated Nov. 30, 2017, from corresponding U.S. Appl. No. 14/550,406. |
Jeannet, Pierre-Yves, et al., “Continuous monitoring and quantification of multiple parameters of daily physical activity in ambulatory Duchenne muscular , dystrophy patients”, Official Journal of the European Paediatric Neurology Society, 2011. |
Notice of Allowance, dated Dec. 13, 2017, from corresponding U.S. Appl. No. 14/610,501. |