The present disclosure relates to the storing and playback of information related to a user's environment, and more particularly to a system and method for storing and playback of information for blind users.
Many individuals are using smart phones in place of having to look up information from a variety of different sources. For example, in the past, many individuals would watch television to obtain weather information, review a calendar to obtain appointment information, study a map to obtain directions to a point of interest and use a camera to take pictures. Now, a single smart phone can be used to obtain all this information.
In addition, smart phones can be used to share this information with one or more users. For example, applications or apps such as Facebook®, Instagram® and Waze® allow users to share directions, news, photos, weather and other information with other users. In particular, the Waze® app allows users to obtain driving directions to a point of interest as well as allows users to upload information about his or her driving experience for others to view.
Unfortunately, for blind users, obtaining and sharing this information is not that simple. For example, a blind user may need to navigate an environment but will find it difficult to use a smart phone for this task. Also, existing apps for smart phones are not well-suited for blind users because they do not provide the features and functions needed for blind users. For example, a blind user generally needs to use different devices to obtain this information than a person that can see. That is, blind users may need other sensory devices to obtain this information.
Thus, there is a need for a system and method for better assisting blind users in navigating environments by allowing this information to be stored and played back for blind users.
A system for providing information to a blind user. The system includes a haptic display including a plurality of sensors positioned thereon, a map positioned on or superimposed onto the haptic display and including points of interest where each point of interest is positioned adjacent to or directly on at least one of the plurality of sensors, and a memory for storing a plurality of breadcrumb files where each of the plurality of breadcrumb files corresponds to at least one of the points of interest and contains information related to at least one of an environment, an object, an obstacle, a path or a location. The system also includes a processor, coupled to the memory, for determining which point of interest has been selected based on contact with at least one of the points of interest and an output device configured to be held or worn by the blind user and configured to provide the information contained in the breadcrumb file corresponding to the selected point of interest to the blind user.
A system for providing information to a blind user. The system includes a haptic device including a plurality of sensors positioned thereon, a map positioned on or superimposed onto the haptic device and including points of interest where each point of interest corresponds to at least one of the plurality of sensors, and a memory for storing a plurality of breadcrumb files where each of the plurality of breadcrumb files corresponds to at least one of the points of interest and contains information related to at least one of an environment, an object, an obstacle, a path or a location. The system also includes a processor, coupled to the memory, for determining which point of interest has been selected based on contact with at least one of the points of interest and a necklace having a speaker and a vibration unit, configured to be worn by the blind user and configured to provide the information contained in the breadcrumb file corresponding to the selected point of interest to the blind user via a combined audio/tactile immersive experience by activating the vibration unit to generate vibrations while simultaneously reproducing audio from the selected breadcrumb file using the speaker.
Other systems, methods, features, and advantages of the present invention will be or will become apparent to one of ordinary skill in the art upon examination of the following figures and detailed description. It is intended that all such additional systems, methods, features, and advantages be included within this description, be within the scope of the present invention, and be protected by the accompanying claims. Component parts shown in the drawings are not necessarily to scale, and may be exaggerated to better illustrate the important features of the present invention. In the drawings, like reference numerals designate like parts throughout the different views, wherein:
The invention provides a user device that allows users, such as a blind person or a person unfamiliar with the surroundings or environment, the ability to save information regarding his or her surroundings or environment for later retrieval and use. The user may access this previously stored information to help him or her navigate the environment during subsequent visits. The information may include information regarding an environment, an object, an obstacle, a path or a location. In addition, third parties can capture or save information regarding an environment, an object, an obstacle, a path or a location and share this information with the user of the user device. For example, a third party may upload this information to a cloud storage for retrieval and use by the user of the user device. Obtaining information from third parties is particularly helpful for blind individuals who generally rely on senses other than sight or other individuals to obtain an understanding of their surroundings or environment. On playback of the information, the processor may create a combined audio/tactile immersive experience for the blind user by activating a vibration unit to generate vibrations while simultaneously reproducing the audio information.
An exemplary system for providing information to a blind user is disclosed. The system includes a haptic display including a plurality of sensors positioned thereon, a map positioned on or superimposed onto the haptic display and including points of interest where each point of interest is positioned adjacent to, directly on or directly in contact with at least one of the plurality of sensors, and a memory for storing a plurality of breadcrumb files where each of the plurality of breadcrumb files corresponds to at least one of the points of interest and contains information related to at least one of an environment, an object, an obstacle, a path or a location. In one embodiment, each point of interest on the map may correspond to at least one of the plurality of sensors.
The term “breadcrumb file” will be used herein to refer to an electronic audio and/or video file that stores the information that has been recorded by a user or a third party. The electronic file may be stored in an uncompressed audio format such as WAV, AIFF, AU, raw header-less PCM, etc. or a compressed audio format such as FLAC, MP3, etc. One or more breadcrumb files may be saved for a particular environment, object, obstacle, path or location. Additional information that is captured and stored in the breadcrumb file may include trajectories (e.g., a path from an entrance of the mall to the bathroom) and annotations of the environment tip (e.g., lots of tables here, be careful). The breadcrumb file may include map data, image data, audio data, video data, text data, location data, vibration data and any other data capable of being stored on an electronic device. The breadcrumb file may be stored in the memory 112, the cloud storage 120 and/or the third party device 150.
The system also includes a processor for determining which point of interest has been selected based on contact with at least one of the points of interest and an output device configured to be held or worn by the blind user and configured to provide the information contained in the breadcrumb file corresponding to the selected point of interest to the blind user.
The user device 110 may include a processor 111, a memory 112, stereo cameras 113, a speaker 114, a vibration unit 115, buttons 116, a global positioning system (GPS) 117, a transceiver 118 and a microphone 119. The user device 110 may also include a battery (not shown) for powering the components of the user device 110. The processor 111 may be a computer processor such as an ARM processor, DSP processor, distributed processor, microprocessor, controller or other processing device. The memory 112 may be one or any combination of the following: a RAM, a ROM or other volatile or nonvolatile memory, a non-transitory memory or a data storage device, such as a hard disk drive, a solid state disk drive, a hybrid disk drive or other appropriate data storage. The memory 112 may further store machine-readable instructions which may be loaded into the memory 112 and executed by the processor 111. As with the processor 111, the memory 112 may be part of or integrated into the wearable or portable smart device 200 of
The stereo cameras 113 may be a stereo camera pair including two cameras offset by a stereo distance. The stereo cameras 113 may be used to take video or capture images of the surrounding environment. The speaker 114 may be one or more output devices or other devices capable of producing sounds or sound waves. The vibration unit 115 may be one or more vibration motors or actuators capable of providing haptic and tactile output. In some implementations, the vibration unit 115 may also be capable of producing sounds, such that the speaker 114 and the vibration unit 115 may be the same or integrated. In some implementations, the speaker 114 and the vibration unit 115 can be activated simultaneously to create a combined audio/tactile immersive experience.
The buttons 116 may be used to turn on and turn off the various components of the user device 110. For example, a first button may be used to turn on and off the stereo cameras 113, a second button may be used to turn on and off the speaker 114, a third button may be used to turn on and off the vibration unit 115 and a fourth button may be used to turn on and off the microphone 119. The buttons 116 may also be configured to operate as inputs to the wearable or portable smart device 200 such that a mode or operation of the wearable or portable smart device 200 can be changed by selecting one or more of the buttons 116. The GPS 117 may be one or more GPS units that can be used to determine the location of the user device 110.
The transceiver 118 can be a receiver and/or transmitter configured to receive and transmit data (e.g., breadcrumb files 140) from the cloud storage 120, the third party device 150 and/or the transceiver 133. The transceiver 118 may include an antenna (not shown) capable of transmitting and receiving wireless communications. For example, the antenna may be a Bluetooth or WiFi antenna, may be a radio frequency identification (RFID) antenna or reader and/or a near field communication (NFC) unit.
The microphone 119 may be a microphone or other device capable of detecting and receiving sounds, such as voice activation/commands or other voice actions from the user, from the surrounding environment and may be part of or integrated into the wearable or portable smart device 200 or external to the wearable or portable smart device 200.
Referring back to
As another example, the user may be visiting the Golden Gate Bridge and may want to save that the Golden Gate Bridge is 1.7 miles in length for later retrieval and use. In addition, third parties can capture or save information regarding an environment, an object, an obstacle, a path or a location and share this information with the user of the user device 110. For example, a third party may upload this information to the cloud storage 120 for retrieval and use by the user of the user device 110. Obtaining information from third parties is particularly helpful for blind individuals who generally rely on senses other than sight or other individuals to obtain an understanding of their surroundings or environment.
The user of the user device 110 may encounter something of interest and may want to save this information for later retrieval and use. Using the user device 110, the user may activate the stereo cameras 113 and/or the microphone 119 to capture information or data related to something of interest such as an environment, an object, an obstacle, a path, a point of interest or a location (step 605). For example, the user may push the button 116 to begin an audio recording. Thereafter, the user can speak into the microphone 119 and state that the Golden Gate Bridge is made of steel. After the audio recording is completed, the user can stop the audio recording by pushing the button 116. The information (e.g., the audio recording) can be saved in the memory 112.
Also, the user may push the button 116 to begin a video recording. Thereafter, the stereo cameras 113 will record information (e.g., a crack in the sidewalk) in the direction of the stereo cameras 113 (e.g., in front of the user) for later retrieval and use. For example, the user may record or tag the crack in the sidewalk that she encounters so that she may be warned of it the next time she comes across it. After the video recording is completed, the user can stop the video recording by pushing the button 116. The information (e.g., the video recording) can be saved in the memory 112 with its corresponding geo coordinates. The information from the user is stored in the memory 112 as a breadcrumb file 140 (step 615).
The information may also be provided to the cloud storage 120 by crowdsourcing or a third party using the third party device 150. For example, the third party can upload information about their retail store such as the store name, store location, hours of operation and current sales. The information from the third party can be uploaded and stored in the cloud storage 120 as a breadcrumb file 140. The breadcrumb file 140 from the cloud storage 120 can be shared by users and can be periodically downloaded to the memory 112 or can be downloaded or accessed upon request by the user or physical location of the user (step 610). For example, as the user physically approaches the Golden Gate Bridge or touches San Francisco on the map 132, the processor 111 may instruct the speaker 114 and/or the vibration unit 115 to reproduce the information contained on the breadcrumb file 140 and corresponding to the Golden Gate Bridge and/or San Francisco. The processor 111 can also automatically deliver and play the breadcrumb file 140 to the user when the GPS 117 determines that the user is physically in an environment that has a corresponding breadcrumb file 140.
The GPS unit 117 may determine or receive the geo coordinates of the location of the recording of the breadcrumb file 140 (step 625). For example, if the user records the information in San Francisco, the geo coordinates can be of San Francisco. Alternatively, the geo coordinates can be of the location of the information contained on the recording. For example, if the user records the information in Los Angeles but the information is about San Francisco, the geo coordinates can be of San Francisco. Based on the title and/or the geo coordinates, the processor 111 can determine the closest landmark or city and populate the location on the map field.
For indoor mapping, the GPS unit 117 may not be able to determine or receive the geo coordinates of the location of the recording of the breadcrumb file 140. In this situation, other localization and mapping algorithms are used by the processor 111 to determine or receive the geo coordinates or other information representing the location of the recording of the breadcrumb file 140. For example, a WiFi indoor triangulation system (ITS), an algorithm such as simultaneous localization and mapping (SLAM) used with the stereo cameras 113 or lasers, or other devices and algorithms can be used to determine geo coordinates or other information to accurately identify the location of walls, rooms, objects, obstacles, or other indoor items.
In one embodiment, the map 132 includes Braille markings describing what is shown on the map 132. For example, the map 132 may include the Braille representation of San Francisco on the map 132 so that when the user runs his or her fingers over the map 132, the user will know what he or she is touching. The map 132 can also have sensors thereon to indicate what the user is touching (step 630). If no corresponding breadcrumb file exists, then the sensors can be non-existent to the blind user's touch indicating to the blind user that no breadcrumb files exists for that particular point of interest.
In one embodiment, the stereo cameras 113 can determine what the user is touching using optical character recognition or recognition or reading of the Braille characters (step 630). The processor 111 can then determine whether a breadcrumb file corresponding to what the user is touching exists and then automatically playing back the breadcrumb file 140 corresponding to what the user is touching (step 635). Alternatively, the processor 111 will notify the user via audio or vibration that a breadcrumb file 140 exists. The user can then push the button 116 to play the breadcrumb file 140 associated with what the user is touching. Hence, in certain embodiments, the haptic device 133 may be optional.
As an example, if the map 132 of
The memory 112 stores a table with data indicating the relationship between the map 132 and the haptic device 133. Hence, if the user were touching San Francisco and/or pin 14, the processor 111 can determine whether a breadcrumb file 140 exists that has information about San Francisco or the geo coordinates corresponding to San Francisco. The information as to what the user is touching is transmitted from the map 132 and/or the haptic device 133 to the user device 110 via transceivers 131 and 118.
The processor 111 may transfer information about the breadcrumb files 140 to the map 132 and/or the haptic device 133. For example, the information may indicate which locations on map 132 or which pins on the haptic device 133 have corresponding breadcrumb files. In one embodiment, only those pins that have corresponding information will be active or cause a bump on the map 132 that can be felt by a user's finger. For example, if no breadcrumb file exists for San Francisco, then pin 14 will be inactive or no bump will appear on the map 132 (e.g., pin is not protruding). Similarly, if a breadcrumb file exists for San Francisco, then pin 14 is active or a bump will appear on the map 132. This advantageously allows the blind user to immediately determine which items or locations on the map 132 have additional information. The pins can also have different shapes to allow a blind user to differential between the different items or locations on the map 132.
On playback of the breadcrumb file 140, the processor 111 may also create an audio, tactile or combined audio/tactile immersive experience for the blind user by activating the vibration unit 115 to generate vibrations while simultaneously reproducing the audio from the breadcrumb file 140 (step 640).
The terminology employed throughout should be read in a non-limiting manner. Although minor modifications to the teachings herein will occur to those well versed in the art, it shall be understood that what is intended to be circumscribed within the scope of the patent warranted hereon are all such embodiments that reasonably fall within the scope of the advancement to the art hereby contributed, and that that scope shall not be restricted, except in light of the appended claims and their equivalents.
Number | Name | Date | Kind |
---|---|---|---|
4520501 | DuBrucq | May 1985 | A |
4586827 | Hirsch et al. | May 1986 | A |
4786966 | Hanson | Nov 1988 | A |
5047952 | Kramer | Sep 1991 | A |
5097856 | Chi-Sheng | Mar 1992 | A |
5129716 | Holakovszky et al. | Jul 1992 | A |
5233520 | Kretsch et al. | Aug 1993 | A |
5265272 | Kurcbart | Nov 1993 | A |
5463428 | Lipton et al. | Oct 1995 | A |
5508699 | Silverman | Apr 1996 | A |
5539665 | Lamming et al. | Jul 1996 | A |
5543802 | Villevieille | Aug 1996 | A |
5544050 | Abe | Aug 1996 | A |
5568127 | Bang | Oct 1996 | A |
5636038 | Lynt | Jun 1997 | A |
5659764 | Sakiyama | Aug 1997 | A |
5701356 | Stanford | Dec 1997 | A |
5733127 | Mecum | Mar 1998 | A |
5807111 | Schrader | Sep 1998 | A |
5872744 | Taylor | Feb 1999 | A |
5953693 | Sakiyama | Sep 1999 | A |
5956630 | Mackey | Sep 1999 | A |
5982286 | Vanmoor | Nov 1999 | A |
6009577 | Day | Jan 2000 | A |
6055048 | Langevin et al. | Apr 2000 | A |
6067112 | Wellner et al. | May 2000 | A |
6199010 | Richton | Mar 2001 | B1 |
6229901 | Mickelson et al. | May 2001 | B1 |
6230135 | Ramsay | May 2001 | B1 |
6230349 | Silver et al. | May 2001 | B1 |
6285757 | Carroll et al. | Sep 2001 | B1 |
6307526 | Mann | Oct 2001 | B1 |
6323807 | Golding et al. | Nov 2001 | B1 |
6349001 | Spitzer | Feb 2002 | B1 |
6466232 | Newell | Oct 2002 | B1 |
6477239 | Ohki | Nov 2002 | B1 |
6542623 | Kahn | Apr 2003 | B1 |
6580999 | Maruyama et al. | Jun 2003 | B2 |
6594370 | Anderson | Jul 2003 | B1 |
6603863 | Nagayoshi | Aug 2003 | B1 |
6619836 | Silvant et al. | Sep 2003 | B1 |
6701296 | Kramer | Mar 2004 | B1 |
6774788 | Balfe | Aug 2004 | B1 |
6825875 | Strub et al. | Nov 2004 | B1 |
6826477 | Ladetto et al. | Nov 2004 | B2 |
6834373 | Dieberger | Dec 2004 | B2 |
6839667 | Reich | Jan 2005 | B2 |
6857775 | Wilson | Feb 2005 | B1 |
6920229 | Boesen | Jul 2005 | B2 |
D513997 | Wilson | Jan 2006 | S |
7027874 | Sawan et al. | Apr 2006 | B1 |
D522300 | Roberts | Jun 2006 | S |
7069215 | Bangalore | Jun 2006 | B1 |
7106220 | Gourgey et al. | Sep 2006 | B2 |
7228275 | Endo | Jun 2007 | B1 |
7299034 | Kates | Nov 2007 | B2 |
7308314 | Havey et al. | Dec 2007 | B2 |
7336226 | Jung et al. | Feb 2008 | B2 |
7356473 | Kates | Apr 2008 | B2 |
7413554 | Kobayashi et al. | Aug 2008 | B2 |
7417592 | Hsiao et al. | Aug 2008 | B1 |
7428429 | Gantz et al. | Sep 2008 | B2 |
7463188 | McBurney | Dec 2008 | B1 |
7496445 | Mohsini | Feb 2009 | B2 |
7501958 | Saltzstein et al. | Mar 2009 | B2 |
7525568 | Raghunath | Apr 2009 | B2 |
7564469 | Cohen | Jul 2009 | B2 |
7565295 | Hernandez-Rebollar | Jul 2009 | B1 |
7598976 | Sofer et al. | Oct 2009 | B2 |
7618260 | Daniel et al. | Nov 2009 | B2 |
D609818 | Tsang et al. | Feb 2010 | S |
7656290 | Fein et al. | Feb 2010 | B2 |
7659915 | Kurzweil et al. | Feb 2010 | B2 |
7743996 | Maciver | Jun 2010 | B2 |
D625427 | Lee | Oct 2010 | S |
7843351 | Bourne | Nov 2010 | B2 |
7843488 | Stapleton | Nov 2010 | B2 |
7848512 | Eldracher | Dec 2010 | B2 |
7864991 | Espenlaub et al. | Jan 2011 | B2 |
7938756 | Rodetsky et al. | May 2011 | B2 |
7991576 | Roumeliotis | Aug 2011 | B2 |
8005263 | Fujimura | Aug 2011 | B2 |
8035519 | Davis | Oct 2011 | B2 |
D649655 | Petersen | Nov 2011 | S |
8123660 | Kruse et al. | Feb 2012 | B2 |
D656480 | McManigal et al. | Mar 2012 | S |
8138907 | Barbeau et al. | Mar 2012 | B2 |
8150107 | Kurzweil et al. | Apr 2012 | B2 |
8177705 | Abolfathi | May 2012 | B2 |
8239032 | Dewhurst | Aug 2012 | B2 |
8253760 | Sako et al. | Aug 2012 | B2 |
8300862 | Newton et al. | Oct 2012 | B2 |
8325263 | Kato et al. | Dec 2012 | B2 |
D674501 | Petersen | Jan 2013 | S |
8359122 | Koselka et al. | Jan 2013 | B2 |
8395968 | Vartanian et al. | Mar 2013 | B2 |
8401785 | Cho et al. | Mar 2013 | B2 |
8414246 | Tobey | Apr 2013 | B2 |
8418705 | Ota et al. | Apr 2013 | B2 |
8428643 | Lin | Apr 2013 | B2 |
8483956 | Zhang | Jul 2013 | B2 |
8494507 | Tedesco et al. | Jul 2013 | B1 |
8494859 | Said | Jul 2013 | B2 |
8538687 | Plocher et al. | Sep 2013 | B2 |
8538688 | Prehofer | Sep 2013 | B2 |
8571860 | Strope | Oct 2013 | B2 |
8583282 | Angle et al. | Nov 2013 | B2 |
8588464 | Albertson et al. | Nov 2013 | B2 |
8588972 | Fung | Nov 2013 | B2 |
8591412 | Kovarik et al. | Nov 2013 | B2 |
8594935 | Cioffi et al. | Nov 2013 | B2 |
8606316 | Evanitsky | Dec 2013 | B2 |
8610879 | Ben-Moshe et al. | Dec 2013 | B2 |
8630633 | Tedesco et al. | Jan 2014 | B1 |
8676274 | Li | Mar 2014 | B2 |
8676623 | Gale et al. | Mar 2014 | B2 |
8694251 | Janardhanan et al. | Apr 2014 | B2 |
8704902 | Naick et al. | Apr 2014 | B2 |
8718672 | Xie et al. | May 2014 | B2 |
8743145 | Price | Jun 2014 | B1 |
8750898 | Haney | Jun 2014 | B2 |
8768071 | Tsuchinaga et al. | Jul 2014 | B2 |
8786680 | Shiratori | Jul 2014 | B2 |
8797141 | Best et al. | Aug 2014 | B2 |
8797386 | Chou et al. | Aug 2014 | B2 |
8803699 | Foshee et al. | Aug 2014 | B2 |
8805929 | Erol et al. | Aug 2014 | B2 |
8812244 | Angelides | Aug 2014 | B2 |
8814019 | Dyster et al. | Aug 2014 | B2 |
8825398 | Alexandre | Sep 2014 | B2 |
8836532 | Fish, Jr. et al. | Sep 2014 | B2 |
8836580 | Mendelson | Sep 2014 | B2 |
8836910 | Cashin et al. | Sep 2014 | B2 |
8902303 | Na'Aman et al. | Dec 2014 | B2 |
8909534 | Heath | Dec 2014 | B1 |
D721673 | Park et al. | Jan 2015 | S |
8926330 | Taghavi | Jan 2015 | B2 |
8930458 | Lewis et al. | Jan 2015 | B2 |
8981682 | Delson et al. | Mar 2015 | B2 |
8994498 | Agrafioti | Mar 2015 | B2 |
D727194 | Wilson | Apr 2015 | S |
9004330 | White | Apr 2015 | B2 |
9025016 | Wexler et al. | May 2015 | B2 |
9042596 | Connor | May 2015 | B2 |
9053094 | Yassa | Jun 2015 | B2 |
9076450 | Sadek | Jul 2015 | B1 |
9081079 | Chao et al. | Jul 2015 | B2 |
9081385 | Ferguson | Jul 2015 | B1 |
D736741 | Katz | Aug 2015 | S |
9111545 | Jadhav et al. | Aug 2015 | B2 |
D738238 | Pede et al. | Sep 2015 | S |
9137484 | DiFrancesco et al. | Sep 2015 | B2 |
9137639 | Garin et al. | Sep 2015 | B2 |
9140554 | Jerauld | Sep 2015 | B2 |
9148191 | Teng et al. | Sep 2015 | B2 |
9158378 | Hirukawa | Oct 2015 | B2 |
D742535 | Wu | Nov 2015 | S |
D743933 | Park et al. | Nov 2015 | S |
9185489 | Gerber et al. | Nov 2015 | B2 |
9190058 | Klein | Nov 2015 | B2 |
9104806 | Stivoric et al. | Dec 2015 | B2 |
9230430 | Civelli et al. | Jan 2016 | B2 |
9232366 | Charlier et al. | Jan 2016 | B1 |
9267801 | Gupta et al. | Feb 2016 | B2 |
9269015 | Boncyk | Feb 2016 | B2 |
9275376 | Barraclough et al. | Mar 2016 | B2 |
9304588 | Aldossary | Apr 2016 | B2 |
D756958 | Lee et al. | May 2016 | S |
D756959 | Lee et al. | May 2016 | S |
9335175 | Zhang et al. | May 2016 | B2 |
9341014 | Oshima et al. | May 2016 | B2 |
9355547 | Stevens et al. | May 2016 | B2 |
9812165 | Ullrich | Nov 2017 | B2 |
20010023387 | Rollo | Sep 2001 | A1 |
20020067282 | Moskowitz et al. | Jun 2002 | A1 |
20020071277 | Starner et al. | Jun 2002 | A1 |
20020075323 | O'Dell | Jun 2002 | A1 |
20020173346 | Wang | Nov 2002 | A1 |
20020178344 | Bourguet | Nov 2002 | A1 |
20030026461 | Hunter | Feb 2003 | A1 |
20030133008 | Stephenson | Jul 2003 | A1 |
20030133085 | Tretiakoff | Jul 2003 | A1 |
20030179133 | Pepin et al. | Sep 2003 | A1 |
20040056907 | Sharma | Mar 2004 | A1 |
20040186713 | Gomas | Sep 2004 | A1 |
20040232179 | Chauhan | Nov 2004 | A1 |
20040267442 | Fehr et al. | Dec 2004 | A1 |
20050020845 | Suzuki et al. | Jan 2005 | A1 |
20050221260 | Kikuchi | Oct 2005 | A1 |
20050259035 | Iwaki | Nov 2005 | A1 |
20050283752 | Fruchter | Dec 2005 | A1 |
20060004512 | Herbst | Jan 2006 | A1 |
20060028550 | Palmer | Feb 2006 | A1 |
20060029256 | Miyoshi | Feb 2006 | A1 |
20060129308 | Kates | Jun 2006 | A1 |
20060171704 | Bingle et al. | Aug 2006 | A1 |
20060177086 | Rye et al. | Aug 2006 | A1 |
20060184318 | Yoshimine | Aug 2006 | A1 |
20060292533 | Selod | Dec 2006 | A1 |
20070001904 | Mendelson | Jan 2007 | A1 |
20070052672 | Ritter et al. | Mar 2007 | A1 |
20070173688 | Kim | Jul 2007 | A1 |
20070182812 | Ritchey | Aug 2007 | A1 |
20070202865 | Moride | Aug 2007 | A1 |
20070230786 | Foss | Oct 2007 | A1 |
20070296572 | Fein | Dec 2007 | A1 |
20080024594 | Ritchey | Jan 2008 | A1 |
20080068559 | Howell | Mar 2008 | A1 |
20080120029 | Zelek et al. | May 2008 | A1 |
20080144854 | Abreu | Jun 2008 | A1 |
20080145822 | Bucchieri | Jun 2008 | A1 |
20080174676 | Squilla et al. | Jul 2008 | A1 |
20080194175 | Last | Aug 2008 | A1 |
20080198222 | Gowda | Aug 2008 | A1 |
20080198324 | Fuziak | Aug 2008 | A1 |
20080208455 | Hartman | Aug 2008 | A1 |
20080251110 | Pede | Oct 2008 | A1 |
20080260210 | Kobeli | Oct 2008 | A1 |
20090012788 | Gilbert | Jan 2009 | A1 |
20090040215 | Afzulpurkar | Feb 2009 | A1 |
20090058611 | Kawamura | Mar 2009 | A1 |
20090106016 | Athsani | Apr 2009 | A1 |
20090118652 | Carlucci | May 2009 | A1 |
20090122161 | Bolkhovitinov | May 2009 | A1 |
20090122648 | Mountain et al. | May 2009 | A1 |
20090157302 | Tashev et al. | Jun 2009 | A1 |
20090177437 | Roumeliotis | Jul 2009 | A1 |
20090189974 | Deering | Jul 2009 | A1 |
20090210596 | Furuya | Aug 2009 | A1 |
20100041378 | Aceves | Feb 2010 | A1 |
20100080418 | Ito | Apr 2010 | A1 |
20100109918 | Liebermann | May 2010 | A1 |
20100110368 | Chaum | May 2010 | A1 |
20100179452 | Srinivasan | Jul 2010 | A1 |
20100182242 | Fields et al. | Jul 2010 | A1 |
20100182450 | Kumar | Jul 2010 | A1 |
20100198494 | Chao | Aug 2010 | A1 |
20100199232 | Mistry et al. | Aug 2010 | A1 |
20100241350 | Cioffi et al. | Sep 2010 | A1 |
20100245585 | Fisher et al. | Sep 2010 | A1 |
20100267276 | Wu | Oct 2010 | A1 |
20100292917 | Emam et al. | Nov 2010 | A1 |
20100298976 | Sugihara et al. | Nov 2010 | A1 |
20100305845 | Alexandre et al. | Dec 2010 | A1 |
20100308999 | Chornenky | Dec 2010 | A1 |
20110066383 | Jangle | Mar 2011 | A1 |
20110071830 | Kim | Mar 2011 | A1 |
20110092249 | Evanitsky | Apr 2011 | A1 |
20110124383 | Garra et al. | May 2011 | A1 |
20110125735 | Petrou | May 2011 | A1 |
20110181422 | Tran | Jul 2011 | A1 |
20110187640 | Jacobsen | Aug 2011 | A1 |
20110211760 | Boncyk | Sep 2011 | A1 |
20110216006 | Litschel | Sep 2011 | A1 |
20110221670 | King, III et al. | Sep 2011 | A1 |
20110234584 | Endo | Sep 2011 | A1 |
20110246064 | Nicholson | Oct 2011 | A1 |
20110260681 | Guccione | Oct 2011 | A1 |
20110307172 | Jadhav et al. | Dec 2011 | A1 |
20120016578 | Coppens | Jan 2012 | A1 |
20120053826 | Slamka | Mar 2012 | A1 |
20120062357 | Slamka | Mar 2012 | A1 |
20120069511 | Azera | Mar 2012 | A1 |
20120075168 | Osterhout et al. | Mar 2012 | A1 |
20120082962 | Schmidt | Apr 2012 | A1 |
20120085377 | Trout | Apr 2012 | A1 |
20120092161 | West | Apr 2012 | A1 |
20120092460 | Mahoney | Apr 2012 | A1 |
20120123784 | Baker et al. | May 2012 | A1 |
20120136666 | Corpier et al. | May 2012 | A1 |
20120143495 | Dantu | Jun 2012 | A1 |
20120162423 | Xiao et al. | Jun 2012 | A1 |
20120194552 | Osterhout et al. | Aug 2012 | A1 |
20120206335 | Osterhout et al. | Aug 2012 | A1 |
20120206607 | Morioka | Aug 2012 | A1 |
20120207356 | Murphy | Aug 2012 | A1 |
20120214418 | Lee | Aug 2012 | A1 |
20120220234 | Abreu | Aug 2012 | A1 |
20120232430 | Boissy et al. | Sep 2012 | A1 |
20120249797 | Haddick et al. | Oct 2012 | A1 |
20120252483 | Farmer et al. | Oct 2012 | A1 |
20120316884 | Rozaieski et al. | Dec 2012 | A1 |
20120323485 | Mutoh | Dec 2012 | A1 |
20120327194 | Shiratori | Dec 2012 | A1 |
20130002452 | Lauren | Jan 2013 | A1 |
20130044005 | Foshee et al. | Feb 2013 | A1 |
20130046541 | Klein et al. | Feb 2013 | A1 |
20130066636 | Singhal | Mar 2013 | A1 |
20130079061 | Jadhav | Mar 2013 | A1 |
20130090133 | D'Jesus Bencci | Apr 2013 | A1 |
20130115578 | Shiina | May 2013 | A1 |
20130115579 | Taghavi | May 2013 | A1 |
20130116559 | Levin | May 2013 | A1 |
20130127980 | Haddick | May 2013 | A1 |
20130128051 | Velipasalar et al. | May 2013 | A1 |
20130131985 | Weiland | May 2013 | A1 |
20130141576 | Lord et al. | Jun 2013 | A1 |
20130144629 | Johnston | Jun 2013 | A1 |
20130155474 | Roach et al. | Jun 2013 | A1 |
20130157230 | Morgan | Jun 2013 | A1 |
20130184982 | DeLuca | Jul 2013 | A1 |
20130201344 | Sweet, III | Aug 2013 | A1 |
20130202274 | Chan | Aug 2013 | A1 |
20130204605 | Illgner-Fehns | Aug 2013 | A1 |
20130211718 | Yoo et al. | Aug 2013 | A1 |
20130218456 | Zelek et al. | Aug 2013 | A1 |
20130228615 | Gates et al. | Sep 2013 | A1 |
20130229669 | Smits | Sep 2013 | A1 |
20130243250 | France | Sep 2013 | A1 |
20130245396 | Berman et al. | Sep 2013 | A1 |
20130250078 | Levy | Sep 2013 | A1 |
20130250233 | Blum et al. | Sep 2013 | A1 |
20130253818 | Sanders et al. | Sep 2013 | A1 |
20130265450 | Barnes, Jr. | Oct 2013 | A1 |
20130271584 | Wexler et al. | Oct 2013 | A1 |
20130290909 | Gray | Oct 2013 | A1 |
20130307842 | Grinberg et al. | Nov 2013 | A1 |
20130311179 | Wagner | Nov 2013 | A1 |
20130328683 | Sitbon et al. | Dec 2013 | A1 |
20130332452 | Jarvis | Dec 2013 | A1 |
20140009561 | Sutherland | Jan 2014 | A1 |
20140031081 | Vossoughi | Jan 2014 | A1 |
20140031977 | Goldenberg et al. | Jan 2014 | A1 |
20140032596 | Fish et al. | Jan 2014 | A1 |
20140037149 | Zetune | Feb 2014 | A1 |
20140055353 | Takahama | Feb 2014 | A1 |
20140071234 | Millett | Mar 2014 | A1 |
20140081631 | Zhu et al. | Mar 2014 | A1 |
20140085446 | Hicks | Mar 2014 | A1 |
20140098018 | Kim et al. | Apr 2014 | A1 |
20140100773 | Cunningham et al. | Apr 2014 | A1 |
20140125700 | Ramachandran | May 2014 | A1 |
20140132388 | Alalawi | May 2014 | A1 |
20140160250 | Pomerantz | Jun 2014 | A1 |
20140184384 | Zhu et al. | Jul 2014 | A1 |
20140184775 | Drake | Jul 2014 | A1 |
20140204245 | Wexler | Jul 2014 | A1 |
20140222023 | Kim et al. | Aug 2014 | A1 |
20140233859 | Cho | Aug 2014 | A1 |
20140236932 | Ikonomov | Aug 2014 | A1 |
20140249847 | Soon-Shiong | Sep 2014 | A1 |
20140251396 | Subhashrao et al. | Sep 2014 | A1 |
20140253702 | Wexler | Sep 2014 | A1 |
20140278070 | McGavran | Sep 2014 | A1 |
20140281943 | Prilepov | Sep 2014 | A1 |
20140287382 | Villar Cloquell | Sep 2014 | A1 |
20140309806 | Ricci | Oct 2014 | A1 |
20140313040 | Wright, Sr. | Oct 2014 | A1 |
20140335893 | Ronen | Nov 2014 | A1 |
20140343846 | Goldman et al. | Nov 2014 | A1 |
20140345956 | Kojina | Nov 2014 | A1 |
20140347265 | Aimone | Nov 2014 | A1 |
20140368412 | Jacobsen | Dec 2014 | A1 |
20140369541 | Miskin | Dec 2014 | A1 |
20140379251 | Tolstedt | Dec 2014 | A1 |
20140379336 | Bhatnager | Dec 2014 | A1 |
20150002808 | Rizzo, III et al. | Jan 2015 | A1 |
20150016035 | Tussy | Jan 2015 | A1 |
20150058237 | Bailey | Feb 2015 | A1 |
20150109107 | Gomez et al. | Apr 2015 | A1 |
20150120186 | Heikes | Apr 2015 | A1 |
20150125831 | Chandrashekhar Nair et al. | May 2015 | A1 |
20150135310 | Lee | May 2015 | A1 |
20150141085 | Nuovo et al. | May 2015 | A1 |
20150142891 | Haque | May 2015 | A1 |
20150154643 | Artman et al. | Jun 2015 | A1 |
20150196101 | Dayal et al. | Jul 2015 | A1 |
20150198454 | Moore et al. | Jul 2015 | A1 |
20150198455 | Chen | Jul 2015 | A1 |
20150199566 | Moore et al. | Jul 2015 | A1 |
20150201181 | Moore et al. | Jul 2015 | A1 |
20150211858 | Jerauld | Jul 2015 | A1 |
20150219757 | Boelter et al. | Aug 2015 | A1 |
20150223355 | Fleck | Aug 2015 | A1 |
20150256977 | Huang | Sep 2015 | A1 |
20150257555 | Wong | Sep 2015 | A1 |
20150260474 | Rublowsky | Sep 2015 | A1 |
20150262509 | Labbe | Sep 2015 | A1 |
20150279172 | Hyde | Oct 2015 | A1 |
20150324646 | Kimia | Nov 2015 | A1 |
20150330787 | Cioffi et al. | Nov 2015 | A1 |
20150336276 | Song | Nov 2015 | A1 |
20150338917 | Steiner et al. | Nov 2015 | A1 |
20150341591 | Kelder et al. | Nov 2015 | A1 |
20150346496 | Haddick et al. | Dec 2015 | A1 |
20150356345 | Velozo | Dec 2015 | A1 |
20150356837 | Pajestka | Dec 2015 | A1 |
20150364943 | Vick | Dec 2015 | A1 |
20150367176 | Bejestan | Dec 2015 | A1 |
20150375395 | Kwon | Dec 2015 | A1 |
20160007158 | Venkatraman | Jan 2016 | A1 |
20160042228 | Opalka | Feb 2016 | A1 |
20160078289 | Michel | Mar 2016 | A1 |
20160098138 | Park | Apr 2016 | A1 |
20160156850 | Werblin et al. | Jun 2016 | A1 |
20160198319 | Huang | Jul 2016 | A1 |
20160350514 | Rajendran | Dec 2016 | A1 |
Number | Date | Country |
---|---|---|
201260746 | Jun 2009 | CN |
101527093 | Sep 2009 | CN |
201440733 | Apr 2010 | CN |
101803988 | Aug 2010 | CN |
101647745 | Jan 2011 | CN |
102316193 | Jan 2012 | CN |
102631280 | Aug 2012 | CN |
202547659 | Nov 2012 | CN |
202722736 | Feb 2013 | CN |
102323819 | Jun 2013 | CN |
103445920 | Dec 2013 | CN |
102011080056 | Jan 2013 | DE |
102012000587 | Jul 2013 | DE |
102012202614 | Aug 2013 | DE |
1174049 | Sep 2004 | EP |
1721237 | Nov 2006 | EP |
2368455 | Sep 2011 | EP |
2371339 | Oct 2011 | EP |
2127033 | Aug 2012 | EP |
2581856 | Apr 2013 | EP |
2751775 | Jul 2016 | EP |
2885251 | Nov 2006 | FR |
2287535 | Sep 1995 | GB |
2401752 | Nov 2004 | GB |
1069539 | Mar 1998 | JP |
2001304908 | Oct 2001 | JP |
201012529 | Jan 2010 | JP |
2010182193 | Aug 2010 | JP |
4727352 | Jul 2011 | JP |
2013169611 | Sep 2013 | JP |
100405636 | Nov 2003 | KR |
20080080688 | Sep 2008 | KR |
20120020212 | Mar 2012 | KR |
1250929 | Apr 2013 | KR |
WO1995004440 | Feb 1995 | WO |
WO 9949656 | Sep 1999 | WO |
WO 0010073 | Feb 2000 | WO |
WO 0038393 | Jun 2000 | WO |
WO 179956 | Oct 2001 | WO |
WO 2004076974 | Sep 2004 | WO |
WO 2006028354 | Mar 2006 | WO |
WO 2006045819 | May 2006 | WO |
WO 2007031782 | Mar 2007 | WO |
WO 2008008791 | Jan 2008 | WO |
WO 2008015375 | Feb 2008 | WO |
WO 2008035993 | Mar 2008 | WO |
WO 2008096134 | Aug 2008 | WO |
WO2008127316 | Oct 2008 | WO |
WO 2010062481 | Jun 2010 | WO |
WO 2010109313 | Sep 2010 | WO |
WO 2012040703 | Mar 2012 | WO |
WO2012163675 | Dec 2012 | WO |
WO 2013045557 | Apr 2013 | WO |
WO 2013054257 | Apr 2013 | WO |
WO 2013067539 | May 2013 | WO |
WO 2013147704 | Oct 2013 | WO |
WO 2014104531 | Jul 2014 | WO |
WO 2014138123 | Sep 2014 | WO |
WO 2014172378 | Oct 2014 | WO |
WO 2015065418 | May 2015 | WO |
WO2015092533 | Jun 2015 | WO |
WO 2015108882 | Jul 2015 | WO |
WO2015127062 | Aug 2015 | WO |
Entry |
---|
Greenberg, Maya Delgado et al.; “Finding Your Way a Curriculum for Teaching and Using the Braillenote with Sendero GPS 2011”; 190 pages; 2011. |
Helal, Abdelsalam et al.; “Drishti: An Integrated Navigation System for Visually Impaired and Disabled”; Fifth International Symposium on Wearable Computers 2001 Proceedings; pp. 149-156; 2001. |
Parkes, Don; “Audio Tactile Systems for Designing and Learning Complex Environments as a Vision Impaired Person: Static and Dynamic Spatial Information Access”; EdTech-94 Proceedings; 8 pages; 1994. |
Zeng, Limin et al.; “Audio-Haptic Browser for a Geographical Information System”; ICCHP 2010, Part II, LNCS 6180; pp. 466-476; 2010. |
AlZuhair, Mona Suliman et al.; “NFC Based Applications for Visually Impaired People—A Review”; Multimedia and Expo Workshops (ICMEW), 2014 IEEE International Conference; 6 pages; Jul. 14, 2014. |
Graf, Christian; “Verbally Annotated Tactile Maps—Challenges and Approaches”; Spatial Cognition VII, LNAI 6222, pp. 308-318; 2010. |
Hamid, Nazatul Naquiah Abd; “Facilitating Route Learning using Interactive Audio-Tactile Maps for Blind and Visually Impaired People”; CHI 2013 Extend Abstracts; pp. 37-42; Apr. 27, 2013. |
Ramya, V. et al.; “Voice Assisted Embedded Navigation System for the Visually Impaired”; International Journal of Computer Applications; pp. 42-48; Feb. 2013. |
The Nex Band; http://www.mightycast.com/#faq; May 19, 2015; 4 pages. |
Cardonha et al.; “A Crowdsourcing Platform for the Construction of Accessibility Maps”; W4A'13 Proceedings of the 10th International Cross-Disciplinary Conference on Web Accessibility; Article No. 26; 2013; 5 pages. |
Bujacz et al.; “Remote Guidance for the Blind—A Proposed Teleassistance System and Navigation Trials”; Conference on Human System Interactions; May 25-27, 2008; 6 pages. |
Rodriguez et al; “CrowdSight: Rapidly Prototyping Intelligent Visual Processing Apps”; AAAI Human Computation Workshop (HCOMP); 2011; 6 pages. |
Chaudary et al.; “Alternative Navigation Assistance Aids for Visually Impaired Blind Persons”; Proceedings of ICEAPVI; Feb. 12-14, 2015; 5 pages. |
Garaj et al.; “A System for Remote Sighted Guidance of Visually Impaired Pedestrians”; The British Journal of Visual Impairment; vol. 21, No. 2, 2003; 9 pages. |
Coughlan et al.; “Crosswatch: A System for Providing Guidance to Visually Impaired Travelers at Traffic Intersections”; Journal of Assistive Technologies 7.2; 2013; 17 pages. |
Sudol et al.; “LookTel—A Comprehensive Platform for Computer-Aided Visual Assistance”; Computer Vision and Pattern Recognition Workshops (CVPRW), 2010 IEEE Computer Society Conference; Jun. 13-18, 2010; 8 pages. |
Kammoun et al.; “Towards a Geographic Information System Facilitating Navigation of Visually Impaired Users”; Springer Berlin Heidelberg; 2012; 8 pages. |
Bigham et al.; “VizWiz: Nearly Real-Time Answers to Visual Questions” Proceedings of the 23nd annual ACM symposium on User interface software and technology; 2010; 2 pages. |
Guy et al; “CrossingGuard: Exploring Information Content in Navigation Aids for Visually Impaired Pedestrians” Proceedings of the SIGCHI Conference on Human Factors in Computing Systems; May 5-10, 2012; 10 pages. |
Zhang et al.; “A Multiple Sensor-Based Shoe-Mounted User Interface Designed for Navigation Systems for the Visually Impaired”; 5th Annual ICST Wireless Internet Conference (WICON); Mar. 1-3, 2010; 9 pages. |
Shoval et al.; “Navbelt and the Guidecane—Robotics-Based Obstacle Avoidance Systems for the Blind and Visually Impaired”; IEEE Robotics & Automation Magazine, vol. 10, Issue 1; Mar. 2003; 12 pages. |
Dowling et al.; “Intelligent Image Processing Constraints for Blind Mobility Facilitated Through Artificial Vision”; 8th Australian and NewZealand Intelligent Information Systems Conference (ANZIIS); Dec. 10-12, 2003; 7 pages. |
Heyes, Tony; “The Sonic Pathfinder an Electronic Travel Aid for the Vision Impaired”; http://members.optuszoo.com.au/aheyew40/pa/pf_blerf.html; Dec. 11, 2014; 7 pages. |
Lee et al.; “Adaptive Power Control of Obstacle Avoidance System Using Via Motion Context for Visually Impaired Person.” International Conference on Cloud Computing and Social Networking (ICCCSN), Apr. 26-27, 2012 4 pages. |
Wilson, Jeff, et al. “Swan: System for Wearable Audio Navigation”; 11th IEEE International Symposium on Wearable Computers; Oct. 11-13, 2007; 8 pages. |
Borenstein et al.; “The GuideCane—A Computerized Travel Aid for the Active Guidance of Blind Pedestrians”; IEEE International Conference on Robotics and Automation; Apr. 21-27, 1997; 6 pages. |
Bhatlawande et al.; “Way-finding Electronic Bracelet for Visually Impaired People”; IEEE Point-of-Care Healthcare Technologies (PHT), Jan. 16-18, 2013; 4 pages. |
Blenkhorn et al.; “An Ultrasonic Mobility Device with Minimal Audio Feedback”; Center on Disabilities Technology and Persons with Disabilities Conference; Nov. 22, 1997; 5 pages. |
Mann et al.; “Blind Navigation with a Wearable Range Camera and Vibrotactile Helmet”; 19th ACM International Conference on Multimedia; Nov. 28, 2011; 4 pages. |
Shoval et al.; “The Navbelt—A Computerized Travel Aid for the Blind”; RESNA Conference, Jun. 12-17, 1993; 6 pages. |
Kumar et al.; “An Electronic Travel Aid for Navigation of Visually Impaired Persons”; Communications Systems and Networks (COMSNETS), 2011 Third International Conference; Jan. 2011; 5 pages. |
Pawar et al.; “Multitasking Stick for Indicating Safe Path to Visually Disable People”; IOSR Journal of Electronics and Communication Engineering (IOSR-JECE), vol. 10, Issue 3, Ver. II; May-Jun. 2015; 5 pages. |
Pagliarini et al.; “Robotic Art for Wearable”; Proceedings of EUROSIAM: European Conference for the Applied Mathematics and Informatics 2010; 10 pages. |
Caperna et al.; “A Navigation and Object Location Device for the Blind”; Tech. rep. University of Maryland College Park; May 2009; 129 pages. |
Burbey et al.; “Human Information Processing with the Personal Memex”; ISE 5604 Fall 2005; Dec. 6, 2005; 88 pages. |
Ghiani, et al.; “Vibrotactile Feedback to Aid Blind Users of Mobile Guides”; Journal of Visual Languages and Computing 20; 2009; 13 pages. |
Guerrero et al.; “An Indoor Navigation System for the Visually Impaired”; Sensors vol. 12, Issue 6; Jun. 13, 2012; 23 pages. |
Nordin et al.; “Indoor Navigation and Localization for Visually Impaired People Using Weighted Topological Map”; Journal of Computer Science vol. 5, Issue 11; 2009; 7 pages. |
Hesch et al.; “Design and Analysis of a Portable Indoor Localization Aid for the Visually Impaired”; International Journal of Robotics Research; vol. 29; Issue 11; Sep. 2010; 15 pgs. |
Joseph et al.; “Visual Semantic Parameterization—To Enhance Blind User Perception for Indoor Navigation”; Multimedia and Expo Workshops (ICMEW), 2013 IEEE International Conference; Jul. 15, 2013; 7 pages. |
Katz et al; “NAVIG: Augmented Reality Guidance System for the Visually Impaired”; Virtual Reality (2012) vol. 16; 2012; 17 pages. |
Rodríguez et al.; “Assisting the Visually Impaired: Obstacle Detection and Warning System by Acoustic Feedback”; Sensors 2012; vol. 12; 21 pages. |
Treuillet; “Outdoor/Indoor Vision-Based Localization for Blind Pedestrian Navigation Assistance”; WSPC/Instruction File; May 23, 2010; 16 pages. |
Ran et al.; “Drishti: An Integrated Indoor/Outdoor Blind Navigation System and Service”; Proceeding PERCOM '04 Proceedings of the Second IEEE International Conference on Pervasive Computing and Communications (PerCom'04); 2004; 9 pages. |
Wang, et al.; “Camera-Based Signage Detection and Recognition for Blind Persons”; 13th International Conference (ICCHP) Part 2 Proceedings; Jul. 11-13, 2012; 9 pages. |
Krishna et al.; “A Systematic Requirements Analysis and Development of an Assistive Device to Enhance the Social Interaction of People Who are Blind or Visually Impaired”; Workshop on Computer Vision Applications for the Visually Impaired; Marseille, France; 2008; 12 pages. |
Lee et al.; “A Walking Guidance System for the Visually Impaired”; International Journal of Pattern Recognition and Artificial Intelligence; vol. 22; No. 6; 2008; 16 pages. |
Ward et al.; “Visual Experiences in the Blind Induced by an Auditory Sensory Substitution Device”; Journal of Consciousness and Cognition; Oct. 2009; 30 pages. |
Merino-Garcia, et al.; “A Head-Mounted Device for Recognizing Text in Natural Sciences”; CBDAR'11 Proceedings of the 4th International Conference on Camera-Based Document Analysis and Recognition; Sep. 22, 2011; 7 pages. |
Yi, Chucai; “Assistive Text Reading from Complex Background for Blind Persons”; CBDAR' 11 Proceedings of the 4th International Conference on Camera-Based Document Analysis and Recognition; Sep. 22, 2011; 7 pages. |
Yang, et al.; “Towards Automatic Sign Translation”; The Interactive Systems Lab, Carnegie Mellon University; 2001; 5 pages. |
Meijer, Dr. Peter B.L.; “Mobile OCR, Face and Object Recognition for the Blind”; The vOICe, www.seeingwithsound.com/ocr.htm; Apr. 18, 2014; 7 pages. |
OMRON; Optical Character Recognition Sensor User's Manual; 2012; 450 pages. |
Park, Sungwoo; “Voice Stick”; www.yankodesign.com/2008/08/21/voice-stick; Aug. 21, 2008; 4 pages. |
Rentschler et al.; “Intelligent Walkers for the Elderly: Performance and Safety Testing of VA-PAMAID Robotic Walker”; Department of Veterans Affairs Journal of Rehabilitation Research and Development; vol. 40, No. 5; Sep./Oct. 2013; 9pages. |
Science Daily; “Intelligent Walker Designed to Assist the Elderly and People Undergoing Medical Rehabilitation”; http://www.sciencedaily.com/releases/2008/11/081107072015.htm; Jul. 22, 2014; 4 pages. |
Glover et al.; “A Robotically-Augmented Walker for Older Adults”; Carnegie Mellon University, School of Computer Science; Aug. 1, 2003; 13 pages. |
OrCam; www.orcam.com; Jul. 22, 2014; 3 pages. |
Eccles, Lisa; “Smart Walker Detects Obstacles”; Electronic Design; http://electronicdesign.com/electromechanical/smart-walker-detects-obstacles; Aug. 20, 2001; 2 pages. |
Graft, Birgit; “An Adaptive Guidance System for Robotic Walking Aids”; Journal of Computing and Information Technology—CIT 17; 2009; 12 pages. |
Frizera et al.; “The Smart Walkers as Geriatric Assistive Device. The SIMBIOSIS Purpose”; Gerontechnology, vol. 7, No. 2; Jan. 30, 2008; 6 pages. |
Rodriquez-Losada et al.; “Guido, The Robotic Smart Walker for the Frail Visually Impaired”; IEEE International Conference on Robotics and Automation (ICRA); Apr. 18-22, 2005; 15 pages. |
Kayama et al.; “Outdoor Environment Recognition and Semi Autonomous Mobile Vehicle for Supporting Mobility of the Elderly and Disabled People”; National Institute of Information and Communications Technology, vol. 54, No. 3; Aug. 2007; 11 pages. |
Kalra et al.; “A Braille Writing Tutor to Combat Illiteracy in Developing Communities”; Carnegie Mellon University Research Showcase, Robotics Institute; 2007; 10 pages. |
Blaze Engineering; “Visually Impaired Resource Guide: Assistive Technology for Students who use Braille”; Braille 'n Speak Manual; http://www.blaize.com; Nov. 17, 2014; 5 pages. |
AppleVis; An Introduction to Braille Screen Input on iOS 8; http://www.applevis.com/guides/braille-ios/introduction-braille-screen-input-ios-8, Nov. 16, 2014; 7 pages. |
Dias et al.; “Enhancing an Automated Braille Writing Tutor”; IEEE/RSJ International Conference on Intelligent Robots and Systems; Oct. 11-15, 2009; 7 pages. |
D'Andrea, Frances Mary; “More than a Perkins Brailler: A Review of the Mountbatten Brailler, Part 1”; AFB AccessWorld Magazine; vol. 6, No. 1, Jan. 2005; 9 pages. |
Trinh et al.; “Phoneme-based Predictive Text Entry Interface”; Proceedings of the 16th International ACM SIGACCESS Conference on Computers & Accessibility; Oct. 2014; 2 pgs. |
Merri et al.; “The Instruments for a Blind Teacher of English: The challenge of the board”; European Journal of Psychology of Education, vol. 20, No. 4 (Dec. 2005), 15 pages. |
Kirinic et al.; “Computers in Education of Children with Intellectual and Related Developmental Disorders”; International Journal of Emerging Technologies in Learning, vol. 5, 2010, 5 pages. |
Campos et al.; “Design and Evaluation of a Spoken-Feedback Keyboard”; Department of Information Systems and Computer Science, INESC-ID/IST/Universidade Tecnica de Lisboa, Jul. 2004; 6 pages. |
Wu et al. “Fusing Multi-Modal Features for Gesture Recognition”, Proceedings of the 15th ACM on International Conference on Multimodal Interaction, Dec. 9, 2013, ACM, pp. 453-459. |
Pitsikalis et al. “Multimodal Gesture Recognition via Multiple Hypotheses Rescoring”, Journal of Machine Learning Research, Feb. 2015, pp. 255-284. |
Shen et al. “Walkie-Markie: Indoor Pathway Mapping Made Easy” 10th USENIX Symposium on Networked Systems Design and Implementation (NSDI'13); pp. 85-98, 2013. |
Tu et al. “Crowdsourced Routing II D2.6” 34 pages; 2012. |
De Choudhury et al. “Automatic Construction of Travel Itineraries Using Social Breadcrumbs” pp. 35-44; Jun. 2010. |
Bharathi et al.; “Effective Navigation for Visually Impaired by Wearable Obstacle Avoidance System;” 2012 International Conference on Computing, Electronics and Electrical Technologies (ICCEET); pp. 956-958; 2012. |
Pawar et al.; “Review Paper on Multitasking Stick for Guiding Safe Path for Visually Disable People;” IJPRET; vol. 3, No. 9; pp. 929-936; 2015. |
Ram et al.; “The People Sensor: A Mobility Aid for the Visually Impaired;” 2012 16th International Symposium on Wearable Computers; pp. 166-167; 2012. |
Singhal; “The Development of an Intelligent Aid for Blind and Old People;” Emerging Trends and Applications in Computer Science (ICETACS), 2013 1st International Conference; pp. 182-185; Sep. 13, 2013. |
Aggarwal et al.; “All-in-One Companion for Visually Impaired;” International Journal of Computer Applications; vol. 79, No. 14; pp. 37-40; Oct. 2013. |
“Light Detector” EveryWare Technologies; 2 pages; Jun. 18, 2016. |
Arati et al. “Object Recognition in Mobile Phone Application for Visually Impaired Users;” IOSR Journal of Computer Engineering (IOSR-JCE); vol. 17, No. 1; pp. 30-33; Jan. 2015. |
Yabu et al.; “Development of a Wearable Haptic Tactile Interface as an Aid for the Hearing and/or Visually Impaired;” NTUT Education of Disabilities; vol. 13; pp. 5-12; 2015. |
Mau et al.; “BlindAid: An Electronic Travel Aid for the Blind;” The Robotics Institute Carnegie Mellon University; 27 pages; May 2008. |
Shidujaman et al.; “Design and navigation Prospective for Wireless Power Transmission Robot;” IEEE; Jun. 2015. |
Zhang, Shanjun; Yoshino, Kazuyoshi; A Braille Recognition System by the Mobile Phone with Embedded Camera; 2007; IEEE. |
Diallo, Amadou; Sep. 18, 2014; Apple iOS8: Top New Features, Forbes Magazine. |
N. Kalar, T. Lawers, D. Dewey, T. Stepleton, M.B. Dias; Iterative Design of a Braille Writing Tutor to Combat Illiteracy; Aug. 30, 2007; IEEE. |
Number | Date | Country | |
---|---|---|---|
20160275817 A1 | Sep 2016 | US |