One-size-fits-all data glove

Information

  • Patent Grant
  • 10884496
  • Patent Number
    10,884,496
  • Date Filed
    Friday, June 21, 2019
    5 years ago
  • Date Issued
    Tuesday, January 5, 2021
    4 years ago
Abstract
Assemblies and techniques are described herein for use with a data glove. The data glove includes sensors configured to translate movement and forces associated with a human hand to the digital domain.
Description
INCORPORATION BY REFERENCE

An Application Data Sheet is filed concurrently with this specification as part of this application. Each application to which this application claims benefit or priority as identified in the concurrently filed Application Data Sheet is incorporated by reference herein in its entirety and for all purposes.


BACKGROUND

Demand is rapidly rising for technologies that bridge the gap between computing devices and the physical world. Such interfaces typically require some form of sensor technology that translates information from the physical domain to the digital domain. One type of interface provides for the translation of the movements of a human operator, as well as forces exerted or experienced by the human operator, to digital information.


SUMMARY

According to a particular class of implementations, an apparatus includes a sensor assembly including a plurality of digit sensor assemblies. Each digit sensor assembly including one or more sensors on an elongated substrate. The one or more sensors of each digit sensor assembly is configured to generate one or more signals representing bending of the corresponding substrate. A textile assembly includes a plurality of digit textile assemblies. Each digit textile assembly is configured to align with a corresponding finger of a hand and includes a fingertip portion for securing the digit textile assembly to a fingertip of the corresponding finger. Each digit textile assembly also includes a sleeve in which a corresponding one of the digit sensor assemblies is contained. The sleeve is connected at a first end of the sleeve to the fingertip portion of the digit textile assembly and is configured to align the corresponding digit sensor assembly with a back side of the corresponding finger. The textile assembly is configured such that a middle portion of each finger is exposed. A base assembly is connected to each of the sleeves of the digit textile assemblies at a second end of the sleeve opposing the first end. The base assembly is configured to secure the apparatus to the hand. Sensor circuitry configured to receive the signals from the digit sensor assemblies and to generate digital representations of the signals.


According to a specific implementation, each digit sensor assembly is secured in the corresponding sleeve such that the digit sensor assembly slides relative to the sleeve when the digit sensor assembly bends.


According to another specific implementation, each sleeve comprises an elastic fabric.


According to another specific implementation, the apparatus includes a plurality of haptic devices. Each haptic device is integrated with the fingertip portion of a corresponding one of the digit textile assemblies. According to a more specific implementation, each sleeve includes a first chamber in which the corresponding digit sensor assembly is contained, and a second chamber in which one or more conductors connected to the haptic device of the corresponding fingertip portion are contained.


According to another specific implementation, each of the sleeves includes friction material on an outside surface of the sleeve configured for contacting the back side of the corresponding finger.


According to another specific implementation, each digit sensor assembly includes at least two sensors. Each of the sensors is configured to align with a corresponding knuckle of the corresponding finger.


According to another specific implementation, each digit sensor assembly includes one or more stiffeners, Each stiffener is aligned with a corresponding one of the one or more sensors of the digit sensor assembly and is configured to support a particular dynamic range of the corresponding sensor.


According to another specific implementation, each sleeve includes one or more stiffeners. Each stiffener is aligned with a corresponding one of the one or more sensors of the digit sensor assembly contained in the sleeve and is configured to support a particular dynamic range of the corresponding sensor.


According to another specific implementation, each of the one or more sensors of each digit sensor assembly includes either a piezoresistive material or a piezoelectric material.


According to another specific implementation, the base assembly is configured to secure the apparatus to the hand with a strap. The strap has a haptic device integrated therewith such that, when the apparatus is secured to the hand with the strap, the haptic device is aligned with a palm of the hand.


According to another class of implementations, a glove includes a glove body configured to be secured to a human hand and a plurality of finger components extending from the glove body. Each finger component is configured to enclose at least a portion of a corresponding finger of the human hand. Each finger component has a haptic device integrated therewith. The glove also includes circuitry configured to control the haptic devices. The circuitry includes one or more class D amplifiers configured to drive the haptic devices, memory configured to store one or more waveform files, and a controller configured to control the one or more class D amplifiers using the one or more waveform files.


According to a specific implementation, the one or more waveform files are a plurality of waveform files stored in the memory, and the controller is configured to use each of the waveform files with each of the haptic devices.


According to another specific implementation, the controller is configured to control the one or more class D amplifiers using the one or more waveform files according to a control protocol. The control protocol corresponds to one of the Musical Instrument Digital Interface (MIDI) protocol, the MIDI Polyphonic Expression (MPE) protocol, or the Open Sound Control (OSC) protocol. According to a more specific implementation, each of the haptic devices corresponds to a different channel in the control protocol.


According to another specific implementation, the controller is configured to control the one or more class D amplifiers by one or more of sampling, scrubbing, or playing back the one or more waveform files.


According to another specific implementation, the glove also includes a wireless interface configured to stream the one or more waveform files from a remote device in real time.


According to another specific implementation, each of the one or more waveform files is configured for simulation of a corresponding sensory effect.


According to another specific implementation, control of the one or more class D amplifiers by the controller is characterized by an amplitude and a frequency, and the controller is configured to modify the amplitude and the frequency in real time.


According to another specific implementation, each of the one or more waveform files is characterized by a timeline, and the controller is configured to control the one or more class D amplifiers by traversing each of the one or more waveform files backward and forward relative to the corresponding timeline.


According to another specific implementation, control of the one or more class D amplifiers by the controller is characterized by spectral brightness, and the controller is configured to modify the spectral brightness in real time.


According to another specific implementation, control of the one or more class D amplifiers by the controller is characterized by one or more signal parameters, and the controller is configured to modify the one or more signal parameters in real time in response to input representing objects or surfaces in a virtual or real environment.


According to another specific implementation, the controller is configured to control the one or more class D amplifiers using a first waveform file to drive a first haptic device and a second haptic device. The driving of the first haptic device is temporally offset from the driving of the second haptic device to simulate a spatial relationship.


According to another specific implementation, the controller is configured to control the one or more class D amplifiers using the one or more waveforms files in a sequence that represents a haptic gesture.


A further understanding of the nature and advantages of various implementations may be realized by reference to the remaining portions of the specification and the drawings.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1A is a perspective view of an example of sensor assembly for use with implementations enabled by the present disclosure.



FIG. 1B is a simplified block diagram of sensor circuitry suitable for use with various implementations enabled by the present disclosure.



FIGS. 2 and 3 are two different perspective views of an apparatus for use with a human hand as enabled by the present disclosure.



FIG. 4 is a cross-sectional view of sleeve for enclosing components associated with an apparatus enabled by the present disclosure.



FIG. 5 is an exploded view of a haptic device for use with apparatus enabled by the present disclosure.



FIG. 6 includes various views of an assembly incorporating a haptic device enabled by the present disclosure.



FIG. 7 is a schematic diagram of circuitry for use with a haptic device enabled by the present disclosure.



FIG. 8 is another perspective view of an apparatus for use with a human hand as enabled by the present disclosure.



FIG. 9 includes various views of a portion of an apparatus for use with a human hand as enabled by the present disclosure.





DESCRIPTION OF SPECIFIC IMPLEMENTATIONS

Reference will now be made in detail to specific implementations. Examples of these implementations are illustrated in the accompanying drawings. It should be noted that these examples are described for illustrative purposes and are not intended to limit the scope of this disclosure. Rather, alternatives, modifications, and equivalents of the described implementations are included within the scope of this disclosure. In addition, specific details may be provided in order to promote a thorough understanding of the described implementations. Some implementations within the scope of this disclosure may be practiced without some or all of these details. Further, well known features may not have been described in detail for the sake of clarity.


U.S. Patent Publication No. 2017/0303853 (U.S. patent application Ser. No. 15/621,935) entitled Sensor System Integrated With a Glove (the entirety of which is incorporated herein by reference for all purposes) describes various implementations of sensor assemblies that are compatible with implementations of a “one-size-fits-all” data glove enabled by the present disclosure. See, for example, the implementation described with reference to FIGS. 12-14C of that publication. A particular implementation of a sensor assembly 100 is depicted in FIG. 1A of the present application. The depicted sensor assembly may be implemented similarly to the implementation depicted in FIGS. 12-14C of the publication incorporated by reference above with the exception that the pads and conductors for haptic devices are not included on the substrate with the knuckle bend sensors. As discussed below, the pads and conductors for the haptics, if included, may be routed separately.


As shown in FIG. 1A, each knuckle bend sensor includes a strip of piezoresistive fabric (e.g., 102) in contact with a pair of conductive traces (e.g., 104 and 106) on a substrate 108. Substrate 108 may be, for example, a polyethylene terephthalate (PET) substrate. The conductive traces are routed along substrate 108 and are connected at edge 110 to a printed circuit board or PCB (not shown) on which the sensor circuitry of the glove is located. It should be noted that sensor assembly 100 is merely one example of a sensor assembly that may be used in connection with a data glove enabled by the present disclosure, and that other types of force sensing technology including a wide range of other sensor types (e.g., other piezoresistive materials, piezoelectric materials, etc.) may be used without departing from the scope of the present disclosure.



FIG. 1B is a simplified diagram of sensor circuitry that may be provided on a PCB for use with implementations described herein. For example, in the sensor assembly described above with reference to FIG. 1A, such sensor circuitry could be connected to the conductive traces at edge 110. When pressure is applied to one of the sensors, a resulting signal (captured via the corresponding traces) is received and digitized (e.g., via multiplexer 152 and A-to-D converter 154) and may be processed locally (e.g., by processor 156) and/or transmitted to a connected device (e.g., via a USB or Bluetooth connection). The sensors may be selectively energized by the sensor circuitry (e.g., under the control of processor 156 via D-to-A converter 158 and multiplexer 160) to effect the generation of the sensor signals.


In addition to transmission of data to and from a connected device, power may be provided to the sensor circuitry via a USB connection. Alternatively, systems that transmit data wirelessly (e.g., via Bluetooth) may provide power to the sensor circuitry using any of a variety of mechanisms and techniques including, for example, using one or more batteries, solar cells, and/or mechanisms that harvest mechanical energy. The LTC3588 (provided by Linear Technology Corporation of Milpitas, Calif.) is an example of an energy harvesting power supply that may be used with at least some of these diverse energy sources. Other suitable variations will be appreciated by those of skill in the art. And as will be appreciated, the sensor circuitry shown in FIG. 1B is merely an example. A wide range of sensor circuitry components, configurations, and functionalities are contemplated. An example of a device suitable for implementing processor 156 is the C8051F380-GM controller provided by Silicon Labs of Austin, Tex.


As will be understood, the responses of the sensors in arrays suitable for use with implementations enabled by the present disclosure may exhibit variation relative to each other. Therefore, calibrated sensor data may be stored (e.g., in memory 157 of processor 156) representing the response of each of the sensors. Such data may be used for ensuring consistency in the way the sensor outputs are processed and/or used to represent applied forces. During calibration, the output of each sensor (e.g., as captured by ADC 154) is measured for a range of known input forces. This may be done, for example, by placing each sensor on a scale, applying force to that sensor, and recording a value in memory for each of a plurality of ADC values that represents a corresponding value reported by the scale. In this way, a set of data points for each sensor is captured (e.g., in a table in memory 157) associating ADC values with corresponding forces (e.g., weights in grams or kilograms). The data set for each sensor may capture a force value for every possible value of the ADC output. Alternatively, fewer data points may be captured and the sensor circuitry may use interpolation to derive force values for ADC outputs not represented in the data set.


Glove Textile Design


As shown in the perspective views of FIGS. 2 and 3, a data glove 200 is designed in a way that omits much of the material of a conventional full-fingered glove from around the finger. The design includes fingertip covers 202 connected to the main body 204 of the glove by sleeves 206 that ride along the back of each finger in which the portion of the sensor assembly (e.g., sensor assembly 100 of FIG. 1A) for that finger (also referred to as a digit assembly) is situated.


Each sleeve is constructed using a stretchable material that allows for bending of the finger. In addition, the longitudinal stretching of each sleeve enables the ability to fit a wide range of finger lengths, giving data glove 200 its “one-size-fits-all” character. According to a particular implementation, the primary fabric of most of the textile portions of the glove is a neoprene-type material and the finger sleeves are constructed using a type of spandex that allows for adequate stretch. Providing the sleeve only along one side of the finger also allows the glove to accommodate a wide range of finger widths.


In some implementations, one or more stiffeners (not shown) are integrated with the sleeves and/or the sensor assemblies to enhance knuckle sensor operation by achieving a desired dynamic range for each sensor. These stiffeners may be implemented in a variety of ways. For example, the thickness of the PET substrate of the digit assemblies may be selectively controlled to achieve the desired dynamic range. Alternatively, additional material may be introduced that is aligned with and/or in the vicinity of each sensor to achieve the desired dynamic range. This might take the form of one or more pieces of such stiffening material aligned with each sensor that is integrated with the digit assembly (on either side) or even the sleeve material. For example, stiffening material (e.g., PET strips) may be coupled to the digit assemblies in alignment with each of the knuckle sensors. According to one such implementation, PET strips are affixed to the piezoresistive fabric strips (e.g., strips 102 in FIG. 1A) on the other side of the piezo strips from the side that is in contact with the corresponding pair of conductive traces (e.g., 104 and 106). Alternatively, the stiffening material may be another piece of fabric aligned with each sensor and attached to the sensor substrate or the glove fabric. As yet another alternative, a stiffening material such as, for example, a dielectric ink may be silk-screened or printed on the flexible substrate, on the piezoresistive strips, or the glove fabric in the vicinity of each sensor.


According to a particular implementation illustrated by the cross-sectional view of FIG. 4, sleeve 206 includes one chamber for that finger's portion of the digit assembly (e.g., the PET sensor print) and another chamber for the routing of wires (e.g., haptic wires) to haptic devices at the fingertips. Each digit assembly slides relatively freely in its sleeve. This may be important for some applications in that, when the finger is bent and then straightened, if the exterior fabric is slow to release, the pressure registered on the sensor may lag the finger movement in an unacceptable way. The second chamber allows for routing of the haptic wiring in a way that does not interfere with the freedom of movement of the digit assembly relative to its chamber.


According to some implementations, silicone rails (not shown) may be provided on the underside of the sleeve (adjacent the finger) to help keep the sleeve and its digit assembly aligned with the underlying finger. The stiffness of the PET in the digit assemblies may also be adjusted to ensure the desired level of flexibility as well as durability, i.e., so they can handle many bends without fatiguing.


As shown in FIG. 3, grip-enhancing features 208 may be provided at the fingertips. Such features (which may be, for example, friction TPU or screened silicone) are shown in this example to be circular and concentric but a wide variety of shapes and patterns are contemplated.


According to some implementations, leaving out the material that typically surrounds each finger in a conventional glove design provides mechanical isolation of each finger, thereby reducing cross-talk between the sensors of the different fingers. That is, movement of a finger in a conventional glove stretches or distorts the fabric associated with adjacent and even more remote fingers causing any sensors associated with those other fingers to register force. Reducing the fabric as described herein isolates each digit and correspondingly reduces the transmission of forces between fingers.


It should be noted that, while the design described herein is not shown as including abduction sensors between fingers (i.e., sensors that generate signals representing the fingers being spread apart), designs are contemplated in which such sensors are included. It should also be noted that the term “one-size-fits-all” contemplates the fact that, while most human hands can be accommodated by the glove designs described herein, there are outliers (both large and small) that may not be accommodated.


Some knuckle sensor implementations may not be able to distinguish between the bending of a knuckle and a hyperextension of that knuckle, i.e., the bend sensor signal may not indicate the direction of the bend. Providing each digit assembly in a sleeve aligned with the back of the finger may address this issue at least in part because, when the finger is hyperextended, the sleeve and the digit assembly sliding inside that sleeve lift off the finger with the result that the digit assembly remains substantially flat. Therefore, when a sufficiently strong bend signal is actually registered by one or more of the sensors of a digit assembly, there can be a high degree of confidence that the signal represents a bend of the finger rather than a hyperextension.


Haptic Devices


Haptic devices may be provided (e.g., at each of the fingertips and the palm) to simulate various sensory experiences. As shown in the exploded view of FIG. 5, haptic device 500 may be an electromagnetic haptic device that includes a reaction mass 502 sitting on top of a coil-magnet assembly 504. The device includes a plastic cover 506 that, with assembly 504, forms a cavity in which reaction mass 502 vibrates. According to a particular implementation, the reaction mass is a copper plate on a Mylar diaphragm in an enclosure that allows it freedom of movement. When the device is driven, the mass moves up and down causing vibration that is sensed by the fingertip or palm.


Some implementations include a palm haptic that fits in the palm of the glove. As shown in FIG. 6, the palm haptic 602 is mounted in a molded plastic component 604 that conforms to the palm and is included (as indicated by the dashed line) in the strap 210 that secures the main body of the glove to the hand.


According to a particular implementation depicted in FIG. 7, each haptic 702 is driven by a simple class D amplifier 704 using any of a library of waveform files 706 stored in the onboard sensor circuitry and shared among the haptic devices. The depicted example represents a pulse width modulation DAC that drives a FET; a simple and inexpensive solution. In some cases, pre-distortion of the waveform files may be introduced to cancel at least a portion of the distortion caused by the amplifier circuit.


According to some implementations, the glove behaves like a synthesizer, sampler, and/or playback device using a corresponding control protocol, e.g. Musical Instrument Digital Interface (MIDI), MIDI Polyphonic Expression (MPE), Open Sound Control (OSC), or a custom specification. Such a system is referred to herein as a Haptic Engine. A Haptic Engine may be configured and used in a variety of ways. Waveform files can be uploaded into the engine and sampled, scrubbed, or played back. Oscillators and synthesis techniques may also be used. Waveforms signals may also be streamed from a host device connected to the glove. Such modes of operation may be adjustable in real-time via the control protocol.


A suitable API or control protocol may operate like a MIDI synthesizer sampler. Multiple different “notes” or commands may be provided per finger to simulate different actions, e.g., a short signal to simulate tapping of a fingertip, or a sustained signal to simulate dragging of a fingertip across a surface. For each fingertip haptic, the sensor circuitry generates a “note on” signal with amplitude and frequency attributes depending on what is being simulated. In this mode of operation, amplitude and frequency can be adjusted in real time via the control protocol to add continuous variations. New libraries of waveform files can be uploaded to the glove (e.g., via Bluetooth or USB), e.g., for use with a new game.


Different waveform synthesis techniques may be mapped to different sets of activities. For example, a waveform file can be scrubbed or traversed in both directions to simulate changing direction when dragging a fingertip across a virtual surface. In another example, as a fingertip is getting close to something in the virtual space, the vibration could get brighter spectrally. The output generated for a given file can be modified (e.g., frequency, amplitude, etc.) depending on input representing characteristics of the objects or surfaces in the virtual environment with which the user is interacting. For example, if the user immerses his virtual hand in virtual water, that information might be used to modify the subsequent playback of waveform files or synthesis to account for the “wetness” of the fingertips.


Each of the haptic devices (fingertips and palm) may be treated as a unique channel in which a different “track” is played for each haptic, i.e., multi-channel playback and control of the array of haptic devices. Playback of the tracks may also be spatially related. For example, for a given texture, there might only be one waveform for sustained contact that is played back for each fingertip, but playback for the respective fingertips may be spatially offset so that if, for example, there is a localized irregularity on the virtual surface, one fingertip might “experience” the irregularity before another as the hand moves across the surface.


Additionally, the waveforms produced by the Haptic Engine may be used to trigger and control “haptic gestures” consisting of waveforms or sequences of waveforms across the array of haptic devices, analogous to audio sound effects generated by a computer operating system upon certain events. These haptic gestures or signifiers, for example, may be associated with certain events in gameplay or other applications, such as typing on a virtual keyboard, picking up or putting down objects, etc., and may not necessarily represent a simulation of a “real” experience.


According to some implementations, a waveform file may be generated using a piezo-based pickup mounted in a silicone fingertip constructed to mimic the density of a human fingertip. The fingertip may be engaged in various forms of contact (e.g., tapping or rubbing) with one or more surfaces having different textures to generate waveform files. The waveform files generated this way may be downsampled (because humans typically can't feel above 2 kHz) using a 4 kHz sampling rate, as well as normalized for signal level.


Glove Translation


According to some implementations, mechanisms are provided that enable determining the position and movements of the glove in the real world and mapping those to a virtual environment or another context in the real world. An inertial measurement unit (IMU) on the assembly's PCB measures pitch, roll, and yaw of the glove. The IMU includes an accelerometer, a gyroscope, and a magnetometer, and performs sensor fusion to generate the pitch, roll, and yaw. In addition, a double integration of the raw accelerometer output is used to get dead reckoning information.


Fiducial Tracking


According to a particular implementation illustrated in FIG. 8, the glove has one or more concentric circular fiducials 802 on the main body of the glove that are located by a fish-eye lens camera (not shown). Using the fiducials, the position of the glove may be determined with reference to the camera. The IMU raw accelerometer output is precise enough to support dead reckoning from the positions captured by the camera for a sufficient period of time between captured positions.


The center of a fiducial is used to determine an XY position in the plane of the image (e.g., in pixels), and the radius of the outside circle of the fiducial (e.g., in pixels) is used to determine a Z position normal to the plane of the image. Even if the image of the fiducial is skewed, because the outside portion of the fiducial is circular, an accurate measurement of its radius can be determined. An occasional determination of the position of the glove based on the fiducials coupled with the use of the accelerometer-based dead reckoning provides an ongoing solid estimate of the position of the glove in space relative to the camera.


The camera can have a fixed position in the room. Alternative, the camera can be associated with the user, e.g., mounted on a virtual reality headset. The hemispherical image from the fish-eye lens is flattened and computer-vision is used to recognize the fiducials. According to some implementations, colors may be used to make the fiducials easier to recognize. In cases where the camera is part of the headset, an IMU in the headset (e.g., in the smart phone in the headset) may be configured to do the head tracking. In such cases, the position of the glove is relative to the position/location of the headset IMU as determined by the head tracking.


LED Tracking


According to another implementation illustrated in FIG. 9, the glove includes one or more LEDs that illuminate a light pipe in the rectangular clip on the back of the glove surrounding the USB connector. Modulation and detection of the LEDs can result in fast operation of the visual tracking system. That is, using a camera to detect and identify a pure color from an LED is very fast. The XY position in the plane of the image (e.g., in pixels) is identified from the location of the LED array (e.g., its center).


The LEDs may be modulated so they are blank (i.e., not be visible to the camera) at some sub-frame multiple (e.g., once every 8 frames). Detection of the specific modulation scheme allows for detection of the specific LED array and, therefore, the corresponding glove. That is, the left glove may be modulated differently from the right so that the glove for which the LED light is detected can be identified.


According to a specific implementation, a distance estimation is based on light emitted from the USB connector frame/light-pipe 902. The light may be directly from 4 LEDs arranged around the frame, or from a light pipe illuminated from one or more light sources (e.g., LEDs) on the assembly's PCB. The distance estimation is based on the distances between each pair of LEDs or the horizontal and vertical edges of the light pipe (e.g., in pixels). An accurate measurement can be determined even if the connector frame is skewed with respect to the camera because the skew can be determined based on the ratio of the distance between the horizontal LEDs (or light pipe segments) and the distance between the vertical LEDs (or light pipe segments). This ratio can be used to correct for the skew to get a more accurate measurement of at least one of the two distances which can then be used to determine the distance from the camera. The apparent brightness of the LEDs or light pipe might also be used to determine distance from the camera, either in conjunction with or instead of these distance measurements.


Position Correction


Information from a virtual environment may also be used for position correction. For example, when the user is reaching for and/or possessing something (e.g., a weapon in a game) or pushing something in the virtual environment, there are constraints on allowable positions (e.g., based on virtual object collisions) that will allow for an understanding of where the glove is in the virtual environment and use of that information to make corrections. Bounding spheres, i.e., limits on perceptual orientation and range of motion, may provide additional constraints that may be used in determining position. For example, positions can be eliminated that are impossible (e.g., your hand can't be ten feet from your head).


It will be understood by those skilled in the art that changes in the form and details of the implementations described herein may be made without departing from the scope of this disclosure. In addition, although various advantages, aspects, and objects have been described with reference to various implementations, the scope of this disclosure should not be limited by reference to such advantages, aspects, and objects.

Claims
  • 1. An apparatus, comprising: a sensor assembly including a plurality of digit sensor assemblies, each digit sensor assembly including one or more sensors on an elongated substrate, the one or more sensors of each digit sensor assembly being configured to generate one or more signals representing bending of the corresponding substrate;a textile assembly including a plurality of digit textile assemblies, each digit textile assembly being configured to align with a corresponding finger of a hand and including a fingertip portion for securing the digit textile assembly to a fingertip of the corresponding finger, each digit textile assembly also including a sleeve in which a corresponding one of the digit sensor assemblies is contained, the sleeve being connected at a first end of the sleeve to the fingertip portion of the digit textile assembly and being configured to align the corresponding digit sensor assembly with a back side of the corresponding finger, wherein the textile assembly is configured such that a middle portion of each finger is exposed, and wherein each digit sensor assembly is secured in the corresponding sleeve such that the digit sensor assembly slides relative to the sleeve when the digit sensor assembly bends;a base assembly connected to each of the sleeves of the digit textile assemblies at a second end of the sleeve opposing the first end, the base assembly being configured to secure the apparatus to the hand; andsensor circuitry configured to receive the signals from the digit sensor assemblies and to generate digital representations of the signals.
  • 2. The apparatus of claim 1, wherein each sleeve comprises an elastic fabric.
  • 3. The apparatus of claim 1, further comprising a plurality of haptic devices, each haptic device being integrated with the fingertip portion of a corresponding one of the digit textile assemblies.
  • 4. The apparatus of claim 3, wherein each sleeve includes a first chamber in which the corresponding digit sensor assembly is contained, and a second chamber in which one or more conductors connected to the haptic device of the corresponding fingertip portion are contained.
  • 5. The apparatus of claim 1, wherein each of the sleeves includes friction material on an outside surface of the sleeve configured for contacting the back side of the corresponding finger.
  • 6. The apparatus of claim 1, wherein each digit sensor assembly includes at least two sensors, each of the sensors being configured to align with a corresponding knuckle of the corresponding finger.
  • 7. The apparatus of claim 1, wherein each digit sensor assembly includes one or more stiffeners, each stiffener being aligned with a corresponding one of the one or more sensors of the digit sensor assembly and being configured to support a particular dynamic range of the corresponding sensor.
  • 8. The apparatus of claim 1, wherein each sleeve includes one or more stiffeners, each stiffener being aligned with a corresponding one of the one or more sensors of the digit sensor assembly contained in the sleeve and being configured to support a particular dynamic range of the corresponding sensor.
  • 9. The apparatus of claim 1, wherein each of the one or more sensors of each digit sensor assembly comprises either a piezoresistive material or a piezoelectric material.
  • 10. The apparatus of claim 1, wherein the base assembly is configured to secure the apparatus to the hand with a strap, the strap having a haptic device integrated therewith such that, when the apparatus is secured to the hand with the strap, the haptic device is aligned with a palm of the hand.
  • 11. An apparatus, comprising: a sensor assembly including a plurality of digit sensor assemblies, each digit sensor assembly including one or more sensors on an elongated substrate, the one or more sensors of each digit sensor assembly being configured to generate one or more signals representing bending of the corresponding substrate;a textile assembly including a plurality of digit textile assemblies, each digit textile assembly being configured to align with a corresponding finger of a hand and including a fingertip portion for securing the digit textile assembly to a fingertip of the corresponding finger, each digit textile assembly also including a sleeve in which a corresponding one of the digit sensor assemblies is contained, the sleeve being connected at a first end of the sleeve to the fingertip portion of the digit textile assembly and being configured to align the corresponding digit sensor assembly with a back side of the corresponding finger, wherein the textile assembly is configured such that a middle portion of each finger is exposed, and wherein each sleeve includes a first chamber in which the corresponding digit sensor assembly is contained, and a second chamber in which one or more conductors connected to the haptic device of the corresponding fingertip portion are contained;a base assembly connected to each of the sleeves of the digit textile assemblies at a second end of the sleeve opposing the first end, the base assembly being configured to secure the apparatus to the hand; andsensor circuitry configured to receive the signals from the digit sensor assemblies and to generate digital representations of the signals.
  • 12. The apparatus of claim 11, wherein each digit sensor assembly is secured in the corresponding sleeve such that the digit sensor assembly slides relative to the sleeve when the digit sensor assembly bends.
  • 13. The apparatus of claim 11, wherein each sleeve comprises an elastic fabric.
  • 14. The apparatus of claim 11, further comprising a plurality of haptic devices, each haptic device being integrated with the fingertip portion of a corresponding one of the digit textile assemblies.
  • 15. The apparatus of claim 11, wherein each of the sleeves includes friction material on an outside surface of the sleeve configured for contacting the back side of the corresponding finger.
  • 16. The apparatus of claim 11, wherein each digit sensor assembly includes at least two sensors, each of the sensors being configured to align with a corresponding knuckle of the corresponding finger.
  • 17. The apparatus of claim 11, wherein each digit sensor assembly includes one or more stiffeners, each stiffener being aligned with a corresponding one of the one or more sensors of the digit sensor assembly and being configured to support a particular dynamic range of the corresponding sensor.
  • 18. The apparatus of claim 11, wherein each sleeve includes one or more stiffeners, each stiffener being aligned with a corresponding one of the one or more sensors of the digit sensor assembly contained in the sleeve and being configured to support a particular dynamic range of the corresponding sensor.
  • 19. The apparatus of claim 11, wherein each of the one or more sensors of each digit sensor assembly comprises either a piezoresistive material or a piezoelectric material.
  • 20. The apparatus of claim 11, wherein the base assembly is configured to secure the apparatus to the hand with a strap, the strap having a haptic device integrated therewith such that, when the apparatus is secured to the hand with the strap, the haptic device is aligned with a palm of the hand.
US Referenced Citations (340)
Number Name Date Kind
4294014 Baumann et al. Oct 1981 A
4438291 Eichelberger et al. Mar 1984 A
4489302 Eventoff Dec 1984 A
4515404 Nishimura et al. May 1985 A
4693530 Stillie et al. Sep 1987 A
4734034 Maness et al. Mar 1988 A
4745301 Michalchik May 1988 A
4790968 Ohkawa et al. Dec 1988 A
4852443 Duncan et al. Aug 1989 A
5033291 Podoloff et al. Jul 1991 A
5079949 Tamori Jan 1992 A
5128880 White Jul 1992 A
5131306 Yamamoto Jul 1992 A
5159159 Asher Oct 1992 A
5219292 Dickirson et al. Jun 1993 A
5237520 White Aug 1993 A
5288938 Wheaton Feb 1994 A
5316017 Edwards et al. May 1994 A
5386720 Toda et al. Feb 1995 A
5429092 Kamei Jul 1995 A
5571973 Taylot Nov 1996 A
5578766 Kondo Nov 1996 A
5624132 Blackburn et al. Apr 1997 A
5659395 Brown et al. Aug 1997 A
5695859 Burgess Dec 1997 A
5729905 Mathiasmeier et al. Mar 1998 A
5822223 Genest Oct 1998 A
5866829 Pecoraro Feb 1999 A
5878359 Takeda Mar 1999 A
5943044 Martinelli et al. Aug 1999 A
5989700 Krivopal Nov 1999 A
6029358 Mathiasmeier et al. Feb 2000 A
6032109 Ritmiller, III Feb 2000 A
6049327 Walker et al. Apr 2000 A
6087930 Kulka et al. Jul 2000 A
6121869 Burgess Sep 2000 A
6141643 Harmon Oct 2000 A
6155120 Taylor Dec 2000 A
6215055 Saravis Apr 2001 B1
6216545 Taylor Apr 2001 B1
6304840 Vance et al. Oct 2001 B1
6331893 Brown et al. Dec 2001 B1
6360615 Smela Mar 2002 B1
6388556 Imai et al. May 2002 B1
6452479 Sandbach Sep 2002 B1
6486776 Pollack et al. Nov 2002 B1
6490515 Okamura et al. Dec 2002 B1
6531951 Serban et al. Mar 2003 B2
6609054 Wallace Aug 2003 B2
6626046 Taguchi et al. Sep 2003 B2
6687523 Jayaramen et al. Feb 2004 B1
6763320 Kimble Jul 2004 B2
6815602 De Franco Nov 2004 B2
6822635 Shahoian et al. Nov 2004 B2
6829942 Yanai et al. Dec 2004 B2
6964205 Papakostas et al. Nov 2005 B2
7037268 Sleva et al. May 2006 B1
7066887 Flesch et al. Jun 2006 B2
7109068 Akram et al. Sep 2006 B2
7113856 Theiss et al. Sep 2006 B2
7138976 Bouzit et al. Nov 2006 B1
7157640 Baggs Jan 2007 B2
7162344 Kojima et al. Jan 2007 B2
7302866 Malkin et al. Dec 2007 B1
7311009 Kotovsky Dec 2007 B2
7332670 Fujiwara et al. Feb 2008 B2
7409256 Lin et al. Aug 2008 B2
7439465 Parkinson Oct 2008 B2
7483866 Luo Jan 2009 B2
7493230 Schwartz et al. Feb 2009 B2
7536794 Hay et al. May 2009 B2
7584666 Kim et al. Sep 2009 B2
7608776 Ludwig Oct 2009 B2
7719007 Tompkins et al. May 2010 B2
7726199 Shkel et al. Jun 2010 B2
7754956 Gain et al. Jul 2010 B2
7780541 Bauer Aug 2010 B2
7855718 Westerman Dec 2010 B2
7928312 Sharma Apr 2011 B2
7984544 Rosenberg Jul 2011 B2
8109149 Kotovsky Feb 2012 B2
8117922 Xia et al. Feb 2012 B2
8120232 Daniel et al. Feb 2012 B2
8127623 Son et al. Mar 2012 B2
8161826 Taylor Apr 2012 B1
8162857 Lanfermann et al. Apr 2012 B2
8250934 Sakurai Aug 2012 B2
8274485 Liu et al. Sep 2012 B2
8346684 Mirbach et al. Jan 2013 B2
8368505 Deppiesse et al. Feb 2013 B2
8448530 Leuenberger et al. May 2013 B2
8479585 Shaw-Klein Jul 2013 B2
8536880 Philipp Sep 2013 B2
8571827 Jang et al. Oct 2013 B2
8587422 Andrews et al. Nov 2013 B2
8661917 Jheng et al. Mar 2014 B2
8680390 McMillen et al. Mar 2014 B2
8813579 Aufrere Aug 2014 B2
8857274 Mamigonians Oct 2014 B2
8880358 Cunningham Nov 2014 B2
8884913 Saynac et al. Nov 2014 B2
8892051 Yi et al. Nov 2014 B2
8893565 White et al. Nov 2014 B2
8904876 Taylor et al. Dec 2014 B2
8925392 Esposito et al. Jan 2015 B2
8925393 Cannard et al. Jan 2015 B2
8928014 Tischler et al. Jan 2015 B2
8945328 Longinotti-Buitoni et al. Feb 2015 B2
8947889 Kelley et al. Feb 2015 B2
8950265 Dunn et al. Feb 2015 B2
8964205 Shimizu Feb 2015 B2
8970513 Kwon et al. Mar 2015 B2
9032804 Granado et al. May 2015 B2
9038482 Xia et al. May 2015 B2
9075404 McMillen et al. Jul 2015 B2
9076419 McMillen et al. Jul 2015 B2
9112058 Bao et al. Aug 2015 B2
9116569 William et al. Aug 2015 B2
9164586 Zellers et al. Oct 2015 B2
9182302 Lim et al. Nov 2015 B2
9271665 Sarrafzadeh et al. Mar 2016 B2
9413376 Lowe et al. Aug 2016 B2
9417693 Seth Aug 2016 B2
9442614 McMillen Sep 2016 B2
9480582 Lundborg Nov 2016 B2
9529433 Shankar et al. Dec 2016 B2
9546921 McMillen et al. Jan 2017 B2
9582035 Connor Feb 2017 B2
9612102 Reese et al. Apr 2017 B2
9652101 McMillen et al. May 2017 B2
9682856 Whitesides et al. Jun 2017 B2
9696223 Lisseman et al. Jul 2017 B2
9696833 McMillen Jul 2017 B2
9710060 McMillen et al. Jul 2017 B2
9721553 McMillen et al. Aug 2017 B2
9753568 McMillen Sep 2017 B2
9756895 Rice et al. Sep 2017 B2
9827996 McMillen Nov 2017 B2
9836151 McMillen Dec 2017 B2
9851267 Ma et al. Dec 2017 B1
9863823 McMillen Jan 2018 B2
9891718 Connor Feb 2018 B2
9965076 McMillen May 2018 B2
9970832 Hong et al. May 2018 B2
9993921 Lessing et al. Jun 2018 B2
10046671 Seiller et al. Aug 2018 B2
10076143 Marriott et al. Sep 2018 B2
10082381 McMillen et al. Sep 2018 B2
10114493 McMillen et al. Oct 2018 B2
10268315 McMillen et al. Apr 2019 B2
10282011 McMillen et al. May 2019 B2
10288507 McMillen et al. May 2019 B2
10352787 McMillen et al. Jul 2019 B2
10362989 McMillen et al. Jul 2019 B2
10654486 McMillen et al. May 2020 B2
10753814 McMillen et al. Aug 2020 B2
10802641 McMillen et al. Oct 2020 B2
20020078757 Hines et al. Jun 2002 A1
20020180578 Sandbach Dec 2002 A1
20040031180 Ivanov Feb 2004 A1
20040093746 Varsallona May 2004 A1
20040118619 Gray et al. Jun 2004 A1
20040183648 Weber et al. Sep 2004 A1
20040189145 Pletner et al. Sep 2004 A1
20040249536 Hattori Dec 2004 A1
20040252007 Lussey et al. Dec 2004 A1
20050072249 Maeda et al. Apr 2005 A1
20050109095 Sinnett May 2005 A1
20050220673 Thaysen Oct 2005 A1
20060103192 Norton May 2006 A1
20060150752 Lorenz et al. Jul 2006 A1
20060192417 Ellinger et al. Aug 2006 A1
20060209050 Serban Sep 2006 A1
20060274055 Reynolds et al. Dec 2006 A1
20070063992 Lundquist Mar 2007 A1
20070129776 Robins et al. Jun 2007 A1
20070151348 Zdeblick et al. Jul 2007 A1
20070188179 Deangelis et al. Aug 2007 A1
20070188180 Deangelis et al. Aug 2007 A1
20070202765 Krans et al. Aug 2007 A1
20070234888 Rotolo de Moraes Oct 2007 A1
20080046152 Ohtake et al. Feb 2008 A1
20080069412 Champagne et al. Mar 2008 A1
20080158145 Westerman Jul 2008 A1
20080189827 Bauer Aug 2008 A1
20080254824 Moraes Oct 2008 A1
20090013793 Kim et al. Jan 2009 A1
20090049980 Sharma Feb 2009 A1
20090134966 Baker May 2009 A1
20090237374 Li et al. Sep 2009 A1
20090272197 Granado et al. Nov 2009 A1
20090301190 Ross, Jr. et al. Dec 2009 A1
20090303400 Hou et al. Dec 2009 A1
20100066572 Dietz et al. Mar 2010 A1
20100123686 Klinghult et al. May 2010 A1
20100134327 Dinh et al. Jun 2010 A1
20100149108 Hotelling et al. Jun 2010 A1
20100179724 Weston Jul 2010 A1
20100199777 Hooper et al. Aug 2010 A1
20100242274 Rosenfeld et al. Sep 2010 A1
20100274447 Stumpf Oct 2010 A1
20100286951 Danenberg et al. Nov 2010 A1
20100292945 Reynolds et al. Nov 2010 A1
20100315337 Ferren et al. Dec 2010 A1
20110005090 Lee et al. Jan 2011 A1
20110088535 Zarimis Apr 2011 A1
20110088536 McMillen et al. Apr 2011 A1
20110107771 Crist et al. May 2011 A1
20110141052 Bernstein et al. Jun 2011 A1
20110153261 Jang et al. Jun 2011 A1
20110199284 Davis et al. Aug 2011 A1
20110221564 Deppiesse et al. Sep 2011 A1
20110241850 Bosch et al. Oct 2011 A1
20110246028 Lisseman et al. Oct 2011 A1
20110260994 Saynac et al. Oct 2011 A1
20110271772 Parks et al. Nov 2011 A1
20110279409 Salaverry et al. Nov 2011 A1
20110292049 Muravsky Dec 2011 A1
20110302694 Wang et al. Dec 2011 A1
20120007831 Chang et al. Jan 2012 A1
20120024132 Wallace et al. Feb 2012 A1
20120026124 Li et al. Feb 2012 A1
20120055257 Shaw-Klein Mar 2012 A1
20120090408 Jheng et al. Apr 2012 A1
20120143092 Xia et al. Jun 2012 A1
20120191554 Xia et al. Jul 2012 A1
20120197161 Xia et al. Aug 2012 A1
20120198949 Xia et al. Aug 2012 A1
20120222498 Mamigonians Sep 2012 A1
20120234105 Taylor Sep 2012 A1
20120246795 Scheffler et al. Oct 2012 A1
20120283979 Bruekers et al. Nov 2012 A1
20120296528 Wellhoefer et al. Nov 2012 A1
20120297885 Hou et al. Nov 2012 A1
20120299127 Fujii et al. Nov 2012 A1
20120312102 Alvarez et al. Dec 2012 A1
20120323501 Sarrafzadeh et al. Dec 2012 A1
20130009905 Castillo et al. Jan 2013 A1
20130055482 D'Aprile et al. Mar 2013 A1
20130082970 Frey et al. Apr 2013 A1
20130085394 Corbett, III et al. Apr 2013 A1
20130113057 Taylor May 2013 A1
20130113704 Sarrafzadeh et al. May 2013 A1
20130165809 Abir Jun 2013 A1
20130192071 Esposito et al. Aug 2013 A1
20130203201 Britton et al. Aug 2013 A1
20130211208 Varadan et al. Aug 2013 A1
20130214365 Schlarmann et al. Aug 2013 A1
20130239787 McMillen et al. Sep 2013 A1
20130248024 Dunn et al. Sep 2013 A1
20130274985 Lee et al. Oct 2013 A1
20130275057 Perlin et al. Oct 2013 A1
20130327560 Ichiki Dec 2013 A1
20130340598 Marquez et al. Dec 2013 A1
20140007704 Granado et al. Jan 2014 A1
20140007706 Aufrere et al. Jan 2014 A1
20140013865 White et al. Jan 2014 A1
20140026678 Cannard et al. Jan 2014 A1
20140033829 Xia et al. Feb 2014 A1
20140090488 Taylor et al. Apr 2014 A1
20140104776 Clayton et al. Apr 2014 A1
20140104792 Jeziorek Apr 2014 A1
20140107966 Xia et al. Apr 2014 A1
20140107967 Xia et al. Apr 2014 A1
20140107968 Xia et al. Apr 2014 A1
20140125124 Verner May 2014 A1
20140130593 Ciou et al. May 2014 A1
20140150573 Cannard et al. Jun 2014 A1
20140182170 Wawrousek et al. Jul 2014 A1
20140195023 Statham et al. Jul 2014 A1
20140215684 Hardy et al. Aug 2014 A1
20140222173 Giedwoyn et al. Aug 2014 A1
20140222243 McMillen et al. Aug 2014 A1
20140238153 Wood et al. Aug 2014 A1
20140240214 Liu et al. Aug 2014 A1
20140264407 Tischler et al. Sep 2014 A1
20140318699 Longinotti-Buitoni et al. Oct 2014 A1
20140347076 Barton et al. Nov 2014 A1
20150035743 Rosener Feb 2015 A1
20150084873 Hagenbuch et al. Mar 2015 A1
20150086955 Poniatowski et al. Mar 2015 A1
20150130698 Burgess May 2015 A1
20150168238 Raut et al. Jun 2015 A1
20150177080 Esposito et al. Jun 2015 A1
20150231991 Yetukuri et al. Aug 2015 A1
20150248159 Luo et al. Sep 2015 A1
20150261372 McMillen et al. Sep 2015 A1
20150316434 McMillen et al. Nov 2015 A1
20150317964 McMillen et al. Nov 2015 A1
20150328492 Marriott et al. Nov 2015 A1
20150330855 Daniecki et al. Nov 2015 A1
20150331512 McMillen et al. Nov 2015 A1
20150331522 McMillen et al. Nov 2015 A1
20150331523 McMillen et al. Nov 2015 A1
20150331524 McMillen et la. Nov 2015 A1
20150331533 McMillen et al. Nov 2015 A1
20150370396 Hotelling et al. Dec 2015 A1
20160052131 Lessing et al. Feb 2016 A1
20160054798 Messingher et al. Feb 2016 A1
20160070347 McMillen Mar 2016 A1
20160073539 Driscoll et al. Mar 2016 A1
20160147352 Filiz et al. May 2016 A1
20160162022 Seth Jun 2016 A1
20160169754 Kowalewski et al. Jun 2016 A1
20160175186 Shadduck Jun 2016 A1
20160187973 Shankar et al. Jun 2016 A1
20160209441 Mazzeo et al. Jul 2016 A1
20160238547 Park et al. Aug 2016 A1
20160246369 Osman Aug 2016 A1
20160252412 McMillen et al. Sep 2016 A1
20160270727 Berg et al. Sep 2016 A1
20160278709 Granado et al. Sep 2016 A1
20160313798 Connor Oct 2016 A1
20160318356 McMillen et al. Nov 2016 A1
20160340534 Wijesundara et al. Nov 2016 A1
20160358849 Jur et al. Dec 2016 A1
20160375910 McMillen et al. Dec 2016 A1
20170000369 Hyde et al. Jan 2017 A1
20170038881 McMillen Feb 2017 A1
20170056644 Chahine et al. Mar 2017 A1
20170059426 Choi et al. Mar 2017 A1
20170086519 Vigano et al. Mar 2017 A1
20170108929 Sinko et al. Apr 2017 A1
20170110103 McMillen et al. Apr 2017 A1
20170127736 Roberts et al. May 2017 A1
20170167931 McMillen et al. Jun 2017 A1
20170176267 Keller et al. Jun 2017 A1
20170212638 McMillen Jul 2017 A1
20170215495 Okumiya et al. Aug 2017 A1
20170303853 McMillen et al. Oct 2017 A1
20170305301 McMillen et al. Oct 2017 A1
20180015932 McMillen et al. Jan 2018 A1
20180077976 Keller Mar 2018 A1
20180094991 McMillen et al. Apr 2018 A1
20180263563 McMillen et al. Sep 2018 A1
20190034019 McMillen et al. Jan 2019 A1
20190219465 McMillen et al. Jul 2019 A1
20200150761 Hogbin May 2020 A1
20200200621 McMillen et al. Jun 2020 A1
20200292399 McMillen et al. Sep 2020 A1
Foreign Referenced Citations (49)
Number Date Country
200980381 Nov 2007 CN
201920728 Aug 2011 CN
102551728 Jul 2012 CN
202396601 Aug 2012 CN
203234132 Oct 2013 CN
102406280 Mar 2014 CN
102 12 023 Oct 2003 DE
11 2010 004 038 Sep 2012 DE
0 014 022 Nov 1984 EP
0 211 984 Mar 1987 EP
2 682 724 Jan 2014 EP
S47-18925 May 1972 JP
H04-011666 Jan 1992 JP
H06-323929 Nov 1994 JP
H08-071978 Mar 1996 JP
H08-194481 Jul 1996 JP
H10-198503 Jul 1998 JP
2000-267664 Sep 2000 JP
2006-503350 Jan 2006 JP
2007-503052 Feb 2007 JP
2008-515008 May 2008 JP
2009-543030 Dec 2009 JP
2010-503113 Jan 2010 JP
2011-502313 Jan 2011 JP
2012-521550 Sep 2012 JP
2012-220315 Nov 2012 JP
2014-077662 May 2014 JP
2017-518338 Dec 2017 JP
10-2007-0008500 Jan 2007 KR
100865148 Oct 2008 KR
10-1362742 Feb 2014 KR
10-2014-0071693 Jun 2014 KR
8900820 Nov 1990 NL
2 533 539 Nov 2014 RU
WO 99020179 Apr 1999 WO
WO 2007024875 Mar 2007 WO
WO 2008032661 Mar 2008 WO
WO 2009155891 Dec 2009 WO
WO 2011047171 Apr 2011 WO
WO 2013116242 Aug 2013 WO
WO 2013181474 Dec 2013 WO
WO 2014058473 Apr 2014 WO
WO 2015175317 Nov 2015 WO
WO 2016070078 May 2016 WO
WO 2016138234 Sep 2016 WO
WO 2016176307 Nov 2016 WO
WO 2016210173 Dec 2016 WO
WO 2017066096 Apr 2017 WO
WO 2017184367 Oct 2017 WO
Non-Patent Literature Citations (123)
Entry
U.S. Office Action dated Sep. 12, 2012 issued in U.S. Appl. No. 12/904,657.
U.S. Office Action dated Apr. 15, 2013 issued in U.S. Appl. No. 12/904,657.
U.S. Notice of Allowance dated Nov. 8, 2013 issued in U.S. Appl. No. 12/904,657.
U.S. Office Action dated Mar. 12, 2015 issued in U.S. Appl. No. 14/173,617.
U.S. Notice of Allowance dated May 1, 2015 issued in U.S. Appl. No. 14/173,617.
U.S. Office Action dated Mar. 10, 2016 issued in U.S. Appl. No. 14/727,619.
U.S. Final Office Action dated Jul. 18, 2016 issued in U.S. Appl. No. 14/727,619.
U.S. Notice of Allowance dated Sep. 15, 2016 issued in U.S. Appl. No. 14/727,619.
U.S. Office Action dated Dec. 31, 2018 issued in U.S. Appl. No. 15/374,816.
U.S. Notice of Allowance dated Mar. 11, 2019 issued in U.S. Appl. No. 15/374,816.
U.S. Office Action dated Jan. 10, 2020 issued in U.S. Appl. No. 16/362,017.
U.S. Office Action dated Apr. 2, 2015 issued in U.S. Appl. No. 13/799,304.
U.S. Notice of Allowance dated Apr. 24, 2015 issued in U.S. Appl. No. 13/799,304.
U.S. Office Action dated Sep. 1, 2015 issued in U.S. Appl. No. 14/728,872.
U.S. Final Office Action dated Mar. 9, 2016 issued in U.S. Appl. No. 14/728,872.
U.S. Office Action dated Jun. 22, 2016 issued in U.S. Appl. No. 14/728,872.
U.S. Final Office Action dated Oct. 18, 2016 issued in U.S. Appl. No. 14/728,872.
U.S. Advisory Action dated Feb. 10, 2017 issued in U.S. Appl. No. 14/728,872.
U.S. Office Action dated May 19, 2017 issued in U.S. Appl. No. 14/728,872.
U.S. Notice of Allowance dated Oct. 16, 2017 issued in U.S. Appl. No. 14/728,872.
U.S. Office Action dated Jul. 25, 2016 issued in U.S. Appl. No. 14/728,873.
U.S. Office Action dated Dec. 30, 2016 issued in U.S. Appl. No. 14/728,873.
U.S. Final Office Action dated Mar. 31, 2017 issued in U.S. Appl. No. 14/728,873.
U.S. Advisory Action and Examiner initiated interview summary dated May 26, 2017 issued in U.S. Appl. No. 14/728,873.
U.S. Office Action dated Aug. 25, 2017 issued in U.S. Appl. No. 14/728,873.
U.S. Final Office Action dated Dec. 22, 2017 issued in U.S. Appl. No. 14/728,873.
U.S. Office Action dated Mar. 26, 2018 issued in U.S. Appl. No. 14/728,873.
U.S. Notice of Allowance dated Jul. 19, 2018 issued in U.S. Appl. No. 14/728,873.
U.S. Office Action dated Mar. 9, 2016 issued in U.S. Appl. No. 14/299,976.
U.S. Final Office Action dated Jul. 6, 2016 issued in U.S. Appl. No. 14/299,976.
U.S. Office Action dated Oct. 21, 2016 issued in U.S. Appl. No. 14/299,976.
U.S. Final Office Action dated Apr. 19, 2017 issued in U.S. Appl. No. 14/299,976.
U.S. Final Office Action dated Jun. 8, 2017 issued in U.S. Appl. No. 14/299,976.
U.S. Office Action dated Sep. 1, 2017 issued in U.S. Appl. No. 14/299,976.
U.S. Notice of Allowance dated Feb. 22, 2018 issued in U.S. Appl. No. 14/299,976.
U.S. Office Action dated Jan. 13, 2016 issued in U.S. Appl. No. 14/464,551.
U.S. Notice of Allowance dated Jun. 23, 2016 issued in U.S. Appl. No. 14/464,551.
U.S. Office Action dated Sep. 23, 2016 issued in U.S. Appl. No. 14/800,538.
U.S. Notice of Allowance dated Jan. 17, 2017 issued in U.S. Appl. No. 14/800,538.
U.S. Office Action dated Jul. 12, 2018 issued in U.S. Appl. No. 15/483,926.
U.S. Notice of Allowance dated Dec. 31, 2018 issued in U.S. Appl. No. 15/483,926.
U.S. Office Action dated Feb. 22, 2017 issued in U.S. Appl. No. 14/671,821.
U.S. Notice of Allowance dated Jul. 3, 2017 issued in U.S. Appl. No. 14/671,821.
U.S. Office Action dated Jun. 30, 2017 issued in U.S. Appl. No. 15/251,772.
U.S. Final Office Action dated Nov. 15, 2017 issued in U.S. Appl. No. 15/251,772.
U.S. Office Action dated Feb. 22, 2018 issued in U.S. Appl. No. 15/251,772.
U.S. Office Action dated Sep. 4, 2018 issued in U.S. Appl. No. 15/251,772.
U.S. Final Office Action dated Dec. 21, 2018 issued in U.S. Appl. No. 15/251,772.
U.S. Notice of Allowance dated Mar. 5, 2019 issued in U.S. Appl. No. 15/251,772.
U.S. Office Action dated Jun. 28, 2016 issued in U.S. Appl. No. 14/671,844.
U.S. Final Office Action dated Nov. 25, 2016 issued in U.S. Appl. No. 14/671,844.
U.S.Notice of Allowance dated Mar. 13, 2017 issued in U.S. Appl. No. 14/671,844.
U.S. Office Action dated Jan. 26, 2017 issued in U.S. Appl. No. 15/052,293.
U.S. Final Office Action dated May 2, 2017 issued in U.S. Appl. No. 15/052,293.
U.S. Notice of Allowance dated May 24, 2017 issued in U.S. Appl. No. 15/052,293.
U.S. Notice of Allowance [Supplemental Notice of Allowability] dated Jun. 20, 2017 issued in U.S. Appl. No. 15/052,293.
U.S. Notice of Allowance dated Sep. 22, 2017 issued in U.S. Appl. No. 15/052,293.
U.S. Notice of Allowance [Supplemental Notice of Allowability] dated Oct. 19, 2017 issued in U.S. Appl. No. 15/052,293.
U.S. Office Action dated Jul. 24, 2018 issued in U.S. Appl. No. 15/835,131.
U.S. Notice of Allowance dated Dec. 4, 2018 issued in U.S. Appl. No. 15/835,131.
U.S. Office Action dated Mar. 6, 2019 issued in U.S. Appl. No. 15/835,131.
U.S. Notice of Allowance dated May 13, 2019 issued in U.S. Appl. No. 15/835,131.
U.S. Office Action dated May 20, 2016 issued in U.S. Appl. No. 14/928,058.
U.S. Final Office Action dated Jan. 6, 2017 issued in U.S. Appl. No. 14/928,058.
U.S. Notice of Allowance dated Mar. 16, 2017 issued in U.S. Appl. No. 14/928,058.
U.S. Office Action dated Aug. 14, 2018 issued in U.S. Appl. No. 15/621,935.
U.S. Final Office Action dated Feb. 14, 2019 issued in U.S. Appl. No. 15/621,935.
U.S. Notice of Allowance dated May 22, 2019 issued in U.S. Appl. No. 15/621,935.
U.S. Supplemental Notice of Allowance dated Jun. 13, 2019 issued in U.S. Appl. No. 15/621,935.
U.S. Office Action dated Oct. 21, 2019 issued in U.S. Appl. No. 15/621,935.
U.S. Office Action dated Nov. 3, 2017 issued in U.S. Appl. No. 15/138,802.
U.S. Final Office Action dated Mar. 1, 2018 issued in U.S. Appl. No. 15/138,802.
U.S. Advisory Action dated May 16, 2018 issued in U.S. Appl. No. 15/138,802.
U.S. Notice of Allowance dated Jul. 3, 2018 issued in U.S. Appl. No. 15/138,802.
U.S. Office Action dated Jun. 23, 2017 issued in U.S. Appl. No. 15/190,089.
U.S. Notice of Allowance dated Aug. 10, 2017 issued in U.S. Appl. No. 15/190,089.
U.S. Office Action dated Dec. 13, 2018 issued in U.S. Appl. No. 15/690,108.
U.S. Office Action dated Jul. 5, 2019 issued in U.S. Appl. No. 15/690,108.
U.S. Notice of Allowance dated Jan. 29, 2020 issued in U.S. Appl. No. 15/690,108.
U.S. Office Action dated May 24, 2019 issued in U.S. Appl. No. 15/479,103.
U.S. Final Office Action dated Sep. 20, 2019 issued in U.S. Appl. No. 15/479,103.
U.S. Office Action dated Dec. 27, 2016 issued in U.S. Appl. No. 15/287,520.
U.S. Notice of Allowance dated Mar. 27, 2017 issued in U.S. Appl. No. 15/287,520.
PCT International Search Report dated May 27, 2011, issued in PCT/US2010/052701.
PCT International Preliminary Report on Patentability and Written Opinion dated Apr. 26, 2012, issued in PCT/US2010/052701.
Japanese Office Action dated Feb. 25, 2014 issued in JP 2012-534361.
PCT International Search Report and Written Opinion dated Sep. 3, 2015 issued in PCT/US2015/029732.
PCT International Preliminary Report on Patentability and Written Opinion dated Nov. 24, 2016 issued in PCT/US2015/029732.
Japanese Office Action dated Dec. 4, 2018 issued in JP 2016-566814.
PCT International Search Report and Written Opinion dated May 26, 2016 issued in PCT/US2016/019513.
PCT International Preliminary Report on Patentability and Written Opinion dated Sep. 8, 2017 issued in PCT/US2016/019513.
Japanese Office Action dated Jul. 30, 2019 issued in JP 2017-518338.
Japanese Office Action dated Jul. 9, 2019 issued in JP 2018-114012.
PCT International Search Report and Written Opinion dated Apr. 14, 2016 issued in PCT/US2015/058370.
PCT International Preliminary Report on Patentability and Written Opinion dated May 11, 2017 issued in PCT/US2015/058370.
PCT International Search Report and Written Opinion dated Nov. 8, 2018 issued in PCT/US2018/035848.
PCT International Preliminary Report on Patentability and Written Opinion dated Dec. 26, 2019 issued in PCT/US2018/035848.
PCT International Search Report and Written Opinion dated Sep. 15, 2016 issued in PCT/US2016/029528.
PCT International Preliminary Report on Patentability and Written Opinion dated Oct. 31, 2017 issued in PCT/US2016/029528.
PCT International Search Report and Written Opinion dated Sep. 29, 2016 issued in PCT/US2016/039089.
PCT International Preliminary Report on Patentability and Written Opinion dated Dec. 26, 2017 issued in PCT/US2016/039089.
PCT International Search Report and Written Opinion dated Jan. 19, 2017 issued in PCT/US2016/055997.
PCT International Preliminary Report on Patentability and Written Opinion dated Apr. 26, 2018 issued in PCT/US2016/055997.
PCT International Search Report and Written Opinion dated Aug. 14, 2017 issued in PCT/US2017/026812.
PCT International Preliminary Report on Patentability dated Nov. 1, 2018 issued in PCT/US2017/026812.
“Electronic Foot Size Measuring Devices,” Sensatech Research Ltd., Custom Electronic Sensing Solutions, Registered Office: 4 Heath Square, Boltro Road, Haywards Heath, RH16 1BL Company Registration No. 4524018 Cardiff [retrieved at http:www.electronicsarena.co.uk/companies/sensatech-research/products/electronic-foot-size-measureing-devices on Sep. 17, 2015], 3 pages.
“iStep® Digital Foot Scan,” (© 2002-2015) [retrieved at http://www.foot.com/site/iStep on Sep. 17, 2015], 1 page.
“Podotech Elftman,” and Podotech Elftman Brochure (UK Version) [retrieved at http://www.podotech.com/diagnostics/podotech-elftman-2/ on Sep. 17, 2015] podo+tech®, Foot Care Technology Solutions, 7 pages.
Roh, Jung-Sim et al. (2011) “Robust and reliable fabric and piezoresistive multitouch sensing surfaces for musical controllers,” from Alexander Refsum Jensenius, Recorded at: 11th International Conference on New Interfaces for Musical Expression May 30-Jun. 1, 2011, Oslo, Norway, a vimeo download at http://vimeo.com/26906580.
“The Emed®-Systems,” [retrieved at http://www.novel.de/novelcontent/emed on Sep. 17, 2015] novel.de, 4 pages.
U.S. Appl. No. 15/630,840, filed Jun. 22, 2017, McMillen et al.
U.S. Appl. No. 16/692,626, filed Nov. 22, 2019, Lacy et al.
U.S. Notice of Allowance dated May 13, 2020 issued in U.S. Appl. No. 16/362,017.
U.S. Office Action dated Mar. 19, 2020 issued in U.S. Appl. No. 14/728,873.
U.S. Notice of Allowance dated Jun. 15, 2020 issued in U.S. Appl. No. 14/728,873.
U.S. Final Office Action dated Apr. 28, 2020 issued in U.S. Appl. No. 15/621,935.
U.S. Office Action dated Jul. 13, 2020 issued in U.S. Appl. No. 15/621,935.
Japanese Office Action dated May 22, 2020 issued in JP 2017-518338.
Japanese Office Action dated Apr. 22, 2020 issued in JP 2018-114012.
Japanese Office Action dated Aug. 17, 2020 issued in JP 2018-114012.
U.S. Appl. No. 16/806,297, filed Mar. 2, 2020, McMillen.
U.S. Appl. No. 16/947,140, filed Jul. 20, 2020, McMillen et al.
U.S. Appl. No. 16/948,131, filed Sep. 3, 2020, McMillen et al.
Related Publications (1)
Number Date Country
20200012344 A1 Jan 2020 US
Provisional Applications (1)
Number Date Country
62694334 Jul 2018 US