Features of airborne ultrasonic fields

Information

  • Patent Grant
  • 11816267
  • Patent Number
    11,816,267
  • Date Filed
    Tuesday, June 22, 2021
    3 years ago
  • Date Issued
    Tuesday, November 14, 2023
    a year ago
Abstract
A method for specifying desired quantities of the energy flux of the combined waves of acoustic radiation pressure to apply producing a mid-air haptic pushing force, which has the effect of simultaneously reducing the harmonic distortion present is described. Further, a method for communicating only the summaries of acoustic field contributions at the required locations in the form of summed portions of the relatively small matrix whose row and column count depend only on the control point count is described. Further, phased arrays of ultrasonic speakers can produce a relatively large amount of acoustic energy which is usually directed in a specific direction or focused to a particular point depending on the application of the array. Further, to allow the system to be driven more strongly than usual, the complex-valued linear system that governs the drive signal to each control point is solved twice. Further, to achieve mid-air haptics with high precision the radiation pressure generated must be modelled accurately.
Description
FIELD OF THE DISCLOSURE

The present disclosure relates generally to improved techniques in establishing useful and unique features in ultrasonic fields.


BACKGROUND

Phased arrays including ultrasonic phased array systems function on the principle of superposition. Superpositions occur when linear quantities that describe waves add together to create areas of constructive and destructive interference. State-of-the-art ultrasonic systems use these quantities directly for control over the linear function values of points in the acoustic field. The resulting samples can be used in a linear system of complex-valued equations to solve for the actuations of transducer that then generate the desired linear field values at the desired points (known as control points).


By solving many times in rapid succession and generating and supplying the drive signal that corresponds to the solution values to the transducers, waveforms may be modulated onto the ultrasonic carrier generated by the transducer elements in the phased array. This is because as the solved for value changes, the amount of acoustic pressure of the carrier also changes.


This modulation has two key non-linear effects that are exploitable by commercial devices. The first, known as acoustic radiation force, is proportional to the energy in the wave and refers to the force generated when the wave is impeded. This force is at a maximum when a sharp change in acoustic impedance is present, such as the surface of a hand when the ultrasonic waves are travelling through the air. The force generated may be used for haptic feedback in mid-air. The second, known as ‘sound from ultrasound’ is the primary operating mechanism for parametric speaker arrays which also has an effect proportional to the energy in the wave. This effect is responsible for causing audible sound to be seemingly emitted by an ultrasonic array when no audible sound was present as the source of the ultrasonic wave.


High precision control of both of these effects are necessary to create reproducible mid-air haptics with controlled, little or no audible noise produced as a side-effect. However, the momentum and energy of the device cannot be controlled directly as it is a non-linear quantity, as well as being dependent on a number of geometric parameters. It is therefore commercially relevant to develop a method of taking from the ‘user’ of the system a desired level of wave energy, converting to the equivalent level of a linear acoustic quantity and then solving for this linear level using linear methods given that it has already been established that this is equivalent to an amount of wave energy that cannot be sought directly.


Further, when considering the solution to the problem of an ultrasonic field which has been discretized into a number of discrete points, the recreation of such a field via a phased transducer array technology may be achieved by solving for the output field required of each transducer. This solution can be described as, given a reference frequency, a series of complex numbers which can be interpreted as the phase and amplitude required of each transducer. From this data an input signal to the transducer may be inferred as a signal wherein a substantial proportion has the properties of a sinusoidal signal with the given amplitude and phase at the transducer. Given a method to compute these complex-valued coefficients, it is reasonable to assume that these must be transferred to the device in order to activate it and the information on the transducer locations and directions must be transferred to the compute location in order to do work on it. This makes the compute required at the transducer very small, as it only has to import complex coefficients and export sufficient data to describe its produced field. However, the communication with these devices will scale with the number of transducers, which if relatively large will prove too much. On top of this, the synthesis of the field produced by each transducer must be achieved in some centralized location, which is again undesirable.


By determining the samples of the field produced by each transducer locally and using these as a common definition of a basis function each of which is made up of output from all transducing elements, communicating only a complex-valued coefficient per basis function may be achieved. This is because both the compute close to the transducer in the hardware and the software residing in the user interface portion understand the definition of the basis function. These need only communicate enough information to define the individual basis functions and their complex-valued linear superposition, which in turn defines the complex-valued linear superposition of basis-function-defined complex-valued coefficients that drive each transducing element. However, in order for this system to function, both the hardware system close to the transducer and the software residing in the user interface portion of the system must understand how to expand the basis functions out into individual transducer information. This leads to computation that is duplicated in both hardware close to the transducers and the software portion in order to make savings on communication bandwidth. This uses extra compute power and resources to achieve low-bandwidth, but is clearly not yet optimal, as reducing system cost must necessarily involve reducing both communication and computation to a minimum.


While a method that eliminates the dependency of bandwidth requirements on the transducer count of the system is needed, duplicating functionality is clearly undesirable. Overcoming this limitation to enable distributed simulation of the acoustic field, while only requiring summaries of acoustic field contributions that do not depend on the transducer element count to be communicated to solve for the output field, is commercially valuable.


Further, airborne ultrasonic phased arrays can be used to create arbitrary acoustic fields. These can be used for haptic feedback, parametric audio, acoustic levitation, etc. To achieve compelling effects, it is often the case that relatively high levels of ultrasound energy are needed. Objects, microphones, animals, and/or people in the area of the ultrasound field could be sensitive to these levels. In many cases, even if the ultrasound is directed elsewhere, fringing (unintentional) fields can still cause problems. Presented below are several methods/strategies to create points or regions in the field specifically lacking ultrasound, or ‘nulls’ without substantially altering the effect generated by the rest of the field. If the location of the sensitive object is known to some degree, it is possible to direct a null point or region towards it in order to protect it from the ultrasound.


A particularly attractive application of this method is to parametric audio. This is the demodulation of ultrasound into audible sound through the nonlinear properties of the air. This creates a beam-like projection of audio. The audible sound created is directed along the same direction as the ultrasound. High levels of ultrasound can interact negatively with microphones and can even be perceived by the mammalian ear through a nonlinearity at the tympanic membrane. This can mask or distort the parametric audio and lower the quality of the experience.


All sound waves are subject to diffraction. This is an effect whereby waves spread out on a length scale related to their wavelength. Short-wavelength sound such as ultrasound at about 40 “kHz” is able to propagate and maintain features approximately equal to its wavelength, 8.6 “mm”. Regular sound, on the other hand, consists of waves with a much longer wavelength (middle c, 261 “Hz”, has a wavelength, λ, of 1.3 “m”), which spread out readily. Parametric audio forms a beam of sound by exploiting the short-wavelength of ultrasound to create tightly contained regions of sound emission. The sound, once created, spreads out as usual. This opens up the possibility of delivering parametric audio without delivering the associated ultrasound. By creating relatively small regions devoid of ultrasound with nearby high levels of ultrasound, we can create a situation where the parametric audio fills in the ‘gaps’ while the ultrasound does not.


Further, a continuous distribution of sound energy, which will refer to as an “acoustic field”, can be used for a range of applications including haptic feedback in mid-air, sound-from-ultrasound systems and producing encoded waves for tracking systems.


By defining one of more control points in space, the acoustic field can be controlled. Each point can be assigned a value equating to a desired amplitude at the control point. A physical set of transducers can then be controlled to create an acoustic field exhibiting the desired amplitude at the control points.


The transducer elements, being physical devices, have physical limitations. In the case of producing an acoustic field there is a maximum output for each element that cannot be exceeded. The mathematical structure of the system makes it cumbersome to force solutions to respect the power limitations of the physical device, where clean solution methods often produce unphysical drive conditions.


If the transducers are arranged such that grating lobes are a problem, then it is possible to reduce the impact of the grating lobes on the control points by apodizing the transducer amplitudes (creating a tapered set of amplitudes towards the edge of the array). This necessarily reduces the efficiency of the array, limiting the maximum output power available. For multiple points also, while producing the relative amplitudes is always possible when the number of transducer elements is greater than the number of control points, efficiency drops as the number of points increases, and the maximum power drops.


A method to arrest these drops in efficiency which would act to incrementally raise the output level when more output power is required of the device than can be supplied using existing methods would therefore be commercially valuable.


It is also possible to, in the case of a single point, drive all transducers at the same power and influence only their phase. Then the collective drive amplitude of the transducers may be modulated to generate a similar modulation on the signal at the control point. However, this solution does not help when the benefits of apodization and/or multiple points are required.


If higher efficiency or output power is desired then, either the solution is made less accurate by pushing the transducer drive to higher levels than is described by the solution or the benefits of apodization and/or multiple points cannot be used.


Further, phased arrays including ultrasonic phased array systems function on the principle of superposition. Superpositions occur when linear quantities that describe waves add together to create areas of constructive and destructive interference. State-of-the-art ultrasonic systems use these quantities directly for control over the linear function values of points in the acoustic field. The resulting samples can be used in a linear system of complex-valued equations to solve for the actuations of transducer that then generate the desired linear field values at the desired points (known as control points).


As tractable solutions for the linear values to control these points are only computable under the assumption of free-field conditions, these same conditions have in many cases erroneously been used as justification for the assumptions of a bulk acoustic medium. For mid-air haptics with ultrasonic phased arrays and other situations in which an ultrasonic phased array is used to apply a force to a boundary separating two materials with different acoustic properties (such as the boundary between air and a human body part), the boundary condition changes the scenario of the problem to the extent that solutions developed with the bulk scenario in mind will necessarily omit details critical to the accurate reproduction of acoustic force on the aforementioned boundary surface.


In this document, a system that generates reproductions of one or more force vectors on one or more high acoustic impedance boundaries is detailed.


SUMMARY

A method for specifying desired quantities of the energy flux of the combined waves of acoustic radiation pressure to apply producing a mid-air haptic pushing force, which has the effect of simultaneously reducing the harmonic distortion present is described.


Further, a method for communicating only the summaries of acoustic field contributions at the required locations in the form of summed portions of the relatively small matrix whose row and column count depend only on the control point count is described.


Further, phased arrays of ultrasonic speakers can produce a relatively large amount of acoustic energy which is usually directed in a specific direction or focused to a particular point depending on the application of the array. Certain objects (e.g. microphones) can become interfered with by the acoustic field which may reduce their function (see FIGS. 10A, 10B, and 10C). One method to reduce the effect of high intensity ultrasound is to create ‘quiet’ areas where the intensity is significantly lower than the surrounding acoustic field. By creating a lower-pressure region around an object (e.g. a number of lower-pressure focal points, or a lower-pressure volume) the acoustic intensity experienced by the object can be significantly reduced. Our-solution allows us to create these quiet areas without significantly affecting the performance of the array.


Further, to allow the system to be driven more strongly than usual, the complex-valued linear system that governs the drive signal to each control point is solved twice. Once to determine how much drive from each transducer might be necessary and a further time in which each transducer has been scaled back by the overshoot in drive that occurred during the first solution, resulting in a more even distribution of power across the transducers in the solution from the second solve.


The system scales back the coefficients used on the parts of the basis functions that are driven hardest. This is counterintuitive—reducing the effectiveness of high efficiency transducers, boosts general power output overall. This works because then the solution method, uses the now less effective parts less, so the output requirements are more evenly distributed.


Further, most solution methods are intended to generate a predefined nonlinear effect in a bulk medium. In some instances, such as generation of parametric audio, this is justifiable as this describes the body forces exerted through the acoustic medium for which using a free-field bulk medium is an acceptable modelling approach. However, if a body force model is being used to describe the interaction with a boundary, then it will not necessarily reflect reality.


To achieve mid-air haptics with high precision the radiation pressure generated must be modelled accurately. The modelling of radiation pressure in the academic literature is generally approached using one of the two following methodologies.


The first approach is by analogy to electromagnetic waves. In this case, radiation pressure is considered to be a force acting along the acoustic Poynting vector. As this is in terms of the acoustic Poynting vector then it is natural that the energy flux density and thus in acoustic terms the acoustic intensity I describes the magnitude of the force.


The second approach is taken by academic papers that write about acoustophoresis levitating objects in the acoustic medium using standing waves or specially constructed interference patterns. In the field of acoustophoresis, radiation pressure is considered to be a scalar potential defined by the time averaged second-order pressure p2. Then as a scalar potential, the negative gradient of the potential field describes the direction and magnitude of the force vector.


Academic papers that describe radiation force using the first approach ignore the second as they assume far-field conditions where it can be shown that p2=0. Those that describe radiation force using the second approach ignore the first as the acoustic Poynting vector either cancels to zero in a standing wave, or generates few useful degrees of freedom that could aid optimization of an acoustophoretic interference pattern and so tends to be neglected.


Neither of these approaches sufficiently describe the phenomenon of radiation pressure for mid-air haptic systems, as both of the simplifying assumptions are false for using acoustic phased array hardware to generate apparent haptic forces on humans.


Assume that the Poynting vector can be directly converted into a linearly related force vector with the same direction. This may then be assumed (incorrectly) to be split into force components, so given an arbitrary unit vector direction {circumflex over (n)}, I·{circumflex over (n)}∝Fr·{circumflex over (n)}, so in the bulk medium the radiation pressure in a given direction may be directly related to the energy flux density in that same direction {circumflex over (n)}. If this {circumflex over (n)} is given as the normal vector to a non-physical surface that cuts the energy flux vector (acoustic intensity) I, then Fr·{circumflex over (n)} being the force acting at a point across this non-physical surface is given as justification for this being the force generated on a physical surface at this same location. This is not the whole picture because as previous described, this is not a body force so does not act in bulk and this scenario assumes far field behavior—for a focusing mid-air haptic device (where the act of focusing implies near-field behavior) this is an approximation at best.


In this disclosure, determining the scalar linear acoustic quantity to solve for that best represents the apparent haptic pressure to be generated is described.





BRIEF DESCRIPTION OF THE DRAWINGS

The accompanying figures, where like reference numerals refer to identical or functionally similar elements throughout the separate views, together with the detailed description below, are incorporated in and form part of the specification, serve to further illustrate embodiments of concepts that include the claimed invention and explain various principles and advantages of those embodiments.



FIG. 1 shows plots of modulated parametric audio responses.



FIG. 2 shows plots of total harmonic distortion.



FIG. 3 shows a plot of total harmonic distortion and its harmonics.



FIG. 4 shows a plot of total harmonic distortion and its harmonics.



FIGS. 5A, 5B, and 5C show low-pressure area simulations in an acoustic field.



FIGS. 6A, 6B, and 6C show low-pressure area simulations in an acoustic field.



FIGS. 7A, 7B, 7C, and 7D show null point arrangements in space around an array of transducers.



FIGS. 5A and 5B show quiet zone efficiency for a 3D spacing arrangement.



FIG. 9 shows a plot of pressure of a single focal point.



FIGS. 10A, 10B, and 10C show spectrograms of recorded human speech.



FIGS. 11A, 11B, 11C, and 11D show low-pressure area simulations in an acoustic field.



FIGS. 12A and 12B show low-pressure area simulations in an acoustic field.



FIGS. 13A and 13B show low-pressure area simulations in an acoustic field.



FIGS. 14A and 14B show low-pressure area simulations in an acoustic field.



FIG. 15 shows low-pressure area simulations in an acoustic field.



FIG. 16 shows low-pressure area simulations in an acoustic field.



FIG. 17 shows stages of an algorithm manipulating focal point basis functions.



FIG. 18 shows the plotting of equiprobable transducer drive amplitudes.



FIGS. 19A, 19B, and 19C show the outputs of transformation matrices.



FIG. 20 shows a diagram of waves related to a surface radiation pressure for a far field wave.



FIG. 21 shows a diagram of waves related to a surface radiation pressure for a near field wave.





Skilled artisans will appreciate that elements in the figures are illustrated for simplicity and clarity and have not necessarily been drawn to scale. For example, the dimensions of some of the elements in the figures may be exaggerated relative to other elements to help to improve understanding of embodiments of the present invention.


The apparatus and method components have been represented where appropriate by conventional symbols in the drawings, showing only those specific details that are pertinent to understanding the embodiments of the present invention so as not to obscure the disclosure with details that will be readily apparent to those of ordinary skill in the art having the benefit of the description herein.


DETAILED DESCRIPTION
I. Energy-Proportional Interfaces for Describing Acoustic Force

Deflection of the skin and ultimately a feeling of touch from an ultrasonic mid-air haptic device is determined by nonlinear acoustic radiation force found at the interface between air and skin. For a plane wave incident on an infinite planar surface, this is given by (Fluid Mechanics 3rd edition, Landau and Lifshitz, pp. 255-256),

p=E1 sin θ1 cos θ1[(1+R) cot θ1−(1−R)cot θ2],

where θ1 and θ2 are the wave's angle of reflection and refraction, respectively, R is the reflection coefficient, and E1 is the time-averaged energy density in the incident acoustic wave. For normal incidence, the pressure reduces to,







p
=

2




E
1

_

[




ρ
1
2



c
1
2


+


ρ
2
2



c
2
2


-

2


ρ
1



ρ
2



c
1
2





(



ρ
1



c
1


+


ρ
2



c
2



)

2


]



,





where ρ and c refer to the density and speed of sound in each of fluid 1 and 2. While a human hand is certainly not an “infinite planar surface”, it is much larger than ultrasonic wavelengths (the wavelength of 25 kHz ultrasound is ˜1.4 cm at normal temperature and pressure and decreases at increasing frequency). When a focused onto the palm, the region which is applying the nonlinear pressure is typically about 1 wavelength in diameter but can be as much as 3 wavelengths depending on the size of the array relative to the distance it is focusing. Therefore, the above approximation should be correct to first order in most mid-air haptic scenarios.


This illustrates that the acoustic force experienced by the skin to create a haptic effect is determined by the energy density within the sound wave and not the linear pressure.


Traditional ultrasonic mid-air haptic devices control linear sound pressure from a phased array of ultrasonic transducers to generate a haptic effect. While this is effective in creating a haptic effect, it does not reflect the basic physics occurring in the real system. Methods are presented herein to instead control the energy of the ultrasonic sound field in order to precisely control the physical force experienced by the user.


Energy density within an acoustic wave in a volume element involves both the particle velocity and pressure (Fundamentals of Acoustics 4th ed. Kinsler et al, eq (5.8.7)),








E

(
t
)

=


1
2




ρ
0

[





"\[LeftBracketingBar]"


u

(
t
)



"\[RightBracketingBar]"


2

+


(


p

(
t
)



ρ
0


c


)

2


]



,





where ρ0 and c are the density and speed of sound of the fluid, u is the average particle velocity within the element, and p is the pressure. For a monochromatic plane wave, the time-averaged energy density can be reduced to,











E
_

=


P
2


2



ρ
0



c
2




,




(
1
)








were in this case P represents the linear amplitude of the acoustic pressure wave. This quantity is also referred to as the acoustic intensity or I. This equation is valid if the radius of curvature of the acoustic wave is much larger than the wavelength. It is clear then if we wish to specify the acoustic radiation pressure in a mid-air haptic device, we should be solving for a field proportional to P2, not linear pressure.


Constructing the correct phase and amplitude for each transducer in an array (also called the activation coefficients) for a particular field solution can be done in many different ways. One such method is covered in Long et al. Rendering Volumetric Haptic Shapes in Mid-Air using Ultrasound. ACM Transactions on Graphics (Proceedings of SIGGRAPH Asia) Vol 33, Num 6, Art 181, 2014. Most involve first estimating the contribution of each transducer to a point or area of interest. The most common solutions for an ultrasonic phased array involve acoustic pressure due to the simplicity of both measurement on a real device with a pressure microphone and constructing a mathematical model to match that measurement. Using equation 1, a standard pressure model can be converted to acoustic intensity and utilize the same field solution techniques.


In addition to intensity, another quantity which is important in mid-air haptics is the direction of propagation of the energy. For an acoustic point source, this direction is directly away from the source. Real sources deviate from this in the near field (closer than a few wavelengths) and approach point-source behavior in the far field. Both cases can be covered by careful measurement and appropriate modelling.


To then solve for the nonlinear pressure in a specific direction, the contribution of each transducer is corrected by constructing the 3-dimensional vector dot product between the propagation direction and the desired direction of energy. This will appropriately adjust the contribution from transducers to reflect the amount on energy they can contribute in the desired direction. This allows, for instance, the ability to direct nonlinear acoustic pressure at the normal of a user's palm, thereby maximizing the perceived haptic.


Another nonlinear quantity which is proportional to P2 is parametric audio. This is the production of audible sound through nonlinear mixing of inaudible ultrasound in the air. Berktay (1974) derived the first-order contribution of two aligned plane waves to be (in the far field),







p
=




ω
2



P
2


β

S


4



πρ
0



c
0
4


R





exp
[


-

(

α
+
jk

)



R

]



α
T

+

jk
(

1
-

cos

(
θ
)







,





where P is the input ultrasonic level of each plane wave (in this case equal, but could be separated), S is the cross-sectional area, R is the distance from the array, and β is the nonlinearity coefficient. It can be shown that even for converging acoustic fields that there still exists a significant contributing term similar to the one above, and also proportional to P2.


To illustrate that the P2 solution has been correctly implemented, its effect in the resulting parametric audio can be measured. FIG. 1 shows the measured audio output at 50 cm from a square, 256 element phased array projecting a focus point centered and directly normal to the array at 1.9 meters distance for a given modulation frequency. At this distance, the focus point does not have enough nonlinear pressure to create a mid-air haptic, but the nonlinear contribution to parametric audio builds constructively over distance and becomes easily measurable with an adequate microphone. By constructing a modulation envelope proportional to the intensity (and including only the contribution from each transducer in the normal direction) this scheme is able to inject more energy into the nonlinear term—in this case the parametric audio. This extra ˜3 dB in audio reflects a similar gain in nonlinear force when projecting a focus point onto a hand at shorter range. More haptic strength and the resulting headroom allows for more efficient use of transducers relative to the standard pressure solve. This can be utilized to create a wider variety of haptic effects or reduce the number of transducers (and therefore the cost) for a haptic generated by the traditional pressure solve.


Turning to FIG. 1, shown is a graph 100 entitled Modulated Parametric Audio Response. The x-axis 110 is frequency in hertz and the y-axis is sound pressure level in db. The dotted line 140 is a plot of pressure solve and the solid line 130 is the I dot n solve. FIG. 1 shows that measured SPL of baseband audio at 50 cm from an ultrasound phased array when modulated at that frequency. The array solution sets a focus at 1.9 m the direction normal to the array and centered in the perpendicular plane. The dashed curve 140 is modulated by a pure sine wave offset so that the amplitude of modulation remains positive. The solid curve 130 instead sets the intensity dotted with a vector normal to the array to the same offset modulated tone, per transducer. This results in higher output as the intensity in the normal direction is a better approximation of nonlinear contributions to the field and the resulting parametric audio.


Turning to FIG. 2, shown is a graph 200 entitled THD (total harmonic distortion). The x-axis 210 is frequency in hertz and the y-axis 220 is THD in percent. The dashed curve 240 shows the pressure solve and the solid curve 230 is the I dot n Solve. FIG. 2 shows that distortion percentage (measured input frequency magnitude divided by the sum of the first 10 harmonics) versus input frequency of the same setup as in FIG. 1. The pressure solve reveals higher distortion at lower frequencies which is inherent to that modulation technique. Solving for intensity dotted by a normal vector normal to the array represents a better approximation of the nonlinear contribution to the field. At higher frequencies, phase shifts inherent to the mechanical nature of the transducers begin to mitigate the benefit.


The P2 solution also has an effect on distortion products for any modulation frequency. The traditional linear pressure solving results in a modulation scheme which looks like,

P=[0.5+0.5 cos(ωdt)] cos(ωct),

where ωc is the ultrasonic carrier frequency and ωd is the baseband modulation frequency. The estimated frequencies present in parametric audio production can be estimated by squaring the ultrasonic pressure and then omitting any term still ultrasonic. For the linear modulation case this is,

P2∝cos(ωdt)+0.25 cos(2ωdt).


The second term, double that of the input modulation frequency, is not present in the input signal and represents measurable distortion. Higher-order factors also exist in a real system but will be greatly diminished relative to this term. A measurement of this is given in FIG. 3.


Turing to FIG. 3, shown is a graph 300 entitled Distortion Percentage Pressure Solve. The x-axis 310 is frequency in hertz. The y-axis 320 is in percent. The solid line 330 is THD, the sum of the first 10 harmonics. The other curves show the first 3 harmonics. The dashed line 340 is order 1, The dot-dashed line 350 is order 2. The dotted line 360 is order 3. As expected, the overall harmonic distortion is dominated by the first order.


A P2 modulation solution instead looks like (with slight variations per transducer if using only the contribution in the normal direction),

P=√{square root over (0.5+0.5 cos(ωdt))} cos(ωct),

and the resulting parametric audio is estimated by,

P2∝cos(ωdt)cos2ct)∝cos(ωdt)+ . . . ,

where all the terms in the “ . . . ” are ultrasonic. The only term in the audible spectrum left is the input modulation. This is illustrated by the lack of first-order dominance in the distortion shown in FIG. 4.


Turning to FIG. 4, shown is a graph 400 entitled Distortion Percentage I dot n Solve, or distortion percentage of intensity dotted with a normal vector normal to the array versus modulation frequency. The x-axis 410 is frequency in hertz. The y-axis 420 is percent. The solid line 430 is THD, the sum of the first 10 harmonics. The other curves show the first 3 harmonics. The dashed line 440 is order 1, The dot-dashed line 450 is order 2. The dotted line 460 is order 3. Compared to pressure solve, the distortion is reduced and is composed of many orders, not dominated by the first.


The benefit of lower distortion for mid-air haptics is twofold. First, producing a nonlinear force more representative of the input allows gives a haptic designer more control of the haptics. A given force versus time profile that has been proven in another system, say a real button or contact-based haptic actuator for instance, can be more faithfully applied and reproduced by a mid-air haptic device employing this invention. Second, an ultrasound-based mid-air haptic device will always make parametric audio, even when only haptics are desired, and by reducing distortion, the array is able to contain the modulated frequency spectrum to a much narrower range. As illustrated in FIG. 1, parametric audio conversion from ultrasound to sound is more efficient at higher frequencies, while mid-air haptics are effective at lower frequencies (typically less than 200 Hz). The more the nonlinear modulation terms can be kept low, the less parametric audio will be produced.


Additional disclosure includes:


1. A mid-air haptic device consisting of:

    • A. A set of transducers with known relative locations.
    • B. Generating a plurality of ultrasonic waves from the transducers having at least one common focal point.
    • C. A desired haptic force versus time.
    • D. Modulating the generation of the ultrasonic waves so that the nonlinear acoustic force generated at the common focal point is the desired haptic force versus time.


2. A method as in paragraph 1 which includes a direction of the desired force.


3. A method as in paragraph 2 where the modulation of the ultrasonic waves substantially generates the desired nonlinear acoustic force in the desired direction.


II. Matrix Summation for Mid-Air Haptics as a Service

While matrix summation for mid-air haptics requires more bandwidth to achieve, the scalability and re-configurability of such a system is highly desirable in many real-world commercial applications which makes this an important architectural step.


To achieve this, the basis function computation remains entirely on the hardware side and the information is obtained from the hardware devices and the computation completed in an abstract form. Due to the nature of the computation and the latencies required for the system to be acceptable in the field of mid-air haptics, it is expected that this system would be implemented using a high speed interconnect to begin with, but that next generation wireless systems may well provide sufficient bandwidth at low enough latencies to be able to achieve this with remote compute facilities.


For this to function, a low-latency query-response system must be created between a controller device and the individual hardware devices which may physically include transducing elements. This may be implemented over a general computer network. The high-bandwidth network systems required for the connectivity of these hardware devices may be described by, but not limited to embedded low-voltage differential signaling (LVDS) communication channels, high speed wired networking systems or next generation wireless networks.


The first query response required is the matrices which are the number of control points square for given queried acoustic properties. These properties will be one or more of: acoustic pressure, acoustic particle velocity of the medium (which would generate potentially separate matrices in x, y and z velocity directions) and particle velocity in a given direction vector. These matrices are complex-valued. Metadata may be included with the request to configure the transducer field evaluation to allow for both diverging and converging focused scenarios or Bessel beam generation for example.


These matrices, once received by the controller may summed to allow multiple hardware devices to be controlled at once, wherein the hardware devices that are to be used are present in the summation. At this point, either the controller may take the data describing the control point properties such as intended phase (if applicable), intended amplitude and waveform dynamic range and process it into a solution vector for the matrix. This may be achieved on the controller device, or it may be exported to another system for processing.


Once the solution vector has been obtained it can be pushed to each hardware device for creation via the hardware available. However, a second mode is also necessary that runs a second speculative solution vector and returns the power level required of the worst case transducer to drive the speculative solution vector. This may be used to inform a limiting system to ensure that waveforms are always smoothly reproduced on each device.


On each device the basis function definition may include, but is not limited to, position of control point in x, y and z, normal to the control point with x, y and z components (defining the Poynting vector direction necessary to know the action direction of the acoustic radiation force vector), a time windowing function, which may be defined by a center time and a time radius, a complex-valued apodization coefficient.


A simulation query may be posed to the device in the form of the quantification of the acoustic quantities of a point, one or more of: acoustic pressure, acoustic particle velocity of the medium (which would generate potentially separate values in x, y and z velocity directions) and particle velocity in a given direction vector. These quantities are complex-valued.


Other queries may be posed to the individual hardware system to gather data on local conditions, such as for example temperature, humidity and air pressure. This data may then be amalgamated by the controlling device before making decision on what frequencies, wavelengths and period to use (and which are expected at the location of each hardware, and making decisions on which hardware devices to use).


A. Mathematical Background to the Acoustic Phased Array Problem


Writing this in mathematics, αqj) may be used to describe a complex-valued scalar linear acoustic quantity α measured at a position offset from the transducer element q by the translation vector χj, which may evaluate to be acoustic pressure or an acoustic particle velocity in a direction chosen for each j, the matrix A may be written:







A
=

[





α
1

(

χ
1

)








α
N

(

χ
1

)


















α
1

(

χ
m

)








α
N

(

χ
m

)




]


,




As this is matrix A is not square, and the degrees of freedom number more than the constraints, this is termed a ‘minimum norm’ system. It is ‘minimum norm’ because as there are infinitely many solutions, the most expeditious solution is the one which achieve the correct answer using the least ‘amount’ of x—the solution x


with minimum norm. To achieve this, some linear algebra is used to create a square system from the minimum norm system Ax=b:

AHAx=AHb,
(AHA)−1AHAx=x=(AHA)−1AHb,


This AHA is now N columns by N rows and given that the number of transducer is often very large this is an equivalently large matrix, and since any solution method must invert it, with it, this is not an efficient method. A more accessible approach is to create a substitution AHz=x, before applying a similar methodology:

Cz=AAHz=Ax=b,
z=C−1b=(AAH)−1b,


This time around, as C=AAH is a mere m columns by m rows, this result is a much smaller set of linear equations to work through. The vector z can be converted into x at any time so long as AH can be produced.


However, this does not end here. This approach is not just a fortuitous set of symbolic manipulations, the change of variables from the complex-valued vector that describes the drive of individual transducer elements x to the much lower dimensional z has further meaning. Each complex-valued component of z can be viewed as a complex-valued drive coefficient that pre-multiplies a focusing function which generates a focus from all of the individual transducer fields, wherein the focal point is co-located with each individual control point. For in control points therefore, there are in such focussing functions, and they can be viewed as defining a complex vector space custom characterm where points in this space correspond to possible configurations of these in ‘focus points’.


B. Splitting the C Matrix


To account for the possibility that the optimal minimum norm solution is not used to form the C matrix, then this may be expressed via an extra weighting for each transducer and control point as σr,q, with r representing the control point index and q representing the transducer index—this can be viewed as reweighting the final x vector used as the excitation vector for the transducer elements by substituting BHz=x, where:







B
=


σ
×
A

=

[





σ

1
,
1





α
1

(

χ
1

)









σ

1
,
N





α
N

(

χ
1

)



















σ

m
,
1





α
1

(

χ
m

)









σ

m
,
N





α
N

(

χ
m

)





]



,





and × here denotes component-wise multiplication. Defining the set of

αc=[αac), . . . , αqc), . . . , αNc)],

and

βc=[σc,1α1c), . . . , σc,qαqc), . . . , σc,NαNc)],

this adjusted C matrix may be expressed as:







C
=

[





α
1

·


β
1

_









α
1

·


β
r

_









α
1

·


β
m

_

























α
r

·


β
1

_









α
r

·


β
r

_









α
r

·


β
m

_

























α
m

·


β
1

_









α
m

·


β
r

_









α
m

·


β
m

_





]


,





but the dot products for each element may be written as the summation where for instance

αa·βbq=1N=1αqa)σb,qαqb).


If there are multiple devices that have access to disjoint sets of transducer elements, so if there exist M devices such that the global transducer set may be numbered q∈{1, . . . , N1, N1+1, . . . , N2, . . . NM=N}, where device 1 drives transducers 1, . . . , N1, device 2 drives transducers N1+1, . . . , N2 and device M−1 drives transducers NM−1+1, . . . , N, then each dot product in the matrix C may be written:








β
𝒶

·

=


(




q
=
1


N







α
q

(

χ
𝒶

)



σ

𝒶
,
q




)

+

+


(




q
=


N

M
-
1


+
1


N





α
q

(

χ
𝒶

)



σ

𝒶
,
q




)

.







This implies that the C matrix itself may be written in a per-transducer element form as:







C

tx
,
q


=




[





α
q



(

χ
1

)



σ

1
,
q






α
q

(

χ
1

)

_










α
q

(

χ
1

)



σ

1
,
q






α
q

(

χ
r

)

_










α
q

(

χ
1

)



σ

1
,
q






α
q

(

χ
m

)

_

























α
q



(

χ
r

)



σ

r
,
q






α
q

(

χ
1

)

_










α
q

(

χ
r

)



σ

r
,
q






α
q

(

χ
r

)

_










α
q

(

χ
r

)



σ

r
,
q






α
q

(

χ
m

)

_

























α
q



(

χ
m

)



σ

m
,
q






α
q

(

χ
1

)

_










α
q

(

χ
m

)



σ

m
,
q






α
q

(

χ
r

)

_










α
q

(

χ
m

)



σ

m
,
q






α
q

(

χ
m

)

_





]

,







yielding:






C
=





q
=
1

N



C

tx
,
q



=


(




q
=
1


N
1




C

tx
,
q



)

+

+


(




q
=


N

M
-
1


+
1


N



C

tx
,
q



)

.







This implies that the C matrices for individual transducers may be collected together by a recursive or hierarchical process that exploits sum-reduction operators to construct successively more complete representations of a distributed system of transducers, to be solved in a single central location (or the computation repeated in distributed locations for better fault tolerance). However, as the matrix B is required to take the z vector produced and reconstruct the transducer excitations, it is also necessary to express B as a disjoint set of matrices, where the B matrix for a given transducer q may be written:








B

tx
,
q


=

[





σ

1
,
q





α
q

(

χ
1

)













σ

m
,
q





α
q

(

χ
m

)





]


,





so the element of the excitation vector corresponding to transducer element q is written:

Btx,qHz=xq.


Therefore, since each portion of the x vector may be stored locally as it is only necessarily required to drive the transducer elements, no information regarding individual transducer elements or their weightings needs to be communicated globally to obtain the transducer excitations—only the C matrix for each subset of the system is required to be communicated. As the C matrix is small, so long as latencies are carefully managed, the platform on which the C matrix is solved to generate the z vector may be chosen flexibly, which may include but is not limited to edge computing, mobile devices, remote servers etc.


Incarnations of the A matrix may also be constructed to generate linear acoustic quantities, such as the acoustic pressure or the particle velocity of the acoustic medium in a known direction. As these are linear quantities, they may be calculated by applying disjoint portions of the A matrix and transmitting only the computed quantities. Then via a similar sum-reduction process to that required to compute the C matrix for the system, any given acoustic quantity may be calculated from synchronised application of portions of x vectors, as any linear quantity α may be found as:








A

tx
,
q


=

[





α
q

(

χ
1

)












α
q

(

χ
m

)




]


,









α
'


Ω
,
q


=


A

tx
,
q




x
q



,





where {acute over (α)}Ω,q is the contribution of the transducer q to the final simulated linear acoustic quantity field. This is useful to give an added layer of control and certainty to the system to allow the controlling device to accurately gauge the capabilities and constraints of the platform or platforms with which it may be communicating.


III. Null Regions in Airborne Ultrasonic Fields

Various excitation solvers exist to generate an arbitrary acoustic field considering specific points/lines/planes/volumes. Many of these allow the possibility of specifying the desired pressure distributions. By placing a null point (a point with desired pressure of zero or a relatively small value), a null line, a null plane or a null volume at sensitive locations, we can mitigate the amplitude of ultrasound found there.


A. Null-Points, Surfaces, and Volumes


In one embodiment, we can use a mathematical model of each transducer to build a matrix A. In this matrix, each column corresponds to a unique location in space and each row is the transducers complex acoustic output at that location. The mathematical model can be in real units or units which correlate to each transducer's output. Using this design, we can define the problem of solving for each transducer's drive phase and amplitude to the following formula,

Ax=b,

where x is a column vector containing the complex activation of each transducer, and b is the desired complex output at each physical location defined in the columns of A. This is an under-determined linear system, with many possible values of x which yields b.


One solution to this system of equations is given by the Moore-Penrose inverse of A, denoted by A+, with the solution explicitly given by,

x=A+b.


This is a least-squares solution to the linear system and has many beneficial properties useful to ultrasonic arrays which can include minimizing the total magnitude of activation. This approach has been applied to create dynamic high-pressure points for mid-air haptics to good effect (Long et al. Siggraph 2014). But this methodology is more flexible that simply creating high-pressure regions—each entry in b can be an arbitrary value, both high or low pressure and the Moore-Penrose inverse will yield a solution which approximates that field. By setting some values of b to substantially small values and we can create ‘null’ points in the field.


In one embodiment of this invention, we use the Moore-Penrose activation solution with null points placed near sensitive items/animals/people to mitigate any possible effects caused by ultrasound.


Placing multiple null points around a sensitive object can further decrease the pressure at that object. In one illustrative arrangement, placing multiple null points in a circle or plus shape (i.e. +) around the central null can enlarge this region.


Turning to FIGS. 5A, 5B, and 5C, shown are simulations 500 of using the null-point solver to create a low-pressure area in the acoustic field. The Figures show 3 ultrasonic phased array simulations using 40 kHz ultrasound from a 256-element array in a square rectilinear arrangement centered at the origin and oriented along the z-axis. To solve for the activation of each transducer, a simple vibrating disk model is used which matches experimentally measured transducer output. Activation coefficients are solved by a power iteration for optimal phase followed by the Moore-Penrose inverse. This algorithm is detailed elsewhere (Brian Kappus and Ben Long, Spatiotemporal Modulation for Mid-Air Haptic Feedback from an Ultrasonic Phased Array, ICSV25, Hiroshima, 8-12 Jul. 2018).


In FIG. 5A, no nulls are used. Shown is the x-z plane 510A, the x-y plane 510B (shown at z=40 cm), and the arbitrary pressure scale 5100. A focus point is placed at [x,y,z] cm=[0,0,+20] where the origin of the coordinate system is in the centre of the array and the array is oriented in the positive z direction.


In FIG. 5B, there is one null at x=−5 “cm”, y=0 “cm”, z=40 “cm”. Shown is the x-z plane 520A, the x-y plane 520B shown at z=40 cm, and the arbitrary pressure scale 520C. A single null is generated at [−5,0,+40] and the pressure is reduced at the point from 200 Pa to <20 Pa.


In FIG. 5C, 5 nulls are in the z=40 “cm” plane, the first at x=5 “cm”, y=0 “cm”, z=40 “cm” and the other 4 offset by 1 “cm” in +x, −x, +y, and −y, forming a ‘plus’ motif. This achieves a larger effective nulled region. Shown is the x-z plane 530A, the x-y plane 530B (shown at z=40 cm), and the arbitrary pressure scale 530C. This shows the effect of using multiple null points to achieve a larger null region. In this case the null points are a cluster of 5 points, one centred at the desired null, with the other 4 forming a ‘plus’ offset by 0.1 cm. This achieves a lower pressure null at the center and a larger <20 Pa null region. This could be used if the object is larger or there is uncertainty in its position.


In all cases the focal point (at [0,0,+20]) amplitude is substantially unaffected. In this example, the focus point could be used for airborne haptic feedback while shielding a device such as a sensitive microphone at the null region.


More complicated arrangements are possible. Turning to FIG. 6, shown is a simulation 600 having a cluster of 4 focus points at [+5,0,+30], [−5,0,+30], [0,+5,+30], [0,−5,+30]. These could be modulated for haptics or parametric audio. Another simulation example of using the null-point solver on a more sophisticated acoustic field to achieve a null. In this case the field is creating 4 high-pressure points at x,y,z (“cm”)=[+5,0,+30], [−5,0,+30], [0,+5,+30], [0,−5,+30]. The plane of interest is at z=40 “cm”.


In FIG. 6A, shown is the x-z plane 610A, the x-y plane 610B (shown at z=40 “cm”), and the arbitrary pressure scale 610C. No nulls were used.


In FIG. 6B, shown is the x-z plane 620A, the x-y plane 620B (shown at z=40 cm), and the arbitrary pressure scale 620C. One null point is at [0,0,+40].


In FIG. 6C, shown is the x-z plane 630A, the x-y plane 630B (shown at z=40 cm), and the arbitrary pressure scale 630C. This has 5 nulls in a ‘plus’ arrangement with 3 “mm” spacing centered at the [0,0,+40]. The cluster of 5 nulls is able to create a larger low-pressure region. As before, adding multiple points increases the size of the null region without significantly affecting the high-pressure field.


The specific set of nulls presented above represent a single possible solution. Many other solutions exist. Spacing of several null points around a target location can be dependent on the wavelength of sound being used. In a system with the ability to change its carrier wavelength, null points can be adjusted accordingly. Possible volumetric arrangements of nulls include, but are not limited to, rectilinear, hexagonal, regular polygon, and regular polyhedra.


In addition to clustering the null regions directly atop the target location, in another arrangement, the nulls are placed in between the array and the target. This creates a ‘shadow’ region which extends beyond the null point location. In another arrangement, the null points can be placed around the target location. Both methods can yield a larger effective low-pressure region using fewer nulls depending on the array layout and placement.


Possible null arrangements including clustering around the target location as well as shadowing the target are shown in FIGS. 7A-7D. FIGS. 7A-7D show 3-D graphs 1300 as further examples of null point arrangements to create a low pressure (null) region in space around an array of transducers.


In FIG. 7A, the graph 1310A shows a grouping of transducers with a focus 1310C above the transducers. Also shown is an example of cubic packing of nulls points 1310D around a desired null region 1310B.


In FIG. 7B, the graph 1320A shows a grouping of transducers with a focus 1320C above the transducers. Also shown is an example of spherical packing of nulls points 1320D around a desired null region 1320B.


In FIG. 7C, the graph 1330A shows a grouping of transducers with a focus 1330C above the transducers. Also shown is an example of hexagonal packing of null points 1330D around a desired null region 1330B.


In FIG. 7D, the graph 1340A shows a grouping of transducers with a focus 1340C above the transducers. Also shown is a rectilinearly-pack plane of null points 1340D in between the array and a desired null region 1340B.


Other null arrangements with more complex geometry are also possible. Examples of possible strategies, include but are not limited to; packing spheres with pre-defined uniform distributions of null points (whose number may change over time), using phyllotactic spirals or other spiral generated using irrational angular increments to generate sphere surface packings which depending on scaling factor may be of uniform or non-uniform density around a locus, generating meshes of the desired volume to be filled (which may be predefined or take input from a tracking system or meshed point cloud data) and using charged particle-like repulsion mechanisms to generate packings of these meshes or other volumetric shapes, which may be static or dynamic or wherein the packing density may be made non-uniform by varying the ‘charge’ on each null point which may further be parameterized as a function of the spatial location of the null point to allow for non-uniform densities. The mesh data taken from the tracking system may be dynamic and in each case the number of null points may be changed dynamically over time.


If a number of null points are to be introduced into the acoustic field their physical separation and their number can be varied depending on the required ‘quiet zone’ dimensions. The ‘packing arrangement’ has an influence on the effectiveness of using nulls to suppress the average sound pressure within a given volume. A metric may be defined which estimates this effectiveness. An example metric is the percentage of a given volume below a threshold sound pressure. FIG. 8 illustrates this effect through a simulation of a 256 element ultrasonic array with Moore-Penrose inverse solution producing a high-pressure point at z=15 cm and a set of nulls clustered around a target at [x,y,z]=[30,0,30] cm. It shows that the number of nulls, their physical spacing and their spacing geometry influence the predefined metric. Indeed, increasing both spacing and number arbitrarily can have a negative effect on the volume which has the desired low pressure and care must be taken to test a choose a proper arrangement of null point locations. Our example uses a 6λ3 volume as the assessment volume, however the assessment volume and/or geometry, will be application specific.


Turning to FIGS. 8A and 8B, shown are a series of graphs 1400. The metrics were created with the percentage of a 6λ3 volume (centered at x=−30 “cm”, y=0 “cm”, z=+30 “cm”) with pressures greater than 6.3 “Pa” with the number of nulls and their spatial separation being varied. A single focal point was created at x=0 “cm”, y=0 “cm”, z=+15 “cm” at a pressure of 1125 “Pa”.



FIG. 8A shows a plot 1430 having a rectilinear 3D spacing arrangement showing the relationship between quiet zone efficiency (i.e. percentage of volume below a threshold pressure). The x-axis 1410 is null point separator/λ. The y-axis 1420 is a null number. The key 1435 is percentage of volume below a threshold pressure. Shown is the relationship between quiet zone efficiency (i.e., percentage of volume below a threshold pressure) for a 3D spacing arrangement.



FIG. 8B shows a plot 1460 having a hexagonal 3D spacing arrangement showing the relationship between quiet zone efficiency (i.e. percentage of volume below a threshold pressure). The x-axis 1440 is null point separator/λ. The y-axis 1450 is a null number. The key 1470 is percentage of volume below a threshold pressure.


An important consideration for the use of null pressure points is their effect on the general acoustic field, specifically the control points used for applications such as mid-air haptics. A key parameter for focused ultrasound is the focal point pressure. FIG. 9 is a graph 1500 that shows that effective quiet zone may be created without a loss of focal point pressure. The x-axis 1510 is null point separator; k. The y-axis 1520 is a null number. The key 1535 is in db. The plot 1530 shows the pressure of a single focal point created at x=0 “cm”, y=0 “cm”, z=+15 “cm” with a requested pressure of 1125 “Pa”. Each point in the figures represents a specific combination of null point number and null point separation.


As with low-pressure volume, some arrangements can reduce performance by influencing focus pressure and again, simulation and/or careful selection must be implemented for maximum performance. When the number of null points is generally less than the number of transducers, the desired focal point pressure is unaffected by the inclusion of null points. When a relatively large number of null points is used the desired focal point pressure may be lowered.


In another arrangement, experimental measurement of a particular setup can be used to improve null performance. With at least one microphone placed in a target null location, this can be used to evaluate performance of null placement. Null point placement and desired pressure can be adjusted, and microphone measurements used as a performance metric. Any number of search algorithms including gradient of least descent can be used to approach an experimentally optimal solution. In addition, at least one microphone can be placed at various high-pressure focus points or regions to evaluate the influence of null placement on focus performance simultaneously.


In one application of this invention, nulls can be used to shield sensitive microphones from undue ultrasonic pressure. In another arrangement of this invention, at least one of the sensitive microphones can transmit its signal or some metric related to the amount of ultrasound received back to the ultrasonic array or device controlling the array. In this way, this signal can be used as feedback to dynamically adjust null placement for optimal reduction of ultrasonic pressure at the sensitive microphone.


It is important to place nulls in the appropriate location for a given target. If the target is moving, such as a person's head or hand-held devices, tracking that target is important. Adjusting the null point locations according to tracking data can maintain the effectiveness of the null region with a minimal amount of null points needed. In one arrangement, the target location can be added to a matrix of null location vectors resulting in a null arrangement which follows the target. For a shadow arrangement, the tracked target location can be both a distance and normal to build a set of null point locations. In another arrangement, the tracked target could move some nulls but not others, leaving some null regions were expected new targets could arise. Multiple tracked targets can exist which could dynamically add, subtract, and move null points, surfaces, or volumes, as necessary.


Tracked objects need not be identified as sensitive in order to add a null region. In one arrangement, any object in the field which is moving at a speed greater than some threshold, e.g., 1 cm per second, that is not identified as a hand (for haptics), would be identified as a target for a null region. This reduces the tracking requirements to hand-identification, and movement tracking—significantly simpler than identification of specific sensitive targets such as heads, and/or microphones. Once an object has been classified as a target, through movement, it can be tracked even if it stops moving. In another arrangement, target objects identified through movement will continue to be considered targets until they have left a pre-specified volume of interaction.


In another arrangement of this invention, the null positioning can be generated by first simulating the field created by the desired high-intensity focus points without any nulls included. Away from the focus points, where low pressure is desired, ideal placement of nulls can be generated using the simulated information. As an example, nulls can be placed preferentially on the highest-pressure peaks of the fringing field in the target low-pressure region. In another arrangement, the simulated null-free field can be used as a weighting function to adjust the exact placement of null points.


Field simulation can also be used to refine placement of null points. For instance, with sufficient computing power available, different arrangements of nulls can be evaluated for effectiveness at each update cycle of the phased array with only the most effective being passed on to array control for emission. For example, for a given field one can define a set of metric-points where the field is evaluated. These can be at desired null points, desired high-pressure points, or any other important point/line/plane/volume in the field. At each field-compute cycle, an initial position for both high-pressure focus points and nulls are selected (this could be based on the previous cycle or completely new), then the field is evaluated at the metric-points and related to a quality metric. This metric could be simply the absolute difference in simulated pressure from desired pressure, squared pressure or something more sophisticated such as the nonlinear pressure or particle velocity. Next, the null and high-pressure points are adjusted and the field and resulting quality metric is evaluated again. One method of computing this adjustment of the null point positions may be achieved by computing the partial derivative of the quality metric with respect to the change in the spatial positioning of the nulls along the spatial axes. When this refinement achieves an acceptable metric or a maximum number of evaluations has happened, then the best solution is passed on to be produced. Adjusting the positions can be static (a set number of possibilities) or adaptive, updating the null point locations as the field changes over time. As the acoustic field is band-limited in spatial frequency, all partial derivatives may be computed by a finite differencing scheme that operates at sub-wavelength scales. This implies the adaptive method could, for instance, test a small adjustment in each direction for each point and then proceed along the gradient of steepest descent.


In another arrangement of this invention, null surfaces or volumes may be considered instead of points. In this setup, after a null surface or volume is chosen its effect on the transducer activation coefficients can be estimated and applied. In one arrangement, the field is simulated without the null surface or volume. Next, pressure at the surface or within the volume is considered phase inverted. Next, this inverted pressure volume or surface is simulated as a source and its resulting pressure is propagated back to the array. This simulation can be accomplished with traditional discrete or continuous methods. This will yield a set of acoustic pressures and phases at each transducer location. These values can then be used as a partial or full perturbation of the original transducer activation coefficients. Applying this perturbation can be done by direct, sum, partial sum, weighted sum, or some other solution method.


The above simulations consider the regions to be within an infinite volume without reflections. While this is often a good approximation in large, open spaces, in more enclosed spaces, reflections of the ultrasound become a significant contribution to the pressure in the desired null region. If the area around the array is carefully measured, such as through 3D scanning, this information can be used to inform a simulation of the area. With reflections included in an acoustic simulation, this can be used to modify the basis functions (also called transducer model) on a transducer-by-transducer basis, or to a repeated transducer basis function. When included in the transducer model, the Moore-Penrose inverse of the matrix A will be able to compensate for reflections. In another arrangement, the basis functions can be left unmodified and instead null point locations and/or amplitudes can be adjusted based upon an output simulation which includes reflections. By including tracking information, such as a 3D point cloud from a stereo camera or time of flight sensor, the simulation of reflections can be further refined by dynamically updating the current state of the interaction volume.


In another arrangement, instead of purely simulating reflections, experimental data can be taken to characterize the environment for a given array placement and this data included in transducer models or field simulations. This could be measurements of an empty interaction volume or many measurements which include intended use case scenarios such as people standing in front of the array with their hand out. In one arrangement, the volume of interaction is scanned by a microphone and a transducer model is fit to this data out to some interaction distance. Since this measurement includes reflections, the model will more accurately reproduce nulls.


Null placement, amplitude, and phase can be computationally difficult depending on the number of factors included in the modelling. Including dynamically-changing reflections from people moving around the interaction volume, in particular, is difficult. Machine learning is particularly suited to these types of problems and can be used to reduce computational complexity. In one arrangement, a supervised-learning scenario can be built using microphones as feedback. Any number can be placed in the environment, including at null target locations, and non-zero pressure locations. In addition, information about the environment can be used as input including, but not limited to, a dynamic point cloud of the environment. A neural net is then setup to output null point locations and target phase and amplitude. Training this setup would involve capturing the microphone output for a wide array of zero and non-zero-point placements and drive conditions with a variety of environment conditions. Note that this could be real-life microphone and point-cloud data or purely simulated data. This trained machine-learning system could then output best-guess null point arrangements dynamically for significantly less computational power than a full acoustic simulation.


Turning to FIGS. 10A, 10B, and 10C shown are spectrograms 1600. For each spectrogram, the x-axis is time; the y-axis is frequency; and the darker the color, the more the signal FIG. 10A shows human speech 1610 recorded with a MFMS microphone without ultrasound present. FIG. 10B shows human speech 1620 recorded with high-intensity ultrasound revealing that the microphone is overloading and unable to pick up spoken voice. FIG. 10C shows human speech 1630 recorded with high-intensity ultrasound near a haptic focus point surrounded by spherically-packed null pressure points. Thus, without the null points reducing the pressure at the microphone, the speech recorded in FIG. 10B is unintelligible. While not identical to FIG. 10A, the speech in FIG. 10C is intelligible.


B. Helicity


Helicity in the context of acoustics is a traveling wave carrying angular momentum.


Helicity can be given to an acoustic field by adding a phase shift related to the angular position of each transducer,

X=X0eimϕ

where X0 is the original activation coefficient to form the desired field for a given transducer, ϕ is the angle of the transducer on the array (in radians) relative to an arbitrary starting axis, and m is the helicity. m imparts not only a net angular momentum but also can create a null in any region that before had a focus. The size of this region is related to the value of the helicity. Integer values of helicity are often better behaved with arbitrary fields but fractional m is also possible. As with null points, field simulation can be used to refine the value of the helicity dynamically.


This technique is valuable primarily in applications where the acoustic pressure is directed towards a sensitive object which is to be narrowly avoided such as parametric audio. The center of the beam or point can be placed directly on the sensitive target and by adding helicity, the target will receive a reduced amount of ultrasound.



FIGS. 11-14 show examples simulations of a 256 element ultrasonic array operating at 40 kHz using helicity. In FIGS. 11A, 11B, 11C, 11D, 12A, and 12B, helicity is applied to a beam or plane wave. The hole in the ultrasonic field created by the helicity follows the steering of the beam. In FIGS. 13 and 14, helicity is applied to a focused field. Like the beam solution, the null region formed by the helicity follows the focus point as it is steered. In both cases the null region is well defined and widened with increasing m and can be placed on a sensitive target.


Turning to FIGS. 11A, 11B, 11C, and 11D, shown are simulations 700 using helicity to create a null region in the center of a plane wave.



FIG. 11A shows the x-z plane 710A, the x-y plane 710B (shown at z=40 “cm”), and the arbitrary pressure scale 710C. This is formed from a basic plane from a 16 “cm” edge array of 40 “kHz” transducers.



FIG. 11B shows the x-z plane 720A, the x-y plane 720B (shown at z=40 cm), and the arbitrary pressure scale 720C. This is formed by adding helicity of m=1.



FIG. 11C shows the x-z plane 730A, the x-y plane 730B (shown at z=40 cm), and the arbitrary pressure scale 730C. This is formed by adding helicity of m=2.



FIG. 11D shows the x-z plane 740A, the x-y plane 740B (shown at z=40 cm), and the arbitrary pressure scale 740C. This is formed by adding helicity of m=3.


Turning to FIGS. 12A and 12B shown are simulations 800 using helicity to create a null region in the center of a plane wave. This is formed from a basic plane from a 16 “cm” edge array of 40 “kHz” transducers.



FIG. 12A shows the x-z plane 810A, the x-y plane 810B (shown at z=40 cm), and the arbitrary pressure scale 810C. No helicity is added in the steered beam.



FIG. 12B shows the x-z plane 820A, the x-y plane 820B (shown at z=40 cm), and the arbitrary pressure scale 820C. This is formed from a basic plane from a 16 “cm” edge array of 40 “kHz” transducers. Here, the helicity is added (m=2).


Turning to FIGS. 13A and 13B shown are simulations 900 adding helicity to open a null region in a steering focus point.



FIG. 13A shows the x-z plane 910A, the x-y plane 910B (shown at z=40 cm), and the arbitrary pressure scale 910C. Here, there is a single focus point at x=+5 “cm”, y=0 “cm”, z=+40 “cm”.



FIG. 13B shows the x-z plane 920A, the x-y plane 920B (shown at z=40 cm), and the arbitrary pressure scale 920C. Here, helicity is added to create a null in the center of the focus (m=2).


Turning to FIGS. 14A and 14B shown are simulations 1000 adding helicity to open a null region in a Bessel beam.



FIG. 14A shows the x-z plane 1010A, the x-y plane 101013 (shown at z=40 cm), and the arbitrary pressure scale 1010C. Here, there is a steered Bessel beam.



FIG. 14B shows the x-z plane 1020A, the x-y plane 1020B (shown at z=40 cm), and the arbitrary pressure scale 1020C. Here, helicity has been added to open a null in the center of the beam (m=2).


C. Array Division and Null Planes


Another method to create a null region is to divide an array into two or more similar regions and reverse the phase of one field relative to the other. When the fields are largely similar this forms a null line which projects through the middle of the field. This can be used when the null region has high precision in one dimension but not in the other. Division need not be directly down the middle but acoustic pressures produced from each region need to match. This means that regions with a larger number of transducers need to have their amplitude reduced to compensate. Angling each of the two beams apart slightly can increase the width of the null line.



FIGS. 15 and 16 show example simulations of 256 element ultrasonic array operating at 40 kHz using array division to create a null plane. In both cases, the null plane is well-defined and can be directed towards a sensitive target as desired.


Turning to FIG. 15, shown is a simulation 1100 with the x-z plane 1110A, the x-y plane 1110B (shown at z=40 cm), and the arbitrary pressure scale 1110C. Splitting an array in half with each driven out of phase results in opening up a null line along the y-axis.


Turning to FIG. 16, shown is a simulation 1200 with the x-z plane 1210A, the x-y plane 1210B (shown at z=40 cm), and the arbitrary pressure scale 1210C. Shown is angling the field each half of a split array away from the origin by 3 degrees to widen the null region.


D. Null Subspace


Consider an under-determined linear system,

Ax=b

where,

A∈custom characterm×n, x∈custom charactern×1, b∈custom characterm×1,

will have 0 or an ∞ number of solutions. This is the same starting point as previously discussed solving methods.


The LSQ, least squares solution, can be shown to be:








min
x





Ax
-
b



2


=



A
+


b

+
v






where, ν∈ker (A), and,

ker(A)={ν∈custom charactern×1|Aν=0}

is the nullspace of A.


Analogously to the previous section, we start by choosing a set of j null points at positions y′ and z′ which we shall denote,

χN={χN1, . . . , χnj}


We can then construct the complex-valued linear system ANxN=bN where:

ANcustom characterj×n, xNcustom charactern×1, bNcustom characterj×1, with







A
N

=

[





Ψ
1

(

χ
N
1

)








Ψ
n

(

χ
N
1

)


















Ψ
1

(

χ
N
j

)








Ψ
n

(

χ
N
j

)




]








x
N

=


[





A
1
emit



e

i


ϕ
1
emit















A
n
emit



e

i


ϕ
n
emit







]



and









b
N

=

[



0









0



]


,





where, as above, Ψm is the basis function or transducer model for each element in the array.


The vector xN is in the nullspace of the matrix AN,

xN∈ker(AN)


Now we can formulate the problem of finding the complex activation coefficients that produce the desired amplitudes at the control point positions and the desired amplitudes at the null point positions as a pseudo linear program:









min
υ







A
+


b

+
υ



2


=

x
N





where
,


υ


ker



(
A
)









subject to the constraint

(A+b+v)∈ker(AN)


This constraint specifies a new under-determined linear system to solve,

AN(A+b+v)=0


Once again, we appeal to the Moore-Penrose inverse to find the minimum-norm solution as,

v=−AN+ANA+b


The vector v can be thought of as a minimum perturbation required to shift the original solution vector A+b into the nullspace of AN, ker(AN). This perturbation ensures that the null point condition,

ΨΩNk)=0, ∀k∈{1, . . . , j}

is met with the minimum change in the original solution vector A+b.


We can now express the full solution as,

xN=A+b−AN+ANA+b

which can also be expressed as,

xN=A+b=ANH(ANANH)+ANA+b,

where the superscript II represents the Hermitian conjugate.


The advantage of this method is that, for a fixed arrangement of nulls, AN+AN does not change and can be calculated in advance. A+b can change, producing arbitrary fields including haptic points and parametric audio beams, and nulls can be added using the above equation to modify the activation coefficients before emission. In particular, arrangements of many nulls may be added to a system where traditional solving may be computationally difficult. This would allow for incorporation into low-end hardware which would otherwise not be capable of performing the necessary matrix inversion in the required time for low-latent response.


When a new solution (xN) is generated with the method, some entries may result in driving values impossible to realize, such as greater than the maximum drive possible by the device. One solution is to scale every value the same amount and preserve each phase value. This will scale the field by the same amount and preserve both non-zero values (scaled) and the nulls. Another option is to clip large values to the maximum drive and leave the remainder. This has the potential to influence both the non-zero points and the nulls and care must be taken to preserve the benefit of the null application.


We are not limited to a single AN+AN matrix in practice. Indeed, many different null arrangements (with their associated AZAN matrix) may be stored on the device and applied when appropriate. Care must be taken when switching between matrices to avoid audible artifacts caused by abrupt changes in activation coefficients. In one arrangement, two can be transitioned between by linearly interpolating each matrix value between each. In another arrangement, the output is reduced to near zero, the nullspace array switched, then the output is increased back to normal. In another arrangement, each transducer coefficient is low-pass filtered after the array is switched to remove high-frequency components likely to produce audio. This filtering would then be switched off or smoothly removed after some amount of time.


While the discussion to this point has focused on pre-calculation of the nullspace arrays, there is no reason new nullspace arrays cannot be calculated on a host or on the device. Depending on the number of nullspace points, this could take significant time compared to one acoustic cycle. Calculation need not be done all at once and could utilize unused processor, storing results in memory as calculation proceeds. When finished, the nullspace array could be introduced using previously discussed methods.


In another arrangement, it may be beneficial to apply the nullspace correction to the basis functions, Ψn, before taking the Moore-Penrose inverse. In this way, smooth transitions to the activation coefficients can be made as points are translated in space, thus minimizing potential acoustic artifacts such as parametric audio.


E. Iterative Nullspace Solution


The nullspace solution allows for a static arrangement of null points to be added to a system. With some additional computation, however, the method detailed above can be reformulated as an iterative update procedure, allowing for the addition of null points. Following the reasoning of the previous section we begin by specifying a single null point χN1 and construct the corresponding update:

AN=[Ψ1N1) . . . ΨnN1)]


The Moore-Penrose inverse of a rank-1 matrix can be expressed as:







A
N
+

=


1




k
=
1

n





Ψ
k

(

χ
N
1

)

·


Ψ
k
*

(

χ
N
1

)






A
N
H






The expression for the complex activation coefficients,

xN=A+b−AN+ANA+b

can now be expressed in the form,







x
N

=



A
+


b

-


1




k
=
1

n





Ψ
k

(

χ
N
1

)

·


Ψ
k
*

(

χ
N
1

)






A
N
H



A
N



A
+


b








Letting
,





k
=
1

n





Ψ
k

(

χ
N
1

)

·


Ψ
k
*

(

χ
N
1

)



=
C






and factorizing the above expression yields,








x
N

=


(


I
n

-


1
C



A
N
H



A
N



)



A
+


b


,





where In is the identity matrix of matching dimensions with A+b.


This procedure can be repeated for additional null points thus providing an iterative update scheme for adding null points to the field.


F. Additional Disclosure


Using nulls to shield microphones/ears is novel.


Many of the null refinement methods are novel including simulating and measuring reflections.


Using helicity specifically to create a low-pressure region to shield mics/ears is novel.


Splitting the array and driving each section out of phase to create a null region in between is novel.


Many-null arrangements are novel (various packing arrangements).


Null shadows (shielding a region by placing nulls in between the array and region) is novel.


The null subspace solving method is novel.


IV. Predistortion of Phased-Array Basis Functions Applied Via Transducer Gain Control

A. Reduced Representation—Using Per-Focus Basis Functions


Traditionally, the linear system is described in terms of linear combinations of complex-valued transducer generated fields and their drive coefficients. This produces a matrix, where for m control points and N transducers, the matrix A is N columns by m rows and consists of the generated complex valued signal by each transducer q∈{1, . . . , N} at the location of each control point j∈{1, . . . , m}. Previous work (1-US) has generated increased power efficiency by adding regularization to this matrix A, but regularisation increases the size of the matrix and thus the compute requirements to solve the system significantly.


Using αqj) to describe a complex-valued scalar linear acoustic quantity α measured at a position offset from the transducer element q by the translation vector χj, which may evaluate to be acoustic pressure or an acoustic particle velocity in a chosen direction, the matrix A may be written:







A
=

[





α
1

(

χ
1

)








α
N

(

χ
1

)


















α
1

(

χ
m

)








α
N

(

χ
m

)




]


,




This, for a number of control points fewer than the number of acoustically active transducer elements can then be placed into a complex-valued linear system wherein a sample vector b={αC11), . . . , αCmm)} represents the desired total linear scalar complex-valued acoustic quantity where the amplitudes are desired amplitudes of the acoustic quantity and the phases are those taken from the phase oracle (which may have been user-influenced). In this linear system described as Ax=b, the x vector is then the initial field coefficients for each transducer element, which may be used to drive a real transducer element, resulting in the recreation of the acoustic field desired. This may then be solved in a loop to provide a system that changes over time.


As this is matrix A is not square, and the degrees of freedom number more than the constraints, this is termed a ‘minimum norm’ system. It is ‘minimum norm’ because as there are infinitely many solutions, the most expeditious solution is the one which achieve the correct answer using the least ‘amount’ of x—the solution x with minimum norm. To achieve this, some linear algebra is used to create a square system from the minimum norm system Ax=b:

AHAx=AHb,
(AHA)−1AHAx=x=(AHA)−1AHb,


This AHA is now N columns by N rows and given that the number of transducer is often very large this is an equivalently large matrix, and since any solution method must invert it, with it, this is not an efficient method. A more accessible approach is to create a substitution AHz=x, before applying a similar methodology:

Cz=AAHz=Ax=b,
z=C−1b=(AAH)−1b,


This time around, as C=AAH is a mere m columns by in rows, this result is a much smaller set of linear equations to work through. The vector z can be converted into x at any time so long as AH can be produced.


However, this does not end here. This approach is not just a fortuitous set of symbolic manipulations, the change of variables from the complex-valued vector that describes the drive of individual transducer elements x to the much lower dimensional z has further meaning. Each complex-valued component of z can be viewed as a complex-valued drive coefficient that pre-multiplies a focusing function which generates a focus from all of the individual transducer fields, wherein the focal point is co-located with each individual control point. For m control points therefore, there are m such focusing functions, and they can be viewed as defining a complex vector space custom characterm where points in this space correspond to possible configurations of these in. ‘focus points’.


B. Dynamic Ranging—a Mechanism to Ensure that Waveforms are Reproducible


To ensure that at no point the user tries to generate a waveform at any control point that is not possible using the hardware, it is important to limit the level of acoustic output. This is achieved by using a second b vector, termed brange. The components of this vector brange are defined as the largest that the user may ask for at any point. The resulting solution zrange, is then the solution that is most taxing on the hardware. As zrange is a point in the complex vector space custom characterm, and this is linearly related to the acoustic output from the transducer, this must be the most extreme outlier. This is evaluated through pre-multiplication by AH to produce xrange. This xrange tells us how to relate the worst-case performance to the real capabilities of the device, so the maximum amplitude drive from all transducers in xrange is used to divide through the maximum output allocation of each point, as these are linear. The phase oracle (the algorithm that predicts the best set of phases to use) is then applied to brange, to ensure that the worst case phase interactions are captured. This ensures the device is never asked to produce an unrealistic output that could result in undefined behavior.


Shown in FIG. 17 is a diagram 1800 of the stages of the algorithm described. Diagram a) 1810 shows the transducer array and the set of control point positions. Diagram b) shows determining the three focal point basis functions 1820, 1830, 1840 comprised of each transducer drive, here weighted by the utility of each transducer to each point. Diagram c) 1850 shows 1) solving without considering the interactions between points (using a diagonal-only C matrix) yields an estimate of how the phasors cancel in the transducer activation and the approximate amplitude of each drive. Some of these amplitudes are greater than the maximum possible (dotted line), which would mean that the whole array solution would need to be scaled back. 2) Considering the fold level (dashed line) in logarithmic space, the logarithmic amplitude may be folded over the fold level to find the per-transducer scaling factor. Diagram d) 1860 shows applying the scaling factor to the transducers, all of the transducers are now at or under the fold level, which can be seen in 1) real space and 2) logarithmic space but the field is no longer generated accurately, even with the assumption that points do not interact. Diagram e) shows this solution is decomposed into focal point basis functions again, which are comprised of each transducer drive, but this time the individual transducer entries are scaled by the scaling found earlier 1870, 1880, 1890. Diagram f) 1895 shows these new focal point basis functions are now used in the full solution, considering extra interactions. Note that the maximum amplitude of the per-transducer drive is reduced, no longer requiring more than the maximum transducer drive amplitude possible (dotted line) and the relative size of the group of resealed transducer amplitude bars remain much the same, although the absolute amount of drive has changed to compensate for the earlier scaling down. The proportions of the contribution to each control point from the transducer (filled pie segments in the circle) has changed noticeably from the other solutions.


C. Breaking Hermitian Symmetry


Due to the Hermitian symmetry of AAH, the component transducer contributions to each basis function are in turn weighted by how usefully they may contribute. While this is helpful to save power and ensure that transducers are used only so far as they are effective and efficient, or in situations where this symmetry produces a simple geometry independent apodization of the phased array to combat grating lobe aliasing artifacts, this is unhelpful when more output is desired from a given hardware system.


The symmetry of AAH may be broken by undoing the ‘utility’ weighting produced by the multiply. This may be viewed as separating the forward propagation operator from the backward propagation operator of the acoustic wave—the matrix AAH forces the two operators to behave identically, and only breaking the symmetry of the matrix prevents this. Fortunately, the symmetry may be broken by allowing A (which describes the physical behaviour of the waves travelling in space—the forward propagation operator) and AH (which describes how the basis function is defined, as AHz=x—the backward propagation operator) to differ. However, this difference can be confusing and difficult to manage, and breaking the symmetry can quickly eliminate all of the benefits described earlier.


The method described in this document does not require symmetry to be broken, functioning equally effectively on both Hermitian symmetric and matrices which have not been symmetrically scaled.


D. Statistical Distributions of Random Phasors—Consequences for Transducer Drive


Settings aside questions of apodization and generation of appropriate per-transducer usage in each basis function, the fundamental statistical behavior of random phase summations—necessary to model how arbitrary many control points are constructed from phasor contributions of back propagations to each transducer—works against the ability to generate consistent output and efficient transducer utilization.


The process of generating the limiting scenario using king, may then be roughly modelled for the generation of the extremal transducer drive coefficient as the summation of a series of random phasors or complex numbers. When the summation exceeds 100% of output utilized, then this is the amount by which the drive of the whole transducer array must be scaled back. Therefore, there is a clear, if approximate, isomorphic relationship between any extremal samples of random phasor summation distributions and the transducer drive efficiencies in the limiting scenario.


Modelling the summation for the transducer drive or activation coefficient as:







x
=


re

i

θ


=





j
=
1

m



a
j


+

ib
J




,





where aj˜N(0,σ) and bj˜N(0,σ) (normally distributed random variables with zero mean) and r=|x|≥0. The probability density function is then exactly:









f



"\[LeftBracketingBar]"

z


"\[RightBracketingBar]"



(
r
)

=


r

m


σ
2





e


-

r
2


/
2

m


σ
2





,





with a cumulative density function exactly:

F|z|(r)=1−e−r2/2mσ2,

so the probability of the sum having amplitude greater than or equal to r must be e−r2/2mσ2. Assuming unit variance, σ=1, and solving for percentiles allows the plotting of equiprobable ‘transducer drive amplitudes’ as shown in FIG. 18.


Shown in FIG. 18 is a graph 1700 entitled “Mean and percentiles of amplitudes magnitudes”. The x-axis 1710 is the number of phasors summed (in) 1710. The y-axis 1720 is the magnitude of final phasor sum (r). The mean is the long-dashed line 1760. The median 50th percentile is the dotted line 1750. The median 90th percentile is the solid line 1730. The median 99th percentile is the short-dashed line 1740. These lines are the amplitude distribution of a sum of m complex numbers with normally distributed components.


Given that transducer arrays often have many hundreds of individual transducers which represent individual summations. This means that since each transducer represents a drawing from the distribution, it is highly likely that there will be a small proportion of transducer elements driven at much higher amplitudes—often amplitudes many times larger for large m—relative to the mean. In the case that the whole array amplitude must be scaled down to accommodate these outlier elements, this small subset causes a large drop in overall efficiency. Therefore, it is important that at least these transducers are targeted with predistortion to ‘condition’ the amplitude distribution.


E. Two Linear System Solutions Method


The two-step approximation method may proceed by first solving the system as before:

x′range=AHz′range=AH(AAH)−1brange.

Then, a ‘fold’, f, is designated. This is termed a ‘fold’ because this can be viewed as a point of symmetry in a logarithmic space, as the process can be applied efficiently using logarithms. This is the smallest value at which the distortion is applied. Then a diagonal matrix is defined as:







B
=

[




B

1
,
1







0















0






B

N
,
N





]


,





where each of the diagonal elements may be defined as,







B

q
,
q


=

{





1
,








if





"\[LeftBracketingBar]"


x

range
,
q





"\[RightBracketingBar]"



<
f

,







f
/



"\[LeftBracketingBar]"


x

range
,
q





"\[RightBracketingBar]"



,






if





"\[LeftBracketingBar]"


x

range
,
q





"\[RightBracketingBar]"




f

,










where x′range,q is the element of the vector x′range corresponding to the qth transducer. This vector does not have any per-transducer gain or distortion applied.


The second system for the pre-distorted basis functions may be applied as:

xrange=BAHzrange=BAH(ABAH)−1brange,

and

x=BAHz=BAH(ABAH)−1b,

where the per transducer gain matrix B applies distortion to counter or ‘smooth out’ excessively high transducer amplitudes that were generated when solving without gain control. When the new xrange vector is inspected, the peak transducer amplitudes are greatly reduced from what would be seen without this step—as shown from the earlier statistical argument.


It should also be noted that it is possible to skip individual points (i.e. not apply this method to all focal point basis functions) and apply different gain factors or ‘fold’ values to different basis functions.


F. Dynamically Determining the Fold Value


The ‘fold’ value f may also be chosen dynamically, as the plotting of the phasor distribution illustrative of how the transducer driving phasors in the early section shows that the transducer with the largest amplitude is many times larger than the mean. The ‘fold’ value may be larger than unity, as the resealing of the transducer gain is applied to take all transducers driven at over the level represented by this value to be exactly this value. Note that since this is all relative, values being greater than unity only means that the physical limitations of the device will require the whole system to be scaled down, as these later whole system scaling steps are applied separately to ensure physical limits are met.


By approximately tracking the behavior of the distribution of transducer amplitudes, the fold level can track where the extreme amplitudes are and eliminate them. To do this simply we may set the transducer fold level to the mean. This may be either the arithmetic mean or the geometric mean, where the geometric mean is more aggressive limit due to it being generally less than the arithmetic mean. Where the arithmetic mean may be calculated as the mean of each transducer drive amplitude, the logarithm of the geometric mean is the mean of each logarithm of the transducer drive amplitudes. The main reason that no other indicator from the distribution is suitable is simply because the mean is simple to calculate.


Using the mean ensures not everything is resealed. Assuming the illustrative phasor distribution from earlier is a reasonable approximation in most settings, then the mean and median are close, leading to roughly half the transducer amplitudes participating in the generation and application of per-transducer gain. This means that not all of the shape of the basis set is ablated, resulting in modest changes in a minority of transducers, and large changes only in a few egregiously overdriven transducers.


G. Approximating the First Linear System


The first linear system solved is not required to be functional. It merely needs to represent the relative drive and power of the basis functions which when added together make up the complex-valued drive coefficients for the individual transducers. For this reason, in the first solution, we can ignore cross-terms in the acoustic field, anticipating that interactions between control points have only a small impact on the total drive amplitude of each transducer. In this case, equivalent to zeroing all of the elements of the matrix C=AAR away from the main diagonal, the basis function and coefficient needed for each control point may be solved separately as many single-row-and-single-column 1×1 matrices. As inverting non-zero single element matrices can be stated as simply the reciprocal of the value, this means that no linear system need be solved. This is then:








x

range
,
q



=




j
=
1

m




b

ramge
,
j






α
q

(

χ
j

)





q
=
1

N





α
q

(

χ
j

)





α
q

(

χ
J

)

_







,





where brange,j is the component of the vector brange corresponding to the complex-valued scalar linear acoustic quantity at the control point which is intended to represent the maximum output required by the user.


It should be noted that this method does not need to be applied to the ‘range’ system. It may be applied directly to the intended transducer drive, although the fluctuations in efficiency may cause excess signal noise in the case that this is used repeatedly in a loop over time.


H. Transform Matrix


In another embodiment, we can modify the basis for which we are solving using a matrix, referred to from here as a ‘transform matrix’. T. This is applied by solving,

ATx′=b′,

where x′ represents a new activation solution distinct from x and b′ is a distinct output goal. T must be a square matrix with dimension equal to the number of transducers. A solution is generated by taking the pseudoinverse, x′=(AT)+b′.


One way for this solution to be useful is to relate b′ to real-world units. This can be accomplished by relating b′ which uses indefinite units to b which uses units related to the transducer function αqj). As this is an underdetermined problem, there is room to add more restrictions to x′. In particular, we can enforce the original problem on the new solution,

Ax′=b.


This allows us to write the following relationship between b and b′,

b′=ATA+b,

where A+ is the pseudoinverse of A. The transform matrix can therefore modify the activation solution while substantially maintaining the original output goals.


While arbitrary matrices can be used as T, purely diagonal transform matrices are the easiest to understand and predict their effect. For instance, a transform matrix with values near 1.0 for some entries and near 0.0 for others will emphasize the transducers corresponding to the 1.0 values while attenuating the others. A possible use for this would be to rotate the transducers being used by periodically changing the locations of the near-1.0 values thereby reducing the overall duty-cycle of the average transducer, possibly extending the life of the array.


Rather than an infrequently-changed transform matrix, another useful embodiment of this method is to have T be a function of A. This allows for dynamic modification of the apodization of the array (FIG. 19). Possible functions include but are not limited to, forming a diagonal matrix from the diagonal entries of A while applying an operator in the process. For instance, taking the absolute value squared (or larger value) of the diagonal entries of A and using those values for the diagonal entries of T (leaving everything else 0) has the effect of weighting the output for each transducer based upon the square (or larger value) of the transducer function αq j) in magnitude. This increases the apodization of the array related to the exponent used. Another possible operator is a threshold function such as 1.0 for diagonal values over some pre-determined threshold value and a small values (such as 0.1) for values under the threshold. An example of this is given in FIG. 3. A benefit of this approach is the threshold function is simple to calculate and, when no transducers are above the threshold, the solution will be substantially similar to the original A, with the transformation only being heavily applied when circumstances put transducer activation over the pre-determined threshold.


The steps involved is using the transformation matrix are as follows,

    • 1. Calculate A as normal;
    • 2. Determine T;
    • 3. Calculate b′ from b′=ATA+b; and
    • 4. Calculate x′ from x′ (AT)+b′.


It must be noted that while the equations above imply a perfect solution, the pseudoinverse (also known as the Moore-Penrose inverse) is only the true inverse if the matrix is invertible. This is not guaranteed in this formulation. As a result, the final solution x′ may not exactly satisfy Ax′=b, and as a result will fail to reproduce the desired pressures in b. In general, the less extreme transform matrix is (in magnitude and in similarity between entries in T), the closer the resulting solution is to achieving the constraints in b. If a given formulation of T is not achieving sufficient performance, an iterative guess-and-check method can improve the solution. This is done by first calculating x′, then performing the multiplication Ax′ and comparing the results to b. The resulting average proportional difference in each element can be calculated and x′ can be scaled appropriately (being careful to keep maximum activation below or equal to full-drive). Multiple iterative steps can be implemented to further improve performance.


Turning to FIGS. 19A, 19B, and 19C, shown are example outputs 1900 using a transformation matrix using a 256-element rectilinear array of 40 kHz transducers with 10.3 cm pitch. Units for the left 3FIGS. 1910, 1930, 1950 are mm while the right FIGS. 1920, 1940, 1960 are m. The key 1925, 1945, 1965 is the amplitude of the transducer drive relative to maximum drive (1.0).


The right 3FIGS. 1920, 1940, 1960 show activation amplitudes per transducer for different T matrix designs while the left 3FIGS. 1910, 1930, 1950 show simulated pressure at z=2.0 cm with white colors being higher pressure than dark colors. The solutions are attempting to create a high-pressure point at [x,y,z]=[8,0,20] cm. FIG. 19A s show a normal pressure solve without using a transform matrix. FIG. 19B shows the effect of T where T is defined as the diagonal elements of A, magnitude squared. FIG. 19C shows another application of T where T first takes the diagonal elements of A and then thresholds the output such that if the absolute value of the entrée is greater than 0.7, the corresponding value of T is 1.0, if not, it is 0.1. As the apodization is more extreme in each row, the resulting focus points pressure is only diminished slightly whereas the grating lob at approximately x=−0.125 mm is diminished much more.


V. Momentum Tensor Corrections for Mid-Air Haptic Systems

A. Solving for the Phased Array Drive to Generate the Linear Acoustic Quantity


Traditionally, the linear system is described in terms of linear combinations of complex-valued transducer generated fields and their drive coefficients. This produces a matrix, where for m control points and N transducers, the matrix A is N columns by M rows and consists of the generated complex valued signal by each transducer q∈{1, . . . , N} at the location of each control point j∈{1, . . . , m}.


Using αqj) to describe a complex-valued scalar linear acoustic quantity α measured at a position offset from the transducer element q by the translation vector χj, which may evaluate to be acoustic pressure or an acoustic particle velocity in a chosen direction, the matrix A may be written:







A
=

[





α
1

(

χ
1

)








α
N

(


χ


1

)


















α
1

(

χ
m

)








α
N

(

χ
m

)




]


,




This, for a number of control points fewer than the number of acoustically active transducer elements can then be placed into a complex-valued linear system wherein a sample vector b={αC11), . . . , αCmm)} represents the desired total linear scalar complex-valued acoustic quantity (either pi or ui·{circumflex over (n)}) where the amplitudes are desired amplitudes of the acoustic quantity (either ∥pi∥ or ∥ui·{circumflex over (n)}∥) and the phases are those taken from the phase oracle (which may have been user-influenced). In this linear system described as Ax=b, the x vector is then the initial field coefficients for each transducer element, which may be used to drive a real transducer element, resulting in the recreation of the acoustic field desired. This may then be solved in a loop to provide a system that changes over time.


As this is matrix A is not square, and the degrees of freedom number more than the constraints, this is termed a ‘minimum norm’ system. It is ‘minimum norm’ because as there are infinitely many solutions, the most expeditious solution is the one which achieve the correct answer using the least ‘amount’ of x—the solution x with minimum norm. To achieve this, some linear algebra is used to create a square system from the minimum norm system Ax=b:

AHAx=AHb,
(AHA)−1AHAx=x=(AHA)−1AHb,


This AHA is now N columns by N rows and given that the number of transducers is often very large this is an equivalently large matrix, and since any solution method must invert it, with it, this is not an efficient method. A more accessible approach is to create a substitution AHz=x, before applying a similar methodology:

Cz=AAHz=Ax=b,
z=C−1b=(AAH)−1b,


This time around, as C=AAH is a mere m columns by m rows, this result is a much smaller set of linear equations to work through. The vector z can be converted into x at any time so long as AH can be produced.


However, this does not end here. This approach is not just a fortuitous set of symbolic manipulations, the change of variables from the complex-valued vector that describes the drive of individual transducer elements x to the much lower dimensional z has further meaning. Each complex-valued component of z can be viewed as a complex-valued drive coefficient that pre-multiplies a focusing function which generates a focus from all of the individual transducer fields, wherein the focal point is co-located with each individual control point. For m control points therefore, there are m such focusing functions, and they can be viewed as defining a complex vector space custom characterm where points in this space correspond to possible configurations of these m ‘focus points’.


B. Dynamic Ranging—A Mechanism to Ensure that Waveforms are Reproducible


To ensure that at no point the user tries to generate a waveform at any control point that is not possible using the hardware, it is important to limit the level of acoustic output. This is achieved by using a second b vector, termed brange. The components of this vector brange are defined as the largest that the user may ask for at any point. The resulting solution zrange, is then the solution that is most taxing on the hardware. As zrange is a point in the complex vector space custom characterm, and this is linearly related to the acoustic output from the transducer, this must be the most extreme outlier. This is evaluated through pre-multiplication by AH to produce xrange. This xrange tells us how to relate the worst-case performance to the real capabilities of the device, so the maximum amplitude drive from all transducers in xrange is used to divide through the maximum output allocation of each point, as these are linear. The phase oracle (the algorithm that predicts the best set of phases to use) is then applied to brange, to ensure that the worst case phase interactions are captured. This ensures the device is never asked to produce an unrealistic output that could result in undefined behavior.


C. Control Point Relations


A control point specifies both amplitude and phase of the carrier. For many applications, the phase is immaterial, so it can be chosen to maximize the amplitude available at the control points in space.


Representing the control point activation Yc as a complex value can be written:

Yc=custom characterec,


To find the effect that the activation of a control point via its transducer basis set has on other control points, the phase of the control point must be set to a reference point, the most suitable being unit amplitude and zero phase. Denoting this unit amplitude and zero phase activation Y′c0=1, then defining the set of αc=[α1c), . . . , αqc), . . . , αNc)], the transducer drives required to generate Y′c0 may be written:








Y

c

0



=


Y

c

0


·

α
c



,



Y

c

0


=



α
c

_



α
c

·


α
c

_




,





where Yc0 is the vector of N transducer activations Yc0=[Y′1; c0, . . . , Y′q; c0, . . . , Y′N; c0], required to generate the control point activation Y′c0.


Given an activation coefficient for a transducer Y′q; c0 the effect of that on another point γ in the acoustic field can be found as α′q; c0γ)=Y′q; c0αqγ). With this, the total effect of activating one point with a given amplitude custom characterc and phase ec has on another in the field can be found as the scaling of the unit amplitude effect, which when summed over all transducers is:








α

Ω
:


c

0




(

𝒳
γ

)

=


e

i


ϕ
c








α
γ

·


α
c

_




α
c

·


α
c

_



.






D. Control Point Relations Matrix


From this statement of α′Ω; c0γ), a matrix may be constructed such that multiplying from the right is an application of the activations l′=[custom character1e1, . . . , custom charactermem]T of all control points to the acoustic field at a point. Then maximising the efficiency of this by allowing the phases ϕ1, . . . , ϕm to change is the aim. The amplitudes may be taken out of l′ and placed into the matrix, leaving l=[e1, . . . , em]T, then yielding the final eigensystem matrix:







R
=

[







α
1

·


α
1

_




α
1

·


α
1

_













α
1

·


α
r

_




α
r

·


α
r

_













α
1

·


α
m

_




α
m

·


α
m

_





























α
r

·


α
1

_




α
1

·


α
1

_













α
r

·


α
r

_




α
r

·


α
r

_













α
r

·


α
m

_




α
m

·


α
m

_





























α
m

·


α
1

_




α
1

·


α
1

_













α
m

·


α
r

_




α
r

·


α
r

_













α
m

·


α
m

_




α
m

·


α
m

_







]


,





If a diagonal matrix is defined to be:







K
=

[






α
1

·


α
1

_








0















0








α
m

·


α
m

_






]


,





then:

R=KC.

In this case, it is clear that the statement of the eigenproblem Rl=λl satisfies the criteria for the best phases to drive the control points with this-configuration of amplitudes when λ is largest—where l describes the dominant eigenvector of the matrix R.


E. Finding the Dominant Eigenvector Power Iteration


The phase oracle is then mostly comprised of an implementation of the power iteration, a simple method to find the dominant eigenvector which corresponds to largest eigenvalue of a matrix. Where the eigenproblem is stated as Rl=λl, it can be stated as:








l
0

=


lim

y







R
y



l
input






R
y



l
input







,





where linput must not be orthogonal to the dominant eigenvector and λ0 must be strictly greater than the other eigenvalues. As R is a complex-valued matrix and linput is a complex-valued vector, then the jth component of the final vector b and brange can be written as:








b

range
,
j


=



l

0
,
j





"\[LeftBracketingBar]"


l

0
,
j




"\[RightBracketingBar]"





b

user
,
range
,
j




,



b
j

=



l

0
,
j





"\[LeftBracketingBar]"


l

0
,
j




"\[RightBracketingBar]"





b

user
,
j




,





since only the phase is taken from the dominant eigenvector result. As such:







l

0
,
j





"\[LeftBracketingBar]"


l

0
,
j




"\[RightBracketingBar]"







is the result of this exemplar phase oracle.


To summarize, in this disclosure it is shown that existing methods used to derive the exerted haptic force for a mid-air haptic system onto a human body part are based on incorrect physical assumptions. Further, a method is provided to modify a mid-air haptic system to reduce a problem scenario including boundaries at the human body parts to a free-field system that may be formulated as a linear system such that correct force vectors are then produced.


F. Derivation of Acoustic Quantities


Taking the equation of state for an ideal gas:

p=ρRT,

where ρ is the density, p is the pressure, R is the ideal gas constant and T is the temperature. For a constant volume, all three variables depend on the entropy of the system.


Acoustics is often described mathematically by applying a perturbative expansion to the equation of state. Due to acoustics being generally concerned with fluctuations that do not change the entropy of the underlying medium, so an isentropic perturbative expansion that freezes the degree of freedom related to entropy change in the equations of state is generally chosen. In this case, where the subscript zero denotes the state of the system at rest, the pressure may then be related isentropically to density as:








p

p
0


=


(

ρ

ρ
0


)

γ


,





where γ is the ratio of specific heats.


To derive equations for the physics of acoustics, it can be written:

ρ=ρ0+ρ′, p=p0+p′,

where the dashed quantities refer to perturbations. Collecting terms and using a Taylor expansion yields:










p
0

+

p




p
0


=


(

1
+


ρ



ρ
0



)

γ


,


p


=




γ


ρ
0



ρ
0




ρ



+

O

(


(

ρ


)

2

)



,





then using the substitution for the square of the speed of sound c02=γp00,

p′=c02ρ′+O((ρ′)2).

The mass and momentum balance equations (as taken from the Euler equations of fluid dynamics) for this closed volume of an inviscid and adiabatic but compressible fluid may be written:











ρ


(






t



+
u


·


)


=


-
ρ



(


·
u

)



,








ρ


(




u



t


+

u
·

(


u

)



)


=

-


p



,








where u is the velocity of the medium. Considering the first-order effects (with subscript 1) of perturbations of the fluid at rest (with subscript 0) and ignoring terms that are higher than first-order or evaluate to zero gives:














ρ
1




t


=


-

ρ
0




(


·

u
1


)



,









ρ
0






u
1




t



=

-



p
1




,








Taking the derivative with respect to time of the first-order mass balance equation yields:











2


ρ
1





t
2



=


-


·

ρ
0








u
1




t




,





then substituting both the first-order momentum balance equation and the relationship between the perturbations of the pressure and density yields the wave equation in density as:










2


ρ
1





t
2



=


c
0
2






2


ρ
1


.







Further, taking the time derivative of the first-order momentum balance equation and substituting the relationship between the perturbations of the pressure and density and then finally substituting the first-order mass balance equation yields the wave equation in velocity as:










2


u
1





t
2



=


c
0
2






2


u
1


.







All sinusoids are solutions to the wave equation, so writing the solutions in space and time as a general complex exponential gives the set of linear acoustic perturbations as:

ρ1(x, t)=ρ1(x)eiωt,
p1(x, t)=c02ρ1(x)e−iωt,
u1(x, t)=u1(x)e−iωt,

where by convention ω=2πf describes the angular frequency, meaning that each of the quantities for a single frequency ω can be written as a complex-valued field in space that is harmonic in time.


Second-order perturbations may now be defined as any further parts of ρ, p or u that do not describe the fluid at rest or belong to the solutions of the wave equation and so are described as nonlinear acoustics.


Considering the second-order effects (with subscript 2) of perturbations of the fluid at rest and ignoring terms that are higher than second-order, belong to the first-order equations or are zero gives new second-order momentum balance equation as:










ρ
1






u
1




t



+


ρ
0

(


u
1

·



u
1



)


=

-



p
2




,





where terms involving u2 have been removed as if it is non-zero then the fluid cannot be starting at rest. Then using the first-order momentum equation and the relationship between the perturbations of the pressure and density, the equation may be rewritten:









p
2


=




p
1





p
1





c
0
2



ρ
0



-



ρ
0

(


u
1

·



u
1



)

.






Using the product rule the two terms may be rewritten in terms of squares yielding:










p
2


=





p
1
2



2


c
0
2



ρ
0



-



ρ
0







u
1
2





2



,





then time-averaging over the wave period and integrating to remove the gradients gives:










p
2



=





p
1
2




2


c
0
2



ρ
0



-



ρ
0







u
1
2






2



,





where angular brackets denote time-averages as:









X


=


1
τ





0


τ




X

(
t
)


dt




,





where τ is defined such that oscillatory behaviour may be neglected.


G. Derivation of Momentum and Energy Flux


The momentum and energy conservation equations for a volume of an inviscid and adiabatic but compressible fluid can be written in conservation form (with a zero right hand side) when the volume has no sources or sinks as:















ρ


u



t


+


·
Π


=
0

,












ρ


E



t


+


·
ε


=
0

,








where E=ρ(u·u/2+e) and e is the internal energy of the ideal gas, the momentum flux tensor Π and energy flux vector ε may be written:

Π=(ρu⊗u)−pI,
ε=ρ(u·u/2+e)u+pu,

where I is here the identity tensor and ⊗ denotes the outer product.


Considering the general component of the momentum flux tensor Π when time-averaged for an acoustic field in this medium and again discarding terms higher than second-order or zero yields:

custom characterΠcustom characterjkρ0custom characteru1,ju1,lcustom character+custom characterp2custom characterδjk,

where j and k are spatial indices and δjk represents the Kronecker delta.


Considering the energy flux vector ε, again time-averaged for the acoustic field and discarding terms higher than second order or zero gives:

custom characterεcustom character=e0custom characterρ1u1custom character0custom charactere1u1custom character+custom characterp1u1custom character.


As the perturbation in internal energy e1 is a small variation about the internal energy e0 of the fluid at rest, then the fundamental thermodynamic relation de=T ds−p dV may be invoked to take derivatives at constant entropy, giving (since ds=0 and dV=1/ρ):









d

e


d

ρ


=

p

ρ
2



,


e
1

=




d

e


d

ρ




ρ
1


=


p

ρ
2




ρ
1




,





then substituting results in:









ε


=



(


e
0

+


p

ρ
2




ρ
0



)






ρ
1



u
1





+




p
1



u
1






,





but since custom characterρ1custom character1custom character is mass flux and the field is acoustic, then this time-average must be zero yielding finally that the energy flux vector is equivalent to the acoustic intensity I:

custom characterεcustom character=custom character=custom characterp1u1custom character.


H. Potential and Kinetic Energy in Acoustics and Their Role in Momentum Flux


Considering the total energy of the fluid perturbed by the acoustic field and noticing that because the internal energy e is a thermodynamic quantity that proceeds from the equation of state it must be a function of density yields:

ρE=ρu·u/2+ρe(ρ),

then a Taylor expansion for (ρe)(ρ0+p1) is permitted yielding after discarding terms higher than second-order or zero:








ρ

E

=



ρ
0



e
0


+


ρ
1




d

ρ

e


d

ρ




(

ρ
0

)


+


1
2



ρ
1
2





d
2


ρ

e


d


ρ
2





(

ρ
0

)


+


1
2




ρ
0

(

u
·
u

)




,





where the derivatives are evaluated at ρ0 as per the Taylor expansion. Invoking the same thermodynamic relation as before yields for the derivatives:









d

e


d

ρ


=

p

ρ
2



,


dp

d

ρ


=

c
0
2


,









d

ρ

e


d

ρ


=



e



d

ρ


d

ρ



+

ρ



d

e


d

ρ




=

e
+

p
ρ




,










d
2


ρ

e


d


ρ
2



=




d

e


d

ρ


+


1
ρ




d

p


d

ρ



+

p



d


ρ

-
1




d

ρ




=


c
0
2

ρ



,





then substituting and evaluating at ρ0 gives:








ρ

E

=



ρ
0



e
0


+


ρ
1

(


e
0

+


p

(

ρ
0

)


ρ
0



)

+


1
2



ρ
1
2




c
0
2


ρ
0



+


1
2




ρ
0

(

u
·
u

)




,





where ρ0e0 is the internal energy of the volume at rest and ρ1 is the perturbation in density and so change in mass of the volume over the wave period, so time-averaging and giving only terms relevant to the acoustic field results in the total energy of the acoustic field in the volume as:










ρ


E





=




1
2





ρ
1
2






c
0
2


ρ
0



+


1
2



ρ
0







u
1
2







=





p
1
2




2


c
0
2



ρ
0



+



ρ
0







u
1
2






2




,





where E′ denotes the energy with respect to the acoustic wave in particular.


Clearly, the second term is the kinetic energy and therefore the first term must be the potential energy:







KE
=



ρ
0







u
1
2






2


,







PE
=




p
1
2




2


c
0
2



ρ
0




,





where KE is a shorthand for kinetic energy and PE a shorthand for potential energy, respectively.


The second order pressure can then be immediately rewritten as:

custom characterp2custom character=PE−KE.


Further, the general momentum flux tensor element may be considered to be:

custom characterΠcustom characterjk0custom character∥u12custom character{circumflex over (n)}j{circumflex over (n)}k+custom characterp2custom characterδjk=2KE {circumflex over (n)}j{circumflex over (n)}k+(PE−KEjk.


I. Momentum Exchange with Planar Surfaces


At a surface separating two acoustic media there are generally three distinct acoustic fields. These represent the waves incident upon, reflected from and refracted through the surface. Considering each volume element in the incident field (the field as it would be given the free field condition and acoustic sources without the second medium), at the position of the boundary, the reflected field and refracted field may also be evaluated as also distinct free field systems at the boundary, wherein the reflected field is present only in the first medium and refracted only in the second. These notionally non-interacting and unphysical free fields are then composed such that the boundary respects the properties of the two media and conserved quantities, yielding quantitative behavior.


Without loss of generality, the coordinate system may be fixed such that the planar surface separating the two acoustic media has a single unit x-component, such that the surface occupies x=0. In this case, only the (Π)xx component of the momentum flux need be considered, because this is the only direction along which forces may act, yielding the momentum flux for an acoustic field as:









Π


xx

=



ρ
0







u
1
2








n
ˆ

x
2


+




p
1
2




2


c
0
2



ρ
0



-


ρ
0








u
1
2





.







The superposition of two acoustic fields may be obtained by expressing the acoustic pressures p1 and components of the medium particle velocity u1={ux,1,uy,1} as complex numbers pi,ui,x,ui,y that define the spatial harmonic field and then adding them together. As only the real part has physical significance at any value of t, then the time average of say p2 can be expressed relative to the complex-valued harmonic field value (a+ib) as:










p
ι

=

(

a
+

i

b


)



,


p
=


p
ι



exp

(

i

ω

t

)



,










p
2



=



1
τ





0


τ






(


(


(

a
+

i

b


)



exp

(

i

ω

t

)


)

2

)


d

t



=



1
2



(


a
2

+

b
2


)


=



1
2



(

a
+
ib

)



(

a
-
ib

)


=


1
2






p
ι
2



.










so the pressures and the particle velocity values for two acoustic free field systems can then be summed to generate time-averaged values for the squares of the field quantities.


As the incident acoustic field and reflected acoustic field are both present in the first medium, they may be summed using the principle of superposition. The reflected wave undergoes a transformation during this process. This is in two parts, the first assumes that the amplitude decays due to the presence of some transmission into the second medium and a phase offset occurs on reflection from the boundary, so this is modelled with a complex coefficient on the reflected wave, exp(μ+iv), so then exp(μ) is the amplitude of the attenuated wave (as a proportion of unity so it may be expressed relative to the incident wave) and v the angle of the phase shift undergone during reflection. The second change is the particle velocity is reflected across the surface, negating the x-component of the velocity ux,1 and so also ui,x before being added, so this component will be reduced in the superposed field. The momentum flux for this superposed incident and reflected acoustic field may then be written in terms of the complex-valued pressure and particle velocity of only the incident field as:









Π




incident
&


reflected


=



1
2



(






p
ι



2


2


c
a
2



ρ
a



+


ρ
a

(






u
ι



2



cos
2



θ
a


-


1
2






u
ι



2



)


)



(


exp

(

2

μ

)

+
1

)


+


1
2



(






p
ι



2


2


c
a
2



ρ
a



-


1
2



ρ
a






u
ι



2



)



exp

(
μ
)


cos


v
.







Assuming a similar amplitude change and phase shift on the transmitted wave as exp(μ′+iv′), then the wave that underwent refraction may be found as:










Π


refracted

=


1
2



(






p
ι



2


2


c
b
2



ρ
b



+


ρ
b

(






u
ι



2



cos
2



θ
b


-


1
2






u
ι



2



)


)



exp

(

2


μ



)



,





so any phase shift v′ on the transmitted wave will have no effect on the momentum flux. The angle θb may be obtained using Snell's law.


Composing the two free fields for the different materials is valid so long as the conserved quantities on the interface agree, so the two free fields may be evaluated at the same point (such as the origin) using this assumption. Placing momentum flux from the first material a on the left and the second material b on the right gives the apparent pressure papp as:

custom characterΠcustom characterincident & reflected=custom characterΠcustom characterrefracted+papp,

so:

papp=custom characterΠcustom characterincident & reflectedcustom characterΠcustom characterrefracted


In this case, the term apparent pressure is used to distinguish this quantity from the other expressions of pressure that assume free field conditions—the pressure here is contingent on the existence of the surface.


The general case involves spatially varying momentum flux fields that do not have the same value in all space. They may be evaluated point-wise to obtain an infinitesimal force on an infinitesimal surface, which is a pressure only valid at that location, or integrated over a larger area, adjusting the reflected field before superposition if necessary to generate a force on that area.


J. Far Field Approximation


Turning to FIG. 20, shown is a diagram 2000 of the three waves that are involved in the problem of surface radiation pressure for a far field wave. The incident wave (α) 2010, reflected wave (β) 2020 and refracted wave (γ) 2030 may be represented by simple velocity potentials that are defined globally. Although in many cases the reflected wave (β) 2020 or refracted wave (γ) 2030 may have zero amplitude, the problem involving all three may be considered without loss of generality. Angles θα2040 and θb 2050 made by the wave propagation directions with the surface normal 2060 in the two materials are also shown.


When the time difference between wave sources is small compared to the wave period the waves are considered to be in the far field. Waves may then be approximated as plane waves, implying that:

p1=c0ρ0∥u1∥,

but substituting into the expression for custom characterρE′custom character reveals that this approximation makes the terms for potential energy and kinetic energy equal.


As the plane wave carries its energy as it travels, this implies:

c0custom characterρE′custom character{circumflex over (n)}=custom characterεcustom character=I=custom characterp1u1custom character=c0ρ0 custom character∥u12custom character{circumflex over (n)},

meaning that in this case the momentum flux tensor element can be written in terms of the time-averaged energy flux, acoustic Poynting vector or acoustic intensity custom characterεcustom character as:










Π


jk
far

=






ε





c
0





n
^

j




n
ˆ

k



,





shadowing the equivalent result for electromagnetic fields.


Following Landau and Lifshitz (Fluid Mechanics 3rd edition, Landau and Lifshitz, pp. 255-256), the three distinct plane waves representing the wave incident upon, reflected from and refracted through the infinite planar surface separating acoustic media may be given as velocity potentials:

ϕα=Aαexp(iω((1/cα)(y sin θα+x cos θα)−t)),
ϕβ=Aβexp(iω((1/cα)(y sin θα−x cos θα)−t)),
ϕγ=Aγexp(iω((1/cb)(y sin θb+x cos θb)−t)),

from whose definitions it is clear that the wavefront normal vectors may be described as:

{circumflex over (n)}α={+cos θα, +sin θα},
{circumflex over (n)}β={−cos θα, +sin θα},
{circumflex over (n)}γ={+cos θb,+sin θb}.


Using the equivalence between potential and kinetic energy and substituting into the equation for papp yields:







p
app
far

=



1
2



ρ
a






u
ι



2



cos
2




θ
a

(


exp

(

2

μ

)

+
1

)


-


1
2



ρ
b






u
ι



2



cos
2



θ
b




exp

(

2


μ



)

.








which is equivalent to the result in Landau and Lifshitz:

pappfar=custom characterρα(E′α+E′62)custom charactercos2 θαcustom characterρbE′γcustom charactercos2 θb.


This simple result is possible because the momentum flux is constant for the whole of the field hosted by material a.


K. Perfect Reflector Approximation


Turning to FIG. 21, shown is a diagram 2100 of the three waves 2110, 2120, 2130 that are involved in the problem of surface radiation pressure for a focused wave in the near field of a phase array of acoustic sources. This figure shows a 0-degree phase offset from the case where the reflected wave is not modified in phase upon reflection. The incident wave (α) 2110, reflected wave (β) 2120 and refracted wave (γ) 2130 are pictured by ray tracing a phased array of acoustic source elements 2170, each represented by a monopole velocity potential source. Note that crossing into another domain with a different wavelength heavily distorts the focal region in the new material due to spherical aberration. Angles θa 2140 and θb 2150 made by the wave propagation directions with the surface normal 2160 in the two materials are also shown.


If the acoustic near-field is instead incident onto a perfect reflector, then even in this complicated case the equations simplify to:








p

a

p

p

reflect

=



1
2



ρ
a






u
ι



2



cos
2



θ
a


+


1
2



(

1
+

cos

v


)



(






p
ι



2


2


c
a
2



ρ
a



-


1
2



ρ
a






u
ι



2



)




,





so in this case, the appearance of the non-linear term is totally dependent on the phase angle of the reflection. If the reflection undergoes a 0° phase shift, then this may be again simplified to:








p
app

reflect
,

0

°



=



1
2



ρ
a






u
ι



2



(



cos
2



θ
a


-
1

)


+





p
ι



2


2


c
a
2



ρ
a





,





If the reflection is instead 180° phase shifted, then it instead simplifies to:







p
app

reflect
,

180

°



=


1
2



ρ
a






u
ι



2



cos
2




θ
a

.






L. Conversion to a Linear Acoustic Quantity


One useful example of a case wherein a point pressure may be used to infer how the wider solution for the force on a surface behaves is for an acoustic phased array system undergoing focusing, so in the near-field of the array. If the focus has a known profile, then other forces may be inferred from a peak apparent pressure measurement under the assumption that the neighboring wavefield has similar properties.


The ∥ui2 cos2θa term in the calculations above are due to the acoustic wavefront vector and the surface normal vector having a difference, whereby that difference is measured as the angle θa. If we attribute a normal vector {circumflex over (n)} to the surface normal vector and not the wave direction vector and solve in terms of ui·{circumflex over (n)}, then it is clear that:

ui·{circumflex over (n)}∥2=∥ui2 cos2θa,

since ui is by definition parallel to the wavefront normal vector. Then, defining:








c


p
/
u

·
n


=




p
ι





c
a



ρ
a







u
ι

·

n
^








,










c


u
/
u

·
n


=




u
ι








u
ι

·

n
^







,





the momentum flux differences may then be written:







p
app

reflect
,

0

°



=


1
2



ρ
a







u
ι

·

n
^




2




(

1
+

c


p
/
u

·
n

2

-

c


u
/
u

·
n

2


)

.







so therefore in the case of a 0° phase shift on reflection the amplitude ∥ui·{circumflex over (n)}∥ fill of the portion of the medium particle velocity along the normal vector {circumflex over (n)} required as the target amplitude for the linear system to achieve may be related to the desired apparent pressure on the surface by:











u
ι

·

n
^




=



2


p
app

reflect
,

0

°






ρ
a

(

1
+

c


p
/
u

·
n

2

-

c


u
/
u

·
n

2


)




,





or alternatively, for the 180° phase shift case:










u
ι

·

n
^




=




2


p
app

reflect
,

180

°





ρ
a



.





Equally, the conversion constants may be written in terms of pressure, yielding:








c

u
·

n
/
p



=



c
a



ρ
a







u
ι

·

n
^









p
ι





,


c

u
/
p


=



c
a



ρ
a






u
ι








p
ι





,





which in this case makes the momentum flux differences:







p
app

reflect
,

0

°



=


1
2







p
ι



2



c
a
2



ρ
a






(

1
+

c

u
·

n
/
p


2

-

c

u
/
p

2


)

.







so therefore in the case of a 0° phase shift on reflection the amplitude of the acoustic pressure, as this is another linear acoustic quantity that may be solved for, would be required as the target amplitude of the acoustic pressure for the linear system to achieve may be related to the desired apparent pressure on the surface by:










p
ι



=



2


c
a
2



ρ
a



p
app

reflect
,

0

°





1
+

c

u
·

n
/
p


2

-

c

u
/
p

2





,





or alternatively, for the 180° phase shift case:









p
ι



=




2


c
a
2



ρ
a



p
app

reflect
,

0

°





c

u
·

n
/
p


2



.





M. Computing the Value of the Conversion Constants


Finally, the conversion constants cu.n/p, cu/p, cp/u.n, and cp/u must be generated to find the values of ∥pi∥ or ∥ui·{circumflex over (n)}∥ to solve for. As the constants are used as coefficients to multiply time-averaged quantities, they may be purely real. There are a number of different ways to achieve this, one efficient approximation using only the ‘primary’ wave may be found using the summation of the quantity in the construction of the A matrix (using the linear acoustic quantity present in the A matrix as α):








c

u
·

n
/
α













q
=
1




N





u
ι

(

χ
j

)

·


n
^

j









q
=
1




N




α
q

(

χ
j

)






,


c

u
/
α














q
=
1




N




u
ι

(

χ
j

)










q
=
1




N




α
q

(

χ
j

)






,



c

p
/
α












q
=
1




N




p
ι

(

χ
j

)








q
=
1




N




α
q

(

χ
j

)






,





where by ‘primary’ wave it is meant that the superposition of due to other control points j′≠j, j′∈{1, . . . , m} are not considered in this computation. As the effects on the primary wave are purely a function of position, this is efficient to compute. Alternatively, the properties of the solved field may be queried using the solution vector, such as in the case of a simple linear system construction:








c

u
·

n
/
α













k
=
1




m








q
=
1




N




z
k





α
q

(


χ


k

)

_





u
ι

(

χ
j

)

·


n
^

j











k
=
1




m








q
=
1




N




z
k





α
q

(


χ


k

)

_




α
q

(

χ
j

)








,



c

u
/
α














k
=
1




m








q
=
1




N




z
k





α
q

(


χ


k

)

_




u
ι

(

χ
j

)












k
=
1




m








q
=
1




N




z
k





α
q

(


χ


k

)

_




α
q

(

χ
j

)








,



c

p
/
α












k
=
1




m








q
=
1




N




z
k





α
q

(


χ


k

)

_




p
ι

(

χ
j

)










k
=
1




m








q
=
1




N




z
k





α
q

(


χ


k

)

_




α
q

(

χ
j

)








,





where AHz=x, although clearly the forward and inverse steps do not have to be symmetric—the expression zkαqk) is intended to derive the transducer drive coefficient for the kth basis function. Further, this is an exercise in simulating the appropriate quantities, so any suitable simulation model for the acoustic quantities may be used.


However, this is a circular reference as each of these is required to define z, by necessity either this is computed using the z vector from the previous iteration or as an iterative refinement process. These ratios may have filtering applied to them to remove high-frequency variations that could potentially generate audible noise at other locations in the dynamic acoustic field that is produced from the phased array to generate the forces desired.


N. General Use of Conversion Constants


These conversion constants may then be used to cross-convert between units, allowing the apparent pressures to be described in terms of scalar linear acoustic quantities that can be solved for using a linear system and basis functions comprised of relative transducer drives. These can then be converted into transducer drive coefficients. Equally, this can be used to cross-convert between a variety of source units into the solvable scalar linear acoustic quantities and control points driven by surface force vectors can be mixed and matched with control points driven by different units for other ultrasonic use cases.


Potential source units that may be cross-converted are limited to time-averageable quantities, but within this category this includes but is not limited to, the time-averaged momentum flux difference described here, the time-averaged force due to the Gor'kov potential, the time-averaged squared pressure, the time-averaged squared acoustic particle velocity of the medium (scalar speed), the time-averaged squared velocity along a direction (with an effective cos2 θ scaling), the time-averaged wave energy flux (Poynting) vector, the time-averaged amplitudes of the acoustic pressure, the time-averaged amplitude of the spatial components of the acoustic particle velocity vector of the medium and the time-averaged amplitude of the acoustic particle velocity vector of the medium along a given direction vector.


These conversion constants may then be used to create an interface wherein conversions may be applied to values given to the control points sent to the phased array. By indicating the type or intent of each of the control points, the appropriate conversion may be inferred automatically.


Further, the destination scalar linear acoustic quantities may be mixed and matched. As the drive coefficient of the transducer may be used equivalently in both pi and ui·{circumflex over (n)}, it, it is possible to switch between them in the same matrix, effectively making for instance the A matrix:







A
=

[






(

α
1

)

1



(

χ
1

)










(

α
1

)

N



(

χ
1

)




















(

α
m

)

n



(

χ
m

)










(

α
m

)

N



(

χ
m

)





]


,





where now (αj)∈{pi,ui·{circumflex over (n)}j}, so the scalar linear acoustic quantities may be mixed within the same matrix (although scale factors may be required for reasons of numerical stability). This is useful in that differing amounts of natural apodization (due to the minimum norm that is implied by the symmetry of forward and inverse steps) occur in the two quantities—the scalar linear acoustic quantity ui·{circumflex over (n)} has more implied apodization due to falloff in the function as the wavefront moves away from {circumflex over (n)}—the process being minimum norm will regard power used away from {circumflex over (n)} to be wasted, which is not the case for the pi unit. However, more complicated adaptive schemes including those previously described can work around the implied apodization of the unit, reducing the impact of this effect.


O. Additional Disclosure


1. A method comprising:

    • (a) producing an acoustic field from a transducer array having known relative positions and orientations;
    • (b) defining at least one control point wherein each control point
      • (i) has a known spatial relationship relative to the transducer array;
      • (ii) has a control point activation coefficient;
      • (iii) has an indicator describing the type of the control point activation coefficient;
      • (iv) has an optional direction vector of action;
    • (c) wherein for each control point the source units are inferred from the type and a unit conversion applied to convert from a quantity of source units into a scalar linear acoustic quantity.


2. The method of paragraph 1 wherein the vector comprised of scalar linear acoustic quantities contains at least one acoustic pressure.


3. The method of paragraph 1 wherein the vector comprised of scalar linear acoustic quantities contains at least one particle velocity of the medium along the direction vector of action.


4. The method of paragraph 1 wherein a complex wave field sample matrix is calculated that

    • (a) defines the scalar linear acoustic quantity of a first index at control point positions of the first index when:
    • (b) actuations at control point positions of a second index using preset amplitude and zero phase offsets with the scalar linear acoustic quantity of the second index are defined.


5. The method of paragraph 4 wherein the complex wave field sample matrix is used to calculate an eigenvector used to further adjust the phase angle of the control point activation coefficients to produce a final phasor vector


6. The method of paragraph 5 wherein the final phasor vector is used to solve for the linear combination of focusing activation coefficients.


7. The method of paragraph 6 wherein the focusing activation coefficients are convertible into transducer actuation coefficients.


8. The method of paragraph 7 wherein the acoustic waves comprise ultrasound waves.


9. The method of paragraph 7 wherein the acoustic field is produced by a mid-air haptic feedback system.


10. A method comprising:

    • (a) producing an acoustic field from a transducer array having known relative positions and orientations;
    • (b) defining at least one control point wherein each control point
      • (i) has a known spatial relationship relative to the transducer array;
      • (ii) has a control point activation coefficient;
    • (c) wherein at least one control point is indicated as an apparent haptic pressure along a corresponding direction vector of action;
    • (d) wherein for each control point indicated as an apparent haptic pressure a conversion is applied from the difference of momentum fluxes into a scalar linear acoustic quantity.


11. The method of paragraph 10 wherein the vector comprised of scalar linear acoustic quantities contains at least one acoustic pressure.


12. The method of paragraph 10 wherein the vector comprised of scalar linear acoustic quantities contains at least one particle velocity of the medium along the direction vector of action.


13. The method of paragraph 10 wherein a complex wave field sample matrix is calculated that

    • (a) defines the scalar linear acoustic quantity of a first index at control point positions of the first index when:
    • (b) actuations at control point positions of a second index using preset amplitude and zero phase offsets with the scalar linear acoustic quantity of the second index are defined.


14. The method of paragraph 13 wherein the complex wave field sample matrix is used to calculate an eigenvector used to further adjust the phase angle of the control point activation coefficients to produce a final phasor vector.


15. The method of paragraph 14 wherein the final phasor vector is used to solve for the linear combination of focusing activation coefficients.


16. The method of paragraph 15 wherein the focusing activation coefficients are convertible into transducer actuation coefficients.


17. The method of paragraph 16 wherein the acoustic waves comprise ultrasound waves.


18. The method of paragraph 16 wherein the acoustic field is produced by a mid-air haptic feedback system.


VI. Conclusion

In the foregoing specification, specific embodiments have been described. However, one of ordinary skill in the art appreciates that various modifications and changes can be made without departing from the scope of the invention as set forth in the claims below. Accordingly, the specification and figures are to be regarded in an illustrative rather than a restrictive sense, and all such modifications are intended to be included within the scope of present teachings.


Moreover, in this document, relational terms such as first and second, top and bottom, and the like may be used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. The terms “comprises,” “comprising,” “has”, “having,” “includes”, “including,” “contains”, “containing” or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises, has, includes, contains a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. An element proceeded by “comprises . . . a”, “has . . . a”, “includes . . . a”, “contains . . . a” does not, without more constraints, preclude the existence of additional identical elements in the process, method, article, or apparatus that comprises, has, includes, contains the element. The terms “a” and “an” are defined as one or more unless explicitly stated otherwise herein. The terms “substantially”, “essentially”, “approximately”, “about” or any other version thereof, are defined as being close to as understood by one of ordinary skill in the art. The term “coupled” as used herein is defined as connected, although not necessarily directly and not necessarily mechanically. A device or structure that is “configured” in a certain way is configured in at least that way but may also be configured in ways that are not listed.


The Abstract of the Disclosure is provided to allow the reader to quickly ascertain the nature of the technical disclosure. It is submitted with the understanding that it will not be used to interpret or limit the scope or meaning of the claims. In addition, in the foregoing Detailed Description, various features are grouped together in various embodiments for the purpose of streamlining the disclosure. This method of disclosure is not to be interpreted as reflecting an intention that the claimed embodiments require more features than are expressly recited in each claim. Rather, as the following claims reflect, inventive subject matter lies in less than all features of a single disclosed embodiment. Thus, the following claims are hereby incorporated into the Detailed Description, with each claim standing on its own as a separately claimed subject matter.

Claims
  • 1. A mid-air haptic device comprising: (a) a set of transducers with known relative locations;(b) a plurality of ultrasonic waves generated from the set of transducers having at least one common focal point; and(c) a preset function of desired haptic force versus time;
  • 2. The mid-air haptic device as in claim 1, wherein the preset function of desired haptic force versus time includes a desired direction of desired force.
  • 3. The mid-air haptic device as in claim 2, wherein modulating the generation of ultrasonic waves substantially generates the nonlinear acoustic force in the desired direction.
  • 4. A method of producing an acoustic field, comprising: implementing a control point set comprising a plurality of control points, each of which has a control point amplitude and a known relative positions and orientations from a transducer array, and each of which comprises: (a) at least one transducer array tile, each comprising: (i) a local computation unit; and (ii) a set of locally addressable transducers; and (b) at least one global computation unit;each local computation unit computing a wave field sample matrix comprising a summed effect that each control point has on the other control points in the control point set when generated from the set of locally addressable transducers;the set of local computation units combining the wave field sample matrices through a sum-reduction operation over the set of wave field sample matrices and communicating the resulting total wave field sample matrix to the at least one global computation unit;wherein each global computation unit uses the control point set and the total wave field sample matrix to compute and broadcast the set of control point coefficients;wherein each local computation unit computes transducer activation coefficients for the locally attached transducers using a set of control point coefficients; andactuating the locally attached transducers using the transducer activation coefficients computed.
  • 5. The method of claim 4, wherein the global computation unit uses a power iteration of a modified total wave field sample matrix to find a dominant eigenvector, wherein the dominant eigenvector informs phases of the control point set.
  • 6. The method of claim 4, wherein the global computation unit uses a linear system of equations to determine the set of control point coefficients.
  • 7. The method of claim 6, wherein the global computation unit is a virtual device hosted in the cloud.
  • 8. An ultrasonic phased array comprising: a set of transducers with known relative locations;at least one high-pressure point-of-interest relative to the transducer array;at least one low-pressure point-of-interest relative to the transducer array;a desired pressure at the least one high-pressure point-of-interest;a set of phase and amplitude driving conditions that generates the desired pressure at the at least one high-pressure point-of-interest and/or the at least one low-pressure point-of-interest;driving conditions such that ultrasonic acoustic pressure is reduced at the least one low-pressure point-of-interest while leaving the ultrasonic acoustic pressure substantially unchanged at the at the least one high-pressure point of interest.
  • 9. The ultrasonic phased array as in claim 8, wherein modification of the driving conditions occurs by considering zero-pressure points in a field.
  • 10. The ultrasonic phased array as in claim 8, wherein modification of driving conditions is calculated by adding helicity.
  • 11. The ultrasonic phased array as in claim 8, wherein modification of driving conditions is calculated by treating the transducer array as multiple sub-arrays.
  • 12. A method of producing an acoustic field from a transducer array having known relative positions and orientations comprising: defining at least one control point wherein each control point: (a) has a known spatial relationship relative to the transducer array which is used to define a corresponding focal point basis function belonging to a first set of focal point basis functions; and (b) has a control point amplitude;computing potentially estimated transducer activations required to produce an acoustic field corresponding to a control point set with the first set of focal point basis functions comprised of transducer coefficients to produce a first set of potentially estimated transducer activation amplitudes;adjusting each transducer coefficient of the first set of focal point basis functions using the potentially estimated transducer activation amplitude for the transducer in combination with statistical properties of the potentially estimated transducer activation amplitudes to produce a second set of focal point basis functions;computing transducer activations required to produce the acoustic field with the second set of focal point basis functions;using the second set of transducer activations to drive the transducer array.
  • 13. The method as in claim 12, wherein the adjustment applied to each transducer coefficient is a scaling in a form of a per-transducer gain.
  • 14. The method as in claim 12, wherein the mean value is within statistical properties of the first set of potentially estimated transducer activation amplitudes.
  • 15. The method as in claim 12, wherein statistical properties of the first set of potentially estimated transducer activation amplitudes are obtained by estimated evaluation of quantile function for percentiles of a distribution.
  • 16. A method producing an acoustic field from a transducer array having known relative positions and orientations, comprising: producing the acoustic field from a transducer array having known relative positions and orientations;defining at least one control point wherein each control point: (a) has a known spatial relationship relative to the transducer array; and (b) has a control point activation coefficient;wherein at least one control point is indicated as a haptic pressure along a corresponding direction vector of action substantially parallel to a surface normal vector originating at the control point location;wherein, for each control point indicated as a haptic pressure, a conversion is applied from a difference of momentum fluxes into a scalar linear acoustic quantity.
  • 17. The method as in claim 16, where the scalar linear acoustic quantity is an acoustic pressure.
  • 18. The method as in claim 16, where the scalar linear acoustic quantity is a particle velocity of an acoustic medium along a direction vector of action.
  • 19. The method as in claim 16, wherein the acoustic field is produced by a mid-air haptic feedback system.
PRIOR APPLICATIONS

This application claims the benefit of the following five applications, all of which are incorporated by references in their entirety (1) U.S. Provisional Patent Application No. 63/043,093 filed on Jun. 23, 2020;(2) U.S. Provisional Patent Application No. 63/065,997, filed on Aug. 14, 2020;(3) U.S. Provisional Patent Application No. 63/067,314, filed on Aug. 18, 2020;(4) U.S. Provisional Patent Application No. 63/210,486, filed on Jun. 14, 2021; and(5) U.S. Provisional Patent Application No. 63/210,619, filed on Jun. 15, 2021.

US Referenced Citations (347)
Number Name Date Kind
4218921 Berge Aug 1980 A
4760525 Webb Jul 1988 A
4771205 Mequio Sep 1988 A
4881212 Takeuchi Nov 1989 A
5226000 Moses Jul 1993 A
5235986 Maslak Aug 1993 A
5243344 Koulopoulos Sep 1993 A
5329682 Thurn Jul 1994 A
5371834 Tawel Dec 1994 A
5422431 Ichiki Jun 1995 A
5426388 Flora Jun 1995 A
5477736 Lorraine Dec 1995 A
5511296 Dias Apr 1996 A
5729694 Holzrichter Mar 1998 A
5859915 Norris Jan 1999 A
6029518 Oeftering Feb 2000 A
6193936 Gardner Feb 2001 B1
6216538 Yasuda Apr 2001 B1
6436051 Morris Aug 2002 B1
6503204 Sumanaweera Jan 2003 B1
6647359 Verplank Nov 2003 B1
6771294 Pulli Aug 2004 B1
6772490 Toda Aug 2004 B2
6800987 Toda Oct 2004 B2
7107159 German Sep 2006 B2
7109789 Spencer Sep 2006 B2
7182726 Williams Feb 2007 B2
7225404 Zilles May 2007 B1
7284027 Jennings, III Oct 2007 B2
7345600 Fedigan Mar 2008 B1
7487662 Schabron Feb 2009 B2
7497662 Mollmann Mar 2009 B2
7577260 Hooley Aug 2009 B1
7692661 Cook Apr 2010 B2
RE42192 Schabron Mar 2011 E
7966134 German Jun 2011 B2
8000481 Nishikawa Aug 2011 B2
8123502 Blakey Feb 2012 B2
8269168 Axelrod Sep 2012 B1
8279193 Birnbaum Oct 2012 B1
8351646 Fujimura Jan 2013 B2
8369973 Risbo Feb 2013 B2
8594350 Hooley Nov 2013 B2
8607922 Werner Dec 2013 B1
8782109 Tsutsui Jul 2014 B2
8823674 Birnbaum Sep 2014 B2
8833510 Koh Sep 2014 B2
8884927 Cheatham, III Nov 2014 B1
9208664 Peters Dec 2015 B1
9267735 Funayama Feb 2016 B2
9421291 Robert Aug 2016 B2
9612658 Subramanian Apr 2017 B2
9662680 Yamamoto May 2017 B2
9667173 Kappus May 2017 B1
9816757 Zielinski Nov 2017 B1
9841819 Carter Dec 2017 B2
9863699 Corbin, III Jan 2018 B2
9898089 Subramanian Feb 2018 B2
9945818 Ganti Apr 2018 B2
9958943 Long May 2018 B2
9977120 Carter May 2018 B2
10101811 Carter Oct 2018 B2
10101814 Carter Oct 2018 B2
10133353 Eid Nov 2018 B2
10140776 Schwarz Nov 2018 B2
10146353 Smith Dec 2018 B1
10168782 Tchon Jan 2019 B1
10268275 Carter Apr 2019 B2
10281567 Carter May 2019 B2
10318008 Sinha Jun 2019 B2
10444842 Long Oct 2019 B2
10469973 Hayashi Nov 2019 B2
10496175 Long Dec 2019 B2
10497358 Tester Dec 2019 B2
10510357 Kovesi Dec 2019 B2
10520252 Mehdizadeh Dec 2019 B2
10523159 Megretski Dec 2019 B2
10531212 Long Jan 2020 B2
10535174 Rigiroli Jan 2020 B1
10569300 Hoshi Feb 2020 B2
10593101 Han Mar 2020 B1
10657704 Han May 2020 B1
10685538 Carter Jun 2020 B2
10755538 Carter Aug 2020 B2
10818162 Carter Oct 2020 B2
10911861 Buckland Feb 2021 B2
10915177 Carter Feb 2021 B2
10921890 Subramanian Feb 2021 B2
10930123 Carter Feb 2021 B2
10943578 Long Mar 2021 B2
11048329 Lee Jun 2021 B1
11098951 Kappus Aug 2021 B2
11113860 Rigiroli Sep 2021 B2
11169610 Sarafianou Nov 2021 B2
11189140 Long Nov 2021 B2
11204644 Long Dec 2021 B2
11276281 Carter Mar 2022 B2
11531395 Kappus Dec 2022 B2
11543507 Carter Jan 2023 B2
11550395 Beattie Jan 2023 B2
11550432 Carter Jan 2023 B2
11553295 Kappus Jan 2023 B2
20010007591 Pompei Jul 2001 A1
20010033124 Norris Oct 2001 A1
20020149570 Knowles Oct 2002 A1
20030024317 Miller Feb 2003 A1
20030144032 Brunner Jul 2003 A1
20030182647 Radeskog Sep 2003 A1
20040005715 Schabron Jan 2004 A1
20040014434 Haardt Jan 2004 A1
20040052387 Norris Mar 2004 A1
20040091119 Duraiswami May 2004 A1
20040210158 Organ Oct 2004 A1
20040226378 Oda Nov 2004 A1
20040264707 Yang Dec 2004 A1
20050052714 Klug Mar 2005 A1
20050056851 Althaus Mar 2005 A1
20050212760 Marvit Sep 2005 A1
20050226437 Pellegrini Oct 2005 A1
20050267695 German Dec 2005 A1
20050273483 Dent Dec 2005 A1
20060085049 Cory Apr 2006 A1
20060090955 Cardas May 2006 A1
20060091301 Trisnadi May 2006 A1
20060164428 Cook Jul 2006 A1
20070036492 Lee Feb 2007 A1
20070094317 Wang Apr 2007 A1
20070177681 Choi Aug 2007 A1
20070214462 Boillot Sep 2007 A1
20070236450 Colgate Oct 2007 A1
20070263741 Erving Nov 2007 A1
20080012647 Risbo Jan 2008 A1
20080027686 Mollmann Jan 2008 A1
20080084789 Altman Apr 2008 A1
20080130906 Goldstein Jun 2008 A1
20080152191 Fujimura Jun 2008 A1
20080226088 Aarts Sep 2008 A1
20080273723 Hartung Nov 2008 A1
20080300055 Lutnick Dec 2008 A1
20090093724 Pernot Apr 2009 A1
20090116660 Croft, III May 2009 A1
20090232684 Hirata Sep 2009 A1
20090251421 Bloebaum Oct 2009 A1
20090319065 Risbo Dec 2009 A1
20100013613 Weston Jan 2010 A1
20100016727 Rosenberg Jan 2010 A1
20100030076 Vortman Feb 2010 A1
20100044120 Richter Feb 2010 A1
20100066512 Rank Mar 2010 A1
20100085168 Kyung Apr 2010 A1
20100103246 Schwerdtner Apr 2010 A1
20100109481 Buccafusca May 2010 A1
20100199232 Mistry Aug 2010 A1
20100231508 Cruz-Hernandez Sep 2010 A1
20100262008 Roundhill Oct 2010 A1
20100302015 Kipman Dec 2010 A1
20100321216 Jonsson Dec 2010 A1
20110006888 Bae Jan 2011 A1
20110010958 Clark Jan 2011 A1
20110051554 Varray Mar 2011 A1
20110066032 Vitek Mar 2011 A1
20110199342 Vartanian Aug 2011 A1
20110310028 Camp, Jr. Dec 2011 A1
20120057733 Morii Mar 2012 A1
20120063628 Frank Mar 2012 A1
20120066280 Tsutsui Mar 2012 A1
20120223880 Birnbaum Sep 2012 A1
20120229400 Birnbaum Sep 2012 A1
20120229401 Birnbaum Sep 2012 A1
20120236689 Brown Sep 2012 A1
20120243374 Dahl Sep 2012 A1
20120249409 Toney Oct 2012 A1
20120249474 Pratt Oct 2012 A1
20120299853 Dagar Nov 2012 A1
20120307649 Park Dec 2012 A1
20120315605 Cho Dec 2012 A1
20130035582 Radulescu Feb 2013 A1
20130079621 Shoham Mar 2013 A1
20130094678 Scholte Apr 2013 A1
20130100008 Marti Apr 2013 A1
20130101141 Mcelveen Apr 2013 A1
20130173658 Adelman Jul 2013 A1
20130331705 Fraser Dec 2013 A1
20140027201 Islam Jan 2014 A1
20140104274 Hilliges Apr 2014 A1
20140139071 Yamamoto May 2014 A1
20140168091 Jones Jun 2014 A1
20140201666 Bedikian Jul 2014 A1
20140204002 Bennet Jul 2014 A1
20140265572 Siedenburg Sep 2014 A1
20140267065 Levesque Sep 2014 A1
20140269207 Baym Sep 2014 A1
20140269208 Baym Sep 2014 A1
20140269214 Baym Sep 2014 A1
20140270305 Baym Sep 2014 A1
20140320436 Modarres Oct 2014 A1
20140361988 Katz Dec 2014 A1
20140369514 Baym Dec 2014 A1
20150002477 Cheatham, III Jan 2015 A1
20150005039 Liu Jan 2015 A1
20150006645 Oh Jan 2015 A1
20150007025 Sassi Jan 2015 A1
20150013023 Wang Jan 2015 A1
20150019299 Harvey Jan 2015 A1
20150022466 Levesque Jan 2015 A1
20150029155 Lee Jan 2015 A1
20150066445 Lin Mar 2015 A1
20150070147 Cruz-Hernandez Mar 2015 A1
20150070245 Han Mar 2015 A1
20150078136 Sun Mar 2015 A1
20150081110 Houston Mar 2015 A1
20150084929 Lee Mar 2015 A1
20150110310 Minnaar Apr 2015 A1
20150130323 Harris May 2015 A1
20150168205 Lee Jun 2015 A1
20150192995 Subramanian Jul 2015 A1
20150220199 Wang Aug 2015 A1
20150226537 Schorre Aug 2015 A1
20150226831 Nakamura Aug 2015 A1
20150241393 Ganti Aug 2015 A1
20150248787 Abovitz Sep 2015 A1
20150258431 Stafford Sep 2015 A1
20150277610 Kim Oct 2015 A1
20150293592 Cheong Oct 2015 A1
20150304789 Babayoff Oct 2015 A1
20150323667 Przybyla Nov 2015 A1
20150331576 Piya Nov 2015 A1
20150332075 Burch Nov 2015 A1
20160019762 Levesque Jan 2016 A1
20160019879 Daley Jan 2016 A1
20160026253 Bradski Jan 2016 A1
20160044417 Clemen, Jr. Feb 2016 A1
20160124080 Carter May 2016 A1
20160138986 Carlin May 2016 A1
20160175701 Froy Jun 2016 A1
20160175709 Idris Jun 2016 A1
20160189702 Blanc Jun 2016 A1
20160242724 Lavallee Aug 2016 A1
20160246374 Carter Aug 2016 A1
20160249150 Carter Aug 2016 A1
20160291716 Boser Oct 2016 A1
20160306423 Uttermann Oct 2016 A1
20160320843 Long Nov 2016 A1
20160339132 Cosman Nov 2016 A1
20160374562 Vertikov Dec 2016 A1
20170002839 Bukland Jan 2017 A1
20170004819 Ochiai Jan 2017 A1
20170018171 Carter Jan 2017 A1
20170024921 Beeler Jan 2017 A1
20170052148 William Feb 2017 A1
20170123487 Hazra May 2017 A1
20170123499 Eid May 2017 A1
20170140552 Woo May 2017 A1
20170144190 Hoshi May 2017 A1
20170153707 Subramanian Jun 2017 A1
20170168586 Sinha Jun 2017 A1
20170181725 Han Jun 2017 A1
20170193768 Long Jul 2017 A1
20170193823 Jiang Jul 2017 A1
20170211022 Reinke Jul 2017 A1
20170236506 Przybyla Aug 2017 A1
20170270356 Sills Sep 2017 A1
20170279951 Hwang Sep 2017 A1
20170336860 Smoot Nov 2017 A1
20170366908 Long Dec 2017 A1
20180035891 Van Soest Feb 2018 A1
20180039333 Carter Feb 2018 A1
20180047259 Carter Feb 2018 A1
20180074580 Hardee Mar 2018 A1
20180081439 Daniels Mar 2018 A1
20180101234 Carter Apr 2018 A1
20180139557 Ochiai May 2018 A1
20180146306 Benattar May 2018 A1
20180151035 Maalouf May 2018 A1
20180166063 Long Jun 2018 A1
20180181203 Subramanian Jun 2018 A1
20180182372 Tester Jun 2018 A1
20180190007 Panteleev Jul 2018 A1
20180246576 Long Aug 2018 A1
20180253627 Baradel Sep 2018 A1
20180267156 Carter Sep 2018 A1
20180304310 Long Oct 2018 A1
20180309515 Murakowski Oct 2018 A1
20180310111 Kappus Oct 2018 A1
20180350339 Macours Dec 2018 A1
20180361174 Radulescu Dec 2018 A1
20190038496 Levesque Feb 2019 A1
20190091565 Nelson Mar 2019 A1
20190163275 Iodice May 2019 A1
20190175077 Zhang Jun 2019 A1
20190187244 Riccardi Jun 2019 A1
20190196578 Iodice Jun 2019 A1
20190196591 Long Jun 2019 A1
20190197840 Kappus Jun 2019 A1
20190197841 Carter Jun 2019 A1
20190197842 Long Jun 2019 A1
20190204925 Long Jul 2019 A1
20190206202 Carter Jul 2019 A1
20190235628 Lacroix Aug 2019 A1
20190257932 Carter Aug 2019 A1
20190310710 Deeley Oct 2019 A1
20190342654 Buckland Nov 2019 A1
20200042091 Long Feb 2020 A1
20200080776 Kappus Mar 2020 A1
20200082804 Kappus Mar 2020 A1
20200103974 Carter Apr 2020 A1
20200117229 Long Apr 2020 A1
20200193269 Park Jun 2020 A1
20200218354 Beattie Jul 2020 A1
20200257371 Sung Aug 2020 A1
20200294299 Rigiroli Sep 2020 A1
20200302760 Carter Sep 2020 A1
20200320347 Nikolenko Oct 2020 A1
20200327418 Lyons Oct 2020 A1
20200380832 Carter Dec 2020 A1
20210037332 Kappus Feb 2021 A1
20210043070 Carter Feb 2021 A1
20210109712 Oliver Apr 2021 A1
20210111731 Oliver Apr 2021 A1
20210112353 Kappus Apr 2021 A1
20210141458 Sarafianou May 2021 A1
20210165491 Sun Jun 2021 A1
20210170447 Buckland Jun 2021 A1
20210183215 Carter Jun 2021 A1
20210201884 Kappus Jul 2021 A1
20210225355 Long Jul 2021 A1
20210303072 Carter Sep 2021 A1
20210303758 Long Sep 2021 A1
20210334706 Yamaguchi Oct 2021 A1
20210381765 Kappus Dec 2021 A1
20210397261 Kappus Dec 2021 A1
20220035479 Lasater Feb 2022 A1
20220083142 Brown Mar 2022 A1
20220095068 Kappus Mar 2022 A1
20220113806 Long Apr 2022 A1
20220155949 Ring May 2022 A1
20220198892 Carter Jun 2022 A1
20220236806 Carter Jul 2022 A1
20220252550 Catsis Aug 2022 A1
20220300028 Long Sep 2022 A1
20220300070 Iodice Sep 2022 A1
20220329250 Long Oct 2022 A1
20220393095 Chilles Dec 2022 A1
20230036123 Long Feb 2023 A1
20230075917 Pittera Mar 2023 A1
20230117919 Iodice Apr 2023 A1
20230124704 Rorke Apr 2023 A1
Foreign Referenced Citations (71)
Number Date Country
2470115 Jun 2003 CA
2909804 Nov 2014 CA
101986787 Mar 2011 CN
102459900 May 2012 CN
102591512 Jul 2012 CN
103797379 May 2014 CN
103984414 Aug 2014 CN
107340871 Nov 2017 CN
107407969 Nov 2017 CN
107534810 Jan 2018 CN
0057594 Aug 1982 EP
309003 Mar 1989 EP
0696670 Feb 1996 EP
1875081 Jan 2008 EP
1911530 Apr 2008 EP
2271129 Jan 2011 EP
1461598 Apr 2014 EP
3207817 Aug 2017 EP
3216231 Aug 2019 EP
3916525 Dec 2021 EP
2464117 Apr 2010 GB
2513884 Nov 2014 GB
2513884 Nov 2014 GB
2530036 Mar 2016 GB
2008074075 Apr 2008 JP
2010109579 May 2010 JP
2011172074 Sep 2011 JP
2012048378 Mar 2012 JP
2012048378 Mar 2012 JP
5477736 Apr 2014 JP
2015035657 Feb 2015 JP
2016035646 Mar 2016 JP
2017168086 Sep 2017 JP
6239796 Nov 2017 JP
20120065779 Jun 2012 KR
20130055972 May 2013 KR
1020130055972 May 2013 KR
20160008280 Jan 2016 KR
20200082449 Jul 2020 KR
9118486 Nov 1991 WO
9639754 Dec 1996 WO
03050511 Jun 2003 WO
2005017965 Feb 2005 WO
2007144801 Dec 2007 WO
2009071746 Jun 2009 WO
2009112866 Sep 2009 WO
2010003836 Jan 2010 WO
2010139916 Dec 2010 WO
2011132012 Oct 2011 WO
2012023864 Feb 2012 WO
2012104648 Aug 2012 WO
2013179179 Dec 2013 WO
2014181084 Nov 2014 WO
2014181084 Nov 2014 WO
2015006467 Jan 2015 WO
2015039622 Mar 2015 WO
2015127335 Aug 2015 WO
2015194510 Dec 2015 WO
2016007920 Jan 2016 WO
2016073936 May 2016 WO
2016095033 Jun 2016 WO
2016099279 Jun 2016 WO
2016132141 Aug 2016 WO
2016132144 Aug 2016 WO
2016137675 Sep 2016 WO
2016162058 Oct 2016 WO
2017172006 Oct 2017 WO
WO-2018109466 Jun 2018 WO
2020049321 Mar 2020 WO
2021130505 Jul 2021 WO
WO-2021260373 Dec 2021 WO
Non-Patent Literature Citations (339)
Entry
“Welcome to Project Soli” video, https://atap.google.eom/#project-soli Accessed Nov. 30, 2018, 2 pages.
A. Sand, Head-Mounted Display with Mid-Air Tactile Feedback, Proceedings of the 21st ACM Symposium on Virtual Reality Software and Technology, Nov. 13-15, 2015 (8 pages).
Alexander, J. et al. (2011), Adding Haptic Feedback to Mobile TV (6 pages).
Aoki et al., Sound location of stero reproduction with parametric loudspeakers, Applied Acoustics 73 (2012) 1289-1295 (7 pages).
Ashish Shrivastava et al., Learning from Simulated and Unsupervised Images through Adversarial Training, Jul. 19, 2017, pp. 1-16.
Bajard et al., BKM: A New Hardware Algorithm for Complex Elementary Functions, 8092 IEEE Transactions on Computers 43 (1994) (9 pages).
Bajard et al., Evaluation of Complex Elementary Functions / A New Version of BKM, SPIE Conference on Advanced Signal Processing, Jul. 1999 (8 pages).
Benjamin Long et al, “Rendering volumetric haptic shapes in mid-air using ultrasound”, ACM Transactions on Graphics (TOG), ACM, US, (Nov. 19, 2014), vol. 33, No. 6, ISSN 0730-0301, pp. 1-10.
Bortoff et al., Pseudolinearization of the Acrobot using Spline Functions, IEEE Proceedings of the 31st Conference on Decision and Control, Sep. 10, 1992 (6 pages).
B{dot over (o)}zena Smagowska & Małgorzata Pawlaczyk-Łuszczyńska (2013) Effects of Ultrasonic Noise on the Human Body—A Bibliographic Review, International Journal of Occupational Safety and Ergonomics, 19:2, 195-202.
Brian Kappus and Ben Long, Spatiotemporal Modulation for Mid-Air Haptic Feedback from an Ultrasonic Phased Array, ICSV25, Hiroshima, Jul. 8-12, 2018, 6 pages.
Canada Application 2,909,804 Office Action dated Oct. 18, 2019, 4 pages.
Casper et al., Realtime Control of Multiple-focus Phased Array Heating Patterns Based on Noninvasive Ultrasound Thermography, IEEE Trans Biomed Eng. Jan. 2012; 59(1): 95-105.
Christoper M. Bishop, Pattern Recognition and Machine Learning, 2006, pp. 1-758.
Colgan, A., “How Does the Leap Motion Controller Work?” Leap Motion, Aug. 9, 2014, 10 pages.
Corrected Notice of Allowability dated Jan. 14, 2021 for U.S. Appl. No. 15/897,804 (pp. 1-2).
Corrected Notice of Allowability dated Jun. 21, 2019 for U.S. Appl. No. 15/966,213 (2 pages).
Corrected Notice of Allowability dated Oct. 31, 2019 for U.S. Appl. No. 15/623,516 (pp. 1-2).
Damn Geeky, “Virtual projection keyboard technology with haptic feedback on palm of your hand,” May 30, 2013, 4 pages.
David Joseph Tan et al., Fits like a Glove: Rapid and Reliable Hand Shape Personalization, 2016 IEEE Conference on Computer Vision and Pattern Recognition, pp. 5610-5619.
Definition of “lnterferometry”according to Wikipedia, 25 pages., Retrieved Nov. 2018.
Definition of “Multilateration” according to Wikipedia, 7 pages., Retrieved Nov. 2018.
Definition of “Trilateration” according to Wikipedia, 2 pages., Retrieved Nov. 2018.
Diederik P. Kingma et al., Adam: A Method for Stochastic Optimization, Jan. 30, 2017, pp. 1-15.
E. Bok, Metasurface for Water-to-Air Sound Transmission, Physical Review Letters 120, 044302 (2018) (6 pages).
E.S. Ebbini et al. (1991), A spherical-section ultrasound phased array applicator for deep localized hyperthermia, Biomedical Engineering, IEEE Transactions on (vol. 38 Issue: 7), pp. 634-643.
EPO Office Action for EP16708440.9 dated Sep. 12, 2018 (7 pages).
EPSRC Grant summary EP/J004448/1 (2011) (1 page).
Eric Tzeng et al., Adversarial Discriminative Domain Adaptation, Feb. 17, 2017, pp. 1-10.
European Office Action for Application No. EP16750992.6, dated Oct. 2, 2019, 3 pages.
Ex Parte Quayle Action dated Dec. 28, 2018 for U.S. Appl. No. 15/966,213 (pp. 1-7).
Extended European Search Report for Application No. EP19169929.7, dated Aug. 6, 2019, 7 pages.
Freeman et al., Tactile Feedback for Above-Device Gesture Interfaces: Adding Touch to Touchless Interactions ICMI'14, Nov. 12-16, 2014, Istanbul, Turkey (8 pages).
Gavrilov L R et al (2000) “A theoretical assessment of the relative performance of spherical phased arrays for ultrasound surgery” Ultrasonics, Ferroelectrics, and Frequency Control, IEEE Transactions on (vol. 47, Issue: 1), pp. 125-139.
Gavrilov, L.R. (2008) “The Possibility of Generating Focal Regions of Complex Configurations in Application to the Problems of Stimulation of Human Receptor Structures by Focused Ultrasound” Acoustical Physics, vol. 54, No. 2, pp. 269-278.
Georgiou et al., Haptic In-Vehicle Gesture Controls, Adjunct Proceedings of the 9th International ACM Conference on Automotive User Interfaces and Interactive Vehicular Applications (AutomotiveUI '17), Sep. 24-27, 2017 (6 pages).
GitHub—danfis/libccd: Library for collision detection between two convex shapes, Mar. 26, 2020, pp. 1-6.
GitHub—IntelRealSense/hand_tracking_samples: researc codebase for depth-based hand pose estimation using dynamics based tracking and CNNs, Mar. 26, 2020, 3 pages.
Gokturk, et al., “ATime-of-Flight Depth Sensor-System Description, Issues and Solutions,” Published in: 2004 Conference on Computer Vision and Pattern Recognition Workshop, Date of Conference: Jun. 27-Jul. 2, 2004, 9 pages.
Hasegawa, K. and Shinoda, H. (2013) “Aerial Display of Vibrotactile Sensation with High Spatial-Temporal Resolution using Large Aperture Airbourne Ultrasound Phased Array”, University of Tokyo (6 pages).
Hilleges et al. Interactions in the air: adding further depth to interactive tabletops, UIST '09: Proceedings of the 22nd annual ACM symposium on User interface software and technologyOct. 2009 pp. 139-148.
Hoshi et al.,Tactile Presentation by Airborne Ultrasonic Oscillator Array, , Proceedings of Robotics and Mechatronics Lecture 2009, Japan Society of Mechanical Engineers; May 24, 2009 (5 pages).
Hoshi T et al, “Noncontact Tactile Display Based on Radiation Pressure of Airborne Ultrasound”, IEEE Transactions on Haptics, IEEE, USA, (Jul. 1, 2010), vol. 3, No. 3, ISSN 1939-1412, pp. 155-165.
Hoshi, T., Development of Aerial-Input and Aerial-Tactile-Feedback System, IEEE World Haptics Conference 2011, p. 569-573.
Hoshi, T., Handwriting Transmission System Using Noncontact Tactile Display, IEEE Haptics Symposium 2012 pp. 399-401.
Hoshi, T., Non-contact Tactile Sensation Synthesized by Ultrasound Transducers, Third Joint Euro haptics Conference and Symposium on Haptic Interfaces for Virtual Environment and Teleoperator Systems 2009 (5 pages).
Hoshi, T., Touchable Holography, SIGGRAPH 2009, New Orleans, Louisiana, Aug. 3-7, 2009. (1 page).
Hua J, Qin H., Haptics-based dynamic implicit solid modeling, IEEE Trans Vis Comput Graph. Sep.-Oct. 2004;10(5):574-86.
Iddan, et al., “3D Imaging in the Studio (And Elsewhwere . . . ” Apr. 2001, 3DV systems Ltd., Yokneam, Isreal, www.3dvsystems.com.il, 9 pages.
Imaginary Phone: Learning Imaginary Interfaces by Transferring Spatial Memory From a Familiar Device Sean Gustafson, Christian Holz and Patrick Baudisch. UIST 2011. (10 pages).
International Preliminary Report on Patentability and Written Opinion issued in corresponding PCT/US2017/035009, dated Dec. 4, 2018, 8 pages.
International Preliminary Report on Patentability for Application No. PCT/EP2017/069569 dated Feb. 5, 2019, 11 pages.
International Search Report and Written Opinion for Application No. PCT/GB2018/053738, dated Apr. 11, 2019, 14 pages.
International Search Report and Written Opinion for Application No. PCT/GB2018/053739, dated Jun. 4, 2019, 16 pages.
International Search Report and Written Opinion for Application No. PCT/GB2019/050969, dated Jun. 13, 2019, 15 pages.
International Search Report and Written Opinion for Application No. PCT/GB2019/051223, dated Aug. 8, 2019, 15 pages.
International Search Report and Written Opinion for Application No. PCT/GB2019/052510, dated Jan. 14, 2020, 25 pages.
ISR & WO for PCT/GB2020/052545 (dated Jan. 27, 2021) 14 pages.
ISR and WO for PCT/GB2020/050013 (dated Jul. 13, 2020) (20 pages).
ISR and WO for PCT/GB2020/050926 (dated Jun. 2, 2020) (16 pages).
ISR and WO for PCT/GB2020/052544 (dated Dec. 18, 2020) (14 pages).
ISR for PCT/GB2020/052546 (dated Feb. 23, 2021) (14 pages).
ISR for PCT/GB2020/053373 (dated Mar. 26, 2021) (16 pages).
Iwamoto et al. (2008), Non-contact Method for Producing Tactile Sensation Using Airborne Ultrasound, EuroHaptics, pp. 504-513.
Iwamoto et al., Airborne Ultrasound Tactile Display: Supplement, The University of Tokyo 2008 (2 pages).
Iwamoto T et al, “Two-dimensional Scanning Tactile Display using Ultrasound Radiation Pressure”, Haptic Interfaces for Virtual Environment and Teleoperator Systems, 2006 14th Symposium on Alexandria, VA, USA Mar. 25-26, 2006, Piscataway, NJ, USA,IEEE, (Mar. 25, 2006), ISBN 978-1-4244-0226-7, pp. 57-61.
Jager et al., “Air-Coupled 40-KHZ Ultrasonic 2D-Phased Array Based on a 3D-Printed Waveguide Structure”, 2017 IEEE, 4 pages.
Japanese Office Action (with English language translation) for Application No. 2017-514569, dated Mar. 31, 3019, 10 pages.
Jonathan Taylor et al., Articulated Distance Fields for Ultra-Fast Tracking of Hands Interacting, ACM Transactions on Graphics, vol. 36, No. 4, Article 244, Publication Date: Nov. 2017, pp. 1-12.
Jonathan Taylor et al., Efficient and Precise Interactive Hand Tracking Through Joint, Continuous Optimization of Pose and Correspondences, SIGGRAPH '16 Technical Paper, Jul. 24-28, 2016, Anaheim, CA, ISBN: 978-1-4503-4279-87/16/07, pp. 1-12.
Jonathan Tompson et al., Real-Time Continuous Pose Recovery of Human Hands Using Convolutional Networks, ACM Trans. Graph. 33, 5, Article 169, Aug. 2014, pp. 1-10.
K. Jia, Dynamic properties of micro-particles in ultrasonic transportation using phase-controlled standing waves, J. Applied Physics 116, n. 16 (2014) (12 pages).
Kaiming He et al., Deep Residual Learning for Image Recognition, http://image-net.org/challenges/LSVRC/2015/ and http://mscoco.org/dataset/#detections-challenge2015, Dec. 10, 2015, pp. 1-12.
Kamakura, T. and Aoki, K. (2006) “A Highly Directional Audio System using a Parametric Array in Air” WESPAC IX 2006 (8 pages).
Kolb, et al., “Time-of-Flight Cameras in Computer Graphics,” Computer Graphics forum, vol. 29 (2010), No. 1, pp. 141-159.
Konstantinos Bousmalis et al., Domain Separation Networks, 29th Conference on Neural Information Processing Sysgtems (NIPS 2016), Barcelona, Spain. Aug. 22, 2016, pp. 1-15.
Krim, et al., “Two Decades of Array Signal Processing Research—The Parametric Approach”, IEEE Signal Processing Magazine, Jul. 1996, pp. 67-94.
Lang, Robert, “3D Time-of-Flight Distance Measurement with Custom Solid-State Image Sensors in CMOS/CCD—Technology”, A dissertation submitted to Department of EE and CS at Univ. of Siegen, dated Jun. 28, 2000, 223 pages.
Large et al.,Feel the noise: Mid-air ultrasound haptics as a novel human-vehicle interaction paradigm, Applied Ergonomics (2019) (10 pages).
Li, Larry, “Time-of-Flight Camera—An Introduction,” Texas Instruments, Technical White Paper, SLOA190B—Jan. 2014 Revised May 2014, 10 pages.
Light, E.D., Progress in Two Dimensional Arrays for Real Time Volumetric Imaging, 1998 (17 pages).
M. Barmatz et al, “Acoustic radiation potential on a sphere in plane, cylindrical, and spherical standing wave fields”, The Journal of the Acoustical Society of America, New York, NY, US, (Mar. 1, 1985), vol. 77, No. 3, pp. 928-945, XP055389249.
M. Toda, New Type of Matching Layer for Air-Coupled Ultrasonic Transducers, IEEE Transactions on Ultrasonics, Ferroelecthcs, and Frequency Control, vol. 49, No. 7, Jul. 2002 (8 pages).
Mahdi Rad et al., Feature Mapping for Learning Fast and Accurate 3D Pose Inference from Synthetic Images, Mar. 26, 2018, pp. 1-14.
Marco A B Andrade et al, “Matrix method for acoustic levitation simulation”, IEEE Transactions on Ultrasonics, Ferroelectricsand Frequency Control, IEEE, US, (Aug. 1, 2011), vol. 58, No. 8, ISSN 0885-3010, pp. 1674-1683.
Marin, About LibHand, LibHand—A Hand Articulation Library, www.libhand.org/index.html, Mar. 26, 2020, pp. 1-2; www.libhand.org/download.html, 1 page; www.libhand.org/examples.html, pp. 1-2.
Markus Oberweger et al., DeepPrior++: Improving Fast and Accurate 3D Hand Pose Estimation, Aug. 28, 2017, pp. 1-10.
Markus Oberweger et al., Hands Deep in Deep Learning for Hand Pose Estimation, Dec. 2, 2016, pp. 1-10.
Marshall, M ., Carter, T Alexander, J., & Subramanian, S. (2012). Ultratangibles: creating movable tangible objects on interactive tables. In Proceedings of the 2012 ACM annual conference on Human Factors in Computing Systems, (pp. 2185-2188).
Marzo et al., Holographic acoustic elements for manipulation of levitated objects, Nature Communications DOI: I0.1038/ncomms9661 (2015) (7 pages).
Meijster, A., et al., “A General Algorithm for Computing Distance Transforms in Linear Time,” Mathematical Morphology and its Applications to Image and Signal Processing, 2002, pp. 331-340.
Mingzhu Lu et al. (2006) Design and experiment of 256-element ultrasound phased array for noninvasive focused ultrasound surgery, Ultrasonics, vol. 44, Supplement, Dec. 22, 2006, pp. e325-e330.
Mueller, GANerated Hands for Real-Time 3D Hand Tracking from Monocular RGB, Eye in-Painting with Exemplar Generative Adverserial Networks, pp. 49-59 (Jun. 1, 2018).
Nina Gaissert, Christian Wallraven, and Heinrich H. Bulthoff, “Visual and Haptic Perceptual Spaces Show High Similarity in Humans ”, published to Journal of Vision in 2010, available at http://www.journalofvision.org/content/10/11/2 and retrieved on Apr. 22, 2020 (Year: 2010), 20 pages.
Notice of Allowance dated Apr. 20, 2021 for U.S. Appl. No. 16/563,608 (pp. 1-5).
Notice of Allowance dated Apr. 22, 2020 for U.S. Appl. No. 15/671,107 (pp. 1-5).
Notice of Allowance dated Dec. 19, 2018 for U.S. Appl. No. 15/665,629 (pp. 1-9).
Notice of Allowance dated Dec. 21, 2018 for U.S. Appl. No. 15/983,864 (pp. 1-7).
Notice of Allowance dated Feb. 10, 2020, for U.S. Appl. No. 16/160,862 (pp. 1-9).
Notice of Allowance dated Feb. 7, 2019 for U.S. Appl. No. 15/851,214 (pp. 1-7).
Notice of Allowance dated Jul. 31, 2019 for U.S. Appl. No. 15/851,214 (pp. 1-9).
Notice of Allowance dated Jul. 31, 2019 for U.S. Appl. No. 16/296,127 (pp. 1-9).
Notice of Allowance dated Jun. 10, 2021 for U.S. Appl. No. 17/092,333 (pp. 1-9).
Notice of Allowance dated Jun. 17, 2020 for U.S. Appl. No. 15/210,661 (pp. 1-9).
Notice of Allowance dated Jun. 25, 2021 for U.S. Appl. No. 15/396,851 (pp. 1-10).
Notice of Allowance dated May 30, 2019 for U.S. Appl. No. 15/966,213 (pp. 1-9).
Notice of Allowance dated Oct. 1, 2020 for U.S. Appl. No. 15/897,804 (pp. 1-9).
Notice of Allowance dated Oct. 16, 2020 for U.S. Appl. No. 16/159,695 (pp. 1-7).
Notice of Allowance dated Oct. 30, 2020 for U.S. Appl. No. 15/839,184 (pp. 1-9).
Notice of Allowance dated Oct. 6, 2020 for U.S. Appl. No. 16/699,629 (pp. 1-8).
Notice of Allowance dated Sep. 30, 2020 for U.S. Appl. No. 16/401,148 (pp. 1-10).
Notice of Allowance in U.S. Appl. No. 15/210,661 dated Jun. 17, 2020 (22 pages).
Obrist et al., Emotions Mediated Through Mid-Air Haptics, CHI 2015, Apr. 18-23, 2015, Seoul, Republic of Korea. (10 pages).
Obrist et al., Talking about Tactile Experiences, CHI 2013, Apr. 27-May 2, 2013 (10 pages).
Office Action dated Apr. 8, 2020, for U.S. Appl. No. 16/198,959 (pp. 1-17).
Office Action dated Apr. 16, 2020 for U.S. Appl. No. 15/839,184 (pp. 1-8).
Office Action dated Apr. 17, 2020 for U.S. Appl. No. 16/401,148 (pp. 1-15).
Office Action dated Apr. 18, 2019 for U.S. Appl. No. 16/296,127 (pags 1-6).
Office Action dated Apr. 28, 2020 for U.S. Appl. No. 15/396,851 (pp. 1-12).
Office Action dated Apr. 29, 2020 for U.S. Appl. No. 16/374,301 (pp. 1-18).
Office Action dated Apr. 4, 2019 for U.S. Appl. No. 15/897,804 (pp. 1-10).
Office Action dated Aug. 22, 2019 for U.S. Appl. No. 16/160,862 (pp. 1-5).
Office Action dated Dec. 11, 2019 for U.S. Appl. No. 15/959,266 (pp. 1-15).
Office Action dated Dec. 7, 2020 for U.S. Appl. No. 16/563,608 (pp. 1-8).
Office Action dated Feb. 20, 2019 for U.S. Appl. No. 15/623,516 (pp. 1-8).
Office Action dated Feb. 25, 2020 for U.S. Appl. No. 15/960,113 (pp. 1-7).
Office Action dated Feb. 7, 2020 for U.S. Appl. No. 16/159,695 (pp. 1-8).
Office Action dated Jan. 10, 2020 for U.S. Appl. No. 16/228,767 (pp. 1-6).
Office Action dated Jan. 29, 2020 for U.S. Appl. No. 16/198,959 (p. 1-6).
Office Action dated Jul. 10, 2019 for U.S. Appl. No. 15/210,661 (pp. 1-12).
Office Action dated Jul. 26, 2019 for U.S. Appl. No. 16/159,695 (pp. 1-8).
Office Action dated Jul. 9, 2020 for U.S. Appl. No. 16/228,760 (pp. 1-17).
Office Action dated Jun. 19, 2020 for U.S. Appl. No. 16/699,629 (pp. 1-12).
Office Action dated Jun. 25, 2020 for U.S. Appl. No. 16/228,767 (pp. 1-27).
Office Action dated Jun. 25, 2021 for U.S. Appl. No. 16/899,720 (pp. 1-5).
Office Action dated Mar. 11, 2021 for U.S. Appl. No. 16/228,767 (pp. 1-23).
Office Action dated Mar. 20, 2020 for U.S. Appl. No. 15/210,661 (pp. 1-10).
Office Action dated Mar. 31, 2021 for U.S. Appl. No. 16/228,760 (pp. 1-21).
Office Action dated May 13, 2021 for U.S. Appl. No. 16/600,500 (pp. 1-9).
Office Action dated May 14, 2021 for U.S. Appl. No. 16/198,959 (pp. 1-6).
Office Action dated May 16, 2019 for U.S. Appl. No. 15/396,851 (pp. 1-7).
Office Action dated May 18, 2020 for U.S. Appl. No. 15/960,113 (pp. 1-21).
Office Action dated Oct. 17, 2019 for U.S. Appl. No. 15/897,804 (pp. 1-10).
Office Action dated Oct. 31, 2019 for U.S. Appl. No. 15/671,107 (pp. 1-6).
Office Action dated Oct. 7, 2019 for U.S. Appl. No. 15/396,851 (pp. 1-9).
Office Action dated Sep. 18, 2020 for U.S. Appl. No. 15/396,851 (pp. 1-14).
Office Action dated Sep. 21, 2020 for U.S. Appl. No. 16/198,959 (pp. 1-17).
OGRECave/ogre—GitHub: ogre/Samples/Media/materials at 7de80a7483f20b50f2b10d7ac6de9d9c6c87d364, Mar. 26, 2020, 1 page.
Optimal regularisation for acoustic source reconstruction by inverse methods, Y. Kim, P.A. Nelson, Institute of Sound and Vibration Research, University of Southampton, Southampton, SO17 1BJ, UK; 25 pages.
Oscar Martínez-Graullera et al, “2D array design based on Fermat spiral for ultrasound imaging”, Ultrasonics, (Feb. 1, 2010), vol. 50, No. 2, ISSN 0041-624X, pp. 280-289, XP055210119.
Partial International Search Report for Application No. PCT/GB2018/053735, dated Apr. 12, 2019, 14 pages.
Partial ISR for Application No. PCT/GB2020/050013 dated May 19, 2020 (16 pages).
PCT Partial International Search Report for Application No. PCT/GB2018/053404 dated Feb. 25, 2019, 13 pages.
Péter Tamás Kovács et al, “Tangible Holographic 3D Objects with Virtual Touch”, Interactive Tabletops & Surfaces, ACM, 2 Penn Plaza, Suite 701 New York NY 10121-0701 USA, (Nov. 15, 2015), ISBN 978-1-4503-3899-8, pp. 319-324.
Phys.org, Touchable Hologram Becomes Reality, Aug. 6, 2009, by Lisa Zyga (2 pages).
Pompei, F.J. (2002), “Sound from Ultrasound: The Parametric Array as an Audible Sound Source”, Massachusetts Institute of Technology (132 pages).
Rocchesso et al., Accessing and Selecting Menu Items by In-Air Touch, ACM CHItaly'19, Sep. 23-25, 2019, Padova, Italy (9 pages).
Schmidt, Ralph, “Multiple Emitter Location and Signal Parameter Estimation” IEEE Transactions of Antenna and Propagation, vol. AP-34, No. 3, Mar. 1986, pp. 276-280.
Sean Gustafson et al., “Imaginary Phone”, Proceedings of the 24th Annual ACM Symposium on User Interface Software and Techology: Oct. 16-19, 2011, Santa Barbara, CA, USA, ACM, New York, NY, Oct. 16, 2011, pp. 283-292, XP058006125, DOI: 10.1145/2047196.2047233, ISBN: 978-1-4503-0716-1.
Search report and Written Opinion of ISA for PCT/GB2015/050417 dated Jul. 8, 2016 (20 pages).
Search report and Written Opinion of ISA for PCT/GB2015/050421 dated Jul. 8, 2016 (15 pages).
Search report and Written Opinion of ISA for PCT/GB2017/050012 dated Jun. 8, 2017. (18 pages).
Search Report by EPO for EP 17748466 dated Jan. 13, 2021 (16 pages).
Search Report for GB1308274.8 dated Nov. 11, 2013. (2 pages).
Search Report for GB1415923.0 dated Mar. 11, 2015. (1 page).
Search Report for PCT/GB/2017/053729 dated Mar. 15, 2018 (16 pages).
Search Report for PCT/GB/2017/053880 dated Mar. 21, 2018. (13 pages).
Search report for PCT/GB2014/051319 dated Dec. 8, 2014 (4 pages).
Search report for PCT/GB2015/052507 dated Mar. 11, 2020 (19 pages).
Search report for PCT/GB2015/052578 dated Oct. 26, 2015 (12 pages).
Search report for PCT/GB2015/052916 dated Feb. 26, 2020 (18 pages).
Search Report for PCT/GB2017/052332 dated Oct. 10, 2017 (12 pages).
Search report for PCT/GB2018/051061 dated Sep. 26, 2018 (17 pages).
Search report for PCT/US2018/028966 dated Jul. 13, 2018 (43 pages).
Sergey Ioffe et al., Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariat Shift, Mar. 2, 2015, pp. 1-11.
Seungryul, Pushing the Envelope for RGB-based Dense 3D Hand Pose Estimation for RGB-based Desne 3D Hand Pose Estimation via Neural Rendering, arXiv:1904.04196v2 [cs.CV] Apr. 9, 2019 (5 pages).
Shakeri, G., Williamson, J. H. and Brewster, S. (2018) May the Force Be with You: Ultrasound Haptic Feedback for Mid-Air Gesture Interaction in Cars. In: 10th International ACM Conference on Automotive User Interfaces and Interactive Vehicular Applications (AutomotiveUI 2018) (11 pages).
Shanxin Yuan et al., BigHand2.2M Bechmark: Hand Pose Dataset and State of the Art Analysis, Dec. 9, 2017, pp. 1-9.
Shome Subhra Das, Detectioin of Self Intersection in Synthetic Hand Pose Generators, 2017 Fifteenth IAPR International Conference on Machine Vision Applications (MVA), Nagoya University, Nagoya, Japan, May 8-12, 2017, pp. 354-357.
Sixth Sense webpage, http://www.pranavmistry.com/projects/sixthsense/ Accessed Nov. 30, 2018, 7 pages.
Stan Melax et al., Dynamics Based 3D Skeletal Hand Tracking, May 22, 2017, pp. 1-8.
Steve Guest et al., “Audiotactile interactions in roughness perception”, Exp. Brain Res (2002) 146:161-171, DOI 10.1007/00221-002-1164-, Accepted: May 16, 2002/Published online: Jul. 26, 2002, Springer-Verlag 2002, (11 pages).
Sylvia Gebhardt, Ultrasonic Transducer Arrays for Particle Manipulation (date unknown) (2 pages).
Takahashi Dean: “Ultrahaptics shows off sense of touch in virtual reality”, Dec. 10, 2016 (Dec. 10, 2016), XP055556416, Retrieved from the Internet: URL: https://venturebeat.com/2016/12/10/ultrahaptics-shows-off-sense-of-touch-in-virtual-reality/ [retrieved on Feb. 13, 2019] 4 pages.
Takahashi, M. et al., Large Aperture Airborne Ultrasound Tactile Display Using Distributed Array Units, SICE Annual Conference 2010 p. 359-62.
Takayuki et al., “Noncontact Tactile Display Based on Radiation Pressure of Airborne Ultrasound” IEEE Transactions on Haptics vol. 3, No. 3, p. 165 (2010).
Teixeira, et al., “A brief introduction to Microsoft's Kinect Sensor,” Kinect, 26 pages, retrieved Nov. 2018.
Toby Sharp et al., Accurate, Robust, and Flexible Real-time Hand Tracking, CHI '15, Apr. 18-23, 2015, Seoul, Republic of Korea, ACM 978-1-4503-3145-6/15/04, pp. 1-10.
Tom Carter et al, “UltraHaptics: Multi-Point Mid-Air Haptic Feedback for Touch Surfaces”, Proceedings of the 26th Annual ACM Symposium on User Interface Software and Technology, UIST'13, New York, New York, USA, (Jan. 1, 2013), ISBN 978-1-45-032268-3, pp. 505-514.
Tom Nelligan and Dan Kass, Intro to Ultrasonic Phased Array (date unknown) (8 pages).
Vincent Lepetit et al., Model Based Augmentation and Testing of an Annotated Hand Pose Dataset, ResearchGate, https://www.researchgate.net/publication/307910344, Sep. 2016, 13 pages.
Wang et al., Device-Free Gesture Tracking Using Acoustic Signals, ACM MobiCom '16, pp. 82-94 (13 pages).
Wilson et al., Perception of Ultrasonic Haptic Feedback on the Hand: Localisation and Apparent Motion, CHI 2014, Apr. 26-May 1, 2014, Toronto, Ontario, Canada. (10 pages).
Wooh et al., “Optimum beam steering of linear phased arays,” Wave Motion 29 (1999) pp. 245-265, 21 pages.
Xin Cheng et al, “Computation of the acoustic radiation force on a sphere based on the 3-D FDTD method”, Piezoelectricity, Acoustic Waves and Device Applications (SPAWDA), 2010 Symposium ON, IEEE, (Dec. 10, 2010), ISBN 978-1-4244-9822-2, pp. 236-239.
Xu Hongyi et al, “6-DoF Haptic Rendering Using Continuous Collision Detection between Points and Signed Distance Fields”, IEEE Transactions on Haptics, IEEE, USA, vol. 10, No. 2, ISSN 1939-1412, (Sep. 27, 2016), pp. 151-161, (Jun. 16. 2017).
Yang Ling et al, “Phase-coded approach for controllable generation of acoustical vortices”, Journal of Applied Physics, American Institute of Physics, US, vol. 113, No. 15, ISSN 0021-8979, (Apr. 21, 2013), pp. 154904-154904.
Yarin Gal et al., Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning, Oct. 4, 2016, pp. 1-12, Proceedings of the 33rd International Conference on Machine Learning, New York, NY, USA, 2016, JMLR: W&CP vol. 48.
Yaroslav Ganin et al., Domain-Adversarial Training of Neural Networks, Journal of Machine Learning Research 17 (2016) 1-35, submitted May 2015; published Apr. 2016.
Yaroslav Ganin et al., Unsupervised Domain Adaptataion by Backpropagation, Skolkovo Institute of Science and Technology (Skoltech), Moscow Region, Russia, Proceedings of the 32nd International Conference on Machine Learning, Lille, France, 2015, JMLR: W&CP vol. 37, copyright 2015 by the author(s), 11 pages.
Yoshino, K. and Shinoda, H. (2013), “Visio Acoustic Screen for Contactless Touch Interface with Tactile Sensation”, University of Tokyo (5 pages).
Zeng, Wejun, “Microsoft Kinect Sensor and Its Effect,” IEEE Multimedia, Apr.-Jun. 2012, 7 pages.
A. B. Vallbo, Receptive field characteristics of tactile units with myelinated afferents in hairy skin of human subjects, Journal of Physiology (1995), 483.3, pp. 783-795.
Amanda Zimmerman, The gentle touch receptors of mammalian skin, Science, Nov. 21, 2014, vol. 346 Issue 6212, p. 950.
Corrected Notice of Allowability dated Aug. 9, 2021 for U.S. Appl. No. 15/396,851 (pp. 1-6).
Henrik Bruus, Acoustofluidics 2: Perturbation theory and ultrasound resonance modes, Lab Chip, 2012, 12, 20-28.
Hyunjae Gil, Whiskers: Exploring the Use of Ultrasonic Haptic Cues on the Face, CHI 2018, Apr. 21-26, 2018, Montreal, QC, Canada.
India Morrison, The skin as a social organ, Exp Brain Res (2010) 204:305-314.
Jonaschatel-Goldman, Touch increases autonomic coupling between romantic partners, Frontiers in Behavioral Neuroscience Mar. 2014, vol. 8, Article 95.
Kai Tsumoto, Presentation of Tactile Pleasantness Using Airborne Ultrasound, 2021 IEEE World Haptics Conference (WHC) Jul. 6-9, 2021. Montreal, Canada.
Keisuke Hasegawa, Electronically steerable ultrasound-driven long narrow airstream, Applied Physics Letters 111, 064104 (2017).
Keisuke Hasegawa, Midair Ultrasound Fragrance Rendering, IEEE Transactions on Visualization and Computer Graphics, vol. 24, No. 4, Apr. 2018 1477.
Keisuke Hasegawa,,Curved acceleration path of ultrasound-driven airflow, J. Appl. Phys. 125, 054902 (2019).
Line S Loken, Coding of pleasant touch by unmyelinated afferents in humans, Nature Neuroscience vol. 12 [ No. 5 [ May 2009 547.
Mariana von Mohr, The soothing function of touch: affective touch reduces feelings of social exclusion, Scientific Reports, 7: 13516, Oct. 18, 2017.
Mitsuru Nakajima, Remotely Displaying Cooling Sensation via Ultrasound-Driven Air Flow, Haptics Symposium 2018, San Francisco, USA p. 340.
Mohamed Yacine Tsalamlal, Affective Communication through Air Jet Stimulation: Evidence from Event-Related Potentials, International Journal of Human-Computer Interaction 2018.
Notice of Allowance dated Jul. 22, 2021 for U.S. Appl. No. 16/600,500 (pp. 1-9).
Office Action dated Aug. 10, 2021 for U.S. Appl. No. 16/564,016 (pp. 1-14).
Office Action dated Aug. 19, 2021 for U.S. Appl. No. 17/170,841 (pp. 1-9).
Office Action dated Aug. 9, 2021 for U.S. Appl. No. 17/068,825 (pp. 1-9).
Office Action dated Sep. 16, 2021 for U.S. Appl. No. 16/600,496 (pp. 1-8).
Office Action dated Sep. 24, 2021 for U.S. Appl. No. 17/080,840 (pp. 1-9).
Rochelle Ackerley, Human C-Tactile Afferents Are Tuned to the Temperature of a Skin-Stroking Caress, J. Neurosci., Feb. 19, 2014, 34(8):2879-2883.
Ryoko Takahashi, Tactile Stimulation by Repetitive Lateral Movement of Midair Ultrasound Focus, Journal of Latex Class Files, vol. 14, No. 8, Aug. 2015.
Stanley J. Bolanowski, Hairy Skin: Psychophysical Channels and Their Physiological Substrates, Somatosensory and Motor Research, vol. 11. No. 3, 1994, pp. 279-290.
Stefan G. Lechner, Hairy Sensation, Physiology 28: 142-150, 2013.
Supplemental Notice of Allowability dated Jul. 28, 2021 for U.S. Appl. No. 16/563,608 (pp. 1-2).
Supplemental Notice of Allowability dated Jul. 28, 2021 for U.S. Appl. No. 17/092,333 (pp. 1-2).
Takaaki Kamigaki, Noncontact Thermal and Vibrotactile Display Using Focused Airborne Ultrasound, EuroHaptics 2020, LNCS 12272, pp. 271-278, 2020.
Tomoo Kamakura, Acoustic streaming induced in focused Gaussian beams, J. Acoust. Soc. Am. 97 (5), Pt. 1, May 1995 p. 2740.
Uta Sailer, How Sensory and Affective Attributes Describe Touch Targeting C-Tactile Fibers, Experimental Psychology (2020), 67(4), 224-236.
Communication Pursuant to Article 94(3) EPC for EP 19723179.8 (dated Feb. 15, 2022), 10 pages.
EPO ISR and WO for PCT/GB2022/050204 (dated Apr. 7, 2022) (15 pages).
https://radiopaedia.org/articles/physical-principles-of-ultrasound-1?lang=GB (Accessed May 29, 2022).
IN 202047026493 Office Action dated Mar. 8, 2022, 6 pages.
ISR & WO For PCT/GB2021/052946, 15 pages.
Office Action (Final Rejection) dated Mar. 14, 2022 for U.S. Appl. No. 16/564,016 (pp. 1-12).
Office Action (Non-Final Rejection) dated Mar. 15, 2022 for U.S. Appl. No. 16/144,474 (pp. 1-13).
Office Action (Non-Final Rejection) dated Apr. 1, 2022 for U.S. Appl. No. 16/229,091 (pp. 1-10).
Office Action (Non-Final Rejection) dated May 2, 2022 for U.S. Appl. No. 17/068,831 (pp. 1-10).
EPO Communication for Application 18 811 906.9 (dated Nov. 29, 2021) (15 pages).
EPO Examination Report 17 748 4656.4 (dated Jan. 12, 2021) (16 pages).
Gareth Young et al . . . Designing Mid-Air Haptic Gesture Controlled User Interfaces for Cars, PACM on Human-Computer Interactions, Jun. 2020 (24 pages).
ISR and WO for PCT/GB2020/052829 (dated Feb. 10, 2021) (15 pages).
ISR and WO for PCT/GB2021/052415 (dated Dec. 22, 2021) (16 pages).
Mohamed Yacine Tsalamlal, Non-Intrusive Haptic Interfaces: State-of-the Art Survey, HAID 2013, LNCS 7989, pp. 1-9, 2013.
Office Action (Non-Final Rejection) dated Jan. 21, 2022 for U.S. Appl. No. 17/068,834 (pp. 1-12).
Office Action (Non-Final Rejection) dated Jan. 24, 2022 for U.S. Appl. No. 16/228,767 (pp. 1-22).
Office Action (Non-Final Rejection) dated Mar. 4, 2022 for U.S. Appl. No. 16/404,660 (pp. 1-5).
Office Action (Notice of Allowance and Fees Due (PTOL-85)) dated Jan. 18, 2022 for U.S. Appl. No. 16/899,720 (pp. 1-2).
Office Action (Notice of Allowance and Fees Due (PTOL-85)) dated Feb. 11, 2022 for U.S. Appl. No. 16/228,760 (pp. 1-8).
Office Action (Notice of Allowance and Fees Due (PTOL-85)) dated Feb. 28, 2022 for U.S. Appl. No. 17/068,825 (pp. 1-7).
Office Action (Notice of Allowance and Fees Due (PTOL-85)) dated Mar. 7, 2022 for U.S. Appl. No. 16/600,496 (pp. 1-5).
ISR & WO for PCT/GB2022/051388 (dated Aug. 30, 2022) (15 pages).
Office Action (Final Rejection) dated Sep. 16, 2022 for U.S. Appl. No. 16/404,660 (pp. 1-6).
Office Action (Non-Final Rejection) dated Aug. 29, 2022 for U.S. Appl. No. 16/995,819 (pp. 1-6).
Office Action (Non-Final Rejection) dated Sep. 21, 2022 for U.S. Appl. No. 17/721,315 (pp. 1-10).
Office Action (Notice of Allowance and Fees Due (PTOL-85)) dated Aug. 24, 2022 for U.S. Appl. No. 16/198,959 (pp. 1-6).
Office Action (Notice of Allowance and Fees Due (PTOL-85)) dated Aug. 31, 2022 for U.S. Appl. No. 16/198,959 (pp. 1-2).
Office Action (Notice of Allowance and Fees Due (PTOL-85)) dated Sep. 7, 2022 for U.S. Appl. No. 17/068,834 (pp. 1-8).
Office Action (Notice of Allowance and Fees Due (PTOL-85)) dated Sep. 8, 2022 for U.S. Appl. No. 17/176,899 (pp. 1-8).
Office Action (Notice of Allowance and Fees Due (PTOL-85)) dated Sep. 12, 2022 for U.S. Appl. No. 16/734,479 (pp. 1-7).
Al-Mashhadany, “Inverse Kinematics Problem (IKP) of 6-DOF Manipulator By Locally Recurrent Neural Networks (LRNNs),” Management and Service Science (MASS), International Conference on Management and Service Science., IEEE, Aug. 24, 2010, 5 pages. (Year: 2010).
Guez, “Solution to the inverse kinematic problem in robotics by neural networks” In Proceedings of the 2nd International Conference on Neural Networks, 1988. San Diego, California. (Year: 1988) 8 pages.
Invitation to Pay Additional Fees for PCT/GB2022/051821 (dated Oct. 20, 2022), 15 pages.
Mahboob, “Artificial neural networks for learning inverse kinematics of humanoid robot arms.” MS Thesis, 2015. (Year: 2015) 95 pages.
Office Action (Ex Parte Quayle Action) dated Jan. 6, 2023 for U.S. Appl. No. 17/195,795 (pp. 1-6).
Office Action (Final Rejection) dated Jan. 9, 2023 for U.S. Appl. No. 16/144,474 (pp. 1-16).
Office Action (Final Rejection) dated Nov. 18, 2022 for U.S. Appl. No. 16/228,767 (pp. 1-27).
Office Action (Final Rejection) dated Nov. 18, 2022 for U.S. Appl. No. 17/068,831 (pp. 1-9).
Office Action (Final Rejection) dated Dec. 8, 2022 for U.S. Appl. No. 16/229,091 (pp. 1-9).
Office Action (Final Rejection) dated Dec. 15, 2022 for U.S. Appl. No. 16/843,281 (pp. 1-25).
Office Action (Non-Final Rejection) dated Oct. 17, 2022 for U.S. Appl. No. 17/807,730 (pp. 1-8).
Office Action (Non-Final Rejection) dated Nov. 9, 2022 for U.S. Appl. No. 17/454,823 (pp. 1-16).
Office Action (Non-Final Rejection) dated Nov. 16, 2022 for U.S. Appl. No. 17/134,505 (pp. 1-7).
Office Action (Non-Final Rejection) dated Nov. 16, 2022 for U.S. Appl. No. 17/692,852 (pp. 1-4).
Office Action (Non-Final Rejection) dated Dec. 6, 2022 for U.S. Appl. No. 17/409,783 (pp. 1-7).
Office Action (Non-Final Rejection) dated Dec. 22, 2022 for U.S. Appl. No. 17/457,663 (pp. 1-20).
Office Action (Notice of Allowance and Fees Due (PTOL-85)) dated Oct. 31, 2022 for U.S. Appl. No. 17/068,834 (pp. 1-2).
Office Action (Notice of Allowance and Fees Due (PTOL-85)) dated Oct. 31, 2022 for U.S. Appl. No. 17/176,899 (pp. 1-2).
Office Action (Notice of Allowance and Fees Due (PTOL-85)) dated Nov. 1, 2022 for U.S. Appl. No. 16/404,660 (pp. 1-5).
Office Action (Notice of Allowance and Fees Due (PTOL-85)) dated Nov. 2, 2022 for U.S. Appl. No. 16/734,479 (pp. 1-2).
Office Action (Notice of Allowance and Fees Due (PTOL-85)) dated Nov. 10, 2022 for U.S. Appl. No. 16/198,959 (pp. 1-2).
Office Action (Notice of Allowance and Fees Due (PTOL-85)) dated Nov. 16, 2022 for U.S. Appl. No. 16/404,660 (pp. 1-2).
Almusawi et al., “A new artificial neural network approach in solving inverse kinematics of robotic arm (denso vp6242).” Computational intelligence and neuroscience 2016 (2016). (Year: 2016).
Azad et al., Deep domain adaptation under deep label scarcity. arXiv preprint arXiv:1809.08097 (2018) (Year: 2018).
Beranek, L., & Mellow, T. (2019). Acoustics: Sound Fields, Transducers and Vibration. Academic Press.
Boureau et al.,“A theoretical analysis of feature pooling in visual recognition.” In Proceedings of the 27th international conference on machine learning (ICML-10), pp. 111-118. 2010. (Year: 2010).
Bybi, A., Grondel, S., Mzerd, A., Granger, C., Garoum, M., & Assaad, J. (2019). Investigation of cross-coupling in piezoelectric transducer arrays and correction. International Journal of Engineering and Technology Innovation, 9(4), 287.
Certon, D., Felix, N., Hue, P. T. H., Patat, F., & Lethiecq, M. (Oct. 1999). Evaluation of laser probe performances for measuring cross-coupling in 1-3 piezocomposite arrays. In 1999 IEEE Ultrasonics Symposium. Proceedings. International Symposium (Cat. No. 99CH37027) (vol. 2, pp. 1091-1094).
Certon, D., Felix, N., Lacaze, E., Teston, F., & Patat, F. (2001). Investigation of cross-coupling in 1-3 piezocomposite arrays. ieee transactions on ultrasonics, ferroelectrics, and frequency control, 48(1), 85-92.
Chang Suk Lee et al., An electrically switchable visible to infra-red dual frequency cholesteric liquid crystal light shutter, J. Mater. Chem. C, 2018, 6,4243 (7 pages).
Der et al., Inverse kinematics for reduced deformable models. ACM Transactions on graphics (TOG) 25, No. 3 (2006): 1174-1179. (Year: 2006).
DeSilets, C. S. (1978). Transducer arrays suitable for acoustic imaging (No. GL-2833). Stanford Univ CA Edward L Ginzton Lab of Physics.
Duka, “Neural network based inverse kinematics solution for trajectory tracking of a robotic arm.” Procedia Technology 12 (2014) 20-27. (Year: 2014).
Henneberg, J., Geriach, A., Storck, H., Cebulla, H., & Marburg, S. (2018). Reducing mechanical cross-coupling in phased array transducers using stop band material as backing. Journal of Sound and Vibration, 424, 352-364.
Office Action (Non-Final Rejection) dated May 25, 2022 for U.S. Appl. No. 16/843,281 (pp. 1-28).
Office Action (Non-Final Rejection) dated Jun. 9, 2022 for U.S. Appl. No. 17/080,840 (pp. 1-9).
Office Action (Non-Final Rejection) dated Jun. 27, 2022 for U.S. Appl. No. 16/198,959 (pp. 1-17).
Office Action (Non-Final Rejection) dated Jun. 27, 2022 for U.S. Appl. No. 16/734,479 (pp. 1-13).
Oikonomidis et al., “Efficient model-based 3D tracking of hand articulations using Kinect.” In BmVC, vol. 1, No. 2, p. 3. 2011. (Year: 2011).
Patricio Rodrigues, E., Francisco de Oliveira, T., Yassunori Matuda, M., & Buiochi, F. (Sep. 2019). Design and Construction of a 2-D Phased Array Ultrasonic Transducer for Coupling in Water. In INTER-NOISE and NOISE-CON Congress and Conference Proceedings (vol. 259, No. 4, pp. 5720-5731). Institute of Noise Control Engineering.
Seo et al., “Improved numerical inverse kinematics for human pose estimation,” Opt. Eng. 50(3 037001 (Mar. 1, 2011) https://doi.org/10.1117/1.3549255 (Year: 2011).
Walter, S., Nieweglowski, K., Rebenklau, L., Wolter, K. J., Lamek, B., Schubert, F., . . . & Meyendorf, N. (May 2008). Manufacturing and electrical interconnection of piezoelectric 1-3 composite materials for phased array ultrasonic transducers. In 2008 31st International Spring Seminar on Electronics Technology (pp. 255-260).
Wang et al., Few-shot adaptive faster r-cnn. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 7173-7182. 2019. (Year: 2019).
EPO Examination Search Report 17 702 910.5 (dated Jun. 23, 2021).
Office Action dated Oct. 29, 2021 for U.S. Appl. No. 16/198,959 (pp. 1-7).
Notice of Allowance dated Nov. 5, 2021 for U.S. Appl. No. 16/899,720 (pp. 1-9).
Corrected Notice of Allowability dated Nov. 24, 2021 for U.S. Appl. No. 16/600,500 (pp. 1-5).
International Search Report and Written Opinion for App. No. PCT/GB2021/051590, dated Nov. 11, 2021, 20 pages.
Anonymous: “How does Ultrahaptics technology work?—Ultrahaptics Developer Information”, Jul. 31, 2018 (Jul. 31, 2018), XP055839320, Retrieved from the Internet: URL:https://developer.ultrahaptics.com/knowledgebase/haptics-overview/ [retrieved on Sep. 8, 2021].
Office Action (Notice of Allowance and Fees Due (PTOL-85)) dated Dec. 14, 2021 for U.S. Appl. No. 17/170,841 (pp. 1-8).
Office Action (Non-Final Rejection) dated Dec. 20, 2021 for U.S. Appl. No. 17/195,795 (pp. 1-7).
EPO Application 18 725 358.8 Examination Report dated Sep. 22, 2021.
EPO 21186570.4 Extended Search Report dated Oct. 29, 2021.
Aksel Sveier et al.,Pose Estimation with Dual Quaternions and Iterative Closest Point, 2018 Annual American Control Conference (ACC) (8 pages).
JP Office Action for JP 2020-534355 (dated Dec. 6, 2022) (8 pages).
Ken Wada, Ring Buffer Basics (2013) 6 pages.
Notice of Allowance dated Feb. 23, 2023 for U.S. Appl. No. 18/060,556 (pp. 1-10).
Office Action (Final Rejection) dated Mar. 21, 2023 for U.S. Appl. No. 16/995,819 (pp. 1-7).
Office Action (Non-Final Rejection) dated Mar. 1, 2023 for U.S. Appl. No. 16/564,016 (pp. 1-10).
Office Action (Non-Final Rejection) dated Apr. 19, 2023 for U.S. Appl. No. 18/066,267 (pp. 1-11).
Office Action (Non-Final Rejection) dated Apr. 27, 2023 for U.S. Appl. No. 16/229,091 (pp. 1-5).
Office Action (Non-Final Rejection) dated May 8, 2023 for U.S. Appl. No. 18/065,603 (pp. 1-17).
Office Action (Non-Final Rejection) dated May 10, 2023 for U.S. Appl. No. 17/477,536 (pp. 1-13).
Office Action (Notice of Allowance and Fees Due (PTOL-85)) dated Mar. 8, 2023 for U.S. Appl. No. 17/721,315 (pp. 1-8).
Office Action (Notice of Allowance and Fees Due (PTOL-85)) dated Mar. 15, 2023 for U.S. Appl. No. 17/134,505 (pp. 1-5).
Office Action (Notice of Allowance and Fees Due (PTOL-85)) dated Mar. 24, 2023 for U.S. Appl. No. 17/080,840 (pp. 1-8).
Office Action (Notice of Allowance and Fees Due (PTOL-85)) dated Apr. 4, 2023 for U.S. Appl. No. 17/409,783 (pp. 1-5).
Office Action (Notice of Allowance and Fees Due (PTOL-85)) dated Apr. 6, 2023 for U.S. Appl. No. 17/807,730 (pp. 1-7).
Office Action (Notice of Allowance and Fees Due (PTOL-85)) dated Apr. 28, 2023 for U.S. Appl. No. 17/195,795 (pp. 1-7).
Office Action (Notice of Allowance and Fees Due (PTOL-85)) dated May 12, 2023 for U.S. Appl. No. 16/229,091 (pp. 1-8).
Office Action (Notice of Allowance and Fees Due (PTOL-85)) dated May 24, 2023 for U.S. Appl. No. 16/229,091 (pp. 1-2).
Office Action dated Feb. 9, 2023 for U.S. Appl. No. 18/060,556 (pp. 1-5).
Office Action dated Mar. 3, 2023 for U.S. Appl. No. 18/060,525 (pp. 1-12).
Office Action dated Apr. 19, 2023 for U.S. Appl. No. 18/066,267 (pp. 1-11).
Partial ISR for PCT/GB2023/050001 (dated Mar. 31, 2023) 13 pages.
Rakkolainen et al., A Survey of Mid-Air Ultrasound Haptics and Its Applications (IEEE Transactions on Haptics), vol. 14, No. 1, 2021, 18 pages.
Related Publications (1)
Number Date Country
20210397261 A1 Dec 2021 US
Provisional Applications (5)
Number Date Country
63210619 Jun 2021 US
63210486 Jun 2021 US
63067314 Aug 2020 US
63065997 Aug 2020 US
63043093 Jun 2020 US