The present invention relates to the field of the detection of objects based on sensor signals. More specifically, it relates to the detection of objects in synthetic-antenna signals, carried out by combining distance measurements at various points.
A sonar is a measuring device that is widely used in underwater navigation for detecting/locating objects in the water and measuring their distance. An active sonar operates as follows:
An object in a sonar image may be detected using the shape of the echo of the object in the image, but also using its shadow, that is to say the shape of the portion of the seabed that is not reached by the sound wave emitted by the sonar, because it is masked by the object.
In order to improve the resolution of a sonar, what is known as a synthetic-antenna sonar system may be used. Synthetic-antenna sonar aims to improve resolution at a given range without increasing the physical linear dimension of the reception antenna. The principle of synthetic-antenna sonar consists in using a composite physical antenna formed by a linear array of N transducers. In this type of sonar, when the carrier is moving forward, an emitter, or emission antenna, emits M successive pulses in an elementary sector that is fixed with respect to the carrier. The signals received by the N transducers of the physical reception antenna at M instants, and therefore at M successive locations, are used to form the beams of the synthetic antenna. The resolution of the images that are obtained, that is to say the resolution of the beams of the synthetic antennas (“array beam resolution”), is substantially equivalent to that of a virtual antenna the length of which corresponds to the length traveled by the physical antenna during these M successive instants.
Synthetic-antenna sonars are widely used because they make it possible to significantly improve sonar resolution without having to make any hardware changes.
However, the shadows of objects perceived by synthetic-antenna sonars are affected by a penumbra effect, also known as a parallax effect: since the angles of emission of the wave and of reception of the echoes are modified between each image capture, the directions of the shadows are modified between each image capture. When the synthetic image is generated, the shadows are thus blurred.
In some cases, for example when an object has its height stretched, or is floating between two bodies of water, the shadow may even become practically undetectable. This may be the case for example with a school of fish.
A similar problem may be encountered for other types of synthetic antennas, that is to say sensors operating on the principle of emitting and receiving waves at various points, and generating a synthetic image based on reflected waves received at the various points. This is the case for example for synthetic-aperture radars, or certain types of ultrasound.
There is therefore a need for improved detection, using synthetic antennas based on the emission of waves and the reception of waves reflected at various points, of objects the shadow cast by which is subject to penumbra effects.
To this end, one subject of the invention is a computer-implemented method comprising: receiving a series of distance measurements generated, from a plurality of respectively different positions, by a detection system that operates by: emitting a wave; receiving waves reflected by the environment; determining distances by computing differences between the time of emission of the wave and the times of reception of the reflected waves; generating, based on said series of distance measurements, a synthetic image representing the distances of the environment from a reference position; for each focusing distance of a plurality of focusing distances: generating, based on said series of distance measurements or said synthetic image, a synthetic image focused at said focusing distance by applying penumbra effect compensation; detecting the presence of an object in said focused synthetic image.
Advantageously, the detection system is a sonar system, and the generation of the synthetic image defines a synthetic-antenna sonar.
Advantageously, detecting the presence of an object in said focused synthetic image comprises applying a supervised machine learning engine trained with a learning base comprising focused images of shadows of objects of the same type as said object.
Advantageously, the method comprises: prior to the detection: computing, for each pixel of the focused synthetic image, a ratio between the intensities of the pixel in the synthetic image and the synthetic image; thresholding the pixels of the synthetic image for which this ratio is greater than a threshold; applying a mathematical morphology operation to the thresholded pixels; applying said detection to the output of said mathematical morphology operation.
Advantageously, the plurality of focusing distances comprises a plurality of initial focusing distances defined by a first distance pitch over a first range of focusing distances, the method comprising: a step of defining a plurality of refined focusing distances, which are defined by: a second range of focusing distances, narrower than the first, around a first focusing distance of said plurality of initial focusing distances, at which the presence of an object has been detected; a second distance pitch, less than the first; for each focusing distance from among said plurality of refined focusing distances, said generating, based on said series of distance measurements, a synthetic image focused at said focusing distance by applying penumbra effect compensation.
Advantageously, the step of generating, based on said synthetic image, a synthetic image focused at said focusing distance by applying penumbra effect compensation is carried out by applying a one-dimensional filter to the synthetic image.
Advantageously, the method as claimed in claim 1 comprises: generating a modified focused synthetic image by adding a shadow associated with a label to said synthetic image focused at said focusing distance; generating a modified synthetic image by applying an inverse filter of said one-dimensional filter to the modified focused synthetic image; enriching a learning base for detecting the presence of objects with the modified focused synthetic image, said focusing distance and said label.
Advantageously, the method comprises, in the event of detection of the presence of an object in said focused synthetic image, generating a composite image based on said synthetic image and said focused synthetic image.
Advantageously, said generating the composite image comprises: detecting shadows in the focused synthetic image; assigning the following for each pixel of the composite image: the intensity value of the corresponding pixel of the focused synthetic image for each pixel belonging to a shadow; the intensity value of the corresponding pixel of the synthetic image otherwise.
Advantageously, said generating the composite image comprises assigning, for each pixel of the composite image, an intensity value equal to the weighted sum of the intensity value of the corresponding pixel of the focused synthetic image and of the intensity value of the corresponding pixel of the synthetic image, where, for each pixel, the relative weight of the intensity value of the corresponding pixel of the focused synthetic image increases with an index of belonging to a shadow of the pixel.
Advantageously, the method comprises defining a region of interest of the synthetic image, wherein the steps of generating, based on said series of distance measurements, a synthetic image focused at said focusing distance by applying penumbra effect compensation and detecting the presence of an object in said focused synthetic image are carried out only in the region of interest.
Advantageously, the region of interest of the synthetic image comprises: displaying the synthetic image on a graphical interface; the user drawing a rectangle defining the region of interest.
Advantageously, the method comprises displaying the focused or composite image inside said rectangle.
Another subject of the invention is a computer program product comprising computer code instructions that, when the program is executed on a computer, cause said computer to execute the method according to one of the embodiments of the invention.
Another subject of the invention is a data processing system comprising a processor configured to implement the method according to one of the embodiments of the invention.
Another subject of the invention is a computer-readable recording medium comprising instructions that, when they are executed by a computer, cause said computer to implement the method according to one of the embodiments of the invention.
Other features, details and advantages of the invention will become apparent on reading the description given with reference to the appended drawings, which are given by way of example and in which, respectively:
The invention will be illustrated by examples relating to the detection of objects using a synthetic-antenna sonar. However, it is more generally applicable to any type of synthetic antenna based on the emission of waves at various points, the reception of reflected waves, and the generation of a synthetic image based on the reflected waves.
In this example, a sonar S, mounted on a vehicle, moves along an arbitrary but generally straight trajectory above the seabed. At regular intervals, S emits a pulse for imaging the background along a line of sight that generally aims to the side of the trajectory. {right arrow over (Xn)} denotes the position 110 of the sonar during the pulse, {right arrow over (rn)} denotes the unit vector 120 giving the pointing direction of the sonar, and {right arrow over (xn)} denotes the vector 130 tangent to the trajectory of the vehicle at the nth emitted pulse.
The pulse is described as the complex function p, a function either of the time of flight t of the pulse or of the sonar range r at the time of emission, the two variables being linked by the relationship r=c×t, where c is the speed of sound. The pulse is narrowband, modulating a carrier of wavelength λ.
The operating principle of the synthetic-antenna sonar will be explained taking the example of an arbitrary, fixed target A (that is to say a point for which it is sought to ascertain whether or not it is occupied) in the scene, the position of which is denoted {right arrow over (A)}.
The sonar S is formed of an emission antenna and a reception antenna. The emission antenna is configured to emit energy in an antenna lobe. For example, the lobes 141 and 142 respectively represent the antenna lobes of the emission antenna for pulses of index n−1 and n−2. The antenna lobe has an aperture β in the horizontal plane and an aperture with a bearing large enough to illuminate a large area on the bottom of the body of water. If this lobe is large enough, the object A may be visible on multiple consecutive pulses, with maximum energy when the point is exactly in the line of sight of the sonar.
Assuming that A is the only target, the raw acoustic signal received by the sonar at the nth pulse, for the range r, is modeled as being:
Where:
In reality, the scene is composed of a multitude of targets 160. The synthetic-antenna integration process is carried out as the pulses progress, as follows: at emitted pulse number n, generating a grid of reference points that are uniformly spaced by a pitch δr*{right arrow over (rn)} on the axis orthogonal to the trajectory and a pitch δr*{right arrow over (xn )} on the axis of the trajectory; the set of targets is therefore defined by:
In which smax is chosen with respect to the maximum range of the sonar.
With a fixed index k, the points {pn(s, k)} form what is referred to as the kth beam of the pulse n (or the kth beam of ping n) 160; the coordinates k and s are called beam index and sample index, respectively. The points {pn(s, k)} are notional reference targets.
A subset of beams Kn={kn,min . . . kn,max} is then selected, such that the points of these beams are those that have a greater gain at the pulse n (with respect to the position and the attitude of the sonar at this pulse) than the gain obtained for these same points in space at the position and the attitude of the sonar at any other pulse (in other words, these are the points that are able to be imaged best at the pulse n).
For each point (s, k), k∈Kn, I(n, s, k), the pulse interval, where Pn (s, k) is illuminated significantly by the sonar, is determined as:
The synthetic-antenna integration for the point {right arrow over (Pn)}(i, k) is then carried out by forming the coherent sum of the raw signals obtained for this point, for all of the pulses where this is able to be perceived by the sonar:
The datum SSAS (n, s, k) for the beams Kn={kn,min . . . Kn,max} is called SAS antenna for the pulse n, SAS being the acronym for synthetic-aperture sonar. Equation 4 is known as the generalized backpropagation equation.
Next, a function wSAS (i, b), called waterfall function, is defined such that:
In which the notation |Km| designates the cardinal of the set Km. The waterfall may be seen as a complex 2D image, with the axis of the beam indices b being called the azimuth axis and the axis of the indices i being the radial axis. It is also possible to parameterize the waterfall in terms of oblique distance r=s×δr and in terms of curvilinear abscissa x=b×δx.
Shadows are elements that, in an SAS image, contribute greatly to the understanding of the scene for operators or to automatic object detection. Indeed, by the very nature of the imaging process, it appears that echoes are relatively compacted on the oblique distance axis, thereby greatly hindering shape recognition for a human operator or a machine algorithm. On the contrary, when the shadows are extended, they make it possible to guess the shape of the object.
These two scenes respectively show an object 212a placed on the seabed, and an object 212b, of the same size, floating between two bodies of water and attached to the seabed by a cable 213b (rope) attached to a base 214b (sinker). The image is produced by a sonar located at the point 230.
In the case of the scene 210a, the sonar will perceive the echo 221a and the shadow 222a of the object 212a. In the case of the scene 210b, the sonar will perceive the echo 221b of the object 212b, of the cable 213b and of the base 214b, and also the shadow 222b of the object 212b and of the cable 213b.
The shadows are therefore defined as areas for which no echo is perceived, because they are masked by an object. It may be seen in
A penumbra effect, or parallax effect, is obtained when a target is illuminated only during some of the image captures of a synthetic-antenna sonar.
For example, in the case of
The target A is therefore illuminated by the sonar in iterations n−2 and n+1, but not in iteration n−1 or in iteration n.
This penumbra effect may remain acceptable for an object close to the seabed, such as the object 212a in
The method 400a is a computer-implemented method that aims to detect objects that are subject to penumbra effects in measurements from synthetic antennas.
The method 400a comprises a first step 410 of receiving a series of distance measurements generated, from a plurality of respectively different positions, by a synthetic-antenna detection system that operates by: emitting a wave; receiving waves reflected by the environment; determining distances by computing differences between the time of emission of the wave and the times of reception of the reflected waves.
The method 400a is thus applicable to any synthetic antenna based on the emission of waves at various points, and the reception of reflected waves. It is applicable for example to synthetic-antenna sonars, synthetic-antenna radars, or scanners.
In one set of embodiments of the invention, the detection system is a sonar system forming a synthetic-antenna sonar.
The method 400a then comprises a step 420 of generating, based on said series of distance measurements, a synthetic image representing the distances of the environment from a reference position.
This step consists in forming a synthetic image of distances for a given position. For example, in the example shown in
This synthetic image will simply be called “synthetic image”, as opposed to the “focused synthetic images” introduced in the remainder of this document. It may also be called “reference synthetic image”.
The method 400a then comprises, for each focusing distance of a plurality of focusing distances:
In other words, for each distance of a given set of focusing distances, the method 400a, in step 430, will generate a synthetic image focused at the desired distance, and then, in step 440, will detect the presence of an object in the focused synthetic image.
Step 430 makes it possible to obtain a synthetic image for which the shadows located at the distance under study are sharp. This therefore makes it possible, in step 440, to carry out object detection based on an image in which the shadows are sharp at a given distance, and therefore to improve object detection.
Steps 430 and 440 may be iterated for each distance of the plurality of distances. Sharp shadows may therefore be obtained for detection, for all desired distances.
The method 400a thus enables improved object detection in synthetic images from synthetic antennas that are subject to the penumbra effect.
For step 430, it is possible to use various penumbra effect compensation methods that have the effect of producing sharp shadows at a given distance.
For example, the penumbra effect compensation is carried out using a method known as “FFSE” (fixed focus shadow enhancement), described for example by Groen, J., Hansen, R. E., Callow, H. J., Sabel, J. C., & Sabo, T. O. (2008). Shadow enhancement in synthetic aperture sonar using fixed focusing. IEEE journal of oceanic engineering, 34(3), 269-284. That method is presented in that publication in the case of an underwater vehicle moving on a straight trajectory, with a sonar aiming at 90° from the trajectory. In this case, applying FFSE consists in replacing the synthetic-antenna sonar integration equation with the following equation, for a distance rt from an object:
In this case, the image associated with a point Pn (s, k) becomes blurry, but the transition between the image and the shadow is sharp. This operation is tantamount to focusing, at the distance rT, all of the points located further away in the range, hence the term fixed focus. The waterfall, or synthetic image associated with the FFSE image, has the same definition (that is to say number of pixels and resolution) as the waterfall, or synthetic image, of the synthetic-antenna sonar.
Another possible penumbra compensation method is what is known as the HVPC (high variant phase compensation) method. The HVPC algorithm may conceptually be seen as a refinement of FFSE, taking into account the height of the object that generated the shadow. This height may be determined in various ways, for example be determined using interferometry, or by a third-party sonar, such as a volume sonar. The height of the object that generated the shadow may also be obtained by trigonometry based on the length of the shadow projected onto the background by the object and knowing the altitude of the sonar (and possibly the local bathymetry, determined using interferometry or by a dedicated sonar).
Two images of a seabed scene depicting a shipwreck are shown in
On the left, the image 510 is obtained using a synthetic-antenna sonar. In this image, the wreck 511 and the shadow 512 generated by the wreck are blurred.
On the right, the image 520 is obtained using the same synthetic-antenna sonar, and has also benefited from the application of a penumbra compensation method, in this example FFSE, parameterized by the distance between the reference point of the image capture and the wreck. In this image, the wreck 521 and the shadow 522 generated by the wreck have become sharp.
This example demonstrates the ability of a penumbra effect compensation method to generate sharp shadows for a given obstacle distance in one set of embodiments of the invention. The shadow that is thus obtained, which is much sharper, may thus be used to carry out object detection more efficiently.
According to various embodiments of the invention, the plurality of focusing distances may be obtained in various ways. For example, a plurality of predefined distances may be used. The distances may be defined by applying a distance pitch over a given range. The distance pitch may be defined so as to ensure that a sufficiently sharp shadow will be obtained for all possible distances within the range.
The focusing distances may for example be defined as distances regularly spaced by a pitch 41 over a distance range [rmin; rmax], that is to say the focusing distances are defined by the set {rmin; rmin+Δ1; rmin+2*Δ1 . . . rmax}. The parameters rmin, rmax and Δ1 may be defined in various ways. For example, in one set of embodiments:
where:
The values of rmin and rmax may also be defined as a function of the minimum and maximum expected heights of the objects, and also their minimum and maximum expected distances. Indeed, the minimum and maximum heights and distances make it possible to identify the minimum and maximum distances of the shadows cast for the objects. The values of rmin and rmax may thus be defined as the minimum and maximum focusing distances at which a shadow is expected to be identified in the rest of the sonar range.
More generally, the values rmin and rmax may be defined so as to determine the minimum and maximum focusing distances at which it is relevant to search for a cast shadow, thus making it possible to limit the computation to distances at which a shadow could be identified. For example, rmax may be defined as the minimum out of the maximum range of the sonar and the maximum focusing distance at which a shadow is expected to be identified in the rest of the sonar range.
These values provide a good compromise between an objective of limiting the number of focusing distances to be tested (and therefore computational complexity) and a detection efficiency objective. Indeed, the distance rmin corresponds to a distance below which the parallax effect is not significant, and objects may be detected directly without processing; rmax is the maximum range of the sonar, and detecting shadows beyond this distance therefore does not make sense; Δ1=rmin/2 is a good compromise in terms of granularity for the focusing distances. These parameters therefore make it possible to test the entire range of relevant distances, while at the same time limiting the complexity of the method.
According to various embodiments of the invention, step 440 of detecting the presence of an object in the focused synthetic image may be carried out in various ways.
In general, any shadow shape detection method may be used. For example, it is possible to use a machine learning method, which may comprise for example the use of artificial neural networks and/or deep learning.
In one set of embodiments of the invention, detecting the presence of an object in said focused synthetic image comprises applying a supervised machine learning engine trained with a learning base comprising focused images of shadows of objects of the same type as said object.
The supervised machine learning engine may thus have been trained with a training base formed of focused shadows of the desired object. For example, if the purpose of the method is to detect the presence of schools of fish, a supervised machine learning engine may have been trained with a database of focused images of shadows of schools of fish.
Such a supervised machine learning engine has the advantage of being able to be trained to detect any type of object, and to provide very efficient detection.
In one set of embodiments of the invention, the method comprises:
If a pixel belongs to a shadow at the focusing distance, it will be darker in the focused image, and therefore its intensity value in the focused synthetic image will be lower than its intensity value in the synthetic image. Therefore, the ratio of the intensity value in the synthetic image divided by the intensity value in the focused synthetic image will be greater than 1. This pixel will therefore be chosen as a potential shadow pixel before the mathematical morphology is applied. In one set of embodiments of the invention, the chosen threshold is however greater than 1, in order to avoid generating false alarms.
In other words, for each of the focusing distances, step 440 may comprise, prior to the detection itself, pre-processing consisting, first of all, in thresholding the pixels corresponding to a shadow, and in then retaining only significant shadows. The detection itself, for example the application of a supervised machine learning engine, then applies only to these pixels belonging to significant shadows.
This makes it possible to make the detection more robust by limiting the detection to the most significant shadows at a given focusing distance.
Once the detection has been carried out, it may be carried out in various ways. By way of non-limiting example:
An alert may be raised;
More generally, automatic object detection may be used in a large number of fields, and any action able to be associated with automatic image detection may be implemented at the end of the detection.
In one set of embodiments of the invention, the focused synthetic images are generated, for each focusing distance, directly based on the series of distance measurements. For example, in the case of an application of FFSE to measurements from a sonar, this means that the FFSE technique is applied directly, for each focusing distance, to the sensor signals forming SAS beams (SAS beamforming).
Although this solution naturally provides a solution for obtaining focused images, it may prove costly in terms of computing time when focused synthetic images have to be obtained for a large number of focusing distances.
In order to limit computational complexity, in one set of embodiments of the invention, the step of generating, based on said synthetic image, a synthetic image focused at said focusing distance by applying penumbra effect compensation is carried out by applying one-dimensional filtering to the synthetic image.
The costly step of generating a synthetic image based on distance measurements is thus carried out just once, and all of the focused synthetic images are then obtained by 1-dimensional filtering, which is far less resource-intensive.
This method for computing focused synthetic images will now be described in one exemplary application to sonar images using FFSE.
In the case of a sonar orientation at 90° on a linear trajectory, given the SAS waterfall wSAS (b, s) obtained using the SAS process with sampling at a pitch δx on the azimuth axis (finer than the physical resolution of the synthetic antenna) and a pitch or on the distal axis, the FFSE image may be obtained, to within a constant multiplicative coefficient, based on said waterfall by 1D (1-dimensional) matched filtering (that is to say a 1D correlation), on each column of constant index s, using the following signal described in the spatial domain and centered around b=0:
Where:
The purpose of this apodization function is to limit the spatial support of the filter to the spatial domain where the target at the range s.δr is illuminated, this illumination taking place over a length of typical standard deviation s.ϵr·tan (β/2). This result may be applied generally to the case where depointing is no more than 90°, at the expense of making the equations slightly more complicated. In this case, it is indeed necessary to consider a depointing angle θs and to take this depointing into account for the computing of the convolution filter hr
FFSE may then be applied, to obtain a focused synthetic image, by way of 1D filtering of the waterfall, which may be implemented by way of 1D correlation for each column s:
In equivalent fashion, this operation may be implemented by a product in the spectral domain:
Where WsSAS (ζ, s), Hr
The refocusing of the SAS beams, that is to say the transformation of the synthetic image into a focused synthetic image, may thus be carried out by:
The refocusing may thus be applied to an already formed SAS image and be efficiently carried out by way of Fast Fourier Transforms (FFT). The focusing may therefore be carried out quickly and inexpensively over a large number of focusing distances.
It is also possible to carry out the inverse operation, that is to say to switch from a focused synthetic image (for example an FFSE focused image) to a synthetic image (for example a reference SAS image), by convoluting the focused synthetic image using the inverse filter hr
In one set of embodiments, this operation may be used to enrich a learning database for object detection.
To this end, in one set of embodiments of the invention, the method 400 comprises:
In other words, a sharp shadow corresponding to a known object is added to an image focused at the focusing distance, and then inverse 1D filtering is applied in order to retrieve a reference synthetic image (for example reference SAS image) comprising the unfocused shadow. This image is added to a training base for detecting the presence of objects, with an object label (defining the type of object) and the focusing distance. The combination of the modified synthetic image, the focusing distance and the label corresponding to the type of object thus makes it possible to train the object detection at various focusing distances (the focusing distances at which the object of said type should or should not be detected are then known).
This makes it possible to efficiently construct a training base, since this offers the possibility of generating a large number of images for the training base, with a large number of shadow images, for a large number of focusing distances. This therefore makes it possible to train the object detection efficiently and quickly, for example by training a supervised learning engine for object detection. This also limits the need to obtain real images, thereby greatly simplifying the construction of the training base.
The method 400b comprises all of the steps of the method 400a, and also three optional steps 450b, 460b and 470b. It should be noted that, although the three steps are shown in
In step 460b, the focusing distances are refined, that is to say focusing distances are tested more finely around a distance considered to be relevant.
In one set of embodiments of the invention, the plurality of focusing distances thus comprises a plurality of initial focusing distances defined by a first distance pitch over a first range of focusing distances, and the method comprising:
In other words, initial focusing distances may be the set of distances of a range [rmin; rmax] with a coarse pitch Δ1 (that is to say the initial distances are defined by the set {rmin; rmin+Δ1; rmin+2*Δ1 . . . rmax}; a first focusing distance rfest is selected from among the focusing distances, and then a range of refined distances is defined around the first focusing distance rfest at which the presence of an object has been detected, with a finer pitch Δ2<Δ1. For example, the refined focusing distances may be selected with the pitch Δ2 over the interval [rfest−Δ1; rfest+Δ1].
Next, the focusing step 430 is carried out for each refined focusing distance.
This makes it possible to test the focusing distances with a finer pitch around a focusing distance at which the presence of an object has been detected, and thus to enable focusing at a distance potentially closer to the distance of the object.
This makes it possible to obtain accurate results while at the same time limiting the computational load required by the method, since the focusing distances are tested with a coarse pitch over the entire possible range, but with a finer pitch for a range of focusing distances of interest.
The first focusing distance is thus a focusing distance for which the presence of an object has been detected.
In other words, as soon as an object is detected in step 440, the focusing distances are refined around the focusing distance that led to the detection of the object.
This makes it possible to refine the detection of objects around distances generating a detection, and therefore around the most relevant distances for detecting objects.
In this case, the refinement may be referred to as autofixed focus, because it consists in automatically determining the focusing distance that produces the sharpest shadows.
Various sharpness indices may be used. By way of non-limiting example, the metrics introduced by A. Buffington, F. S. Crawford, R. A. Muller, A. J. Schwemin and R. C. Smits, “Correction of atmospheric distortion with an image-sharpening telescope”, Jour. Acoust. Soc. Am., Vol. 67, no. 3, pp. 298-303, March 1977 may be used.
In one set of embodiments of the invention, the method 400b comprises, in the event of detection of the presence of an object in said focused synthetic image, a step 470b of generating a composite image based on said synthetic image and said focused synthetic image.
Indeed, as mentioned above, the focused synthetic image makes it possible to obtain sharp echoes and shadows for a given focusing distance, but while blurring the rest of the image. On the contrary, the unfocused synthetic image is sharper over the entire image, but exhibits the penumbra effect. The composite image produced based on the synthetic image and the focused synthetic image may therefore, at the same time, be sharp overall over the entire image and not exhibit a penumbra effect but, on the contrary, exhibit sharp shadows at the focusing distance.
Compositing step 470b therefore makes it possible to obtain an image that is as relevant as possible to be presented to an operator, in which the shadows are sharp at a given focusing distance, without sacrificing the sharpness of the rest of the image. This step may be carried out in various ways.
In one set of embodiments of the invention, step 470b of generating the composite image comprises:
In other words, shadows are detected in the focused synthetic image, and the pixels of the composite image come from:
This provides a simple and effective solution for obtaining a composite image with sharp shadows, without sacrificing the general quality of the image.
Other ways of generating the composite image may be envisaged.
For example, in one set of embodiments of the invention, said generating the composite image comprises assigning, for each pixel of the composite image, an intensity value equal to the weighted sum of the intensity value of the corresponding pixel of the focused synthetic image and of the intensity value of the corresponding pixel of the synthetic image, where, for each pixel, the relative weight of the intensity value of the corresponding pixel of the focused synthetic image increases with an index of belonging to a shadow of the pixel.
Thus, in this case, the pixel intensity values are not selected exclusively from one image or the other, but are defined as a weighted average of the intensity values in both images, where the relative weight of the intensity of the corresponding pixel of the focused synthetic image is greater when the pixel is considered to belong to a shadow.
This makes it possible to benefit from the sharpness of the shadow from the focused synthetic image, while at the same time benefiting from a more gradual transition between shadows and the rest of the image.
The relative weight of the pixels from the two images (synthetic image and focused synthetic image) may be computed in various ways. For example, a low-pass filter may be applied to the focused synthetic image. Indeed, the low-pass filter makes it possible to efficiently determine the pixels forming or not forming part of a shadow.
For example, if the synthetic image is denoted IA (b, s), the focused synthetic image is denoted IB (b, s), the composite image is denoted IC (b, s), and the focusing distance under consideration is denoted rT, then the composite image may be generated as follows:
For the other pixels (that is to say those such that s.δr≥rT), there is a process of:
for each pixel, where T is a shadow threshold in dB and t is a filter adjustment parameter (log-sigmoidal weighting);
Each pixel of the composite image is thus a weighted average of the corresponding pixels of the synthetic image and the focused synthetic image, giving more weight to the focused synthetic image in shadow areas, and more weight to the synthetic image in other areas, while at the same time ensuring a smooth transition between these two types of areas.
In one set of embodiments of the invention, the method 400b comprises defining 450b a region of interest of the synthetic image, and steps 430 of generating the focused synthetic image and 440 of detecting an object are carried out only in the region of interest.
This makes it possible to limit the complexity of the method, since steps 430 and 440 are carried out only on a region of interest considered to be relevant.
In one set of embodiments of the invention, the region of interest comprises the entire image based on the focusing distance. In other embodiments of the invention, the region of interest comprises only a sub-part of the image, resulting from a division of the total image into blocks of beams or defined by a user, for example.
In one set of embodiments of the invention, defining the region of interest of the synthetic image comprises:
In a first state 610, the graphical interface represents a synthetic image, or waterfall, of a sonar. In this representation, the range is defined by the horizontal axis: the pixels furthest to the right correspond to the points furthest from the sensor; the beams are defined by the vertical axis: the lowest points correspond to the beams acquired at the oldest dates and the highest points correspond to the beams acquired at the most recent dates.
The user is able, in the interface, to trigger the definition of the region of interest, for example by pressing a button 630, and trigger the compositing, for example by pressing the button 631. However, these buttons are provided only by way of example, and other means may be used to trigger the definition of the region of interest and/or the compositing, such as keyboard shortcuts or voice commands, for example.
In the example of
In this example, the user may thus manually define the region of interest directly on the synthetic image. The user may modify the region of interest by modifying the rectangle, for example by modifying the rectangle using a moving cursor (which may for example be manipulated using an input interface such as a mouse or touch sensor): for example, the user may drag a corner, drag an edge, or drag-and-drop the entire rectangle.
In one embodiment of the invention, the method 400b comprises displaying the focused synthetic image or the composite image inside said rectangle.
In the example of
This allows the user to directly visualize the effect of the focusing on the image.
The composite image may be generated and displayed when the user presses the button 631, or makes another compositing command (keyboard shortcut, voice command, etc.).
As an alternative, the compositing and the display may be carried out in real time, when the rectangle 621 is displayed, and as soon as the user modifies the rectangle 621. This allows the user to visualize the result of the focusing and the compositing in real time.
In the example of
The user is thus able, in real time, to modify the shadowing parameters and visualize the obtained result.
The above examples demonstrate the ability of the invention to improve the detection of objects in images from a synthetic antenna. However, they are given only by way of example and in no way limit the scope of the invention as defined in the claims below.
Number | Date | Country | Kind |
---|---|---|---|
FR2112194 | Nov 2021 | FR | national |
This application is a National Stage of International patent application PCT/EP2022/081984, filed on Nov. 15, 2022, which claims priority to foreign French patent application No. FR 2112194, filed on Nov. 18, 2021, the disclosures of which are incorporated by reference in their entirety.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/EP2022/081984 | 11/15/2022 | WO |