The invention generally relates to video processing, and particularly to a method and a device for enhancing radar video.
In many technological fields, there is a need for detecting specific objects (also referred to as “targets”) and generating a video rendering of the detected targets and of the environment for “situational awareness”.
In particular, radar systems are used to detect targets by transmitting a Radio Frequency (RF) signal with specific characteristics and receiving the emitted signal perturbed by the environment after reflection from the target in this environment. The transmitted RF signal can be bundled in a specific direction, either by using a particular antenna design or by signal processing (i.e., digital beamforming). A similar approach can be used for the received signal. Bundling the signals in specific directions results in a higher gain, which allows the detection of objects at larger ranges.
Signal and data processing, consisting of a plot extraction process (position, Doppler speed, Radar Cross Section estimation of a target) and a tracking process, is generally performed after reception of the radar signals to detect targets while ensuring robustness against interference and clutter signals. Every target is classified based on its properties such as Doppler speed, range, height and Radar Cross Section (RCS). As a result, a large and non-exclusive set of objects can be obtained in the environment. Radar operators generally need an overview of the clutter in the environment for situational awareness. However, data processing does not allow to efficiently distinguish low RCS and slow moving targets from stationary clutter. To be able to analyze the environment and all objects in it, the radar operators thus need to display the plots and tracks over a display video image which refreshes over time.
The display video represents the received signal strength at all ranges and all bearings to render a 2D projected image on the horizontal plane. An extraction process (also called “plot creation process”) and a tracking process (also called “track creation process”) are used to find and track moving objects from which the amplitude of the received signal is above certain thresholds. In most radar systems, objects that are substantially stationary are quickly deemed clutter and are therefore not tracked by the system. However, slow moving objects may be of particular interest for radar operators. Indeed, a surface object, such as for example a swimmer in the sea, might be missed as a wanted target, due to its characteristics not being tracked by the radar system. In this case, the radar operator must visually distinguish its radar video blips, the received signal strength at a certain range, and the bearing where a target is present from the clutter environment, which in the example of a swimmer can be the sea and waves.
From scan to scan, a radar signal from a specific object may be subject to large fluctuations due to a variety of causes from the environment, to the radar system and to the object itself. From scan to scan, a radar display video image thus fluctuates heavily, delivering a quite busy and flickering image to the radar operator. Discerning a small target in a highly fluctuating and noisy environment becomes therefore difficult.
This limitation is especially significant for surveillance radar systems which have scan times of one second or a few seconds. Accordingly, the refresh rate of the image hinders the operators' ability to distinguish small and slow moving targets from clutter and noise. This may be still emphasized by the eye strain which is due to the low image refresh rate.
A known approach to the problem of enhancing the display video refresh rate consists in coupling the processing results (i.e. plots and tracks) to the display video. By identifying which part of the display video image represents a target, the display video underlying the track update can be predicted and moved smoothly to the coordinates of the next predicted track update. The solution consisting of coupling track information to the display video presents a number of drawbacks so that very few radar systems resort to this approach and even keep the display video unprocessed and displayed at the low refresh rate.
A major limitation of this approach lies on the fact that the tracking is primarily designed for non-stationary objects (such as for example jets, airliner, missiles, shells) and is not optimized for low RCS and slow moving objects. As a result, these specific objects are not recognized and tracked.
Additionally, implementing a feedback loop from the tracking to the display video can only be done based on the premises that the display video can be predicted accurately from the tracking. However, in case of strong maneuvering targets or high noise conditions, the predictions are not precise enough, thereby jeopardizing the cross-checking capacity.
Another limitation of this approach is due to the fact that a feedback loop implies constraints on implementation and timing. Even in situations where the tracking process is capable of tracking slow moving and low RCS objects, a fundamental problem thus exists. More specifically, since radar signals are highly fluctuating and very disturbed by noise, the resulting video signal from a low RCS object may change a lot from scan to scan. Therefore, just smoothly moving the video blip from scan “n” to scan “n+1” would lead to a significant change from the final prediction of scan “n+1” to the actual video measurement of scan “n+1”. This generally causes eye strain and cannot result in a smooth image over multiple scans due to the jump at scan changes. Such solution thus only helps the operator within a scan time interval and not over multiple scans.
In order to address these and other problems, there is provided a method for enhancing video display as defined in the appended independent claim 1, and a device for enhancing video display as defined in appended claim 18. Preferred embodiments are defined in the dependent claims.
The invention accordingly provides a time-extrapolated easeful video enhancing device capable of creating a smooth and low eye-strain display video stream from radar signal data to help the radar operator assess the environment and more easily discern low RCS and slow moving objects which may be of interest.
Further advantages of the present invention will become clear to the skilled person upon examination of the drawings and detailed description. It is intended that any additional advantages be incorporated herein.
Embodiments of the present invention will now be described by way of example with reference to the accompanying drawings in which like references denote similar elements, and in which:
Additionally, the detailed description is supplemented with Exhibit A. This Exhibit is placed apart for the purpose of clarifying the detailed description, and of enabling easier reference. It nevertheless forms an integral part of the description of the present invention. This applies to the drawings as well.
Referring to
The video enhancing device 100 is configured to receive two images In and In−1 corresponding to two successive scans captured by a video surveillance system 1. More specifically, the video enhancing device 100 is further configured to output the next scan In+1 from the two successive scans corresponding to the previous scan In−1 and the current scan In.
Even if the invention is not limited to such applications, the invention has particular advantages for the surveillance and detection of slow moving objects, such as for example a swimmer in the sea. In a preferred embodiment of the invention, the video surveillance system 1 may be a radar surveillance system. The following description will be made with reference to such a radar application. However, the skilled person will readily understand that the invention may be used in other applications such as for example in video surveillance systems, time lapse movies or video games systems
The radar surveillance system 1 may include radar detection units 11 providing raw data from a surveillance area (e.g. a sea area) to a signal processing and imaging unit 12. The radar detection units 11 may comprise radar transmitters to send RF pulses in the surveillance area and radar receivers to receive the response provided by the reflected pulses (scan data). The radar transmitters and receivers interact with a scanning radar antenna respectively for transmitting the pulses and receiving the reflected pulses. The scanning radar antenna is placed at a location at which the surveillance area is to be monitored. The scan data are fed to the signal processing and rendering unit 12, in the form of an analog signal. The signal processing and rendering unit 12 then processes the signal received from the radar detection units 11 to generate a digitized representation of a target detected in the scanned area on a radar display 14.
The radar display 14 may be any suitable display such as, for example, a Liquid Crystal Display (LCD) or a Cathode Ray Tube (CRT) display.
The signal processing and rendering unit 12 includes a radar processor 120 for performing digital video processing. The radar processor 120 may be in communication with a graphics card 122, which may include its own processing element such as a graphics processing unit (GPU) 1220 for presenting selected image data processed at the graphics card 122 to be rendered at the display 14. The graphics card 122 may be configured to map a texture to a plurality of vertex buffers in order to generate a display of the scan data on the radar display 14.
The radar processor 120 may include an analog filter 1200 for filtering analogue radar signal feed received from the antenna and an analogue-to-digital converter 1202 for conversion of the filtered signal into a digital data stream. The converted digital data stream may then be stored in buffers 1203. Data from the buffers 1203 may then be transmitted to a conversion unit 1204 (scan converter) configured to convert the digital polar coordinate radar data, defined by range and bearing, into rectangular coordinate pixel radar image data (Cartesian coordinates), defined by x=range*cosinus(bearing), and y=range*sinus(bearing), for use by the graphics card 122. The graphics card 122 may then render the rectangular coordinate pixel radar image data in the display 14 based on a plurality of textures calculated by the GPU 1220, depending on the display mode.
Additionally, to allow control of radar image display by an operator, a graphical user interface (GUI) may be provided.
The video enhancing device 100 according to certain embodiments of the invention may process two images, such as Cartesian Images, corresponding to successive scans provided by the conversion unit 1204. By processing pairs of successive images In and In−1 (real images) and the last provided predicted image Jn (previous prediction), the video enhancing device 100 can create a smooth and low eye-strain display video stream from the radar signal data, which helps the radar operator to better assess the environment, the objects and targets in it.
To facilitate a better understanding of the detailed description, there follow definitions of certain notations corresponding to parameters used in relation with the described embodiments of the present invention:
Theoretically, Ln+1(1)=Jn+1(1). However in practice this may not be the case. In addition, Ln+1(f)≠Jn+1(f) for all intermediate images where f≠1, since the starting images are both different.
The video enhancing device 100 is configured to calculate on per-pixel basis the optical flow in an image. As used herein, the optical flow designates the displacement in 2D for all pixels from image to image. According to the invention, the video input may be enhanced without additional information about image content (such as tracks or plots).
The video enhancing device 100 may use an algorithm based on the Lucas-Kanade image morphing function which is used twice.
The Lucas Kanade based algorithm may be the standard Lucas-Kanade or alternatively a variation of the Lucas-Kanade (LK) image morphing algorithm fitted to radar images. The variation of the LK algorithm will be referred to as “Range Dependent Accuracy Reduction” or “RADAR-LK” in the following description. The RADAR-LK algorithm is based on the standard LK algorithm but comprises additional steps to reduce the number of pixels for which the LK algorithm is calculated.
In the following description, the notation LK( ) will be used to designate both the standard LK algorithm and the RADAR_LK algorithm. The difference between both approaches will be detailed in the description. In the following description, both algorithms will be referred to commonly as “Lucas Kanade based algorithm” or “LK based algorithm”. The standard LK algorithm is described for example in the article by by Simon Baker and lain Matthews, “Lucas-Kanade 20 Years On: A Unifying Framework”, International Journal of Computer Vision, February 2004, Volume 56, Issue 3, pp 221-255.
The video enhancing device 100 may comprise a displacement vector calculation unit 11 which on the LK based algorithm.
The displacement vector calculation unit 11 is called twice for each optical flow determination. In the first call, the displacement vector calculation unit calculates a linear backwards displacement vector Dn per pixel from image In, corresponding to radar scan “n”, to image In−1, corresponding to radar scan “n−1”.
More specifically, the vector calculation unit 11 uses the LK based algorithm to determine the linear backwards displacement vector Dn per pixel from image In to image In−1 according to the following equation:
D
n=LK(In,In−1) (equation 1),
where LK( ) designates the LK based algorithm (standard LK algorithm or Radar-LK algorithm).
The video enhancing device 100 further comprises a prediction vector calculation unit 12 to determine the prediction vector Pn+1 per pixel for scan n+1 based on the displacement vector Dn determined from image In to image In−1. The prediction vector calculation unit 12 extrapolates the prediction vector Pn+1 per pixel via a negation of the displacement vector Dn, according to equation 2:
P
n+1
=D
n (equation 2)
The video enhancing device 100 further comprises a displaced image calculation unit 13 which is called three times during the optical flow determination process. In the first call, the displaced image calculation unit 13 is called for predicting what the coming image Jn+1 should be from the prediction vector Pn+1. The displacement image generation unit 13 then uses a backwards linear extrapolation from In according to equation 3:
J
n+1=INTERP2(In,Pn+1) (equation 3)
In equation 3 INTERP2 designates the function corresponding to the standard 2D bilinear interpolation algorithm in which all pixels in In are displaced by the 2D values in Pn+1.
In theory, the predicted image Jn+1 is equal to the actual next scan measured image In+1. However, in most cases, the predicted image Jn+1 will not be equal to the actual image In+1 due to heavy measurement fluctuations inherent to radar signals and non-constant movement of objects in the environment.
The new measurement In+1 can be used similarly to determine the next predicted image Jn+2:
the new measurement In+1 is provided as an input together with the previous measurement In to the vector displacement calculation unit 11 to determine Dn+1,
the prediction vector calculation unit 12 then determines Pn+2 from Dn+1
the displaced image generation unit 13 estimates the full predicted image Jn+1(1) from In+1 and Pn+2.
The use of only the actual measured images In and In+1 to determine Jn+2 serves as a sort of a negative feedback loop that keeps the processing stable and not diverging from the measurements.
In order to guarantee smooth transitions from the predicted image Jn+1 to the predicted image Jn+2, the displacement vector calculation unit 11 is called again to determine a second set of displacement vectors (referred to as flow fading vectors). The second set of displacement vectors Fn+1 is determined using the LK based algorithm between Jn+1 and Jn+2, according to equation 4:
F
n+1=LK(Jn+1,Jn+2) (equation 4)
The video enhancing device 100 is capable of generating k intermediate images between the last updated predicted image Jn and Jn+1. However, depending on the starting image (either Jn or In), the intermediates are different and need to be blended to guarantee a smooth video output over successive scans.
The video enhancing device 100 therefore also includes a linear mixing unit 15 for perform linear mixing of intermediate images Jn+1(f) (originating from In) and Ln+1(f) (originating from Jn). The linear mixing unit 15 linearly mixes the images to render the combined intermediate outputted image Mn+1(f).
This linear mixing of intermediate predicted 2D images is performed in such a way that the image starts at Jn according to equations 5 of Exhibit A. The linear mixing unit 15 calls twice the displaced image generation unit 13 in addition to the first call to calculate Jn+1 as described above. In the second call, the displaced image generation unit 13 is called for calculating the displaced intermediate image originating from the previous predicted image Jn:
L
n+1
(f)=INTERP2(Jn,f·Fn+1)
The resulting image Ln+1(f) is multiplied by f.
In the third call, the displaced image generation unit 13 is called for calculating the displaced image originating from the previous measured image In:
J
n+1
(f)=INTERP2(In,f·Pn+1)
The resulting image is multiplied by (1−f). The images thus obtained are the added according to equations 5 from Exhibit A and the result is displayed on a suitable display 14, such as a radar system screen in radar applications of the invention.
Accordingly, the predicted image from the previous prediction is weighed heavily at the first few intermediates, but gradually the prediction from the measurements is taking over.
The generation of intermediate images and the mixing scheme used to blend I and J intermediate images allow for a smooth transition from the previous prediction to the next prediction.
This guarantees a smooth display video image from scan to scan.
As illustrated in
It should be noted that
In the example of
The predicted optical flow vector P_4 between scan 3 and 4 leads to the predicted image J_4 of scan 4 from measured image I_3 of scan 3.
The flow fading vector F_4 between scan 3 and 4 leads to the predicted image L_4 from previous predicted image J_3 of scan 3.
At the full prediction (f=1) the two predicted images L_4 and J_4 are theoretically equal.
The flow fading vector F_5 of scan 5 leads to the predicted image L_5 from previous predicted image J_4 of scan 4.
The predicted optical flow P_3 between scan 2 and 3 is determined from the backwards displacement D_2 between scan 2 and 1 from measured image I_2 to measured image I_1.
The predicted optical flow P_4 between scan 3 and 4 is determined from the backwards displacement D_3 between scan 3 and 2 from measured image I_3 to measured image I_2.
The predicted optical flow P_5 between scan 4 and 5 is determined from the backwards displacement D_4 between scan 4 and 3 from measured image I_4 to measured image I_3.
The flow fading vector F_4 between scan 3 and 4 is calculated from the predicted image J_4 of scan 4 and the predicted image J_3 of scan 3.
The flow fading vector F_5 between scan 4 and 5 is calculated from the predicted image J_5 of scan 5 and the predicted image J_4 of scan 4.
Referring to
In step 401, a burst video from the radar video surveillance system is received. The buffer memories 1203 are previously initialized to zeros at step 400.
In step 402, the burst video is added in a Current Scan Memory referred to as “CSM”.
In step 403, it is determined if the current scan memory contains a full image. If so, step 404 is performed. Otherwise, the video enhancing device 100 waits until a full image In is received.
In step 404, displacement vectors Dn (pixel displacements) are calculated from image In−1 stored in a Previous Scan Memory (referred to as “PSM”) and current image In stored in Current Scan Memory.
In step 405, the displacement vectors Dn are multiplied with coefficient “−1”. This turns the displacement vectors into prediction vectors Pn+1. The prediction vectors Pn+1 are then stored in a Prediction Vectors Memory PVM.
In step 406, a displaced image Jn+1 is generated based on the starting image In from current scan memory CSM and the set of displacement vectors Pn+1. The displaced image Jn+1 is then stored in a current prediction memory CPM.
In step 407, flow fading vectors Fn+1 are computed to determine the pixel displacements from the previous predicted image Jn stored in a Previous Prediction Memory PPM to a newly predicted image Jn stored in the CPM. The resulting displacement vectors are stored in a flow fading memory FFM.
For each index k of the intermediate image, k being initialized to 0 in step 408, and a corresponding number Nk of intermediate images, step 409 is performed to calculate the ratio f=k/Nk.
The displacement vectors Pn+1 from Prediction Vectors Memory PVM are then multiplied by f in step 410 and a displaced image is generated from the resulting vectors in step 412 (INTERP2(In,f·Pn1). The intensity of the displaced image thus obtained is multiplied by f in step 413.
Further, the displacement vectors Fn+1 from flow fading memory FFM are multiplied by f in step 414 and a displaced image (INTERP2(Jn,f·Fn+1) is generated from the resulting vectors in step 415. The intensity of the displaced image thus obtained is multiplied by (1-f) in step 416.
The images obtained in step 413 and 416 are then added in step 418 and displayed on the radar display 14 in step 419.
If parameter f is not equal to 1 (step 420), k is incremented in step 422 and steps 410 to 418 are iterated.
Otherwise, if f is equal to 1, the Previous Prediction Memory (PPM) is overwritten with Current Prediction Memory (CPM) content, in step 421. The enhancement video method then returns to step 401.
Accordingly, the video enhancing method according to the embodiments of the invention is based on two calls of a displacement vectors calculation function, in step 404 and 407 (the function is implemented by the displacement vectors calculation unit 11 in
The flow fading features involve a gradual mixing of the flow from the previous prediction to the new prediction via the flow fading vector F, and from the current image to the new prediction via the prediction vector P.
The LK based algorithm according to the described embodiments of the invention can thus be used to morph any two images in a radar video surveillance system, the RADAR-LK algorithm may be specifically used. The LK based algorithm as used by the video enhancing method (steps 404 and 407) further allows finding a per-pixel displacement such that two images can be morphed from one to the other. More specifically, it provides per pixel a displacement vector such that when all pixels are moved, the resulting image fits the second image. However, since the LK based algorithm is provided to determine sub-pixel displacements, a pyramidal and layered approach may be additionally applied according to the invention. More specifically, according to one aspect of the invention, the highest-resolution image is blurred a few times such that the real-world displacements the LK algorithm can handle fit within the pixel resolution. According to the embodiments of the invention, the core LK algorithm operates iteratively within a pyramid level to determine the displacement vectors with increasingly high accuracy. The maximum number of iterations is tunable and dependent on the processing available and the accuracy needed.
In certain embodiment of the invention, the LK-based algorithm may be the standard LK algorithm. In particular, three approaches of the LK standard algorithm can be applied by the displacement vector calculation function:
a single-level and single-iteration LK approach which refers to the way the pixel displacement vectors operate for a single iteration and on a single pyramidal level;
a single-level and iterative LK approach which refers to the way the accuracy increases by iteratively refining the found displacement vectors in the single-iteration LK scheme;
a multi-level and iterative LK approach which refers to the way the found displacement transfer to higher resolution levels and the way the highest-resolution optical flow is found.
The single level and single iteration LK approach allows calculation of displacement vectors. It relies on a single level and single iteration LK method that independently operates on every pixel. More specifically, it determines the sub-pixel displacement based on the specific pixel value and its surrounding pixel values. All pixels in the surrounding of the pixel of interest are said to be within the correlation window of the LK algorithm. This correlation window size is tunable and depends on the resolution and the object sizes, which depends on the specific radar. The correlation window may be a square window, or any other type of suitable windows.
The single level and single iteration LK method uses derivatives between pixels in the surrounding correlation window, in horizontal and vertical directions, between an image I and an image Ĩ, where I designates the starting image and Ĩ designates the final image.
In an initial step, the derivative of I may be calculated in both Cartesian directions, according to equations 8 and 9 in which x and y represent pixel coordinates and the subscripts indicate the direction of the derivative. With the above derivative images, a spatial derivative matrix G is constructed per pixel position, according to equation 10. In equation 10, (px, py) designates the position of any specific pixel, and w designates half of the window size in pixels (for example, a 5×5 window would mean w=2).
Then, the intensity/value difference between image I and image I is calculated according to equation 11.
Another vector called “image mismatch vector” is constructed according to equation 12.
At this point, the optical flow vector
The optical flow vector or displacement vector is also denoted Dn(x, y) or Fn(x,y) in the present description, depending on the call of the function (step 404 or 407 of
To make the estimation function more accurate, an iterative approach may be taken instead of the single-layer, single iteration LK method.
Since the estimation is quite coarse, via multiple iterations, a better estimate can be obtained using a single level, iterative LK method.
Iteratively, the refinement lies in equation 11 where, at every iteration, the new difference image is taking into account the found displacements
In complement, the new optical flow vectors may be added to the already existing ones from the previous iteration, thus gradually converging to a minimized error.
A process parameter may be used to define when to stop the iterative process. The process parameter may be defined depending on the noisiness of the images and the needed accuracy. Comparing the difference image between two highly fluctuating images does not need a high accuracy but merely a roughly estimation of the optical flow vectors, and the iteration process may be limited or even omitted. This then ends the single layer, iterative LK method.
The standard LK algorithm can only be used to find subpixel displacements for a single layer. A problem occurs when there is a need to find larger displacements, such as for a radar system having a very low data refresh rate (at least compared to camera-type systems). In such situation, it is proposed to apply the pyramidal or multi-level breakdown of the images from a high-resolution to a lower resolution, according to the third approach. A pyramidal implementation of Lucas Kanade algorithm is described for example in J.-Y. Bouguet, “Pyramidal implementation of the lucas kanade feature tracker”, Intel Corporation, Microprocessor Research Labs, 2000.
In the multi-level and iterative LK approach, the pyramidal breakdown of the images is applied from a high-resolution to a lower resolution. The pyramidal breakdown is created via image convolution with a blurring matrix K and a down-sampling operation. These operations result in a blurred image with pixel values consisting of a combination of intensities from its neighbors. The pixel resolution is reduced by a factor 2, going through the layers in the pyramid.
To go from an image IL at level L to an image IL-1 at level L−1, the multi-level and iterative LK method initially performs the convolution of image IL with the blurring matrix K. K is defined according to equation 15.
With the defined blurring matrix K, the image at level L can be constructed according to equation 16, where IBlurL(x,y) is obtained from the convolution of IL and M. At this stage, IBlurL(x,y) designates a blurred image with the same resolution as image IL. To downsample and create the lower level image IL-1, every second pixel per row and column is taken. This reduces the resolution by a factor of 2 and terminates the process. The lower level image IL-1 is obtained according to equation 17, where NxL (or respectively NyL) designates the number of pixels in level L in the x (or respectively y) direction.
When stepping down as described above, a pyramid consisting of reduced resolution images can be constructed. The LK method thus obtained can work on every level, starting with the lowest-resolution image. The number of levels L required in the pyramid depends on the starting resolution in combination with the number of pixels objects that are expected to move in the radar video.
The initial optical flow vectors to start the estimation with are then equal to twice the found optical flow vectors at a lower level according to equation 18.
The initial values are used in equation 14 to start the iterative LK for a specific pyramidal level.
The displacement vector function may use the optimized RADAR-LK approach (Range Dependent Accuracy Reduction-LK), as the LK-based algorithm, to overcome the problems related to the accuracy with which the optical flow vectors are calculated for all pixels. For a radar system, the measured polar resolution depends on the size of the object, and the design of the system. However, the Cartesian resolution (i.e., in x and y) does depend on the range and the bearing of the object. This implies that the accuracy with which the optical flow vectors are to be calculated at shorter ranges for a radar system must be higher than at large ranges. This is because the same object, being seen by a radar system at short range, will be displayed on a Cartesian grid by a far smaller form (“blob”) than at long ranges. The accuracy of the measurement thus decreases with range, and so is the needed accuracy for the optical flow vectors.
The inventor thus modified the pyramidal approach to take into account the radar-specific video features. As the desired optical flow vector accuracy reduces with range and it is not required to process all pixels similarly, processor load can be reduced dramatically, when only refining the pixels in subsequently smaller boxes around the radar position.
The number of pixels for which the pyramidal LK algorithm needs to be calculated is thus significantly reduced. For example, for a radar display video grid, the cell sizes are defined to be square. In the lowest resolution (i.e., the lowest pyramidal level), the number of pixels is N·N=N2. For the pyramidal LK approach, when moving to a higher resolution, the number of pixels grows by a factor of 4. This leads to a total number of pixels NtotalNormal to evaluate according to equation 20. For a 4 level system as shown in
The total number of pixels number grows rapidly with number of pyramidal levels and the number of pixels in the lowest resolution level. However, with the RADAR-LK approach according to the invention, the total number of pixel NtotalRADAR-LK as given by equation 22 is improved. According to the preceding example, this leads to a total number of pixels NtotalRADAR-LK to process equal to 4N2, according to equation 23.
It should be noted that this number does not increase with the number of levels squared anymore but just linearly increases, which can be managed more easily. Especially, since the RADAR-LK process is performed twice (step 404 and 407 of
The optimized LK algorithm (RADAR-LK) thus improves the standard LK algorithm by reducing processing and data loads needed for practical reasons and/or based on data density inherent to radar systems.
According to one aspect of the invention, a displaced image generation function (implemented by the displaced image generation unit 13 of
When a prediction vector is available for all pixels in an image, the pixels are moved to their predicted locations as shown in part A of
The video enhancing device according to the embodiments of the invention obviates the need for a feedback loop between the tracking and the display video, thereby removing the constraint of a perfectly functioning tracking and the need for specific hardware and/or software to implement the feedback loop. Further, the video enhancing device 100 according to the invention does not require track information for performing the display video prediction from scan n to scan n+1. The proposed video enhancing device can work stand-alone with only the display video images of two subsequent scans as an input, which therefore greatly reduces complexity.
In addition, the video enhancing device according to the invention provides a smooth display video image over multiple scans while still using the highly fluctuating measured-real-display video images as an input. It eliminates fluctuations from scan to scan, thus removing radar operator eye strain, increasing situational awareness, and optimizing the ability to distinguish low RCS and slow moving targets.
The proposed video enhancing device can work stand-alone with only the display video images of two subsequent scans as an input.
Even if the invention has particular advantages for a radar system application, it is not limited to such application. In particular, the invention may be applied to video surveillance systems, time lapse movies or video games. As the need for any prior knowledge on the content of any image/environment is not required, the invention can be used on any stream or set of images. The invention can be also implemented on any image rendering system for which the time interval between subsequent images is too large to create a smooth video stream.
The invention can take the form of an embodiment containing both hardware and software elements implementing the invention.
The invention can also take the form of a computer program product accessible from a computer-usable or computer-readable medium providing program code for use by or in connection with a computer or any instruction execution system. For the purposes of this description, a computer-usable or computer readable medium can be any apparatus that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device.
The foregoing is considered as illustrative only of the principles of the invention. Further, since numerous modifications and changes will readily occur to those skilled in the art, it is not desired to limit the invention to the exact construction and operation shown and described, and accordingly, all suitable modifications and equivalents may be resorted to, falling within the scope of the invention. For instance, the video enhancing device 100 may alternatively use a non-linear image mixing scheme, or set the range borders of the RADAR-LK method at different distances. Further, it should be noted that although the displacement vector calculation unit 11 may be based on the RADAR-LK algorithm to reduce computational load, when processing power is relatively high (to be able to process high resolution pixels at all levels), in situations where the processing power does not need to be conserved, the displacement vector calculation unit 11 can be based on the standard Lucas Kanade algorithm alternatively.
Equation 1
D
n=LK(In,In−1) (eq. 1)
Equation 2
P
n+1
=−D
n (eq. 2)
Equation 3
J
n+1
=I
n
+P
n+1 (eq. 3)
Equation 4
F
n+1=LK(Jn+1,Jn+2) (eq. 4)
Equation 6
Dn(x,y)=LK(In(x,y),In−1(x,y)) (Eq. 6)
Equation 7
F
n+1(x,y)=LK(Jn(x,y),Jn+1(x,y)) (Eq 7.1)
F
n+2(x,y)=LK(Jn+1(x,y),Jn+2(x,y)) (Eq 7.2)
Equation 11
ΔI(x,y)=I(x,y)−{tilde over (I)}(x,y) (eq. 11)
Equation 13
Equation 14
ΔI(x,y)=I(x,y)−{tilde over (I)}(x+
Equation 16
I
blur
L(x,y)=ILM (eq. 16)
Equation 17
I
L-1(x,y)=IblurL([2:2:NxL],[2:2:NyL]) (eq. 17)
Equation 18
init
L=2·
Equation 19
L=1=2L-1·
Equation 20
N
total
normal=ΣL=1N
Equation 21
N
total
normal=ΣL=14(2L-1·N)2=N2((20)2+(21)2+(22)2+(23)2)=85N2 (eq. 21)
Equation 22
N
total
RADAR-LK=ΣL=1N
Equation 23
N
total
RADAR-LK=ΣL=14N2=4·N2 (eq. 23)
Number | Date | Country | Kind |
---|---|---|---|
13197557.5 | Dec 2013 | EP | regional |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/EP2014/077645 | 12/12/2014 | WO | 00 |