The present invention relates to object detection, and specifically to visualizing movement of a detected object in a scene.
In prior art camera systems for object detection, a map view in a horizontal plane of a scene is sometimes derived and displayed in a display e.g. for viewing by an operator. In systems where the locations of detected objects in a horizontal plane of the scene can be determined, e.g. by means of radar, lidar etc, the location of a detected object may be identified on the map. When the detected object moves, the location of the detected object may then be updated on the map to visualize the movement of the detected object. However, many camera systems do not include any means for determining the location of the detected objects in the horizontal plane of the scene. In such systems an exact location of the detected objects cannot be identified on the map and hence, movement of a detected object cannot be easily visualized without addition of further means for identifying locations of the detected objects in the horizontal plane of the scene.
An object of the present invention is to provide a method and a system for visualizing movement of a detected object in a scene which overcome or mitigate issues in prior art.
The present disclosure makes use of a recognition that visualization in a 2-dimensional horizontal plane of a detected object in a field of view of a camera is advantageous to a user, such as an operator, from a perceptional point of view as compared to a sequence of images from a camera, and that information indicating a location in horizontal direction in an image of the detected object in each image of the sequence of images from the camera is sufficient to visualize movement in the 2-dimensional horizontal plane even without access to any information regarding the distance from the camera to the detected object.
According to a first aspect, a method for visualizing movement of a detected object in a scene is provided. For each image of a sequence of images captured by a camera and depicting the scene information is obtained indicating a location in horizontal direction in the image of the detected object, and a graphical shape is displayed in a display in a 2-dimensional graphical representation of a field of view of the camera in a horizontal plane, wherein the graphical shape has a location in the 2-dimensional graphical representation, which location represents the location in horizontal direction in the image of the detected object.
The method of the first aspect is advantageous in a system without means for determining locations of detected objects in a horizontal plane of a scene. The act of obtaining information indicating a location in horizontal direction in the image of the detected object may comprise determining a location in horizontal direction in the image of the detected object by means of analysis of the image.
By knowing a location in horizontal direction in the image, locations in the 2-dimensional graphical representation corresponding to the known location in horizontal direction in the image can be derived. Hence, the graphical shape can be displayed on a display at a location in the 2-dimensional graphical representation that represents the location in horizontal direction of the detected object in the image. By this, a movement of the detected object in horizontal direction between two images in the sequence of images will be visualized by the location of the graphical shape in the 2-dimensional graphical representation changing between the two images.
The information obtained may further indicate an extension in horizontal direction in the image of the detected object. The graphical shape displayed may then be a circle sector having an angle representing the extension in horizontal direction in the image of the detected object and a location representing the location in horizontal direction in the image of the detected object.
By this, a movement of the detected object in horizontal direction over the sequence of images will be visualized by the location of the circle sector in the 2-dimensional graphical representation changing to represent the new location in horizontal direction in the image of the detected object. Furthermore, a movement of the object away from the camera over the sequence of images will be visualized by the angle of the circle sector changing to represent the new decreased extension in horizontal direction of the detected object due to the movement away from the camera.
The circle sector may be such that a first quotient between the angle of the circle sector and a total angle of the field of view of the camera in the 2-dimensional graphical representation is equal to, or at least proportional to, a second quotient between the extension in horizontal direction in the image of the detected object and a horizontal length of the image.
By this, a movement of the detected object away from the camera over the sequence of images will be visualized by a decrease of the angle of the circle sector, which decrease is proportional to the decreased extension in horizontal direction of the detected object due to the movement away from the camera.
The circle sector may be such that a third quotient between a first angle from a closest edge in angular direction of the field of view of the camera in the 2-dimensional graphical representation to the centre of the circle sector and a total angle of the field of view of the camera in the 2-dimensional graphical representation is equal to a fourth quotient between a distance from a closest edge in horizontal direction of the image to the centre of the detected object in the image and a horizontal length of the image.
In alternative, the graphical shape may be a line having a location in the 2-dimensional graphical representation representing the location in horizontal direction in the image of the detected object.
By this, a movement of the detected object in horizontal direction over the sequence of images will be visualized by the location of the line in the 2-dimensional graphical representation changing to represent the new location in horizontal direction in the image of the detected object.
The 2-dimensional graphical representation may be included in a 2-dimensional map of a site in which the camera is mounted.
The act of displaying may be performed without access to information related to a distance between the camera and the detected object in the scene.
In embodiments, the graphical shape has an appearance such that it is distinguishable in the 2-dimensional graphical representation.
In embodiments, the method of the first aspect further comprises for each further image of a further sequence of images captured by a further camera and depicting the scene from a different angle than the camera obtaining information indicating a location in horizontal direction in the further image of the detected object and an extension in horizontal direction in the further image of the detected object, and displaying on the display, in the 2-dimensional graphical representation of the field of view of the camera in the horizontal plane, a further circle sector. The further circle sector has an angle representing the extension in horizontal direction in the image of the detected object and a location representing the location in horizontal direction in the image of the detected object. The further circle sector is further displayed concurrently with a circle sector relating to a corresponding image of the sequence of images, and any overlap between the further circle sector and the circle sector has an appearance such that it is distinguishable in the 2-dimensional graphical representation.
By this, the possible location of the detected object in the 2-dimensional graphical representation can be limited to the overlap between the further circle sector and the circle sector. Furthermore, a movement of the detected object in horizontal direction between two images in the sequence of images and/or in horizontal direction between two corresponding further images in the further sequence of images will be visualized by the location of the overlap in the 2-dimensional graphical representation changing between the two images and the corresponding to further images due to the location of the circle sector changing to represent the new location in horizontal direction of the detected object between the two images and/or the location of the further circle sector changing to represent the new location in horizontal direction of the detected object between the two further images.
According to a second aspect, a non-transitory computer-readable storage medium is provided having stored thereon instructions for implementing the method according to the first aspect when executed by a system having processing capabilities and a display.
The above-mentioned optional features of the method according to the first aspect, when applicable, apply to the non-transitory computer-readable storage medium of the second aspect as well. To avoid undue repetition, reference is made to the above.
According to a third aspect a system for visualizing movement of a detected object in a scene is provided. The system comprises circuitry configured to execute an obtaining function and a displaying function for each image of a sequence of images captured by a camera and depicting the scene. The obtaining function is configured to obtain information indicating a location in horizontal direction in the image of the detected object. The displaying function is configured to display on the display, in a 2-dimensional graphical representation of the field of view of the camera in a horizontal plane, a graphical shape having a location representing the location in horizontal direction in the image of the detected object.
The above-mentioned optional features of the method according to the first aspect, when applicable, apply to the system of the third aspect as well. To avoid undue repetition, reference is made to the above.
A further scope of applicability of the present invention will become apparent from the detailed description given below. However, it should be understood that the detailed description and specific examples, while indicating preferred embodiments of the invention, are given by way of illustration only, since various changes and modifications within the scope of the invention will become apparent to those skilled in the art from this detailed description.
Hence, it is to be understood that this invention is not limited to the particular component parts of the system described or acts of the methods described as such system and method may vary. It is also to be understood that the terminology used herein is for purpose of describing particular embodiments only and is not intended to be limiting. It must be noted that, as used in the specification and the appended claim, the articles “a,” “an,” “the,” and “said” are intended to mean that there are one or more of the elements unless the context clearly dictates otherwise. Furthermore, the words “comprising”, “including”, “containing” and similar wordings does not exclude other elements or steps.
The above and other aspects of the present invention will now be described in more detail, with reference to appended figures. The figures should not be considered limiting but are instead used for explaining and understanding.
The present invention will now be described hereinafter with reference to the accompanying drawings, in which currently preferred embodiments of the invention are illustrated. This invention may, however, be embodied in many different forms and should not be construed as limited to the embodiments set forth herein.
In the following, embodiments of a method 100 for visualizing movement of a detected object in a scene will be discussed with reference to
The detected object may have been detected by means of any type of method for object detection, such as using a machine learning model e.g. using a neural network.
For each image of a sequence of images captured by a camera and depicting the scene, information is obtained S110 indicating a location in horizontal direction in the image of the detected object. Furthermore, a graphical shape is displayed S120 on a display in a 2-dimensional graphical representation of a field of view of the camera in a horizontal plane. The graphical shape has a location in the 2-dimensional graphical representation, which location represents the location in horizontal direction in the image of the detected object. As the detected object has an extension in horizontal direction in the image, the location in horizontal direction in the image of the detected object may for example be determined as the objects centre in horizontal direction. However, as it sometimes is of less interest to visualize the exact location of the detected object at a specific time and of more interest to visualize movement, the location in horizontal direction in the image of the detected object may be selected as any portion of the object. Preferably, but not necessarily, the same portion is selected in each image of the sequence of images, such as a rightmost end or leftmost end of the detected object.
The method 100 may for example be performed in real time or near real time, i.e. the information is received in real time when the image of the sequence of images is captured by the camera with only a delay due to determining of the information and the displaying of the graphical shape in the 2-dimensional graphical representation of the field of view of the camera in a horizontal plane is performed as soon as possible thereafter. By this, the movement of the detected object over the sequence of images is visualized in real time or near real time by the movement of the graphical shape in the 2-dimensional graphical representation over the sequence of images.
The camera may be a fixed camera having a fixed field of view or it may be a pant tilt zoom (PTZ) camera. In the latter case the 2-dimensional graphical representation of a field of view of the camera in a horizontal plane may differ between images of the sequence of images if the camera pans, tilts, and/or zooms between the images.
By displaying the graphical shape at a location in the 2-dimensional graphical representation representing the location in horizontal direction in the image of the detected object on a display in a 2-dimensional graphical representation of a field of view of the camera in a horizontal plane in each image of the sequence of images, the graphical shape will move in the 2-dimensional graphical representation if the detected object moves in horizontal direction between images in the sequence of images.
Using visualization in a 2-dimensional horizontal plane of a detected object in a field of view of a camera is advantageous to a user from a perceptional point of view as compared to viewing only the sequence of images from the camera. For example, it is easier for a user to identify how a detected object, e.g. a person, is moving in relation to portions of interest in a site where the camera is located when viewing the graphical shape moving in the 2-dimensional graphical representation as compared when viewing only the movement of the detected object in the sequence of images from the camera.
Using visualization in a 2-dimensional horizontal plane of a detected object in a field of view of a camera is further advantageous to a user in scenarios where the user is not allowed to view the sequence of images from the camera but is allowed to view the 2-dimensional representation of the field of view in the horizontal plane.
Based on visualization of the movement, a user may perform one or more of a multitude of actions or events such as:
The 2-dimensional graphical representation of the field of view of the camera in a horizontal plane does not have to be an exact representation. An approximation is sufficient in order to enable visualization of the movement of the detected object in horizontal direction in the image also as movement of the graphical shape in the 2-dimensional graphical representation of the field of view.
The method 100 is for example advantageous in scenarios where there are no means for determining locations of detected objects in a horizontal plane of a scene. Such scenarios are for example scenarios where there is only the sequence of images available from the camera, and hence, there are no information available regarding the distance from the camera to the detected objects. According to the method 100, information indicating a location in horizontal direction in an image of the detected object in each image of a sequence of images from the camera can be used to visualize movement even without any information regarding the distance from the camera to the detected object.
The act of obtaining S110 information indicating a location in horizontal direction in the image of the detected object may comprise determining a location in horizontal direction in the image of the detected object by means of analysis of the image.
Furthermore, the graphical shape may be visually adapted based on a speed with which the detected object moves in horizontal direction between images in the sequence of images, such that the speed of the detected object in horizontal direction between images in the sequence of images can be visually distinguished from the graphical shape. For example, the graphical shape may have different colours or other visual features depending on the speed of the detected object in horizontal direction between images in the sequence of images. The speed can be calculated by determining the difference between locations in horizontal direction in two images divided by the difference in time between the capturing of the two images.
Turning to
Furthermore,
The 2-dimensional graphical representation of the portion of the site may further include graphical representations of other features in the image 200, of the camera, and other features in a site where the camera is located. For example,
The 2-dimensional graphical representation of the field of view of the camera in a horizontal plane can be included in a 2-dimensional graphical representation of the portion of a site either manually by a user or automatically or as a combination of both. In case of manual inclusion by a user, tools may be provided such that the user clicks on a point in the 2-dimensional graphical representation of the portion of a site where the camera is located, and a default circle sector is the included which the user may then adapt the circle sector, e.g. by adapting its direction and (central) angle. It is further to be noted that the 2-dimensional graphical representation of the field of view of the camera does not have to be a circle sector but may have other shapes representing the field of view, such as a triangular shape. Sensors like a compass and/or gyroscope may give automatic readings of camera direction, if the compass direction is predefined I the 2-dimensional graphical representation of the portion of the site, and installed camera view angle that may be used during an automatic creation and placement of the 2-dimensional graphical representation of the field of view of the camera. The camera may well also know its location, which provides the possibility for a fully automatic generation and placement of the 2-dimensional graphical representation of the field of view. The 2-dimensional graphical representation of the field of view of the camera may further be modified by the user or automatically based on physical limitations in the scene, such as walls, floor and ceiling. Information regarding camera type, lens parameters, height and view angle at which the camera is mounted may be used to further adapt the 2-dimensional graphical representation of the field of view of the camera.
Turning back to
Turning to
The line relating to a detected object should be specifically distinguished from the 2-dimensional graphical representation of the field of view of the camera in a horizontal plane. For example, it may be indicated by means of a specific colour or any other way of visually identifying or distinguishing it from the 2-dimensional graphical representation of the field of view of the camera in a horizontal plane. Furthermore, the line may be solid, dashed, dotted, have different widths, end with an arrow etc.
Furthermore, if the information obtained in the act of obtaining S110 further indicates an extension in horizontal direction in the image of the detected object, a visual indication may be added to the line representing the extension in horizontal direction in the image of the detected object, such as an added line, arrow or other graphical shape perpendicular to the line.
Turning back to
The information obtained in the act of obtaining S110 may include a bounding box in relation to the detected object. In such a case the extension in horizontal direction in the image of the detected object may be determined as the extension in horizontal direction (width) of the bounding box.
Turning to
The angle of a circle sector in relation to a detected object may be determined such that a first quotient between the angle of the circle sector and a total angle of the field of view of the camera in the 2-dimensional graphical representation is equal to a second quotient between the extension in horizontal direction in the image of the detected object and a horizontal length of the image. As can be seen in
The circle sector relating to a detected object should be specifically distinguished from the 2-dimensional graphical representation of the field of view of the camera in a horizontal plane. For example, it may be indicated by means of a specific colour such as a grey tone or it may be indicated by lines along its periphery or any other way of visually identifying or distinguishing it from the 2-dimensional graphical representation of the field of view of the camera in a horizontal plane.
A circle sector relating to a detected object may be located in the 2-dimensional graphical representation of the field of view of the camera in a horizontal plane such that a third quotient between a first angle from the closest edge in angular direction of the field of view of the camera in the 2-dimensional graphical representation to the centre of the circle sector and the total angle of the field of view of the camera in the 2-dimensional graphical representation is equal to a fourth quotient between the distance from a closest edge in horizontal direction of the image to the centre of the detected object in the image and the horizontal length of the image.
Turning back to
By a corresponding image of the sequence of images is meant an image of the sequence of images that is captured at the same time as the further image of the further sequence of images to which the further circle sector relates.
In the first embodiment, the second embodiment, and the third embodiment, the information obtained in the act of obtaining S110 may further indicate an object class of the detected object. Furthermore, in the act of displaying S120, the graphical shape may then further include a representation of the object class, such as an icon illustrating the object class. Additionally or alternatively, the object class may be represented by a different colour of the graphical shape depending on object class, by a different shape of the graphical shape depending on object class, by a different type of edges of the graphical shape depending on object class, etc.
Turning to
The locations and angles of the circle sectors 253, 254, 255 relating to the camera and the circle sectors 263, 264 relating to the further camera according to the third embodiment may be determined in the same way as the locations and angles the circle sectors 243, 244, 245 relating to the camera according to the second embodiment.
As can be seen in
As the third detected object 205 is located such that it is obscured by the second stationary object 209 illustrated by the third graphical representation 229, this object is not detected in the further image captured by the further camera. Hence, from the combined information relating to the camera and the further camera, it is not possible to determine in which overlapping portion between the circle sectors 253, 254, 255 relating to the camera and the circle sectors 263, 264 relating to the further camera the detected objects 203, 204, 205 are located. For example, since the third detected object 205 is only detected in relation to the camera and not the further camera, it is not possible to say whether the third detected object 205 is obscured in or outside the field of view of the further camera, or if it is located at a position such that it is in front of and hence obscuring one of the other detected objects 203, 204 in relation to the further camera.
On condition that information regarding object class is available for the detected objects 203, 204, 205, such information in relation to the camera and the further camera, the overlap in which a detected object is located may be determined in some instances. For example, if the first detected object 203 in relation to the camera is of a known object class different from the object classes of the other detected objects, 204, 205, it can be determined that the first detected object is located in the overlap between the leftmost circle sector 253 in relation to the camera and the leftmost circle sector 263 in relation to the further camera.
Furthermore, if the second detected object 204 and the third detected object 205 are of the same known object class, it is not possible to say whether the third detected object 205 is obscured in or outside of the field of view of the further camera, or if it is located at a position such that it is in front of and hence obscuring the second detected object 204 in relation to the further camera. On the other hand, if the second detected object 204 and the third detected object 205 are of different but known object classes, it is possible to say whether the third detected object 205 is obscured in or outside the field of view of the further camera, or if it is located at a position such that it is in front of and hence obscuring the second detected object 204 in relation to the further camera. Specifically, if the object class of the detected object in relation to the rightmost circle sector 264 of the further camera is the same as the object class of the detected object in relation to the middle circle sector 226 of the camera, the third detected object 205 is obscured in or outside the field of view of the further camera. On the other hand, if the object class of the detected object in relation to the rightmost circle sector 264 of the further camera is not the same as the object class of the detected object in relation to the middle circle sector 226 of the camera, the third detected object 205 is located at a position such that it is in front of and hence obscuring the second detected object 204 in relation to the further camera.
In the following, embodiments of a system 300 for visualizing movement of a detected object in a scene will be discussed with reference to
The system 300 comprises circuitry 310 and a display 320. The circuitry 310 is configured to carry out functions of the system 300. The circuitry 310 may include a processor 312, such as a central processing unit (CPU), microcontroller, or microprocessor. The processor 312 is configured to execute program code. The program code may for example be configured to carry out the functions of the system 300.
The system 300 may further comprise a memory 330. The memory 330 may be one or more of a buffer, a flash memory, a hard drive, a removable media, a volatile memory, a non-volatile memory, a random access memory (RAM), or another suitable system. In a typical arrangement, the memory 330 may include a non-volatile memory for long term data storage and a volatile memory that functions as system memory for the circuitry 310. The memory 330 may exchange data with the circuitry 310 over a data bus. Accompanying control lines and an address bus between the memory 330 and the circuitry 310 also may be present.
Functions of the system 300 may be embodied in the form of executable logic routines (e.g., lines of code, software programs, etc.) that are stored on a non-transitory computer readable medium (e.g., the memory 330) of the system 300 and are executed by the circuitry 310 (e.g., using the processor 312). Furthermore, the functions of the system 300 may be a stand-alone software application or form a part of a software application that carries out additional tasks related to the system 300. The described functions may be considered a method that a processing unit, e.g. the processor 312 of the circuitry 310 is configured to carry out. Also, while the described functions may be implemented in software, such functionality may as well be carried out via dedicated hardware or firmware, or some combination of hardware, firmware and/or software.
The circuitry 310 is configured to, for each image of a sequence of images captured by a camera and depicting the scene, execute an obtaining function 331, and a displaying function 332.
The obtaining function 331 is configured to obtain information indicating a location in horizontal direction in the image of the detected object.
The displaying function 332 is configured to display on the display, in a 2-dimensional graphical representation of a field of view of the camera in a horizontal plane, a graphical shape having a location in the 2-dimensional graphical representation representing the location in horizontal direction in the image of the detected object.
The information obtained in the obtaining function 331 may further indicate an extension in horizontal direction in the image of the detected object. In the displaying function 332, the graphical shape may then be a circle sector having an angle representing the extension in horizontal direction in the image of the detected object and a location in the 2-dimensional graphical representation representing the location in horizontal direction in the image of the detected object. The circle sector may then be such that a first quotient between the angle of the circle sector and a total angle of the field of view of the camera in the 2-dimensional graphical representation is equal to a second quotient between the extension in horizontal direction in the image of the detected object and a horizontal length of the image.
The circle sector may further be such that a third quotient between a first angle from a closest edge in angular direction of the field of view of the camera in the 2-dimensional graphical representation to the centre of the circle sector and a total angle of the field of view of the camera in the 2-dimensional graphical representation is equal to a fourth quotient between a distance from a closest edge in horizontal direction of the image to the centre of the detected object in the image constitutes of a horizontal length of the image.
In the displaying function 331, the graphical shape may in alternative be a line having a location in the 2-dimensional graphical representation representing the location in horizontal direction in the image of the detected object.
The 2-dimensional graphical representation may further be included in a 2-dimensional map of a site in which the camera is mounted.
The displaying function may be performed without access to information related to a distance between the camera and the detected object in the scene.
The graphical shape may have an appearance such that it is distinguishable in the 2-dimensional graphical representation.
The circuitry 310 may be further configured to, for each further image of a further sequence of images captured by a further camera and depicting the scene from a different angle than the camera, execute a further obtaining function 333, and a further displaying function 334.
The further obtaining function 333 is configured to obtain information indicating a location and extension in horizontal direction in the further image of the detected object.
The further displaying function 334 is configured to display on the display, in the 2-dimensional graphical representation of the field of view of the camera in the horizontal plane, a further circle sector having an angle representing the extension in horizontal direction in the further image of the detected object and a location in the 2-dimensional graphical representation representing the location in horizontal direction in the further image of the detected object, wherein the further circle sector is displayed concurrently with a circle sector relating to a corresponding image of the sequence of images, wherein any overlap between the further circle sector and the circle sector has an appearance such that it is distinguishable in the 2-dimensional graphical representation.
Further details regarding the functions the circuitry 310 is configured to execute are provided in relation to the corresponding acts of the method 100 for visualizing movement of a detected object in a scene as described hereinabove in relation to
A person skilled in the art realizes that the present invention is not limited to the embodiments described above. On the contrary, many modifications and variations are possible within the scope of the appended claims. Such modifications and variations can be understood and effected by a skilled person in practicing the claimed invention, from a study of the drawings, the disclosure, and the appended claims.
Number | Date | Country | Kind |
---|---|---|---|
23156497.2 | Feb 2023 | EP | regional |