The present invention relates to a system for analyzing a video in which a moving object was photographed.
A video monitoring system that realizes security and safety by finding a suspicious person and a suspicious object in real time from a video of a monitoring camera has been widely used. In addition, a video monitoring system that monitors not only a single place, but also plural places and entire facilities comprehensively has appeared.
In such a situation, in order to search for and present a scene in which a specific person or object appears from an enormous amount of videos, various video search techniques have been proposed. Further, it has been required not only to search for a person or object separately, but also to monitor a relation between a person and an object in a video such as a person and baggage possessed by the person. Accordingly, it is possible to realize finding of a person who possesses specific baggage and determination of delivery and leaving of baggage.
Japanese Unexamined Patent Application Publication No. 2011-227647 is a relevant technique related to this technique. Japanese Unexamined Patent Application Publication No. 2011-227647 discloses a technique in which plural persons are detected and tracked in a camera video in a predetermined area, a relative position (distance) between a first person and a second person is calculated, and a case in which a distance between the persons is closer for a prescribed time or over is determined as a suspicious person.
In Japanese Unexamined Patent Application Publication No. 2011-227647, relationship between persons is determined using a single camera. Thus, in the case where relationship between a person and baggage photographed by a monitoring camera is determined, the following problems exist.
In the case where a relation between a person and baggage possessed by the person is grasped using a single camera, the baggage is hidden (occlusion) in some cases. Thus, there is a case that the baggage does not necessarily appear in all the frames in which the person appears. The reason is that the baggage cannot be detected because the baggage is hidden behind the person, or the baggage is hidden behind another person or baggage because the position of the baggage is low. Therefore, it is difficult to determine a relation between a person and baggage on the basis of inter-frame tracking such as determination of a relation between persons.
Further, in order to further strengthen ownership between a person and baggage, only a single camera is insufficient. For example, only a camera with an angle of view that photographs a person from the front cannot detect baggage carried on person's back such as a backpack. Thus, the possession of baggage carried on person's back cannot be determined by only the corresponding camera. In such a case, it is necessary to analyze a camera video with various angles of view placed at plural locations to comprehensively determine the ownership.
In addition, it is required to detect a suspicious person on the basis of a change in the ownership of baggage such as delivery or leaving of the baggage. However, in the determination using a single camera video, it is difficult to make a decision in the case where a video of an act itself is not photographed by a camera. In such a case, it is necessary to determine the possibility that the suspicious behavior occurs by using videos between plural cameras to clarify a change in the ownership of baggage.
Accordingly, an object of the present invention is to enable to comprehensively determine ownership between a person and baggage between plural cameras and to enable to grasp a change in the ownership between a person and baggage between plural cameras.
As an example of the present invention, in a system that analyzes videos photographed by plural cameras, a detection/tracking process is performed for first and second objects using videos of plural cameras, and a relationship degree between the first and second objects is determined on the basis of the types of the first and second objects and a distance between the objects to be stored in a database.
According to the present invention, it is possible to provide a video analysis system and a video analysis method in which ownership between a person and baggage can be comprehensively determined.
Hereinafter, embodiments of the present invention will be described with reference to the drawings.
An entire configuration diagram of a video monitoring system in the embodiment is shown in
The video analysis/database server 1000 is a normal PC server, and is configured using a CPU 1010, a memory 1011, a network interface 1012, and a storage 1013. Each function to be described below is expanded on the memory 1011, and is implemented as a program 1200 operated on the CPU 1010. A work area 1220 used by the program is further secured on the memory.
The storage 1013 has a database 1300 and a frame image storage area 1380 for a video of the camera.
In the video monitoring system, a video photographed by each of the video photographing devices 1101 to 1110 is analyzed by the video analysis/database server 1000, the result is stored in the database 1300, and an analysis/search result is output to the terminal PC 1400 to be provided to a monitoring person (operator).
The database 1300 includes a person table 1310 in which information of a detected person is input, a baggage table 1330 in which information of detected baggage is input, and a person/baggage relationship table 1350 in which information indicating a relationship (ownership and the like) between the person and the baggage is input.
As a function of the video analysis/database server 1000, a time-series image acquisition functional unit 1201 acquires images of the latest video frames of time-series images from each of the video photographing devices 1101 to 1110, and the images are stored in the frame image storage area 1380 of the storage 1013. Further, the real-time video analysis functional unit 1210 conducts a real-time video analysis for each video frame. Namely, the real-time video analysis functional unit 1210 allows a relationship determination functional unit 1213 to determine the relationship between a person and baggage detected after processes by a person detection tracking functional unit 1211 and a baggage detection/tracking functional unit 1212 are performed. Here, the person and the baggage may be detected by one multi-class detector. Each of the detection/tracking result of the person and the baggage and the determined relationship is stored into the person table 1310, the baggage table 1330, and the person/baggage relationship table 1350 in the database 1300.
A user interface functional unit 1240 in the program 1200 of the video analysis/database server 1000 displays a baggage possession search screen and the like on a display 1410 using a browser 1401 of the terminal PC 1400. The operator operates the screen using a mouse 1430 and a keyboard 1420 to perform a baggage possession search and the like. The operation by the operator (for example, a baggage possession search operation) is transmitted to a search functional unit 1230 through the browser 1401 and the user interface functional unit 1240 of the video analysis/database server 1000, data on the database 1300 is searched, and the search result is displayed on the display 1410. The video of the search result to be displayed is read from the frame image storage area 1380.
The embodiment is characterized in that the relationship determination functional unit 1213 of the video analysis/database server 1000 determines a relationship between the person and the baggage detected and tracked in real time to be stored in the person/baggage relationship table 1350 in the database 1300, and the operator searches the database 1300 in accordance with an operation such as a baggage possession search input from the terminal PC 1400 to determine baggage possession on the basis of the time-series person/baggage relationship obtained from plural video photographing devices.
A detailed configuration of the person table 1310 in the database 1300 is shown in
In
(1) Camera ID 1311: the ID of a camera that detected the person
(2) Date and time 1312: the date and time when the person was detected (although hours, minutes, and seconds are shown in the drawing for the sake of simplicity, UNIX time and the like are recorded in reality)
(3) Frame ID 1313: the ID of a video frame in which the person was detected
(4) Person rectangle ID 1314: the ID of a person rectangle (a rectangle surrounding the detected person) in a video frame. The ID is a unique value in a frame.
(5) Rectangle information 1315: a coordinate value (two-dimensional coordinates in an image) of a person rectangle in a video frame. For example, the information is represented by coordinates of the central point of the rectangle and an array holding the longitudinal/lateral lengths of the rectangle.
(6) In-camera person track ID 1316: the ID of the person given as a result of person tracking in a camera. The ID is a unique value for each camera.
(7) Global person ID 1317: the ID of the person given as a result of person tracking between cameras. The ID is a unique value in the entire system.
(8) Person feature 1318: a vector value storing the feature of the appearance of the person. Persons who are similar in appearance have close features. The feature is used for in-camera person tracking, an inter-camera person search, and the like.
(9) Person attribute 1319: a vector value representing the attributes (the gender, age, cloth color, and the like) of the person. The attribute is used for an attribute search.
A detailed configuration of the baggage table 1330 in the database 1300 is shown in
In
(1) Camera ID 1331: the ID of a camera that detected the baggage
(2) Date and time 1332: the date and time when the baggage was detected
(3) Frame ID 1333: the ID of a video frame in which the baggage was detected
(4) Baggage rectangle ID 1334: the ID of a baggage rectangle in a video frame
(5) Rectangle information 1315: a coordinate value of a baggage rectangle in a video frame
(6) In-camera baggage track ID 1336: the ID of the baggage given as a result of baggage tracking in a camera. The ID is a unique value for each camera. In the case where the baggage tracking interrupted due to an occlusion, plural IDs are assigned to the same baggage in some cases. These plural tracks are grouped by the following track cluster ID 1337.
(7) In-camera baggage track cluster ID 1337: the ID given by grouping baggage tracking results (plural in-camera baggage track IDs) that were interrupted due to an occlusion on the same baggage basis. The same value is assigned to baggage clusters determined as the same. The ID is a unique value for each camera.
(8) Global baggage ID 1338: the ID of the baggage given as a result of baggage tracking between cameras
(9) Baggage feature 1339: a vector value storing the feature of the appearance of the baggage
(10) Baggage attribute 1340: a vector value representing the attributes (the type, color, and the like) of the baggage
A configuration of the person/baggage relationship table 1350 in the database 1300 is shown in
In
(1) Camera ID 1351: the camera ID that determined the relationship between the person and the baggage
(2) In-camera person track ID 1352: the track ID of the person determined to have relationship
(3) In-camera baggage track cluster ID 1353: the track cluster ID (a set of baggage tracks) of the baggage determined to have relationship
(4) Baggage possession determination indicator value 1354: an indicator value indicating the possibility that the baggage belongs to the person (hereinafter, described as “indicator value”)
(5) Representative frame number 1355: the frame number of a video having the highest possibility that the baggage belongs to the person. The number is used when a search result is displayed on a GUI.
Here, a basic policy of a baggage possession determination of the embodiment will be described using
A processing flow of an in-camera baggage possession determination performed by the real-time video analysis functional unit 1210 is shown in
In the detection process, rectangles of the person and the baggage are extracted from a still image. In the case where a common object detector is used, the process is divided in accordance with the type (person/baggage) of the detected object. The tracking process is a known technique, and from objects (person/baggage) detected in continuous frame images, the same objects are linked on the basis of the similarity of the appearance and the position in the video. For example, in the case of persons, each person detected in frame images processed this time is linked to one detected in previous frames in consideration of the similarity of the appearance and the position. The in-camera person track ID 1316 is newly given to a person rectangle that has newly appeared. The tracked result is managed as a track (trace). The track ID is assigned to each object determined as the same in the continuous images in the camera by the above-described process.
After the person features, baggage features, person attributes, and baggage attributes of all the persons and all the pieces of baggage detected in the corresponding frames are obtained (Step S104), the rectangle positions, attributes, and features of the detected persons and pieces of baggage are written into the corresponding areas of the person table 1310 and the baggage table 1330 of the DB. The track IDs given by tracking are stored in an in-camera person track ID 1316—area of the person table and an in-camera track ID 1336—area of the baggage table (Step S105).
Next, the possibility that the persons have the pieces of baggage is determined. Indicator values (to be described later) are calculated for all the pieces of baggage located within a fixed distance (for example, three times the width of the person rectangle) from all the persons detected in the images, and are temporarily (until the tracking is finished) stored in the work area 1220 together with the frame IDs (Step S106).
Here, a calculation method of the indicator value will be described using
In the embodiment, as the indicator values, values obtained by dividing distances (the numbers of pixels) (L1, L2, and L3 in the drawing) on the image between the central points of the person and baggage rectangles shown in
The possibility that the person 901 has the baggage 903: L1/W1
The possibility that the person 901 has the baggage 904: L2/W1
The possibility that the person 902 has the baggage 904: L3/W2
Here, as the indicator values, values from 0 to 3 are used. As the indicator value is smaller, the distance is nearer, and it is determined that there is a high possibility that the corresponding person is the owner of the corresponding baggage.
With reference to
If there is completed person tracing, all the indicator values for the pieces of baggage linked to the corresponding person in all the frames linked to the corresponding person tracking result (track ID), namely, a local relationship degree between the objects in plural frame images is acquired from the work area 1220 (Step S108). Next, similar tracks of baggage are grouped on the basis of the similarity of the appearance of baggage using the baggage feature 1339 of each baggage to create a track cluster (Step S109). A new track cluster ID is allocated to each track cluster, and is written into the in-camera baggage track cluster ID field 1337 of the baggage table. By grouping the baggage tracks that are linked to the selected person and are similar in appearance into a track cluster, it is possible to integrate the baggage tracking that was interrupted due to an occlusion and the like.
Here, the baggage track cluster will be described using
Person track HT1: person rectangles 911, 912, 913, and 914
Baggage track BT1: baggage rectangles 915 and 916
Baggage track BT2: baggage rectangle 918
Here, the baggage is hidden by the person and cannot be seen at the time corresponding to the person rectangle 913, and thus the baggage track is interrupted once (divided into two baggage tracks BT1 and BT2). Accordingly, as a result of a determination that BT1 and BT2 are of the same baggage, BT1 and BT2 are grouped to create a baggage track cluster BTG1 in the process of Step S109. On the other hand, the baggage track is not interrupted in the case of camera 12, and thus a baggage track cluster BTG3 is configured using only the baggage track BT3.
With reference to the flow of
By the above-described process, a combination of a person and baggage that possibly belongs to the person and the indicator value are obtained in an in-camera video, and can be stored in the database.
Next, a baggage possession searching method between plural cameras will be described. First, an outline of a searching process will be described. When an operator designates an attentional person (the owner of baggage) in an in-camera baggage search, the corresponding person is first tracked between the cameras from the information of the database 1300. Namely, a person who is similar in appearance to the corresponding person is searched for in all the cameras, and the results are arranged in the order or time. On the basis of the person tracking information between plural cameras, the database is further searched, and all the pieces of baggage that possibly belong to the corresponding person are obtained in each camera to be displayed in a time-series manner. Further, on the basis of the baggage possession information in plural cameras, the pieces of baggage that possibly belong to the corresponding person are comprehensively determined and displayed.
On the basis of the video displayed by designating the camera ID and time or the video obtained from a search result to be described later, the operator selects a person or baggage from the video display area 2003 to be added to a tracking target object selection area 2100 by dragging. The tracking target object selection area can display plural objects (2111, 2112, and the like) to switch a tracking target object to be described later. The object displayed in the corresponding field is selected using a selection button 2111a as a tracking target between the cameras. In
A baggage possession search result is displayed in the lower half of the display screen 2000. The reference numeral 2200 denotes a tracking information display area. In an inter-camera person tracking result area 2210, videos (2211, 2212, 2213, and 2214) of the respective cameras for the corresponding person (M1 in the drawing) detected from the inter-camera person tracking result are arranged in the order of time. The video is a representative frame image of each track detected between the cameras to be described later. The camera ID and the time of the representative video are also displayed. In an area 2220, rectangle images of baggage candidates determined by the respective cameras and the indicator values are displayed on a track (camera) basis described above. This screen displays information obtained by tracking the person between the cameras in a time-series manner. Namely, information determined by the camera 1 is displayed in an area 2301, and information determined by the camera 2 is displayed in an area 2302 (the same applies to an area 2303 and the like). In the case where there are plural pieces of baggage of each camera as candidates, the pieces of baggage are displayed from the top in ascending order of the indicator value. Namely, baggage B1 and baggage B2 are displayed in the camera 2 displayed in the area 2302. The baggage B1 whose indicator value (0.3) is smaller is displayed in first rank, and the baggage B2 whose indicator value (1.0) is larger is displayed in second rank. Although the pieces of baggage up to the second rank are displayed in the example, more pieces of baggage can be displayed or the pieces of baggage can be scrolled.
A possession determination overall ranking is displayed in an area 2400 of the display screen 2000. Namely, the results obtained by determining the possibility of the possession of the person M1 by plural cameras are comprehensively determined (to be described later), and the pieces of baggage that possibly belong to the corresponding person are displayed. In
Next, a determination processing flow of an inter-camera baggage possession in the embodiment will be described using
Thereafter, the baggage information (the baggage track cluster ID, indicator value, and baggage feature) of all the pieces of baggage (all the pieces of baggage that are registered in the person/baggage relationship table 1350 and possibly belong to the corresponding person) linked to the person track obtained as described above is obtained for each camera (Step S123). Specifically, the person/baggage relationship table 1350 is referred to using the camera ID 1351 and in-camera person track ID 1352, and the baggage information (the baggage track cluster ID 1353 and the baggage possession determination indicator value 1354) is obtained. Thereafter, the baggage track clusters of all the pieces of baggage obtained as described above that are similar in appearance are grouped on the basis of the baggage feature 1339 of the baggage table, and are determined as the same baggage (Step S124).
Here, a “baggage track cluster group” will be described using
With reference to
(1) The groups are arranged in descending order of the number of baggage track clusters belonging to each group. Namely, the groups are arranged in descending order of the camera in which the number of times the baggage appears near the person is larger.
(2) In the case where the numbers of times are equal to each other, the groups are arranged in ascending order of the total value of indicator values. Namely, the baggage existing nearer the person is displayed high in rank.
The above-described method of determining the ranks (overall ranking) indicating the possibility that the baggage belongs to the designated person by the inter-camera tracking will be described in detail using
The traces (tracks) of the person determined as the same person obtained in Step S142 are arranged in the order of time (Step S143). Thereafter, Step S144 and processes subsequent thereto are performed for each of the traces (tracks) in order starting from the earliest time. Namely, the baggage track cluster IDs of the baggage candidates are arranged in ascending order of the indicator value (Step S144). Thereafter, the representative frame number 1355 is acquired from the person/baggage relationship table for the baggage track cluster ID having the smallest indicator value, and the entire camera image of the representative frame is displayed in the column 2210 of
Further, Step S146 and processes subsequent thereto are performed for all the baggage track cluster IDs in ascending order of the indicator value. Namely, the representative frame number is acquired (Step S146), and an image of a baggage rectangle is acquired from the image of the representative frame number to be displayed, together with the indicator value, at the position of the corresponding rank in the column 2220 of
Finally, for all the baggage track cluster groups determined as the same baggage, the rectangle images (any of the rectangles of the representative frame numbers) of the baggage are displayed in the possession determination overall ranking area 2400 in accordance with the overall ranking (Step S150). In the case where the baggage track cluster group of the next rank exists, Step S150 is repeated (Step S151).
The display processing flow described in
A screen example used in the case where the owner is determined using the baggage image as a key by switching the tracking target object is shown in
Here, as a utilization example of the embodiment, an example in which leaving or delivery is detected by utilizing the information of the tracking information display area 2200 is shown in
In
Strictly speaking, it is necessary to confirm that other persons do not have the corresponding baggage B5 by performing the baggage tracking. In the case where the other persons have the baggage B5, it is determined as delivery to be described next.
In
As described above, various persons and pieces of baggage in videos are selected from the tracking target object selection area 2100 to determine the possession and owner by tracking, and the results are displayed in the tracking information display area 2200. Accordingly, it is possible to determine the ownership of baggage from various viewpoints such as the possession, owner, leaving, and delivery.
Next, a method of performing a person-baggage cooperation search in the video analysis system of the embodiment will be described using
A screen interface of the person-baggage cooperation search is shown in
The reference numerals 3101 and 3201 denote check boxes for designating a person attribute search result or a baggage attribute search result to be sorted on a priority basis when the search results are displayed. Since the priority of the person attribute search is designated in the example of
Next, for all the frames (all the frames obtained by matching the frame ID 1313 of the person table and the frame ID 1333 of the baggage table with each other) in which the corresponding person and baggage appear, the person/baggage relationship table 1350 is searched for all the combinations of the person and the baggage (all the combinations of the in-camera person ID and the in-camera baggage track cluster ID) for each camera, and the baggage possession determination indicator value 1354 is obtained (Step S163). Thereafter, all the combinations in which the above-described indicator value is smaller than the threshold designated in the area 3002 are acquired (Step S164).
Thereafter, on the basis of the display priority designated in the areas 3101 and 3201, the determination using the evaluation value (the distance between the designated attribute vector and the attribute vector of the database) of the person/baggage attributes and the baggage possession determination indicator value is made to determine the priority degree of the search result display (Step S165). For example, the following determination is made.
(1) A case in which the person attributes are given priority: a descending order of the matching degree of the person attributes. Namely, an ascending order of a distance between the designated attributes and the person attribute vector stored in the DB. It should be noted that in the case where the person attribute determination results are the same, a descending order of the matching degree of the baggage attributes is used.
(2) A case in which the baggage attributes are given priority: a descending order of the matching degree of the baggage attributes. It should be noted that in the case where the baggage attribute determination results are the same, a descending order of the matching degree of the person attributes is used.
(3) A case in which both of the person attributes and the baggage attributes are given priority (both of the areas 3101 and 3201 are checked): an ascending order of the weighting average of the person attributes, the baggage attributes, and the evaluation values. Namely, a descending order of the possibility of the possession because the attributes are comprehensively close.
(4) A case in which neither the person attributes nor the baggage attributes are given priority (neither the areas 3101 nor 3201 are checked): an order of time
With reference to
It should be noted that although an example of the attribute search has been described above, the similar image search of the person and the baggage in cooperation can be also made and displayed using the same screen and algorism. In this case, the person and the baggage that are similar in appearance to the searched person/baggage rectangle images are searched for on the basis of the similarity (the distance of the feature vector) instead of the matching degree of the attribute vector. Further, the determination in which the attributes and the similar images are combined to each other can be also made by the same method. For example, a person who is similar to a particular person and has a blue suitcase can be searched for.
As described above, according to the embodiment, since a first object and a second object are stored in the database while being linked to each other, combinations of plural objects can be monitored, namely, for example, the ownership of the baggage by the person can be determined. In addition, even in the case where the baggage does not appear in all the frames, the similar baggage can be linked and associated using the owner as a key. Further, similar objects (a person and baggage) are determined between cameras, and the determination of the ownership in which information of plural cameras is integrated can be made. Further, a change in the ownership between plural cameras can be determined.
In the first embodiment, the possession/owner determination of the corresponding person/baggage is made on the basis of the time-series results obtained by tracking the person and the baggage similar in appearance features to those selected in the tracking target object selection area 2100 between plural cameras. However, if other persons and other baggage that are similar in appearance are mixed in the similarity (person/baggage) search conducted in the tracking, wrong tracking occurs, and the determination result is possibly incorrect. In particular, there are many pieces of similar baggage, and thus it is conceivable that wrong tracking is likely to occur.
Accordingly, in the embodiment, provided is a function of deleting tracking results (tracks by cameras) of the person/baggage displayed as a result of wrong tracking. Further, a tracking result of a camera related to the deleted tracking result of the camera is also deleted by considering a spatiotemporal restriction on the basis of the special positional relationship between the cameras.
In the example of
In this case, in the case where an operator determines that wrong detection has occurred when viewing the screen, it is necessary to delete the corresponding wrong tracking result by pressing a deletion button 4103 or 4105. Further, when the video of either the camera 52 or 51 is deleted by the deletion button 4103 or 4105 in the embodiment, a spatiotemporal restriction is determined by an algorism to be described later, and the relevant wrong detection result (one of the tracking results of the cameras 52 and 51) is also automatically deleted.
Next, a method of deleting a relevant wrong tracking result on the basis of a positional relation between cameras will be described using
When the tracking screen shown in
Thereafter, a camera (track) that is further reachable from the camera of the selected track is recursively selected (Step S202). Specifically, the camera 51 that is reachable from the camera 52, the camera 55 that is reachable from the camera 54, and the camera 56 are determined. Finally, a list structure (
In
Next, an algorism to remove a wrong tracking object is shown in
For example, in the case where the tracking result of the object B11 appearing on the camera 52 is deleted by an operator in
In the case where an operator deletes the wrong tracking object in
In the above-described embodiments, ownership between a person and baggage is determined. However, the method in the above-described embodiments can be applied to a relation between other objects. For example, the similar method can be applied to a relation between a person and an automobile and a relation between a person and a bicycle.
Accordingly, the baggage possession determination indicator value 1354 in the above-described embodiment serves as an index indicating a possession, use, and grasp between a person and an object in the embodiment.
In addition, between a person and an automobile, the tracking of the person is possibly interrupted due to occurrence of an occlusion. Therefore, plural tracks of a person are connected to each other on the screen using the track cluster ID.
Although the embodiments have been described above, the present invention is not limited to the above-described embodiments, and includes various modified examples. For example, the embodiments have been described in detail to easily understand the present invention, and the present invention is not necessarily limited to those including all the configurations described above. In addition, some configurations of an embodiment can be replaced by a configuration of another embodiment. In addition, a configuration of an embodiment can be added to a configuration of another embodiment. In addition, some configurations of each embodiment can be added to, deleted from, and replaced by other configurations.
In addition, some or all of the above-described configurations, functions, functional units, and the like may be realized using hardware by designing with, for example, integrated circuits.
Number | Date | Country | Kind |
---|---|---|---|
2020-040748 | Mar 2020 | JP | national |