This application claims the benefit of Taiwan application Serial No. 110144217, filed Nov. 26, 2021, the subject matter of which is incorporated herein by reference.
The disclosure relates in general to an image analysis method and image analysis device using the same.
The development of wireless networks or other environmental reasons (such as COVID-19 epidemic) usually changes the model of people participating in sports events. Most people turn to watch the game on-line, and thus physical game (or match) has changed from full cheering to a small number of spectators or even no spectators. Therefore, in order to cope with this changing trend, it is one of the important issues faced by the industry in this technical field to submit an image analysis method for image stream of game.
According to an embodiment, an image analysis method is provided. The image analysis method includes receiving an image stream; analyzing a to-be-analyzed frame of the image stream to obtain a scene type of the to-be-analyzed frame; determining whether the scene type of the to-be-analyzed frame is a classification of needing posture analysis; obtaining a human body posture of a human body image of the to-be-analyzed frame when the scene type of the to-be-analyzed frame is the classification of needing posture analysis; and determining an event type of the to-be-analyzed frame according to the scene type and the human body posture.
According to another embodiment, an image analysis device is provided. The image analysis device includes a scene analysis unit, a posture analysis unit and an event analysis unit. The scene analysis unit is configured to receive an image stream and analyze a to-be-analyzed frame of the image stream to obtain a scene type of the to-be-analyzed frame and determine whether the scene type of the to-be-analyzed frame is a classification of needing posture analysis. The posture analysis unit is configured to obtain a human body posture of a human body image of the to-be-analyzed frame when the scene type of the to-be-analyzed frame is the classification of needing posture analysis. The event analysis unit is configured to determine an event type of the to-be-analyzed frame according to the scene type and the human body posture.
The above and other aspects of the disclosure will become better understood with regard to the following detailed description of the preferred but non-limiting embodiment (s). The following description is made with reference to the accompanying drawings.
In the following detailed description, for purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding of the disclosed embodiments. It will be apparent, however, that one or more embodiments may be practiced without these specific details. In other instances, well-known structures and devices are schematically shown in order to simplify the drawing.
Referring to
The image analysis device 100 includes a scene analysis unit 110, a posture analysis unit 120, an event analysis unit 130, and a processing unit (or processor) 140. The scene analysis unit 110 is configured to receive an image stream S1 and analyzing a to-be-analyzed frame F1 of the image stream S1 to obtain a scene type C1 of the to-be-analyzed frame F1, and determine whether the scene type C1 of the to-be-analyzed frame F1 is “classification of needing posture analysis”. The posture analysis unit 120 is configured to obtain a human body posture P1 in a human body image H1 of the to-be-analyzed frame F1 when the scene type C1 of the to-be-analyzed frame F1 is the “classification of needing posture analysis”. The event analysis unit 130 is configured to determine the event type E1 of the to-be-analyzed frame F1 according to the scene type C1 and the human body posture P1. As a result, the image analysis device 100 could automatically analyze to-be-analyzed frame F1 to determine (or output) the event type E1, without the need for additional manual determining and processing. In addition, after obtaining the event type E1, the image analysis device 100 could perform corresponding steps (or action), such as inserting a virtual advertisement and/or storing a frame (or recording), wherein the processing unit 140 further could post-produce the stored frame into a specific action movie (e.g. pitching, swinging, catching) and/or game highlights, etc. and/or analyze game (or race) data (e.g., ball speed analysis) based on the stored frame.
In an embodiment, the scene analysis unit 110 is configured to determine whether the scene type C1 of the to-be-analyzed frame F1 is the “classification of needing posture analysis” according to a corresponding relationship of the scene type C1 and the posture analysis.
In an embodiment, the posture analysis unit 120 is further configured to obtain the whole body skeleton characteristic of the human body image H1, analyze a number of the to-be-analyzed frames F1 to obtain a skeleton movement of the whole body skeleton characteristic, and obtain the human body posture P1 of the human body image H1 according to the skeleton movement. In an embodiment, the human body posture P1 is, for example, the whole body posture of the human body image H1.
In an embodiment, the processing unit 140 is configured to perform an event operation W1 corresponding to the event type E1 according to the corresponding relationship between the event type E1 and the event operation W1. The corresponding relationship is, for example, pre-stored in a storage unit (not shown), wherein the storage unit could be disposed inside or outside the processing unit 140.
The image analysis device 100 of the embodiment of the present disclosure could be applied to the analysis of the image stream of a baseball game. In the baseball game, the scene type C1 includes, for example, “classification of needing posture analysis” and “classification of not needing posture analysis”, wherein the “classification of not needing posture analysis” includes, for example, a scene having high viewer attention (e.g., a frame having high extent of excitement) such as “outfield”, “pitcher-batter”, “infield”, etc., and “classification of not needing posture analysis”, for example, is a scene having low viewer attention (e.g., a frame having low extent of excitement) such as “panorama of infield and outfield” (offensive-and-defensive exchange) . . . etc. The human body image H1 includes, for example, “pitcher”, “batter (or hitter)”, “outfielder”, “runner” . . . etc. The human body posture P1 includes, for example, the actions performed by the human body image H1, such as standing, clasping palms, striding, pitching, running, raising both hands, swinging, catching a ball, and any other actions that players would make on a baseball field. The event type E1 includes, for example, “pitching preparation”, “pitching”, “hitting preparation”, “strike/hit”, “home run”, “catch-out”, “offensive-and-defensive exchange” . . . etc. The event operation W1 includes, for example, “insert virtual advertisement” and/or “saving frame (or screen)” (or video recording) . . . etc.
In case of the baseball game, the corresponding relationship among the scene type C1, the human body image H1, the human body posture P1, the event type E1 and the event operation W1 is shown in the following TABLE 1. The corresponding relationship could be preset and stored in the storage unit in advance. However, the corresponding relationship among the scene type C1, the human body image H1, the human body posture P1, the event type E1 and the event operation W1 in the embodiment of the present disclosure is not limited by TABLE 1, and it may be other forms of corresponding relationship. In addition, the number of groups of the corresponding relationship is not limited to eight in TABLE 1, and the actual number of groups of the corresponding relationship could be increased or decreased depending on the actual application.
Furthermore, the image analysis method of the image analysis device 100 is further described with
The following is a description of the to-be-analyzed frame F1 in
In step S110, the scene analysis unit 110 receives the image stream S1 including the to-be-analyzed frame F1 shown in
In step S120, the scene analysis unit 110 analyzes at least one to-be-analyzed frame F1 of the image stream S1 (
In step S130, the scene analysis unit 110 determines whether the scene type C1 of the to-be-analyzed frame F1 is “classification of needing posture analysis”. If yes, the process proceeds to step S140; if not, the human body posture P1 of the human body image H1 does not need to be analyzed, and the process directly proceeds to step S150. For example, the scene analysis unit 110 determines that the scene type C1 (“pitcher-batter”) of the to-be-analyzed screen F1 in
The scene analysis unit 110 could determine whether the scene type C1 of the to-be-analyzed frame F1 is “posture analysis type” according to the corresponding relationship between the scene type C1 and the posture analysis. For example, as shown in the corresponding relationship #1 in TABLE 1, “pitcher-batter” belongs to the “classification of needing posture analysis”, and as shown in the corresponding relationship #7 in TABLE 1, if the scene type C1 of the to-be-analyzed frame F1 is “panorama of infield and outfield” (offensive-and-defensive exchange), and it belongs to “classification of not needing posture analysis”.
In step S140, the posture analysis unit 120 obtains the human body posture P1 of the human body image H1 of the to-be-analyzed frame F1.
The human body posture P1 is, for example, the whole body posture of the human body image H1. In detail, the posture analysis unit 120 could obtain, by using image analysis technology, the human body image H1 of each to-be-analyzed frame F1, for example, the human body images H11 to H13 shown in
As shown in
In step S150, the event analysis unit 130 determines the event type E1 of the to-be-analyzed frame F1 according to the scene type C1 and the human body posture P1. For example, the event analysis unit 130 determines the event type E1 of the to-be-analyzed frame F1 according to “pitcher-batter” (scene type C1) and “walking” (human body posture P1). The event analysis unit 130 could determine the result of the event type E1 of the to-be-analyzed frame F1 is “pitching preparation” according to the corresponding relationship #1 in TABLE 1.
In S160, after the event type E1 is generated, the processing unit 140 could perform the event operation W1 corresponding to the event type E1 according to TABLE 1. For example, as shown in
The following is a description of the to-be-analyzed frame F1 in FIG. 3B.
The scene analysis unit 110 analyzes at least one to-be-analyzed frame F1 (
The following is a description of the to-be-analyzed frame F1 in
The scene analysis unit 110 analyzes at least one to-be-analyzed frame F1 (
The following is a description of the to-be-analyzed frame F1 in
The scene analysis unit 110 analyzes at least one to-be-analyzed frame F1 (
In the present embodiment, the image stream S1 includes several to-be-analyzed frames F1. The image analysis device 100 could sequentially analyze the to-be-analyzed frame F1, and generate or output the event type E1 corresponding to one or more of the to-be-analyzed frames F1. In addition, the image analysis device 100 could mark (or insert), by using image insertion/processing technology, the analysis/determination result (e.g., text/wording) of at least one of the corresponding event type E1, the human body posture P1 and scene type C1 in the advertising area and/or the corner area of each to-be-analyzed frame F1. For example, as shown in
In summary, the embodiment of the present disclosure proposes an image analysis device that could determine the event type of the to-be-analyzed frame according to the scene type and the human body posture of at least one to-be-analyzed frame in the image stream. After obtaining the event type, the image analysis device accordingly could perform corresponding steps, such as inserting virtual advertisements and/or storing frames (or video recording). As a result, the image analysis device could automatically analyze at least one to-be-analyzed frame in the image stream without additional manual processing. Furthermore, through the image analysis method of this disclosed embodiment, even if the audience is watching the game online, the image analysis device could insert the virtual advertisement in an appropriate area of the to-be-analyzed frame without affecting the frame viewing, or/and for the to-be-analyzed frame having low extent of excitement, the image analysis device could store the to-be-analyzed frame for post-producing a short video and/or analyzing the event data.
It will be apparent to those skilled in the art that various modifications and variations could be made to the disclosed embodiments. It is intended that the specification and examples be considered as exemplary only, with a true scope of the disclosure being indicated by the following claims and their equivalents.
Number | Date | Country | Kind |
---|---|---|---|
110144217 | Nov 2021 | TW | national |