The present disclosure relates to a data processing method, a data processing device, a program, and a moving body control system, and more particularly, to a data processing method, a data processing device, a program, and a moving body control system capable of easily realizing intended movement.
Patent Document 1 discloses a content sharing device that obtains a subject size by obtaining a subject position of a subject appearing in content on the basis of an imaging position, an imaging direction, and a subject distance of an imaging device and obtaining an imaging range size on the basis of an imaging surface size, a focal length, and the subject distance of the imaging device.
According to the content sharing device disclosed in Patent Document 1, a subject can be reliably identified for shared content, and composition and camerawork can be easily managed.
In recent years, it is becoming possible to capture a dynamic moving image by using a moving body capable of moving in a wide range at a high speed and with a high degree of freedom. However, it is difficult to operate and control the moving body as intended, and it is not easy to capture a desired moving image.
The present disclosure has been made in view of such a situation, and an object thereof is to easily realize intended movement.
A data processing method of the present disclosure is a data processing method in which a data processing device is configured to generate control data corresponding to a moving route of a first moving body for controlling movement of a second moving body on the basis of a first moving image captured from the first moving body.
A data processing device of the present disclosure is a data processing device including a data generation unit that generates control data corresponding to a moving route of a first moving body for controlling movement of a second moving body on the basis of a moving image captured from the first moving body.
A program of the present disclosure is a program for causing a computer to execute processing of generating control data corresponding to a moving route of a first moving body for controlling movement of a second moving body on the basis of a moving image captured from the first moving body.
A moving body control system of the present disclosure is a moving body control system including: a data processing device that generates control data corresponding to a moving route of a first moving body for controlling movement of a second moving body on the basis of a moving image captured from the first moving body; and the second moving body that moves on the basis of the control data.
In the present disclosure, control data corresponding to a moving route of a first moving body for controlling movement of a second moving body is generated on the basis of a moving image captured from the first moving body.
Hereinafter, modes for carrying out the present disclosure (hereinafter referred to as embodiments) will be described. Note that the description will be made in the following order.
In recent years, it is becoming possible to capture a dynamic moving image by using a moving body capable of moving in a wide range at a high speed and with a high degree of freedom. However, it is difficult to operate the moving body so as to smoothly move along the intended moving route, and furthermore, it is not easy to ensure safety while performing imaging. Naturally, the user who captures the moving image does not necessarily have an advanced moving body operation techniques.
For example, a drone or the like on which imaging equipment is mounted is mainly operated manually or controlled according to a programmed template. The difficulty of safely capturing images while manually operating as intended is as described above. Meanwhile, according to the control according to the template, the image can be captured with high reproducibility in a limited pattern, but the moving image tends to be an ordinary moving image.
Therefore, in the technology according to the present disclosure, movement as intended by a moving image creator (user) is easily realized on the basis of a moving image captured by another person, so that a desired moving image can be captured more easily.
A moving body control system 1 illustrated in
The data processing device 10 includes, for example, a cloud server provided on a cloud or a general-purpose computer such as a personal computer (PC).
The data processing device 10 includes a control data generation unit 11. The control data generation unit 11 acquires a captured moving image captured from another moving body (first moving body), and generates control data for controlling the moving body of the moving body 20 (second moving body). The control data is data corresponding to a moving route of another moving body, and includes, in addition to the moving route of the moving body 20, a speed and a posture when the moving body 20 moves according to the moving route, control information for controlling imaging from the moving body 20 to be described later, and the like.
The moving body 20 includes, for example, an autonomous mobile robot such as an unmanned aerial vehicle (UAV) (so-called drone) that autonomously flies, an autonomous traveling vehicle that moves on land, an autonomous navigation vessel that moves on water or under water, and an autonomous mobile vacuum cleaner that moves indoors.
The moving body 20 can move on the basis of the control data generated by the data processing device 10 (control data generation unit 11). Furthermore, a camera 20C is mounted on the moving body 20, and a user who is a moving image creator can capture a desired moving image by the camera 20C.
In step S1, the control data generation unit 11 acquires a captured moving image (first moving image) captured from another moving body.
In step S2, the control data generation unit 11 generates control data on the basis of the captured moving image. Specifically, the control data generation unit 11 generates control data in response to a user's direct editing operation on a moving route of another moving body estimated from the captured moving image, or generates control data reflecting the user's imaging intention or imaging environment regarding the moving image (second moving image) captured by the camera 20C.
The control data generated in this manner controls the movement of the moving body 20 or controls imaging from the moving body 20 (camera 20C). Movement of the moving body 20 and imaging from the moving body 20 (camera 20C) may be controlled on the basis of control data generated in advance, or may be controlled in real time on the basis of control data sequentially generated.
A moving body control system 100 illustrated in
In the moving body control system 100, the server 130 is communicably connected to each of the client terminal 110, the controller 150, and the moving body 170 via a wired or wireless communication path. In addition, the client terminal 110, the server 130, the controller 150, and the moving body 170 constituting the moving body control system 100 may be communicably connected to each other via a network NW such as the Internet.
The client terminal 110 includes a PC or a tablet terminal operated by a user who operates and controls the moving body 170. The client terminal 110 includes a control unit 111, a communication unit 112, an input unit 113, a display unit 114, and a memory 115.
The control unit 111 includes a processor such as a central processing unit (CPU), and controls each unit of the client terminal 110 by executing a predetermined program according to an input signal from the input unit 113 or the like.
The communication unit 112 includes a network interface or the like, and performs wireless or wired communication with the server 130.
The input unit 113 includes a keyboard, a mouse, a microphone, a touch panel, and the like, and supplies an input signal corresponding to a user's operation to the control unit 111.
The display unit 114 includes a liquid crystal display, an organic electro-luminescence (EL) display, or the like, and displays various types of information under the control of the control unit 111.
The memory 115 includes a non-volatile memory such as a flash memory, and stores various types of information under the control of the control unit 111.
In the client terminal 110 configured in this manner, search for a captured moving image, input of an imaging intention and an imaging environment, confirmation and correction of a moving route reproduced by control data generated by the server 130, and the like are executed according to a user's operation. Furthermore, in the client terminal 110, various processing results such as control data generated by the server 130 and a simulation result of movement based on the control data are presented to the user.
The server 130 includes a cloud server provided on a cloud or a general-purpose computer such as a PC. The server 130 includes a control unit 131, a communication unit 132, a display unit 133, and a memory 134.
The control unit 131 includes a processor such as a CPU, and controls each unit of the server 130 by executing a predetermined program.
The communication unit 132 includes a network interface or the like, and performs wireless or wired communication with the client terminal 110, the controller 150, and the moving body 170.
The display unit 133 includes a liquid crystal display, an organic EL display, or the like, and displays various types of information under the control of the control unit 131.
The memory 134 includes a non-volatile memory such as a flash memory, and stores various types of information under the control of the control unit 131.
In the server 130 configured in this manner, generation of control data based on the captured moving image, construction of a neural network used for generation of control data, simulation of movement based on the control data, and the like are executed. These execution results are transmitted to the client terminal 110. The generated control data is transmitted to the moving body 170 via the controller 150 or directly. The constructed neural network may be transmitted to and installed in the moving body 170.
The controller 150 includes a radio control device or a smartphone operated by a user who operates and controls the moving body 170. The controller 150 includes a control unit 151, a communication unit 152, an input unit 153, a display unit 154, and a memory 155.
The control unit 151 includes a processor such as a CPU, and controls each unit of the controller 150 by executing a predetermined program according to an input signal from the input unit 153 or the like.
The communication unit 152 includes a network interface or the like, and performs wireless or wired communication with the server 130. Furthermore, the communication unit 152 can also perform wireless direct communication with the moving body 170.
The input unit 153 includes a touch panel or the like in addition to a stick, a switch, and a button, and supplies an input signal corresponding to a user's operation to the control unit 151.
The display unit 154 includes a liquid crystal display, an organic EL display, or the like in addition to various lamps, and displays various types of information under the control of the control unit 151.
The memory 155 includes a non-volatile memory such as a flash memory, and stores various types of information under the control of the control unit 151.
In the controller 150 configured in this manner, an operation signal corresponding to a user's operation is transmitted to the moving body 170. Furthermore, in the controller 150, the aircraft state of the moving body 170 is presented to the user as necessary.
The moving body 170 includes a UAV (drone), an autonomous mobile robot such as an autonomous traveling vehicle, an autonomous navigation vessel, and an autonomous mobile vacuum cleaner. The moving body 170 includes a control unit 171, a communication unit 172, a camera 173, a sensor 174, and a drive unit 175.
The control unit 171 includes a CPU, a memory, and the like, and controls the communication unit 172, the camera 173, and the drive unit 175 by executing a predetermined program.
The communication unit 172 includes a network interface or the like, and performs wireless or wired communication with the controller 150 and the server 130 for operating the moving body 170.
The camera 173 captures a moving image under the control of the control unit 171. Furthermore, the camera 173 can also be configured to capture a moving image on the basis of control of an imaging processing unit (processor or the like) provided therein, for example, without depending on the control of the control unit 171.
The sensor 174 includes various sensors, and senses each direction around the moving body 170 including the traveling direction of the moving body 170. By sensing the surroundings of the moving body 170, autonomous movement of the moving body 170 is realized.
The drive unit 175 is a mechanism for moving the moving body 170, and includes a flight mechanism, a traveling mechanism, a propulsion mechanism, and the like. In this example, the moving body 170 is configured as a UAV, and the drive unit 175 is configured by a motor, a propeller, or the like as a flight mechanism. Furthermore, in a case where the moving body 170 is configured as an autonomous traveling vehicle, the drive unit 175 is configured by wheels or the like as a traveling mechanism, and in a case where the moving body 170 is configured as an autonomous navigation vessel, the drive unit 175 is configured by a screw propeller or the like as a propulsion mechanism. The drive unit 175 is driven according to the control of the control unit 171 to move the moving body 170.
In the moving body 170 configured in this manner, control of movement of the moving body 170 and control of imaging from the moving body 170 are executed on the basis of the control data generated by the server 130.
Hereinafter, an embodiment realized by the moving body control system 100 described with reference to
The moving body control system in
The search unit 201 can be realized by the client terminal 110 in
The search unit 201 searches the captured moving image captured from another UAV according to the user's operation, and supplies the searched captured moving image to the flight data generation unit 202.
Specifically, the search unit 201 searches for the captured moving image on the basis of imaging feature parameters representing imaging features of the moving image, such as a composition and an angle of view of the moving image, camerawork, a positional relationship with a subject to be gazed, a moving route and a moving speed at the time of imaging, and a structure of an environment in which the moving image is captured, which are input by the user, and other free words. Furthermore, the search unit 201 may search and select a captured moving image from, for example, sample moving images that can be previewed.
The user can find a captured moving image to be referred to in capturing a desired moving image by operating the search unit 201. Note that the search engine that actually searches for the captured moving image may be incorporated in the client terminal 110 or may be arranged on the Web.
The flight data generation unit 202 and the simulation unit 203 can be realized by the server 130 in
The flight data generation unit 202 generates flight control data corresponding to a moving route (flight route) of the UAV on which the captured moving image is captured, for controlling the flight of the UAV 204, on the basis of the captured moving image from the search unit 201, and outputs the flight control data to the simulation unit 203.
The flight data generation unit 202 of
The flight data estimation unit 211 estimates flight data including a flight route of the UAV on which the captured moving image is captured on the basis of the captured moving image from the search unit 201. For example, the flight data estimation unit 211 estimates the flight data by calculating a map of the surrounding environment where the UAV in which the captured moving image is captured flies and a moving route of the UAV in the map by visual simultaneous localization and mapping (SLAM) by the monocular camera.
In a case where the camera parameters of the camera mounted on the UAV are unknown, the influence of distortion remains or the scale becomes indefinite. In addition, in a case where the moving route extends over a long distance, the accuracy of the SLAM decreases. However, in order to generate the flight data in the subsequent stage, it is important to obtain an element used for editing in the subsequent stage, and the influence of distortion and geometric accuracy of SLAM on the final imaging result is minor. Further, the indefinite scale can be resolved by manually adjusting the scale, for example.
The flight data editing unit 212 generates flight data by editing estimated flight data. Specifically, the flight data editing unit 212 generates the flight control data in response to the user's direct editing operation of the estimated flight data.
Here, the flight control data is generated by, for example, directly geometrically editing the flight route corresponding to the estimated flight data displayed on the client terminal 110 by the user's operation. Specifically, control data adapted to the flight scene of the UAV 204 is generated by editing a straight line or a curved line representing the flight route or changing a subject to be gazed.
The simulation unit 203 executes a flight simulation of the UAV 204 on the basis of the flight control data generated by the flight data generation unit 202. The simulation result of the flight is displayed on the client terminal 110. The simulation unit 203 corrects the flight control data according to the user's direct editing operation on the simulation result displayed on the client terminal 110.
In this manner, the flight control data obtained by the user repeating confirmation and correction of the simulation result is output to the UAV 204.
The UAV 204 corresponds to the moving body 170 in
Next, a flow of flight data generation processing by the flight data generation unit 202 in
In step S11, the flight data generation unit 202 acquires the captured moving image searched by the search unit 201.
In step S12, the flight data estimation unit 211 estimates flight data including the flight route of the UAV on which the captured moving image is captured on the basis of the captured moving image from the search unit 201.
In step S13, the flight data editing unit 212 generates the flight control data in accordance with the user's direct editing operation of the estimated flight data.
The flight control data generated in advance in this manner controls the flight of the UAV 204 or controls imaging from the camera mounted on the UAV 204.
According to the above processing, the flight data of the UAV in which the captured moving image is captured is estimated, and the flight control data is generated according to the editing operation of the estimated flight data. As a result, it is possible to easily realize the intended flight similar to the UAV in which the captured moving image is captured by the simple editing operation without requiring the advanced manual operation techniques, and eventually, it is possible to more easily capture a moving image desired by the user.
In the moving body control system of
The imaging feature extraction unit 221 extracts imaging feature parameters of the captured image from the flight data (estimated flight data) estimated by the flight data estimation unit 211, and supplies the imaging feature parameters to the flight data generation model 222. The imaging feature parameter is a human-resolvable parameter representing an imaging feature of a moving image.
The imaging feature extraction unit 221 includes, for example, a machine learning model obtained by performing learning with actual flight data obtained by the UAV actually flying as an input and an imaging feature parameter of a moving image captured by the flight as an output.
The flight data generation model 222 generates flight control data by applying the imaging feature parameters extracted by the imaging feature extraction unit 221 to the imaging intention and the imaging environment input by the user.
The imaging intention input by the user corresponds to an imaging feature represented by an imaging feature parameter, and includes, for example, a composition and an angle of view of a moving image desired by the user, camerawork, a positional relationship with a subject to be gazed, a moving route at the time of imaging, and the like.
The imaging environment input by the user also corresponds to the imaging feature represented by the imaging feature parameter, and includes a structure of an environment (in which the UAV 204 flies) in which the moving image desired by the user is captured, that is, a geometric shape of the imaging target or the capturable target around the UAV 204. For example, the imaging environment includes a roof, a wall, a column, and the like in a building, includes a ground, a cliff, trees, and the like in the outdoors, and includes a large structure such as a building or a bridge in an urban area in particular. Furthermore, the imaging environment may include other moving bodies including a person, a vehicle, and a flying object, or may include an object as an imaging target or an object to be excluded from the imaging target. Furthermore, the imaging environment may include a person or an object that needs to be separated by a predetermined distance or more to generate a flight route during flight of the UAV 204, a person or an object that needs to be avoided during flight, or may include a positional relationship between a light source including the sun or illumination and a subject. Furthermore, the imaging environment may include latitude, longitude, map information including topography, season, time zone, weather, temperature, and the like.
Here, in the client terminal 110, for example, the imaging feature is adjusted using a graphical user interface (GUI) by the user's operation, thereby generating the flight control data.
The flight data generation model 222 includes a using the imaging feature parameter, the imaging intention, and the imaging environment as inputs and using the actual flight data obtained by the UAV actually flying as an output under these conditions.
Next, a flow of flight data generation processing by the flight data generation unit 202 in
In step S21, the flight data generation unit 202 acquires the captured moving image searched by the search unit 201.
In step S22, the flight data estimation unit 211 estimates flight data including the flight route of the UAV on which the captured moving image is captured on the basis of the captured moving image from the search unit 201.
In step S23, the imaging feature extraction unit 221 extracts the imaging feature parameters of the captured image from the estimated flight data estimated by the flight data estimation unit 211.
In step S24, the flight data generation model 222 generates flight control data by applying the imaging feature parameters extracted by the imaging feature extraction unit 221 to the imaging intention/imaging environment input by the user. At this time, the imaging feature parameters extracted by the imaging feature extraction unit 221 may be adjusted by the user and then input to the flight data generation model 222.
The flight control data generated in advance in this manner controls the flight of the UAV 204 or controls imaging from the camera mounted on the UAV 204.
Note that the flight of the UAV 204, the actual flight data and the moving image obtained by imaging from the UAV 204, and the imaging intention and the imaging environment input by the user can be used for learning of the imaging feature extraction unit 221 and the flight data generation model 222.
According to the above processing, the imaging feature parameter that is resolvable to a person is extracted from the flight data estimated on the basis of the captured moving image, and the imaging feature parameter is applied to the imaging intention or the imaging environment of the user, whereby the flight control data is generated. As a result, it is possible to easily realize the intended flight only by adjusting the flight and imaging features that can be interpreted by the user, and eventually, it is possible to more easily capture a moving image desired by the user.
In the moving body control system of
The flight data generation model 231 generates flight control data by adapting the flight data (estimated flight data) estimated by the flight data estimation unit 211 to the imaging intention and the imaging environment input by the user.
Here, in the client terminal 110, for example, a text indicating an imaging intention or an imaging environment is input or the imaging intention or the imaging environment is selected using the GUI by the user's operation, whereby the flight control data is generated.
The flight data generation model 231 includes a machine learning model obtained by performing learning using the estimated flight data, the imaging intention, and the imaging environment as inputs and using the actual flight data obtained by the UAV actually flying as an output under these conditions.
Next, a flow of flight data generation processing by the flight data generation unit 202 in
In step S31, the flight data generation unit 202 acquires the captured moving image searched by the search unit 201.
In step S32, the flight data estimation unit 211 estimates flight data including the flight route of the UAV on which the captured moving image is captured on the basis of the captured moving image from the search unit 201.
In step S33, the flight data generation model 231 generates flight control data by adapting the estimated flight data estimated by the flight data estimation unit 211 to the imaging intention and the imaging environment input by the user.
The flight control data generated in advance in this manner controls the flight of the UAV 204 or controls imaging from the camera mounted on the UAV 204.
Note that the flight of the UAV 204, the actual flight data and the moving image obtained by imaging from the UAV 204, and the imaging intention and the imaging environment input by the user can be used for learning of the flight data generation model 231.
According to the above processing, the flight control data is generated by adapting the flight data estimated on the basis of the captured moving image to the user's imaging intention and imaging environment. As a result, it is possible to easily realize the intended flight only by the user inputting his/her own imaging intention or an imaging environment in which the UAV 204 is desired to fly, and eventually, it is possible to more easily capture a moving image desired by the user.
Note that, in the second and third embodiments described above, in a case where the captured image searched by the search unit 201 is a moving image associated with the actual flight data or the imaging intention, the estimated flight data or the imaging feature parameter based on the captured moving image is the actual flight data or the imaging intention itself. In this case, the flight data estimation and the imaging feature parameter extraction can be skipped, and the accuracy up to the generation of the flight control data can be improved.
In the moving body control system of
The imaging feature extraction unit 241 directly extracts imaging feature parameters from the captured moving image from the search unit 201, and supplies the imaging feature parameters to the flight data generation model 242. As described above, the imaging feature parameter is a human-resolvable parameter representing the imaging feature of the moving image.
The imaging feature extraction unit 241 includes, for example, a machine learning model obtained by performing learning with a moving image captured by actual flight of the UAV as an input and an imaging feature parameter of the moving image as an output.
The flight data generation model 242 generates flight control data by applying the imaging feature parameters extracted by the imaging feature extraction unit 241 to the imaging intention and the imaging environment input by the user.
Here, in the client terminal 110, for example, the imaging feature is adjusted using the GUI by the user's operation, thereby generating the flight control data.
The flight data generation model 242 includes a machine learning model obtained by performing learning using the imaging feature parameter, the imaging intention, and the imaging environment as inputs and using the actual flight data obtained by the UAV actually flying as an output under these conditions.
Next, a flow of flight data generation processing by the flight data generation unit 202 in
In step S41, the flight data generation unit 202 acquires the captured moving image searched by the search unit 201.
In step S42, the imaging feature extraction unit 241 extracts the imaging feature parameters from the captured moving image from the search unit 201.
In step S43, the flight data generation model 242 generates flight control data by applying the imaging feature parameters extracted by the imaging feature extraction unit 241 to the imaging intention/imaging environment input by the user.
The flight control data generated in advance in this manner controls the flight of the UAV 204 or controls imaging from the camera mounted on the UAV 204.
Note that the flight of the UAV 204, the actual flight data and the moving image obtained by imaging from the UAV 204, and the imaging intention and the imaging environment input by the user can be used for learning of the imaging feature extraction unit 241 and the flight data generation model 242.
According to the above processing, the imaging feature parameter that is resolvable to a person is extracted from the captured moving image, and the imaging feature parameter is applied to the imaging intention and the imaging environment of the user, whereby the flight control data is generated. As a result, it is possible to easily realize the intended flight only by adjusting the flight and imaging features that can be interpreted by the user, and eventually, it is possible to more easily capture a moving image desired by the user.
In the moving body control system of
The flight data generation model 251 generates flight control data adapted to the imaging intention and the imaging environment input by the user on the basis of the captured moving image from the search unit 201.
Here, in the client terminal 110, for example, a text indicating an imaging intention or an imaging environment is input or the imaging intention or the imaging environment is selected using the GUI by the user's operation, whereby the flight control data is generated.
The flight data generation model 251 includes a using the moving image, the imaging intention, and the imaging environment as inputs and using the actual flight data obtained by the UAV actually flying as an output under these conditions.
For analysis of flight data based on a moving image, for example, a technique of estimating a trajectory of a camera position from a moving image by structure from motion (SEM) can be applied.
Next, a flow of flight data generation processing by the flight data generation unit 202 in
In step S51, the flight data generation unit 202 acquires the captured moving image searched by the search unit 201.
In step S52, the flight data generation model 251 generates flight control data adapted to the imaging intention and the imaging environment input by the user on the basis of the captured moving image from the search unit 201.
The flight control data generated in advance in this manner controls the flight of the UAV 204 or controls imaging from the camera mounted on the UAV 204.
Note that the flight of the UAV 204, the actual flight data and the moving image obtained by imaging from the UAV 204, and the imaging intention and the imaging environment input by the user can be used for learning of the flight data generation model 251.
According to the above processing, the flight control data adapted to the imaging intention and the imaging environment of the user is generated on the basis of the captured moving image. As a result, it is possible to easily realize the intended flight only by the user inputting his/her own imaging intention or an imaging environment in which the UAV 204 is desired to fly, and eventually, it is possible to more easily capture a moving image desired by the user.
In the moving body control system of
The flight data generation model 261 generates flight control data by applying the imaging feature parameter designated by the user to the imaging intention or the imaging environment input by the user.
Here, in the client terminal 110, for example, a text indicating an imaging intention or an imaging environment is input or the imaging intention or the imaging environment is selected using the GUI by the user's operation, whereby the flight control data is generated.
The flight data generation model 261 includes a machine learning model obtained by performing learning using the imaging feature parameter, the imaging intention, and the imaging environment as inputs and using the actual flight data obtained by the UAV actually flying as an output under these conditions.
According to the above configuration, it is possible to generate the flight control data adapted to the user's imaging intention and imaging environment without preparing the original moving image. In particular, in this case, the user can enjoy the generation of the flight control data in a more exploratory manner, such as finding an unexpected motion of the UAV.
The above-described embodiments can be applied to a case of creating a flight plan (scenario) from the start to the end of flight by generating flight control data in advance and repeating confirmation and correction. However, with the flight plan prepared in advance, it is not possible to cope with imaging for a long time exceeding a schedule, occurrence of an unexpected situation, and the like.
Therefore, in the following, embodiments in which the imaging can be continued as long as the situation continues by sequentially executing the flight control of the UAV will be described.
The moving body control system in
The search unit 301 can be realized by the client terminal 110 in
The search unit 301 searches for a captured moving image captured by another UAV according to the user's operation, and supplies the searched captured moving image to the imaging feature extraction unit 302.
The imaging feature extraction unit 302 can be realized by the server 130 in
The imaging feature extraction unit 302 directly extracts imaging feature parameters from the captured moving image from the search unit 301, and supplies the imaging feature parameters to the flight control unit 303. As described above, the imaging feature parameter is a human-resolvable parameter representing the imaging feature of the moving image.
The flight control unit 303, the drive unit 304, and the camera 305 can be realized by the moving body 170 (UAV) in
The flight control unit 303 controls the flight of the UAV and imaging from the UAV by controlling the drive unit 304 corresponding to the drive unit 175 in
The flight data generation model 311 sequentially generates flight control data by applying the imaging feature parameters extracted by the imaging feature extraction unit 302 to the imaging intention and the imaging environment input by the user. At this time, the imaging feature parameters extracted by the imaging feature extraction unit 302 may be adjusted by the user and then input to the flight data generation model 311.
Here, in the client terminal 110 or the controller 150, for example, the imaging feature is adjusted using the GUI by the user's operation, whereby the flight control data is generated. Note that the imaging feature parameters may be installed in the UAV in advance together with the flight data generation model 311.
The flight data generation model 311 includes a machine learning model obtained by performing learning using the imaging feature parameter, the imaging intention, and the imaging environment as inputs and using the actual flight data obtained by the UAV actually flying as an output under these conditions.
Next, a flow of flight control processing by the flight control unit 303 in
In step S111, the flight control unit 303 acquires the imaging feature parameters from the server 130.
In step S112, the flight data generation model 311 generates flight control data by applying the acquired imaging feature parameters to the imaging intention/imaging environment input by the user.
The flight control data sequentially generated in this manner controls the flight of the UAV or controls imaging from the camera mounted on the UAV.
According to the above processing, the imaging feature parameter that is extracted from the captured moving image and is resolvable to the person is applied to the imaging intention and the imaging environment of the user, so that the flight control data is sequentially generated and the flight is controlled. As a result, it is possible to easily realize the intended flight only by adjusting the flight and imaging features that can be interpreted by the user, and eventually, it is possible to more easily capture a moving image desired by the user.
In the moving body control system of
The search unit 301 searches for the captured moving image captured from another UAV according to the user's operation, and supplies the searched captured moving image to the flight control unit 303 via the server 130.
The flight data generation model 321 generates flight control data adapted to the imaging intention and the imaging environment input by the user on the basis of the captured moving image from the search unit 301.
Here, in the client terminal 110 or the controller 150, for example, a text indicating an imaging intention or an imaging environment is input or the imaging intention or the imaging environment is selected using the GUI by the user's operation, whereby the flight control data is generated.
The flight data generation model 321 includes a machine learning model obtained by performing learning using the moving image, the imaging intention, and the imaging environment as inputs and using the actual flight data obtained by the UAV actually flying as an output under these conditions.
Next, a flow of flight control processing by the flight control unit 303 in
In step S121, the flight control unit 303 acquires the captured image from the search unit 301.
In step S122, the flight data generation model 321 generates flight control data adapted to the imaging intention and the imaging environment input by the user on the basis of the captured moving image from the search unit 201.
The flight control data sequentially generated in this manner controls the flight of the UAV or controls imaging from the camera mounted on the UAV.
According to the above processing, the flight control data adapted to the imaging intention and the imaging environment of the user is sequentially generated on the basis of the captured moving image, and the flight is controlled. As a result, it is possible to easily realize the intended flight only by the user inputting his/her own imaging intention or an imaging environment in which the UAV is desired to fly, and eventually, it is possible to more easily capture a moving image desired by the user.
Although the embodiments in which the flight control data is generated in advance (the flight plan is created in advance) and the embodiments in which the sequential flight control data is generated (the sequential execution of the flight control) have been described above, these may be combined.
For example, a via-point (coordinates) on a flight route is manually designated in advance by creating a flight plan in advance, and between the via-points, flight reflecting an imaging intention can be realized by sequentially executing flight control.
In addition, by setting a flight route that loops the via-point by creating a flight plan in advance, it is possible to realize a flight in which an imaging intention is reflected by sequentially executing flight control while realizing an intended flight.
Furthermore, flight control data generated for a certain environment may be cut and pasted by geometric editing by creating a flight plan in advance, and the same flight may be repeated for a long time by performing partial reverse reproduction, end point connection, and the like.
Hereinafter, cases where the technology according to the present disclosure can be applied will be described.
The upper part of
The lower part of
The captured moving image V510 and the moving image V520 have a common composition in which the notable subject is kept within the angle of view until the moving body passes near the notable subject and route in which the moving body descends while approaching the notable subject from a distance and passes near the notable subject. This can be realized, for example, by the imaging feature parameter directly or indirectly extracted from the captured moving image V510.
Meanwhile, the captured moving image V510 and the moving image V520 are different from each other in the size (height) of the notable subject, and accordingly, the scale of the route is different, so that the distance and the flight speed of the entire route are different. This can be realized, for example, by applying the imaging feature parameter directly or indirectly extracted from the captured moving image V510 to the imaging intention of the user.
At this time, the target object to be the notable subject is selected in the captured moving image V510, and the target object (gaze point wp) to be the notable subject in the moving image V520 needs to be designated by the user at the start of capturing the moving image V520.
The case of
Similarly to
The lower part of
The captured moving image V510 and the moving image captured by the flight along the route R530 have a common composition in which the notable subject is kept within the angle of view until the moving body passes near the notable subject and route in which the moving body descends while approaching the notable subject from a distance and passes near the notable subject. This can be realized, for example, by the imaging feature parameter directly or indirectly extracted from the captured moving image V510.
Meanwhile, the captured moving image V510 and the moving image captured by the flight along the route R530 are different from each other in the size (height) of the notable subject and the scale of the route, so that the distance and the flight speed of the entire route are different from each other. Furthermore, since the route R530 is an indoor flight, a contract is made in the structure of the environment. This can be realized, for example, by applying the imaging feature parameters directly or indirectly extracted from the captured moving image V510 to the user's imaging intention and imaging environment.
The case of
The upper part of
In the lower part of
The captured moving image V610 and the moving image V620 have a common composition of a route (routes R610 and R620) for turning around the notable subject and a composition for keeping the notable subject within the angle of view while turning. This can be realized, for example, by the imaging feature parameter directly or indirectly extracted from the captured moving image V610.
Meanwhile, the captured moving image V610 and the moving image V620 are different from each other in the size of the notable subject, and accordingly, the scale of the route is different, so that the distance and the flight speed of the entire route are different. Furthermore, since the route R620 is a flight avoiding other trees, a contract is made in the structure of the environment. This can be realized, for example, by applying the imaging feature parameters directly or indirectly extracted from the captured moving image V610 to the user's imaging intention and imaging environment. Here, the route R620 may be changed according to the structure of the environment (interval between trees).
The case of
The left part of
The right part of
The captured moving image V710 and the moving image V720 are common in the position of the light source (sun) in the screen (the point that the sun is located in the upper part of the screen) and the scale feeling, the position, and the posture of the person who is the notable subject in the screen (the point that the person moves sideways at the bottom of the screen). Furthermore, in the captured moving image V710 and the moving image V720, the positional relationship between the ground and the camera (point at which the UAV is flying at a low altitude near the ground) and the ratio of the background on the screen (point at which the upper half of the screen is the background) are common. This can be realized, for example, by the imaging feature parameter directly or indirectly extracted from the captured moving image V710.
Meanwhile, the moving direction (up or down the slope) of the person who is the notable subject, the season and place of the environment, the state of the background, and the presence or absence of an obstacle are different between the captured moving image V710 and the moving image V720. Furthermore, the number of notable subjects may be different between the captured moving image V710 and the moving image V720. This can be realized, for example, by applying the imaging feature parameters directly or indirectly extracted from the captured moving image V710 to the user's imaging intention and imaging environment.
The case of
The left part of
On the right side of
The image capturing on the route R810 and the image capturing on the route R820 are common in the appearance (point of bust shot) of the person as the notable subject in the moving image. This can be realized, for example, by the imaging feature parameter directly or indirectly extracted from the captured moving image on the route R810.
Meanwhile, since there is the obstacle B20 in the imaging on the route R820, the route R820 is a route that avoids the obstacle B20 unlike the route R820. This can be realized, for example, by applying the imaging feature parameters directly or indirectly extracted from the captured moving image V810 to the user's imaging intention and imaging environment.
In particular, since the obstacle B20 is near the front of the person, the imaging on the route R820 is performed to achieve the imaging intention to cleanly capture the upper body of the person.
Note that, as illustrated in
As described above, in a case where an obstacle exists in the environment, not only the obstacle is simply avoided, but also a route for avoiding the obstacle while achieving the imaging intention can be set.
The cases of
An example of the imaging feature parameter applied to the imaging intention of the user will be described. Here, the imaging feature parameter and the imaging intention are assumed to be synonymous. The imaging intention includes an imaging intention that does not consider time (is constant regardless of the lapse of time) and an imaging intention that considers time (can change depending on a time result).
It is a composition when the subject in front of the camera is captured in the screen, and examples thereof include the composition of the disc of the sun in which the subject is captured in the center of the screen. The coordinate position of the center of the subject in the screen is set as a parameter. In which region of the regions divided by the three-division method the center of the subject is located may be set as a parameter.
For example, as illustrated in
An example of 90°<θ<270° is illustrated on the left side of
Instead of the angle θ, the user may specify how light such as front lighting or back lighting hits and the position (right side or left side) of the light source LS in the screen.
As the scale feeling of a person, for example, how much of a human skeleton is illustrated is set as a parameter. Specifically, in a case where only the face is illustrated, the face-up imaging is performed, and in a case where the chest or the waist is illustrated from the face, the bust-up imaging is performed. In addition, in a case where the face to the toe is reflected, the whole-body imaging is performed.
As illustrated in
The ratio of the background on the screen is obtained as a ratio α of a region deeper than the subject extracted by determining the depth of the subject using the monocular depth estimation technique to the entire screen (entire pixels). The ratio α is set as a parameter.
An imaging intention whether the upper body of the person is to be captured clearly, that is, whether the composition is a tilt composition or a bird's eye composition is set. Specifically, the posture β of the camera CAM with respect to the subject SB illustrated in
As the orientation of the camera, an angle φ1 in the horizontal direction and an angle φ2 in the vertical direction in the three-dimensional coordinate system may be set. Instead of the angles φ1 and φ2, an angle to look up, an angle to look down, a position directly below, or the like may be designated by the user. The orientation of the camera can also be indirectly set by other parameters.
As illustrated in
As the shake of the entire screen, the frequency ν (Hz) and the shake width s (pixel) may be set by using a motion estimation technology by dense optical flow. Instead of the frequency ν and the shake width s, a preset shake such as walking, running, train, car, or off-road car may be prepared and specified by the user.
A zoom ratio is set as a parameter. The zoom ratio can be defined by, for example, a change in a focal length x (mm) in terms of 35 mm per second by ±y (mm), a relative change in a size of a subject from a moving image at the start of imaging by z % per second, or the like.
The maximum angular velocity and the maximum angular acceleration of the camera in panning and tilting, and the maximum speed and the maximum acceleration of the camera in translational movement are set as how much the camera follows the movement of the subject such as the jump and the start of running.
Whether to avoid a moving obstacle so as not to appear or to allow the obstacle to be hidden by the subject (occlusion) may be selectively set. For example, after a route with minimum occlusion is calculated, a prohibited region around the obstacle is set, and the route is distorted in a direction to avoid the obstacle in accordance with the set prohibited region, whereby a route in which the obstacle does not appear can be set.
For example, slow camerawork may be set in a scene with high importance, and quick camerawork may be set in a scene with low importance.
For example, in a case where a person who is a subject dances in accordance with music, camerawork and movement may be set to be controlled in accordance with rhythm of the music.
As a change in the positional relationship with the subject, a stereoscopic camerawork in a three-dimensional space may be set.
For example, in a scene where an image is captured so as to go around the subject, left and right turning, ascending/descending, angular velocity, and the like of the camera (moving body) are set. In a scene where an image is captured so as to cross in front of a subject, left and right turning, ascending/descending, speed, and the like of the camera (moving body) are set. In a scene in which an image is captured while approaching or moving away from a subject, a closest distance between the camera (moving body) and the subject, ascending/descending, angular velocity, and the like are set.
In a scene where an image is captured while following the movement of the subject or running in parallel, the distance r between the camera CAM and the subject SB and the angle θ of the moving direction of the subject SB are set as the positional relationship with the subject as illustrated in
The series of processing steps described above can be executed by hardware and also can be executed by software. In a case where the series of processing is executed by software, a program constituting the software is installed in a computer. Here, examples of the computer include a computer incorporated in dedicated hardware, and for example, a general-purpose personal computer capable of executing various functions by installing various programs or the like.
In the computer, a CPU 901, a read only memory (ROM) 902, and a random access memory (RAM) 903 are connected to one another through a bus 904.
An input/output interface 905 is further connected to the bus 904. An input unit 906, an output unit 907, a storage unit 908, a communication unit 909, and a drive 910 are connected to the input/output interface 905.
The input unit 906 includes a keyboard, a mouse, a microphone, and the like. The output unit 907 includes a display, a speaker and the like. The storage unit 908 includes a hard disk, a non-volatile memory and the like. The communication unit 909 includes a network interface and the like. The drive 910 drives a removable medium 911 such as a magnetic disk, an optical disk, a magneto-optical disk, or a semiconductor memory.
In the computer configured in the above described manner, the CPU 901 loads the program stored in the storage unit 908, for example, on the RAM 903 through the input/output interface 905 and the bus 904 to execute, and according to this, the above-described series of processing is performed.
The program executed by the computer (CPU 901) can be provided, for example, by being recorded in the removable medium 911 as a package medium and the like. Further, the program can be provided via a wired or wireless transmission medium such as a local area network, the Internet, or digital satellite broadcasting.
In the computer, the program can be installed in the storage unit 908 via the input/output interface 905 by mounting the removable medium 911 on the drive 910. Furthermore, the program may be received by the communication unit 909 via a wired or wireless transmission medium to be installed on the storage unit 908. In addition, the program may be installed in advance on the ROM 902 and the storage unit 908.
Note that the program executed by the computer may be a program that performs processing in a time series according to an order described in the present specification, or may be a program that performs processing in parallel or at necessary timing such as when a call is made.
The embodiments of the present disclosure are not limited to the above-described embodiments, and various modifications can be made without departing from the gist of the present disclosure.
The effects described in the present specification are merely examples and are not limited, and other effects may be provided.
Moreover, the technology according to the present disclosure can have the following configurations.
(1)
A data processing method in which
The data processing method according to (1), in which
The data processing method according to (2), in which
The data processing method according to (2), in which
The data processing method according to (4), in which
The data processing method according to (4) or (5), in which
The data processing method according to (6), in which
The data processing method according to any one of (4) to (7), in which
The data processing method according to any one of (4) to (7), in which
The data processing method according to any one of (4) to (7), in which
The data processing method according to any one of (4) to (7), in which
The data processing method according to (1), in which
The data processing method according to (12), in which
The data processing method according to (13), in which
The data processing method according to (13) or (14), in which
The data processing method according to any one of (13) to (15), in which
The data processing method according to any one of (13) to (15), in which
A data processing device including
A program for causing a computer to execute processing of
A moving body control system including:
| Number | Date | Country | Kind |
|---|---|---|---|
| 2022-053271 | Mar 2022 | JP | national |
| Filing Document | Filing Date | Country | Kind |
|---|---|---|---|
| PCT/JP2023/009499 | 3/13/2023 | WO |