The present invention is related to an information processing method, a non-transitory computer-readable storage medium, and an information processing device.
The deep learning technology has been producing excellent results in the field of computer vision or sound recognition. For example, counting the number of fish present in a fish preserve of a farm has been an issue. However, it is expected that the use of the deep learning method would result in remedying the issue.
The accuracy of the analysis based on deep learning is dependent on the training data. That is, in order to obtain the analysis result having a high degree of accuracy, a machine learning model needs to be trained in a large volume of training data covering a wide range of patterns. For example, in order to accurately estimate the information regarding a school of fish (for example, the number of fish) from a taken image in which the school of fish present in a fish preserve is captured using computer vision, a machine learning model needs to be trained in a large volume of training data covering the combinations of a wide range of conditions such as the fish species and the number of fish.
However, it is not an easy task to create a large volume of training data covering a wide range of patterns. For example, it is extremely difficult to create a large volume of training data by manually assigning correct-solution data to each image among an enormous number of images covering the combinations of a wide range of conditions.
In this way, if it is not possible to obtain a large volume of training data covering a wide range of patterns, then a machine learning model cannot be given a high-quality training. Hence, there are times when it becomes difficult to accurately estimate the information regarding a plurality of target subjects from the actual taken image.
In that regard, in the application concerned, it is an objective to provide an information processing method, a computer program product, and an information processing device that enable accurate estimation of the information regarding a plurality of target subjects.
It is an object of the present invention to at least partially solve the problems in the conventional technology.
According to one aspect of an embodiment, an information processing method according to the application concerned is implemented in a computer; and includes obtaining a two-dimensional simulation image that is formed when a plurality of target subjects present in a three-dimensional simulation space is captured by a virtual camera, and generating the simulation image that visually displays information indicating the degree of overlapping of the plurality of target subjects in the simulation image.
The above and other objects, features, advantages and technical and industrial significance of this invention will be better understood by reading the following detailed description of presently preferred embodiments of the invention, when considered in connection with the accompanying drawings.
An illustrative embodiment (hereinafter, called “embodiment”) of an information processing method, a computer program product, and an information processing device is described below in detail with reference to the accompanying drawings. However, the information processing method, the computer program product, and the information processing device according to the application concerned is not limited by the embodiment described below. In the embodiment described below, identical regions are referred to by the same reference numerals, and their explanation is not repeated.
In recent years, cultivation of fish has been attracting attention as a means for overcoming the global food problem. In order to supply high-quality fish by way of fish cultivation, it is important to precisely know the number of fish, which is deeply related to the feeding of the fish.
However, under the specific environment of the farm that is under water, sometimes it is difficult to utilize the overground IT technology. For that reason, conventionally, after using a net and manually scooping some of the fish for measuring purposes, the number of fish is visually counted. However, in this method, not only there is a large burden on the fish, but there is also the issue of lack of precision.
In that regard, in recent years, the focus is on a method in which the number of fish is automatically counted using computer vision from a taken image in which the school of fish present in a fish preserve is captured. As a specific example, a method is known in which a machine learning model for image recognition is trained to estimate the number of fish present in a school of fish as captured in a taken image. However, in order to estimate the number of fish from the taken image using a machine learning model, the machine learning model needs to be trained in a large volume of training data covering the combinations of a wide range of conditions such as the surrounding environment of the fish (for example, the season, the time, the weather, the water temperature, the light intensity, and the extent of water pollution) and the biological characteristics of the fish (for example, the fish species, the fish size, the speed, and the number of fish).
However, it is not an easy task to create a large volume of training data covering the combinations of a wide range of conditions. For example, it is extremely difficult to create a large volume of training data by manually assigning correct-solution data to each image among an enormous number of images covering the combinations of a wide range of conditions. Besides, in order to accurately estimate the number of fish from a taken image using a machine learning model, the quality of the training data also plays an important role. However, in the case of manually assigning the correct-solution data, it cannot always be said that the training data is of a high quality.
For example, in an image in which the school of fish swimming in a three-dimensional fish preserve is captured, sometimes the image includes a region in which there is overlapping of the fish present on the near side of the camera (hereinafter, also called the near-side fish) and the fish present on the far side (hereinafter, also called the far-side fish). Thus, in a taken image that includes a region in which there is overlapping of the near-side fish and the far-side fish, some part of the bodies of the far-side fish remains hidden, thereby making it difficult to visually identify the positions of the fish captured in the taken image. For that reason, with respect to a taken image that includes a region in which there is overlapping of the near-side fish and the far-side fish, for example, if an annotation task for enclosing the positions of the fish in a bounding box is to be performed manually, then sometimes it is difficult to assign annotations indicating the precise positions of the fish. Besides, since it is difficult to visually identify the positions of the fish captured in a taken image, the annotation result happens to be different according to the person who performed the annotation task.
In contrast, the information processing device according to the embodiment generates a simulation image that is meant for visually displaying the information indicating the degree of overlapping of a plurality of fish in a two-dimensional simulation image, which is formed when a plurality of fish present in a three-dimensional simulation space is captured by a virtual camera (see
In this way, in the information processing device according to the embodiment, instead of using a taken image in which a school of fish actually present in a three-dimensional space (for example, a fish preserve) is captured, a simulation image is used that is formed when the school of fish present in a three-dimensional simulation space is captured by a virtual camera. As a result, the information processing device becomes able to automatically calculate the information indicating the degree of overlapping of a plurality of fish included in the simulation image. Moreover, the information processing device can assign, as the annotation to the simulation image, the information indicating the degree of overlapping of a plurality of fish. Hence, accurate and high-quality training data can be generated with ease. With that, for example, the information processing device enables training of a machine learning model, which is meant for estimating the information regarding the school of fish (for example, the number of fish) from the taken image, in a large volume of high-quality training data. That is, as a result of carrying out the training of a machine learning model, which is meant for estimating the information regarding the school of fish (for example, the number of fish) from the taken image, in a large volume of high-quality training data, the information processing device enables achieving enhancement in the estimation accuracy of the machine learning model. Thus, the information processing device enables accurate estimation of the information regarding the school of fish.
Communication Unit 110
The communication unit 110 is implemented using, for example, a network interface card (NIC). The communication unit 110 is connected to a network in a wired manner or a wireless manner, and sends information to and receives information from a terminal device used by a supervisor of the fish.
Memory Unit 120
The memory unit 120 is implemented using, for example, a semiconductor memory device such as a random access memory (RAM) or a flash memory; or a memory device such as a hard disk or an optical disc. More particularly, the memory unit 120 is used to store various computer programs (an example of an information processing program) to be used in simulation. Moreover, the memory unit 120 is used to store the values of various parameters to be used in simulation.
Thus, based on the information regarding the biological characteristics of the fish, based on the information regarding the surrounding environment of the fish, and based on the characteristics of the fish school behavior exhibited by the fish; the information processing device 100 generates a simulation image that includes the behavior exhibited by a school of fish. That enables accurate reproduction of the behavior exhibited by the school of fish present in an actual fish preserve. Moreover, the information processing device 100 varies the values of a plurality of parameters used in generating simulation images, and thus can easily generate a large number of simulation images covering a wide range of patterns. Furthermore, the information processing device 100 can utilize the values of the parameters, which are used in generating simulation images, as the correct-solution data. Hence, as compared to the case of manually assigning the correct-solution data, a large volume of high-quality data can be generated with ease.
As a result, for example, the information processing device 100 enables training of a machine learning model, which is meant for estimating the information regarding the school of fish (for example, the number of fish) captured in a taken image, in a large volume of high-quality training data. That is, as a result of carrying out the training of a machine learning model, which is meant for estimating the information regarding the school of fish (for example, the number of fish) from the taken image, in a large volume of high-quality training data; the information processing device 100 enables achieving enhancement in the estimation accuracy of the machine learning model. Thus, the information processing device 100 enables accurate estimation of the information regarding the school of fish.
Input Unit 130
The input unit 130 receives input of various operations from the user. For example, the input unit 130 can receive various operations from the user according to touch-sensitive panel functions and via a display screen (for example, the output unit 140). Alternatively, the input unit 130 can receive various operations from buttons provided in the information processing device 100 or from a keyboard or a mouse connected to the information processing device 100.
Output Unit 140
The output unit 140 is a display screen implemented using, for example, a liquid crystal display or an organic electro-luminescence (EL) display, and represents a display device meant for displaying a variety of information. The output unit 140 displays a variety of information under the control of the control unit 150. Meanwhile, if a touch-sensitive panel is used in the information processing device 100, then the input unit 130 and the output unit 140 are integrated together. In the following explanation, the output unit 140 is sometimes referred to as the screen.
Control Unit 150
The control unit 150 is a controller and is implemented when, for example, a central processing unit (CPU) or a micro processing unit (MPU) executes various computer programs (equivalent to an example of an information processing program), which are stored in an internal memory device of the information processing device 100, while using a read only memory (RAM) as the work area. Alternatively, the control unit 150 is a controller and is implemented using, for example, an integrated circuit such as an application specific integrated circuit (ASIC) or a field programmable gate array (FPGA).
The control unit 150 includes, as function units, an obtaining unit 151, a generating unit 152, an estimating unit 153, and an output control unit 154 that can implement or execute the operations in the information processing explained later. Meanwhile, the internal configuration of the control unit 150 is not limited to the configuration illustrated in
Obtaining Unit 151
The obtaining unit 151 obtains the following: the values of the internal parameters regarding the biological characteristics of the fish; the values of the external parameters regarding the characteristics of the surrounding environment of the fish; and the values of fish school parameters regarding the characteristics of the fish school behavior indicating the behavior exhibited by the fish toward each other. More particularly, the obtaining unit 151 obtains the values of the internal parameters, the values of the external parameters, and the values of the fish school parameters as input by the user via the input unit 130.
More particularly, the obtaining unit 151 obtains the values of the internal parameters that are set for every fish belonging to the school of fish included in a simulation image. For example, when there are N number of fish (where N is a natural number) in the school of fish included in a simulation image, the obtaining unit 151 obtains N number of sets of values of internal parameters set for the N number of fish.
Moreover, the obtaining unit 151 obtains the values of the fish school parameters set for the school of fish included in a simulation image. The values of the fish school parameters are set in common to every fish in a school of fish. For example, when there are N number of fish (where N is a natural number) in the school of fish included in a simulation image, the obtaining unit 151 obtains the values of the fish school parameters set in common for all of the N number of fish.
Furthermore, the obtaining unit 151 obtains the values of the external parameters set for each type of surrounding environment of the school of fish included in a simulation image. For example, the obtaining unit 151 obtains the values of the external parameter that are set for the fish preserve in which the school of fish included in a simulation image is present. For example, the obtaining unit 151 obtains the values of the external parameters that are set to be different according to the depth of water in the fish preserve in which the school of fish included in a simulation image is present.
Generating Unit 152
The generating unit 152 controls, based on the values of the internal parameters, the values of the external parameters, and the values of the fish school parameters as obtained by the obtaining unit 151, the behavior exhibited by each individual fish belonging to the school of fish; and generates a simulation image. Explained below in detail with reference to
It is known that the distribution of water temperature in a fish preserve differs according to the season (or the tide of the season) and the depth of the water. In the example illustrated in the upper part in
Meanwhile, as a biological characteristic of fish, it is known that, higher the surrounding water temperature, higher the locomotion speed of the fish; and that, lower the surrounding water temperature, lower is the locomotion speed of the fish. Thus, as the value of an internal parameter, based on the distribution of water temperature of the fish preserve in which the school of fish is present, the generating unit 152 controls the behavior of each individual fish belonging to the school of fish in such a way that that the locomotion speed of that fish increases as the surrounding water temperature becomes higher.
It is known that the distribution of light intensity in a fish preserve differs according to the season, the time, the weather, and the water depth. In the example illustrated in the upper part in
Moreover, as illustrated in
Herein, the fish school behavior exhibited by a fish with respect to other fish implies the behavior following three rules, namely, cohesion, separation, and alignment. The “cohesion” indicates the force by which the fish attempt to remain close to each other in order to not get separated (hereinafter, called the cohesion force). The “separation” indicates the force by which the fish attempt to move away from each other in order to avoid collision (hereinafter, called the separation force). The “alignment” indicates the force for attempting to get lined up to move as a single school of fish (hereinafter, called the alignment force). The behavior exhibited by a fish gets decided by the sum total of these three forces. In the following explanation, the three forces, namely, the cohesion force, the separation force, and the alignment force are collectively referred to as the “fish school parameters”. Each of the three forces is decided by the values of parameters indicating the magnitude of the force, the direction of the force, and the range over which the force is exerted. For example, the fish school parameters are decided by the size of the space in which the school of fish is present (for example, the size of the fish preserve), the size of the fish, and the density of the school of fish. In this way, as the values of the fish school parameters, based on the cohesion force by which the fish attempt to remain close to each other in order to not get separated, based on the separation force by which the fish attempt to move away from each other in order to avoid collision, and based on the alignment force for attempting to get lined up to move as a single school of fish; the generating unit 152 controls the behavior of every fish.
Moreover, as a biological characteristic of fish, it is known that, lower (higher) the water temperature surrounding a fish (or body temperature of a fish), the slower (faster) is the reaction speed of the Meissner's cell of the fish. Thus, as the value of an internal parameter, based on the decision-making duration that increases as the water temperature surrounding a fish becomes lower and based on the time interval between decision making instances, the generating unit 152 can control the behavior of the fish.
In
For example, the generating unit 152 changes the swimming mode of the fish from the “Slow Swim” state to the “C-start Left” state. Then, the generating unit 152 changes the swimming mode from the “C-start Left” state to a “Fast Swim” state. Subsequently, the generating unit 152 changes the swimming mode from the “Fast Swim” state to a “Semi Fast Swim” state. Then, the generating unit 152 changes the swimming mode from the “Semi Fast Swim” state to the “Slow Swim” state. Subsequently, the generating unit 152 changes the swimming mode of the fish from the “Slow Swim” state to a “Rest Swim” state. Meanwhile, at the time of changing the state, the generating unit 152 can perform interpolation between the pre-change state and the post-change state, so that the animation can be smooth.
Moreover, based on the values of the internal parameters, the values of the external parameters, and the values of the fish school parameters as obtained by the obtaining unit 151, the generating unit 152 controls the behavior of the three-dimensional CG of the fish in such a way that the realistic three-dimensional CG of the fish (see
Upon generating the simulation image, the generating unit 152 can generate training data in which two-dimensional bounding boxes are added that indicate the position information of the fish included in the generated simulation image. Herein, as a result of using the position information of every fish that was used at the time of generating a simulation image, the generating unit 152 can easily generate training data in which two-dimensional bounding boxes are added.
Moreover, upon generating the simulation image, the generating unit 152 can generate training data in which three-dimensional bounding boxes are added that indicate the position information of the fish included in the generated simulation image. Herein, as a result of using the position information of every fish that was used at the time of generating a simulation image, the generating unit 152 can easily generate training data in which three-dimensional bounding boxes are added.
Furthermore, upon generating the simulation image, the generating unit 152 can generate training data in which the fish included in the generated simulation image are changed to silhouette. For example, the generating unit 152 generates training data in which the fish that are changed to silhouette are displayed in a color-coded manner. Herein, the color coding is equivalent to assigning color elements. For example, when a plurality of fish is present, the generating unit 152 can assign a different color for the color coding of every fish, or can assign different colors according to the classification information of the fish.
In the example explained above, the generating unit 152 generates a simulation image, which includes the behavior exhibited by every fish belonging to a school of fish, based on the values of the internal parameters related to the biological characteristics of the fish, based on the values of the external parameters related to the characteristics of the surrounding environment of the fish, and based on the values of the fish school parameters related to the characteristics of the fish school behavior indicating the behavior exhibited by a fish with respect to other fish. However, the types of the parameters are not limited to the types explained above. Explained below with reference to
The generating unit 152 generates a simulation image in which the degree of overlapping of a plurality of fish in the two-dimensional simulation image, which is obtained when the fish present in the three-dimensional simulation space are captured by a virtual camera, is visually displayed. More particularly, based on the simulation image obtained by the obtaining unit 151, the generating unit 152 determines whether or not any of the fish included in the simulation image are apparently overlapping. For example, based on the position coordinates of every fish present in the three-dimensional simulation space and based on the position coordinates of the virtual camera, the generating unit 152 determines whether or not any fish are apparently overlapping in the image when viewed from the position of the virtual camera. With reference to
Then, based on the position coordinates of every fish present in the three-dimensional simulation space and based on the position coordinates of the virtual camera, the generating unit 152 calculates the information indicating the degree of overlapping of a plurality of fish included in the simulation image. For example, as the information indicating the degree of overlapping, the generating unit 152 calculates the percentage of the area of the viewable region of every fish. With reference to
When the information indicating the degree of overlapping is calculated, the generating unit 152 visualizes the information indicating the degree of overlapping, and adds it to the simulation image. For example, the generating unit 152 displays, at a position within a predetermined range from every fish, a label indicating the percentage of the area of the viewable region of that fish. With reference to
For example, when the generated simulation image is input to a machine learning model M1, the generating unit 152 trains the machine learning model M1 to output a simulation image in which three-dimensional bounding boxes each of which indicates the position information of an identified fish are displayed in a superimposed manner, and to output the number of identified fish (the number of fish present in the school of fish).
Moreover, when the generated simulation image is input to a machine learning model M2, the generating unit 152 trains the machine learning model M2 to output a simulation image in which two-dimensional bounding boxes each of which indicates the size of an identified fish are displayed in a superimposed manner, and to output the number of identified fish (the number of fish present in the school of fish).
Furthermore, when the generated simulation image is input to a machine learning model M3, the generating unit 152 trains the machine learning model M3 to output a simulation image in which every identified fish is changed to silhouette, and to output the number of identified fish (the number of fish present in the school of fish).
When the generated simulation image is input to each of the machine learning models M1 to M3, the generating unit 152 trains each of the machine learning models M1 to M3 to output a simulation image in which the information obtained by visualizing the information indicating the degree of overlapping of a plurality of fish in the simulation image is displayed in a superimposed manner. For example, when the generated simulation image is input to each of the machine learning models M1 to M3, the generating unit 152 trains each of the machine learning models M1 to M3 to output a simulation image in which labels each of which indicates the percentage of the area of the viewable region of one of a plurality of fish in the simulation image are displayed in a superimposed manner.
In this way, when a simulation image is input to a machine learning model, the generating unit 152 trains the machine learning model to output correct-solution data generated based on the parameter information used in generating the simulation image or to output the information corresponding to the correct-solution data.
Estimating Unit 153
The estimating unit 153 uses the already-trained machine learning model generated by the generating unit 152, and estimates the information related to a plurality of target subjects from a taken image in which the target subjects are captured. For example, the estimating unit 153 uses the already-trained machine learning model generated by the generating unit 152, and estimates the information related to the school of fish from a taken image in which the school of fish is captured.
Moreover, the estimating unit 153 estimates, as the information related to a plurality of target subjects, the number of target subjects captured in the taken image. For example, the estimating unit 153 inputs the taken image in the already-trained machine learning model M2 generated by the generating unit 152. Then, the estimating unit 153 makes the machine learning model M2 output the following: the taken image in which two-dimensional bounding boxes are displayed in a superimposed manner along with the superimposed display of the labels each of which indicates the percentage of the area of the viewable region of a fish captured in the taken image; and the number of identified fish (the number of fish present in the school of fish). The estimating unit 153 treats the information output from the machine learning model M2 as the estimation result.
Output Control Unit 154
The output control unit 154 performs control to display the estimation result, which is obtained by the estimating unit 153, in a screen.
In the example illustrated in
In the example illustrated in
In the example illustrated in
Meanwhile, in the embodiment described above, it is a fish that represents the target subject included in a simulation image, and a plurality of fish present in a school of fish represents a plurality of target subjects. However, the target subjects included in a simulation image are not limited to the fish. Alternatively, for example, other than the fish, the target subjects included in a simulation image can be persons, animals, or vehicles. Moreover, the information processing device 100 can generate a simulation image that visually displays the information indicating the degree of overlapping of a plurality of target subjects (for example, a plurality of persons, a plurality of animals, or a plurality of vehicles) in a two-dimensional simulation image that is obtained when the target subjects present in a three-dimensional simulation space are captured by a virtual camera.
As explained above, the information processing device 100 according to the embodiment includes the obtaining unit 151 and the generating unit 152. The obtaining unit 151 obtains a two-dimensional simulation image when a plurality of target subjects present in a three-dimensional space is captured by a virtual camera. The generating unit 152 generates a simulation image for visually displaying the information that indicates the degree of overlapping of a plurality of target subjects in a two-dimensional simulation image, which is formed when the target subjects present in a three-dimensional simulation space is captured by a virtual camera.
In this way, instead of using a taken image in which a plurality of target objects present in a three-dimensional space is captured, the information processing device 100 uses a simulation image that is formed when a plurality of target subjects present in a three-dimensional space is captured by a virtual camera. As a result, the information processing device 100 becomes able to automatically calculate the information indicating the degree of overlapping of a plurality of target subjects included in the simulation image. Moreover, since the information indicating the degree of overlapping of a plurality of target subjects can be added as the annotation in the simulation image, the information processing device 100 can generate accurate and high-quality training data with ease. As a result, for example, the information processing device 100 enables training of a machine learning model, which is meant for estimating the information related to a plurality of target subjects (for example, the number of target subjects) captured in the taken image, in a large volume of high-quality training data. That is, as a result of carrying out the training using a large volume of high-quality training data, for example, the information processing device 100 becomes able to enhance the estimation accuracy of the machine learning model that is meant for estimating the information related to a plurality of target subjects (for example, the number of target subjects) captured in the taken image. Hence, the information processing device 100 enables accurate estimation of the information related to a plurality of target subjects.
Moreover, based on the position coordinates of each of a plurality of target subjects present in the three-dimensional simulation space and based on the position coordinates of the virtual camera, the generating unit 152 calculates the degree of overlapping of the target subjects in the simulation image.
Thus, the information processing device 100 can add, as the annotation in the simulation image, the information indicating the degree of overlapping of a plurality of target subjects; and hence can easily generate accurate and high-quality training data to be used in the training of a machine learning model meant for estimating the information related to the target subjects captured in a taken image (for example, information such as the number of target subjects).
When a simulation image is input to a machine learning model, the generating unit 152 trains the machine learning model to output the correct-solution data generated based on the parameter information used in generating simulation images or to output the information corresponding to the correct-solution data.
As a result, for example, the information processing device 100 enables training of a machine learning model, which is meant for estimating the information related to a plurality of target subjects (for example, the number of target subjects) captured in the taken image, in a large volume of high-quality training data. That is, as a result of carrying out the training using a large volume of high-quality training data, the information processing device 100 becomes able to, for example, enhance the estimation accuracy of the machine learning model that is meant for estimating the information related to a plurality of target subjects (for example, the number of target subjects) captured in the taken image.
The information processing device 100 further includes the estimating unit 153 that uses the already-trained machine learning model generated by the generating unit 152 and estimates the information related to a plurality of target subjects from a taken image in which the target subjects are captured.
Thus, as a result of carrying out the training using a large volume of high-quality training data, the information processing device 100 becomes able to, for example, enhance the estimation accuracy of the machine learning model that is meant for estimating the information related to a plurality of target subjects (for example, the number of target subjects) captured in the taken image. For example, the information processing device 100 becomes able to enhance the estimation accuracy of the machine learning model that is meant for estimating the information related to a school of fish (for example, the number of fish) captured in a taken image. Thus, the information processing device 100 can accurately estimate the information related to a plurality of target subjects.
Meanwhile, the estimating unit 153 estimates, as the information related to a plurality of target subjects, the number of target subjects captured in the taken image.
Thus, the information processing device 100 can accurately estimate the number of target subjects included in the taken image.
Herein, a target subject implies a fish, and a plurality of target subjects implies a plurality of fish present in a school of fish.
In this way, instead of using a taken image in which a school of fish actually present in a three-dimensional space (for example, a fish preserve) is captured, the information processing device 100 uses a simulation image that is formed when a school of fish present in a three-dimensional simulation space is captured by a virtual camera. As a result, the information processing device 100 becomes able to automatically calculate the information indicating the degree of overlapping of a plurality of fish included in the simulation image. Moreover, since the information indicating the degree of overlapping of a plurality of fish can be added as the annotation in the simulation image, the information processing device 100 can generate accurate and high-quality training data with ease. As a result, for example, the information processing device 100 enables training of a machine learning model, which is meant for estimating the information related to a plurality of target subjects (for example, the number of target subjects) captured in the taken image, in a large volume of high-quality training data. That is, as a result of carrying out the training using a large volume of high-quality training data, the information processing device 100 becomes able to, for example, enhance the estimation accuracy of the machine learning model that is meant for estimating the information related to a plurality of target subjects (for example, the number of target subjects) captured in the taken image. Hence, the information processing device 100 enables accurate estimation of the information related to the school of fish.
The obtaining unit 151 obtains the values of the internal parameters related to the biological characteristics of the fish; obtains the values of the external parameters related to the characteristics of the surrounding environment of the fish; obtains the values of the fish school parameters related to the characteristics of the fish school behavior indicating the behavior exhibited by a fish toward other fish; and obtains the values of the color parameters related to the characteristics of the color of the water in which the fish are present. Then, based on the values of the internal parameters, the values of the external parameters, the values of the fish school parameters, and the values of the color parameters as obtained by the obtaining unit 151; the generating unit 152 generates a simulation image that includes the behavior exhibited by every fish belonging to the school of fish.
In this way, based on the information related to the biological characteristics of the fish, based on the information related to the surrounding environment of the fish, based on the information related to the characteristics of the fish school behavior exhibited by the fish, and based on the information related to the characteristics of the color of the water in which the fish are present; the information processing device 100 generates a simulation image that includes the behavior exhibited by the school of fish. Hence, the behavior exhibited by the school of fish present in an actual fish preserve can be reproduced with accuracy. Moreover, the information processing device 100 varies the values of a plurality of parameters used in generating simulation images, and thus can easily generate a large number of simulation image covering a wide range of patterns. Furthermore, the information processing device 100 can utilize the values of the parameters, which are used in generating simulation images, as the correct-solution data. Hence, as compared to the case of manually assigning the correct-solution data, a large volume of high-quality training data can be generated with ease.
Moreover, based on the values of the internal parameters, the values of the external parameters, the values of the fish school parameters, and the values of the color parameters; the generating unit 152 controls the behavior exhibited by every fish belonging to the school of fish, and generates a simulation image.
Thus, based on the information related to the biological characteristics of the fish, based on the information related to the surrounding environment of the fish, based on the information related to the characteristics of the fish school behavior exhibited by the fish, and based on the information related to the characteristics of the color of the water in which the fish are present; the information processing device 100 controls the behavior exhibited by every fish belonging to the school of fish. As a result, the behavior exhibited by the school of fish present in the actual fish preserve can be reproduced with accuracy. As a result, for example, the information processing device 100 enables training of a machine learning model, which is meant for estimating the information related to a school of fish (for example, the number of fish) captured in the taken image, in a large volume of high-quality training data.
Meanwhile, as the value of a color parameter, based on the concentration of the chlorophyl in the water of the fish preserve in which the fish belonging to the fish school are present, the generating unit 152 generates a simulation image.
Thus, the information processing device 100 becomes able to easily generate a large volume of training data containing simulation images having different appearances according to the concentration of the chlorophyl in the water. Hence, the information processing device 100 enables training of a machine learning model using the training data containing the generated simulation images, thereby enabling achieving enhancement in the estimation accuracy of the machine learning model.
Moreover, as the value of a color parameter, based on the concentration of the floating matter in the water of the fish preserve in which the fish belonging to the fish school are present, the generating unit 152 generates a simulation image.
Thus, the information processing device 100 becomes able to easily generate a large volume of training data containing simulation images having different appearances according to the concentration of the floating matter in the water. Hence, the information processing device 100 enables training of a machine learning model using the training data containing the generated simulation images, thereby enabling achieving enhancement in the estimation accuracy of the machine learning model.
Furthermore, as the values of color parameters, based on the intensity of the sunlight shining on the water of the fish preserve in which the fish belonging to the fish school are present and based on the depth of the sunlight from the water surface of the fish preserve, the generating unit 152 generates a simulation image.
Thus, the information processing device 100 becomes able to easily generate a large volume of training data containing simulation images having different appearances according to the intensity of the sunlight shining on the water and the depth of the sunlight from the water surface of the fish preserve. Hence, the information processing device 100 enables training of a machine learning model using the training data containing the generated simulation images, thereby enabling achieving enhancement in the estimation accuracy of the machine learning model.
Moreover, as the value of a color parameter, based on the depth from the water surface of the fish preserve, in which the fish belonging to the fish school are present, to the position of the virtual camera, the generating unit 152 generates a simulation image.
Thus, the information processing device 100 becomes able to easily generate a large volume of training data containing simulation images having different appearances according to the depth from the water surface to the position of the virtual camera. Hence, the information processing device 100 enables training of a machine learning model using the training data containing the generated simulation images, thereby enabling achieving enhancement in the estimation accuracy of the machine learning model.
Meanwhile, when a simulation image is input to a machine learning model, the generating unit 152 trains the machine learning model to output the correct-solution data generated based on the parameter information used in generating the simulation image or to output the information corresponding to the correct-solution data.
As a result, for example, the information processing device 100 enables training of a machine learning model, which is meant for estimating the information regarding a school of fish (for example, the number of fish) captured in the taken image, in a large volume of high-quality training data. That is, as a result of carrying out the training of a machine learning model, which is meant for estimating the information regarding the school of fish (for example, the number of fish) from the taken image, in a large volume of high-quality training data, the information processing device 100 enables achieving enhancement in the estimation accuracy of the machine learning model.
The information processing device 100 further includes the estimating unit 153 that uses the already-trained machine learning model generated by the generating unit 152 and estimates the information related to the school of fish from a taken image in which the school of fish is captured.
Thus, as a result of carrying out the training using a large volume of high-quality training data, for example, the information processing device 100 becomes able to enhance the estimation accuracy of the machine learning model that is meant for estimating the information related to the school of fish (for example, the number of fish) captured in the taken image. Hence, using the already-trained machine learning model, the information processing device 100 can accurately estimate the information related to the school of fish. Moreover, for example, in the case of estimating the number of fish in the school of fish captured in the taken image, the information processing device 100 implements image recognition and hence becomes able to accurately estimate the information related to the school of fish without directly touching the fish (i.e., in a contactless manner). That enables achieving reduction in the burden on the fish. Moreover, as a result of implementing the image recognition, as compared to the case of manually scooping the fish using a net, the information processing device 100 can obtain a highly accurate estimation result in a short period of time.
The information processing device 100 according to the embodiment described above is implemented using, for example, a computer 1000 having a configuration as illustrated in
The CPU 1100 performs operations based on a computer program stored in the ROM 1300 or in the HDD 1400, and controls the other constituent elements. The ROM 1300 is used to store a boot program that is executed by the CPU 1100 at the time of booting of the computer 1000, and to store computer programs that are dependent on the hardware of the computer 1000.
The HDD 1400 is used to store the computer programs to be executed by the CPU 1100, and to store the data used in those computer programs. The communication interface 1500 receives data from other devices via a predetermined communication network and sends that data to the CPU 1100; as well as sends the data generated by the CPU 1100 to other devices via a predetermined communication network.
The CPU 1100 controls, via the input-output interface 1600, output devices, such as a display and a printer, and input devices, such as a keyboard and a mouse. The CPU 1100 obtains data from an input device via the input-output interface 1600. Moreover, the CPU 1100 outputs the generated data to an output device via the input-output interface 1600.
The media interface 1700 reads the computer programs and the data stored in the recording medium 1800, and provides them to the CPU 1100 via the RAM 1200. The CPU 1100 loads the computer programs from the recording medium 1800 in the RAM 1200 via the media interface 1700, and executes the programs. The recording medium 1800 is, for example, an optical recording medium such as a DVD (Digital Versatile Disc) or a PD (Phase change rewritable Disk); or a magneto-optical recording medium such as an MO (Magneto-Optical disk); or a tape medium; or a magnetic recording medium; or a semiconductor memory.
For example, when the computer 1000 functions as the information processing device 100 according to the embodiment, the CPU 1100 of the computer 1000 executes computer programs loaded in the RAM 1200 and implements the functions of the control unit 150. Herein, the CPU 1100 of the computer 1000 reads the computer programs from the recording medium 1800 and executes them. Alternatively, as another example, the CPU 1100 of the computer 1000 can obtain the computer programs from other devices via a predetermined communication network.
Till now, the embodiment of the application concerned was described in detail based on the accompanying drawings. However, that embodiment is only exemplary, and is to be construed as embodying all modifications and alternative constructions that may occur to one skilled in the art that fairly fall within the basic teaching herein set forth.
Of the processes described in the embodiment, all or part of the processes explained as being performed automatically can be performed manually. Similarly, all or part of the processes explained as being performed manually can be performed automatically by a known method. The processing procedures, the control procedures, specific names, various data, and information including parameters described in the embodiments or illustrated in the drawings can be changed as required unless otherwise specified. For example, a variety of information illustrated in the drawings is not limited to that information.
The constituent elements of the device illustrated in the drawings are merely conceptual, and need not be physically configured as illustrated. The constituent elements, as a whole or in part, can be separated or integrated either functionally or physically based on various types of loads or use conditions.
The embodiments and the modification examples can be appropriately combined without causing any contradictions in the operation details.
Moreover, the terms “section”, “module”, and “unit” mentioned above can be read as “device” or “circuit”. For example, a generating unit can be read as a generating device or a generating circuit.
According to an aspect of the embodiment, the information related to a plurality of target subjects can be estimated with accuracy.
Although the invention has been described with respect to specific embodiments for a complete and clear disclosure, the appended claims are not to be thus limited but are to be construed as embodying all modifications and alternative constructions that may occur to one skilled in the art that fairly fall within the basic teaching herein set forth.
The present application is a divisional of U.S. patent application Ser. No. 17/875,428, filed Jul. 28, 2022, the entire contents of which are incorporated herein by reference.
Number | Date | Country | |
---|---|---|---|
Parent | 17875428 | Jul 2022 | US |
Child | 18418388 | US |