The present invention relates to an image encoding method, an image decoding method, an image processing method, an image encoding device, and an image decoding device.
For example, as disclosed in Patent Literatures 1 and 2, a conventional image encoding system architecture includes a camera or a sensor that captures an image, an encoder that encodes the captured image to a bitstream, a decoder that decodes the image from the bitstream, and a display device that displays the image for human determination. Since the advent of machine learning or neural network-based applications, machines are rapidly replacing humans in determining images because machines outperform humans in scalability, efficiency, and accuracy.
Machines tend to work well only in situations where they are trained. If environment information partially changes on a camera side, the performance of the machines deteriorates, detection accuracy deteriorates, and thus poor determinations occur. In a case where environment information has been taught to machines, the machines can be customized to accommodate changes for achieving better detection accuracy.
An object of the present disclosure is to improve the accuracy of task processing.
An image encoding method according to one aspect of the present disclosure includes: by an image encoding device, encoding an image and generating a bitstream, adding, to the bitstream, one or more parameters that not used for encoding the image, transmitting, to an image decoding device, the bitstream to which the one or more parameters have been added, and outputting the image and the one or more parameters to a first processing device that executes predetermined task processing.
A problem of the above-described background art is that the encoder 3002 does not transmit information necessary for improving the accuracy of task processing to the decoder 3004. The encoder 3002 transmits this information to the decoder 3004, thus providing important data related to an environment of an application or the like that can be used for improving the accuracy of the task processing, from the decoder 3004 to the task processing unit 3005. This information may include the camera characteristics, the size of the object included in the image, or the depth of the object included in the image. The camera characteristics may include a mounting height of the camera, a tilt angle of the camera, a distance from the camera to a region of interest (ROI), a visual field of the camera, or any combination thereof. The size of the object may be calculated from the width and height of the object in the image, or may be estimated by executing a computer vision algorithm. The size of the object may be used to estimate the distance between the object and the camera. The depth of the object may be obtained by using a stereo camera or running the computer vision algorithm. The depth of the object may be used to estimate the distance between the object and the camera.
In order to solve the problems with the background art, the present inventor has introduced a new method for signalizing the camera characteristics, the size of an object contained in an image, the depth of the object contained in the image, or any combination thereof. The concept is to transmit important information to a neural network to make the neural network adaptable with an environment from which the image or characteristics are originated. One or more parameters indicating this important information are encoded together with the image or stored in a header of the bitstream, and are added to the bitstream. The header may be a video parameter set (VPS), a sequence parameter set (SPS), a picture parameter set (PPS), a picture header (PH), a slice header (SH), or a supplemental enhancement information (SEI). One or more parameters may be signalized in a system layer of the bitstream. What is important in this solution is that the transmitted information is intended to improve the accuracy of determination and the like in the task processing including the neural network.
In the task processing units 3105 and 3207, the information signalized as the one or more parameters can be used for changing a neural network model that is being used. For example, a complex or simple neural network model can be selected depending on the size of the object or the mounting height of the camera. The task processing may be executed by using the selected neural network model.
The information signalized as the one or more parameters can be used for changing parameters to be used for adjusting an estimated output from the neural network. For example, the signalized information may be used to set a detection threshold to be used for estimation. The task processing may be executed by using a new detection threshold for estimating the neural network.
The information signalized as the one or more parameters can be used for adjusting scaling of images to be input to the task processing units 3105 and 3207. For example, the signalized information is used for set the scaling size. The input images to the task processing units 3105 and 3207 are scaled to the set scaling size before the task processing units 3105 and 3207 execute the task processing.
Next, each aspect of the present disclosure will be described.
An image encoding method according to one aspect of the present disclosure includes: by an image encoding device, encoding an image to generate a bitstream, adding, to the bitstream, one or more parameters that are not used for encoding the image, transmitting, to an image decoding device, the bitstream to which the one or more parameters have been added, and outputting the image and the one or more parameters to a first processing device that executes predetermined task processing.
According to this aspect, the image encoding device transmits, to the image decoding device, the one or more parameters to be output to the first processing device for execution of the predetermined task processing. As a result, the image decoding device can output the one or more parameters received from the image encoding device to a second processing device that executes task processing which is same as the predetermined task processing. As a result, the second processing device executes the predetermined task processing based on the one or more parameters input from the image decoding device, thereby improving the accuracy of the task processing in the second processing device.
In the above aspect, the image decoding device receives the bitstream from the image encoding device, and outputs the image and the one or more parameters to the second processing device that executes the task processing which is same as the predetermined task processing.
According to this aspect, the second processing device executes the predetermined task processing based on the one or more parameters input from the image decoding device, thereby improving the accuracy of the task processing in the second processing device.
In the above aspect, when executing the predetermined task processing, the first processing device and the second processing device switch at least one of a machine learning model, a detection threshold, a scaling value, and a post-processing method based on the one or more parameters.
According to the this aspect, at least one of the machine learning model, the detection threshold value, the scaling value, and the post-processing method is switched based on the one or more parameters, thereby improving the accuracy of the task processing in the first processing device and the second processing device.
In the above aspect, the predetermined task processing includes at least one of object detection, object segmentation, object tracking, action recognition, pose estimation, pose tracking, and hybrid vision.
According to the aspect, the accuracy of each of the processing can be improved.
In the above aspect, the predetermined task processing includes image processing for improving image quality or image resolution.
According to this aspect, the accuracy of the image processing for improving image quality or image resolution can be improved.
In the above aspect, the image processing includes at least one of morphological transformation and edge enhancement processing for enhancing an object included in an image.
According to the aspect, the accuracy of each of the processing can be improved.
In the above aspect, the one or more parameters include at least one of a mounting height of a camera that outputs the image, a tilt angle of the camera, a distance from the camera to a region of interest, and a visual field of the camera.
According to this aspect, the accuracy of the task processing can be improved by allowing these pieces of information to be included in one or more parameters.
In the above aspect, the one or more parameters include at least one of the depth and the size of an object included in the image.
According to this aspect, the accuracy of the task processing can be improved by allowing these pieces of information to be included in one or more parameters.
In the above aspect, the one or more parameters include boundary information indicating a boundary surrounding an object included in the image, and distortion information indicating presence or absence of distortion in the image.
According to this aspect, the accuracy of the task processing can be improved by allowing these pieces of information to be included in one or more parameters.
In the above aspect, the boundary information includes position coordinates of a plurality of vertices related to a figure defining the boundary.
According to this aspect, even in a case where distortion occurs in the image, the boundary surrounding an object can be accurately defined.
In the above aspect, the boundary information includes center coordinates, width information, height information, and tilt information related to the figure defining the boundary.
According to this aspect, even in a case where distortion occurs in the image, the boundary surrounding an object can be accurately defined.
In the above aspect, the distortion information includes additional information indicating that the image is an image captured by a fisheye camera, a super-wide angle camera, or an omnidirectional camera.
According to the this aspect, a determination is easily made whether the fisheye camera, the super-wide angle camera, or the omnidirectional camera is used depending on whether the additional information is included in the one or more parameters.
An image decoding method according to one aspect of the present disclosure includes: by an image decoding device, receiving a bitstream from an image encoding device, decoding an image from the bitstream, obtaining, from the bitstream, one or more parameters that are not used for decoding the image, and outputs the image and the one or more parameters to a processing device that executes predetermined task processing.
According to this aspect, the image decoding device outputs, to the processing device that executes the predetermined task processing, the one or more parameters received from the image encoding device. As a result, the processing device executes the predetermined task processing based on the one or more parameters input from the image decoding device, thereby improving the accuracy of the task processing in the processing device.
An image processing method according to one aspect of the present disclosure includes: by an image decoding device, receiving, from an image encoding device, a bitstream including an encoded image and one or more parameters that are not used for encoding the image, obtaining the one or more parameters from the bitstream, and outputting the one or more parameters to a processing device that executes predetermined task processing.
According to this aspect, the image decoding device outputs, to the processing device that executes the predetermined task processing, the one or more parameters obtained from the bitstream received from the image encoding device. As a result, the processing device executes the predetermined task processing based on the one or more parameters input from the image decoding device, thereby improving the accuracy of the task processing in the processing device.
An image encoding device according to one aspect of the present disclosure encodes an image to generate a bitstream, adds, to the bitstream, one or more parameters that are not used for encoding the image, transmits, to an image decoding device, the bitstream to which the one or more parameters have been added, and outputs the image and the one or more parameters to a first processing device that executes predetermined task processing.
According to this aspect, the image encoding device transmits, to the image decoding device, the one or more parameters to be output to the first processing device for execution of the predetermined task processing. As a result, the image decoding device can output the one or more parameters received from the image encoding device to a second processing device that executes task processing which is same as the predetermined task processing. As a result, the second processing device executes the predetermined task processing based on the one or more parameters input from the image decoding device, thereby improving the accuracy of the task processing in the second processing device.
An image decoding device according to one aspect of the present disclosure receives a bitstream from an image encoding device, decodes an image from the bitstream, obtains, from the bitstream, one or more parameters that are not used for decoding the image, and outputs the image and the one or more parameters to a processing device that executes predetermined task processing.
According to this aspect, the image decoding device outputs, to the processing device that executes the predetermined task processing, the one or more parameters received from the image encoding device. As a result, the processing device executes the predetermined task processing based on the one or more parameters input from the image decoding device, thereby improving the accuracy of the task processing in the processing device.
In the following, embodiments of the present disclosure will be described in detail with reference to the drawings. Elements denoted by the same corresponding reference numerals in different drawings represent the same or corresponding elements.
Each of the embodiments described below illustrates specific examples of the present disclosure. Numerical values, shapes, components, steps, order of steps, and the like shown in the following embodiments are merely examples, and are not intended to limit the present disclosure. The components in the embodiments below include a component that is not described in an independent claim representing the highest concept and that is described as an arbitrary component. All the embodiments have respective contents that can be combined.
The image encoding device 1101A encodes an input image per block to generate a bitstream. Further, the image encoding device 1101A adds input one or more parameters to the bitstream. The one or more parameters are not used for encoding the image. Further, the image encoding device 1101A transmits, to the image decoding device 2101A, the bitstream to which the one or more parameters have been added. Further, the image encoding device 1101A generates a pixel sample of the image, and outputs a signal 1120A including the pixel sample of the image and the one or more parameters to the first processing device 1102A. The first processing device 1102A executes predetermined task processing such as a neural network task based on the signal 1120A input from the image encoding device 1101A. The first processing device 1102A may input a signal 1121A obtained as a result of executing the predetermined task processing to the image encoding device 1101A.
The image decoding device 2101A receives the bitstream from the image encoding device 1101A. The image decoding device 2101A decodes the image from the received bitstream, and outputs the decoded image to a display device. The display device displays the image. In addition, the image decoding device 2101A acquires one or more parameters from the received bitstream. The one or more parameters are not used for decoding the image. Further, the image decoding device 2101A generates a pixel sample of the image, and outputs a signal 2120A including the pixel sample of the image and the one or more parameters to the second processing device 2102A. The second processing device 2102A executes predetermined task processing which is same as that in the first processing device 1102A based on the signal 2120A input from the image decoding device 2101A. The second processing device 2102A may input a signal 2121A obtained as a result of executing the predetermined task processing to the image decoding device 2101A.
(Processing on Encoder Side)
As illustrated in
As illustrated in
As illustrated in
As illustrated in
The camera characteristics may be dynamically updated via another sensor mounted on the moving body. In a case of the camera mounted on a vehicle, the distance from the camera to the region of interest may be changed depending on a driving situation such as driving on a highway or driving in town. For example, a braking distance is different between driving on a highway and driving in town due to a difference in vehicle speed. Specifically, since the braking distance becomes long during high-speed driving on a highway, a farther object have to be found. On the other hand, since the braking distance becomes short during normal-speed driving in town, a relatively nearby object may be found. Actually, switching a focal length changes the distance from the camera to the ROI. For example, the distance from the camera to the ROI is increased by increasing the focal length. In the case of the camera mounted on a flight vehicle, the mounting height of the camera may be changed based on the flight altitude of the flight vehicle. In the case of the camera mounted on a robot arm, the distance from the camera to the region of interest may be changed depending on a movement of the robot arm.
As another example, the one or more parameters include at least one of the depth and the size of an object included in the image.
With reference to
In a final step S1003A, the image encoding device 1101A outputs the signal 1120A including the pixel sample of the image and the one or more parameters to the first processing device 1102A.
The first processing device 1102A executes predetermined task processing such as a neural network task using the pixel sample of the image and the one or more parameters included in the input signal 1120A. In the neural network task, at least one determination processing may be executed. An example of the neural network is a convolutional neural network. An example of the neural network task is object detection, object segmentation, object tracking, action recognition, pose estimation, pose tracking, machine and human hybrid vision, or any combination thereof.
The first processing device 1102A outputs a signal 1121A indicating the execution result of the neural network task. The signal 1121A may include at least one of a number of detected objects, a confidence level of the detected objects, boundary information or position information about the detected objects, and classification categories of the detected objects. The signal 1121A may be input from the first processing device 1102A to the image encoding device 1101A.
Hereinafter, utilization examples of the one or more parameters in the first processing device 1102A will be described.
Next, an exemplary operation flow will be described. An input image and a predicted image are input to an adder, and an addition value corresponding to a subtraction image between the input image and the predicted image is input from the adder to the transformation unit 1301. The transformation unit 1301 inputs a frequency coefficient obtained by transforming the addition value to the quantization unit 1302. The quantization unit 1302 quantizes the input frequency coefficient and inputs the quantized frequency coefficient to the inverse quantization unit 1303 and the entropy encoding unit 1313. Further, one or more parameters including the depth and the size of an object are input to the entropy encoding unit 1313. The entropy encoding unit 1313 entropy-encodes the quantized frequency coefficient and generates a bitstream. Further, the entropy encoding unit 1313 entropy-encodes the one or more parameters including the depth and the size of the object together with the quantized frequency coefficient or stores the one or more parameters in the header of the bitstream to add the one or more parameters to the bitstream.
The inverse quantization unit 1303 inversely quantizes the frequency coefficient input from the quantization unit 1302 and inputs the frequency coefficient that has been inversely quantized to the inverse transformation unit 1304. The inverse transformation unit 1304 inversely transforms the frequency coefficient to generate a subtraction image, and inputs the subtraction image to the adder. The adder adds the subtraction image input from the inverse transformation unit 1304 and the predicted image input from the intra prediction unit 1307 or the inter prediction unit 1312. The adder inputs an addition value 1320 (corresponding to the pixel sample described above) corresponding to the input image to the first processing device 1102 A, the block memory 1306, and the picture memory 1308. The addition value 1320 is used for further prediction.
The first processing device 1102A executes at least one of the morphological transformation and edge enhancement processing such as the unsharp masking on the addition value 1320 based on at least one of the depth and the size of the object, and enhances characteristics of the object included in the input image corresponding to the addition value 1320. The first processing device 1102A executes object tracking with at least determination processing using the addition value 1320 including the enhanced object and at least one of the depth and the size of the object. The depth and the size of the object improve the accuracy and speed performance of the object tracking. Here, in addition to at least one of the depth and the size of the object, the first processing device 1102A may execute the object tracking using position information indicating the position of the object included in the image (for example, boundary information indicating a boundary surrounding the object). This further improves the accuracy of the object tracking. In this case, the entropy encoding unit 1313 allows the position information to be included in the bitstream in addition to the depth and the size of the object. A determination result 1321 is input from first processing device 1102A to the picture memory 1308, and used for further prediction. For example, object enhancement processing is executed on the input image corresponding to the addition value 1320 stored in the picture memory 1308, based on the determination result 1321, thereby improving the accuracy of the subsequent inter prediction. However, the input of the determination result 1321 to the picture memory 1308 may be omitted.
The intra prediction unit 1307 and the inter prediction unit 1312 search for an image region most similar to the input image for prediction in a reconstructed image stored in the block memory 1306 or the picture memory 1308. The block memory 1309 fetches a block of the reconstructed image from the picture memory 1308 using a motion vector input from the motion vector prediction unit 1310. The block memory 1309 inputs the block of the reconstructed image to the interpolation unit 1311 for interpolation processing. The interpolated image is input from the interpolation unit 1311 to the inter prediction unit 1312 for inter prediction processing.
Thereafter, in step S1202A, the entropy encoding unit 1313 entropy-encodes the image to generate a bitstream, and generates a pixel sample of the image. Here, the depth and the size of the object are not used for the entropy encoding of the image. The entropy encoding unit 1313 adds the depth and the size of the object to the bitstream, and transmits, to the image decoding device 2101A, the bitstream to which the depth and the size of the object have been added.
In step S1203A, then, the first processing device 1102A executes a combination of the morphological transformation and the edge enhancement processing such as the unsharp masking on the pixel sample of the image based on the depth and the size of the object in order to enhance the characteristics of at least one object included in the image. The object enhancement processing in step S1203A improves the accuracy of the neural network task in the first processing device 1102A in next step S1204A.
In a final step S1204A, the first processing device 1102A executes the object tracking involving at least the determination processing, based on the pixel sample of the image and the depth and the size of the object. Here, the depth and the size of the object improve the accuracy and speed performance of the object tracking. The combination of the morphological transformation and the edge enhancement processing such as the unsharp masking may be replaced by another image processing technique.
(Processing on Decoder Side)
Next, in step S2002A, the image decoding device 2101A decodes the image from the bitstream to generate a pixel sample of the image. Here, the one or more parameters are not used for decoding the image. In addition, the image decoding device 2101A acquires the one or more parameters from the bitstream.
In a final step S2003A, the image decoding device 2101A outputs a signal 2120A including the pixel sample of the image and the one or more parameters to the second processing device 2102A.
The second processing device 2102A executes predetermined task processing similar to the processing in the first processing device 1102A using the pixel sample of the image and the one or more parameters included in the input signal 2120A. In the neural network task, at least one determination processing may be executed. An example of the neural network is a convolutional neural network. An example of the neural network task is object detection, object segmentation, object tracking, action recognition, pose estimation, pose tracking, machine and human hybrid vision, or any combination thereof.
The second processing device 2102A outputs a signal 2121A indicating the execution result of the neural network task. The signal 2121A may include at least one of a number of detected objects, confidence levels of the detected objects, boundary information or position information about the detected objects, and classification categories of the detected objects. The signal 2121A may be input from the second processing device 2102A to the image decoding device 2101A.
Hereinafter, utilization examples of the one or more parameters in the second processing device 2102A will be described.
Next, an exemplary operation flow will be described. The encoded bitstream input to the image decoding device 2101A is input to the entropy decoding unit 2301. The entropy decoding unit 2301 decodes the input bitstream, and inputs a frequency coefficient that is a decoded value to the inverse quantization unit 2302. Further, the entropy decoding unit 2301 acquires a depth and a size of an object from the bitstream, and inputs these pieces of information to the second processing device 2102A. The inverse quantization unit 2302 inversely quantizes the frequency coefficient input from the entropy decoding unit 2301, and inputs the frequency coefficient that has been inversely quantized to the inverse transformation unit 2303. The inverse transformation unit 2303 inversely transforms the frequency coefficient to generate a subtraction image, and inputs the subtraction image to the adder. The adder adds the subtraction image input from the inverse transformation unit 2303 and the predicted image input from the intra prediction unit 2306 or the inter prediction unit 2310. The adder inputs the addition value 2320 corresponding to the input image to the display device. As a result, the display device displays the image. In addition, the adder inputs the addition value 2320 to the second processing device 2102A, the block memory 2305, and the picture memory 2307. The addition value 2320 is used for further prediction.
The second processing device 2102A performs at least one of the morphological transformation and the edge enhancement processing such as the unsharp masking on an addition value 2320 based on at least one of the depth and the size of the object, and emphasizes characteristics of the object included in the input image corresponding to the addition value 2320. The second processing device 2102A executes object tracking involving at least determination processing using the addition value 2320 including the emphasized object and at least one of the depth and the size of the object. The depth and the size of the object improve the accuracy and speed performance of the object tracking. Here, in addition to at least one of the depth and the size of the object, the second processing device 2102A may execute the object tracking using position information indicating the position of the object included in the image (for example, boundary information indicating a boundary surrounding the object). This further improves the accuracy of the object tracking. In this case, the position information is included in the bitstream, and the entropy decoding unit 2301 acquires the position information from the bitstream. A determination result 2321 is input from second processing device 2102A to the picture memory 2307, and used for further prediction. For example, object enhancement processing is executed on the input image corresponding to the addition value 2320 stored in the picture memory 2307, based on the determination result 2321, thereby improving the accuracy of the subsequent inter prediction. However, the input of the determination result 2321 to the picture memory 2307 may be omitted.
The analysis unit 2311 parses the input bitstream to input some pieces of prediction information, such as a block of residual samples, a reference index indicating a reference picture to be used, and a delta motion vector, to the motion vector prediction unit 2312. The motion vector prediction unit 2312 predicts a motion vector of a current block based on the prediction information input from the analysis unit 2311. The motion vector prediction unit 2312 inputs a signal indicating the predicted motion vector to the block memory 2308.
The intra prediction unit 2306 and the inter prediction unit 2310 search for an image region most similar to the input image for prediction in a reconstructed image stored in the block memory 2305 or the picture memory 2307. The block memory 2308 fetches a block of the reconstructed image from the picture memory 2307 using the motion vector input from the motion vector prediction unit 2312. The block memory 2308 inputs the block of the reconstructed image to the interpolation unit 2309 for interpolation processing. The interpolated image is input from the interpolation unit 2309 to the inter prediction unit 2310 for inter prediction processing.
Next, in step S2202A, the entropy decoding unit 2301 entropy-decodes the image from the bitstream to generate a pixel sample of the image. Further, the entropy decoding unit 2301 acquires the depth and the size of the object from the bitstream. Here, the depth and the size of the object are not used for the entropy decoding of the image. The entropy decoding unit 2301 inputs the acquired depth and the size of the object to the second processing device 2102A.
In step S2203A, then, the second processing device 2102A executes a combination of the morphological transformation and the edge enhancement processing such as the unsharp masking on the pixel sample of the image based on the depth and the size of the object in order to enhance the characteristics of at least one object included in the image. The object enhancement processing in step S2203A improves the accuracy of the neural network task in the second processing device 2102A in next step S2204A.
In a final step S2204A, the second processing device 2102A executes the object tracking involving at least the determination processing, based on the pixel sample of the image and the depth and the size of the object. Here, the depth and the size of the object improve the accuracy and speed performance of the object tracking. The combination of the morphological transformation and the edge enhancement processing such as the unsharp masking may be replaced by another image processing technique.
According to the present embodiment, the image encoding device 1101A transmits, to the image decoding device 2101A, the one or more parameters to be output to the first processing device 1102A for execution of the predetermined task processing. As a result, the image decoding device 2101A can output the one or more parameters received from the image encoding device 1101A to the second processing device 2102A that executes task processing which is same as the predetermined task processing. As a result, the second processing device 2102A executes the predetermined task processing based on the one or more parameters input from the image decoding device 2101A, thereby improving the accuracy of the task processing in the second processing device 2102A.
A second embodiment of the present disclosure describes a response in a case where a camera that outputs an image with great distortion, such as a fisheye camera, a super-wide angle camera, or an omnidirectional camera, can be used in the first embodiment will be described.
(Processing on Encoder Side)
In step S2002B, the entropy encoding unit 2102B encodes a parameter set included in the one or more parameters into a bitstream. The parameter set includes boundary information indicating a boundary surrounding the object included in the image, and distortion information indicating presence or absence of distortion in the image.
The boundary information includes position coordinates of a plurality of vertices regarding a bounding box that is a figure defining the boundary. Alternatively, the boundary information may include center coordinates, width information, height information, and tilt information regarding the bounding box. The distortion information includes additional information indicating that the image is an image captured by a fisheye camera, a super-wide angle camera, or an omnidirectional camera. The boundary information and the distortion information may be input from the camera or the sensor 3101 illustrated in
The parameter set may be entropy-encoded to be added to the bitstream, or may be stored in a header of the bitstream to be added to the bitstream.
The encoder 2100B transmits, to a decoder 1100B, the bitstream to which the parameter set has been added.
In a final step S2003B, the entropy encoding unit 2102B outputs the image and the parameter set to the first processing device 1102A. The first processing device 1102A executes predetermined task processing such as a neural network task using the input image and the parameter set. In the neural network task, at least one determination processing may be executed. The first processing device 1102A may switch between a machine learning model for a greatly distorted image and a machine learning model for a normal image with small distortion depending on whether the additional information is included in the distortion information in the parameter set.
As illustrated in
As illustrated in
Furthermore, with reference to
(Processing on Decoder Side)
In next step S1002B, the entropy decoding unit 1101B decodes a parameter set from the bitstream received from the encoder 2100B. The parameter set includes boundary information indicating a boundary surrounding the object included in the image, and distortion information indicating presence or absence of distortion in the image.
In a final step S1003B, the entropy decoding unit 1101B outputs the decoded image and the parameter set to the second processing device 2102A. The second processing device 2102A executes predetermined task processing which is same as the task in the first processing device 1102A using the input image and the parameter set. In the neural network task, at least one determination processing may be executed. The second processing device 2102A may switch between a machine learning model for a greatly distorted image and a machine learning model for a normal image with small distortion depending on whether the additional information is included in the distortion information in the parameter set.
According to the present embodiment, even in a case where a camera that outputs a greatly distorted image, such as a fisheye camera, a super-wide angle camera, or an omnidirectional camera, is used, the bounding box surrounding the object can be accurately defined. Further, the encoder 2100B transmits a parameter set including the boundary information and the distortion information to the decoder 1100B. As a result, the decoder 1100B can output the parameter set received from the encoder 2100B to the second processing device 2102A. As a result, the second processing device 2102A executes the predetermined task processing based on the input parameter set, thereby improving the accuracy of the task processing in the second processing device 2102A.
The present disclosure is particularly useful for application to an image processing system including an encoder that transmits an image and a decoder that receives the image.
Number | Date | Country | |
---|---|---|---|
63178798 | Apr 2021 | US | |
63167789 | Mar 2021 | US |
Number | Date | Country | |
---|---|---|---|
Parent | PCT/JP2022/015319 | Mar 2022 | US |
Child | 18372220 | US |