This patent application is based on and claims priority pursuant to 35 U.S.C. § 119(a) to Japanese Patent Application No. 2023-012012, filed on Jan. 30, 2023, in the Japan Patent Office, the entire disclosure of which is hereby incorporated by reference herein.
Embodiments of the present disclosure relate to an object processing apparatus, an object determination method, and a non-transitory computer-executable medium.
Waste sorting apparatuses that sort one or more particular pieces of waste from multiple pieces of waste are known. Such waste sorting apparatuses generate three-dimensional data of waste that is piled up and conveyed on the basis of image data obtained by imaging the waste with X-rays, and sort an object to be removed from the waste.
According to an embodiment of the present disclosure, an object processing apparatus includes a first imaging device to capture a surface of a subject to obtain a surface image. The object processing apparatus includes a second imaging device to capture an internal object in the subject to obtain an internal image. The object processing apparatus includes circuitry configured to determine whether the subject is a battery-containing product that has a built-in battery based on the surface image and the internal image to generate a determination result. The object processing apparatus includes a processing device to operate based on the determination result.
According to an embodiment of the present disclosure, an object determination method includes acquiring a surface image in which a surface of a subject appears. The object determination method includes acquiring an internal image in which an internal object in the subject appears. The object determination method includes determining whether the subject is a battery-containing product that has a built-in battery based on the surface image and the internal image.
According to an embodiment of the present disclosure, a non-transitory computer-executable medium storing a plurality of instructions which, when executed by a processor, causes the processor to perform a method. The method includes acquiring a surface image in which a surface of a subject appears. The method includes acquiring an internal image in which an internal object in the subject appears. The method includes determining whether the subject is a battery-containing product that has a built-in battery based on the surface image and the internal image.
A more complete appreciation of embodiments of the present disclosure and many of the attendant advantages and features thereof can be readily obtained and understood from the following detailed description with reference to the accompanying drawings, wherein:
The accompanying drawings are intended to depict embodiments of the present disclosure and should not be interpreted to limit the scope thereof. The accompanying drawings are not to be considered as drawn to scale unless explicitly noted. Also, identical or similar reference numerals designate identical or similar components throughout the several views.
In describing embodiments illustrated in the drawings, specific terminology is employed for the sake of clarity. However, the disclosure of this specification is not intended to be limited to the specific terminology so selected and it is to be understood that each specific element includes all technical equivalents that have a similar function, operate in a similar manner, and achieve a similar result.
Referring now to the drawings, embodiments of the present disclosure are described below. As used herein, the singular forms “a,” “an,” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise.
Embodiments of an object processing apparatus, an object determination method, and an object determination program are described with reference to the drawings. The technologies of the present disclosure are not limited to those in the following description. In the following description, like reference signs denote like elements, and redundant descriptions thereof are simplified or omitted.
As illustrated in
A conveyance path 6 is formed in the belt conveyor 3. The conveyance path 6 extends along a plane substantially parallel to a horizontal plane. A straight line along which the conveyance path 6 extends is parallel to a conveyance direction 7 that is parallel to the horizontal plane. The belt 5 includes a conveyance path facing portion 8. The conveyance path facing portion 8 is disposed under the conveyance path 6 and is disposed along the conveyance path 6. The belt 5 is supported by the belt driver so that the conveyance path facing portion 8 translates in the conveyance direction 7. The belt driver moves the belt 5 such that the conveyance path facing portion 8 moves toward the downstream end in the conveyance direction 7.
The conveyance path 6 includes an object supply area 11, a surface image capturing area 12, an internal image capturing area 14, and a work area 15. The surface image capturing area 12 is disposed downstream from the object supply area 11 in the conveyance direction 7. The internal image capturing area 14 is disposed downstream from the surface image capturing area 12 in the conveyance direction 7. The work area 15 is disposed downstream from the internal image capturing area 14 in the conveyance direction 7. In other words, the surface image capturing area 12 and the internal image capturing area 14 are disposed between the object supply area 11 and the work area 15. In the work area 15, a worker 16 finds out a battery-containing product from multiple pieces of waste conveyed along the conveyance path 6, and the worker 16 removes the found battery-containing product from the conveyance path 6. The battery-containing product is a product that includes a built-in battery such as a lithium ion secondary battery. The battery is in a container formed from a material that does not transmit x-rays.
The object processing apparatus 1 includes a visible line image sensor 21, an X-ray sensor 22, and a projector 23. The visible line image sensor 21 is disposed above the surface image capturing area 12 such that the surface image capturing area 12 of the conveyance path 6 is disposed between the conveyance path facing portion 8 of the belt 5 and the visible line image sensor 21. The visible line image sensor 21 receives visible light reflected from an area intersecting a surface image reading plane 24 of the surface of a subject that is present in the conveyance path facing portion 8 and disposed in the surface image capturing area 12. The surface image reading plane 24 is substantially perpendicular to the conveyance direction 7 and intersects the surface image capturing area 12. The visible line image sensor 21 further generates multiple surface line images corresponding to multiple times on the basis of the received visible light. In a particular surface line image corresponding to a certain time among the surface line images, a portion of the surface of a subject that intersects the surface image reading plane 24 at the certain time appears.
The X-ray sensor 22 includes a light emitter 25 and a light receiver 26. The light emitter 25 is disposed below the internal image capturing area 14 of the conveyance path 6. The light emitter 25 emits X-rays along an internal image reading plane 27 toward the internal image capturing area 14. The internal image reading plane 27 is substantially perpendicular to the conveyance direction 7 and intersects the internal image capturing area 14. The light receiver 26 is disposed above the internal image capturing area 14 such that the internal image capturing area 14 is disposed between the light emitter 25 and the light receiver 26. The light receiver 26 receives transmitted X-rays transmitted through the internal image capturing area 14 among X-rays emitted from the light emitter 25. The X-ray sensor 22 generates multiple internal line images corresponding to multiple times on the basis of the transmitted X-rays received by the light receiver 26. In a particular internal line image corresponding to a certain time among the surface line images, a portion of a subject that intersects the internal image reading plane 27 at the certain time appears.
The projector 23 is disposed above the work area 15 of the conveyance path 6 such that the work area 15 is disposed between the projector 23 and the conveyance path facing portion 8 of the belt 5. The projector 23 performs projection mapping to project video onto an object that is present in the work area 15.
As illustrated in
The communication interface 35 downloads information to the object recognition control apparatus 31 from another computer connected through a communication line and transmits information from the object recognition control apparatus 31 to another computer under control of the CPU 34. In the media interface 36, a storage medium 41 that is non-transitory and tangible can be loaded. Examples of the storage medium 41 include a semiconductor memory, a magnetic disk, a magneto-optical disk, and an optical disc. The media interface 36, when the storage medium 41 is loaded therein, reads information from the storage medium 41 and records information in the storage medium 41 under control of the CPU 34. Alternatively, the computer program to be installed in the object recognition control apparatus 31 may be downloaded from another computer via the communication interface 35 or may be read from the storage medium 41 via the media interface 36.
The computer program to be installed in the object recognition control apparatus 31 includes multiple computer programs for causing the object recognition control apparatus 31 to implement multiple functions. The multiple functions include an imaging control unit 37, a determination unit 38, a position calculation unit 39, and an output unit 40.
The imaging control unit 37 controls the visible line image sensor 21 to capture multiple surface line images. The imaging control unit 37 generates a surface image on the basis of the surface line images captured by the visible line image sensor 21. The surface line images are laid without any clearance on a surface image so that a subject that is conveyed along the conveyance path 6 and present in the surface image capturing area 12 appears in the surface image. The surface image further includes multiple pixels and indicates multiple pieces of color information corresponding to the pixels. The pixels are arranged in the surface image without any clearance. Color information corresponding to a certain pixel among the multiple pieces of color information indicates a color of an area corresponding to the certain pixel in the surface of the subject appearing in the surface image, and indicates the red gradation value, the green gradation value, and the blue gradation value of the color.
The imaging control unit 37 further controls the X-ray sensor 22 to capture multiple internal line images. The imaging control unit 37 generates an internal image on the basis of the internal line images captured by the X-ray sensor 22. The internal line images are laid without any clearance on an internal image so that a subject that is conveyed along the conveyance path 6 and present in the internal image capturing area 14 appears in the internal image. The internal image further includes multiple pixels and indicates multiple transmittances corresponding to the pixels. The pixels are arranged in the internal image without any clearance. A transmittance corresponding to a certain pixel among the transmittances indicates a ratio of X-ray transmitted through a portion corresponding to the certain pixel in a subject. In other words, when the outer contour of a subject is formed of a material that transmits X-rays, an internal object contained in the subject appears in the internal image together with the outer contour of the subject.
The determination unit 38 determines whether a subject conveyed along the conveyance path 6 is a battery-containing product on the basis of the surface image and the internal image generated by the imaging control unit 37. When the determination unit 38 determines that the subject is a battery-containing product, the position calculation unit 39 performs image processing on the surface image, to generate imaging-time position information on the basis of the surface image. The imaging-time position information indicates an imaging time when the subject was present in the surface image capturing area 12, and indicates a position where the subject was present at the imaging time. When the determination unit 38 determines that the subject is a battery-containing product, the output unit 40 outputs the imaging-time position information generated by the position calculation unit 39 to the projector control apparatus 32.
The projector control apparatus 32 is connected to the projector 23 so as to communicate information with each other. The projector control apparatus 32 is a computer, and includes a memory and a CPU. In the memory, a computer program to be installed in the projector control apparatus 32 is recorded, and information to be used by the CPU is recorded. The CPU executes a computer program installed in the projector control apparatus 32 to perform information processing. The projector control apparatus 32 further controls the projector 23 so as to perform projection mapping, on the basis of the imaging-time position information that is output from the object recognition control apparatus 31.
Examples of the battery-containing product removed from the conveyance path 6 in the work area 15 include a portable electric fan, an electric shaver, and a portable charger. As illustrated in
As illustrated in
As illustrated in
As illustrated in
As illustrated in
As illustrated in
The imaging control unit 37 generates, when the surface image is generated, the internal image such that a subject appearing in surface image appears in the internal image and such that a position where an internal view representing the subject appears in the internal image matches a position where a surface view representing the subject appears in the surface image. For example, the imaging control unit 37 generates, when the surface image 51 is generated, the internal image 61 such that a position where the portable electric fan internal image 62 appears in the internal image 61 matches a position where the portable electric fan surface image 52 appears in the surface image 51. The imaging control unit 37 generates, when the surface image 53 is generated, the internal image 64 such that a position where the electric shaver internal image 65 appears in the internal image 64 matches a position where the electric shaver surface image 54 appears in the surface image 53. The imaging control unit 37 generates, when the surface image 55 is generated, the internal image 67 such that a position where the portable charger internal image 68 appears in the internal image 67 matches a position where the portable charger surface image 56 appears in the surface image 55.
For example, as illustrated in
The battery-containing product determination unit 46 determines whether a subject indicated by the integrated information is a battery-containing product on the basis of the integrated information generated by the integration unit 45, using a learning model for integrated information generated by machine learning.
The operation of the waste sorting apparatus 2 includes a conveyance operation and a battery-containing product removal operation.
Conveyance Operation In the conveying operation, first, the belt conveyor 3 starts in response to an operation by a user to the belt conveyor 3. In response to the start of belt conveyor 3, the belt driver of the belt conveyor 3 moves the belt 5 such that the conveyance path facing portion 8 of the belt 5 translates in the conveyance direction 7 at a predetermined constant conveyance speed. The user further places waste on the conveyance path facing portion 8 of the belt 5 such that the waste is placed in the object supply area 11 of the conveyance path 6.
The waste placed in the object supply area 11 is conveyed in the conveyance direction 7 along the conveyance path 6 at the conveyance speed as the conveyance path facing portion 8 translates. Thus, the waste is placed in the surface image capturing area 12. The waste placed in the surface image capturing area 12 is further conveyed in the conveyance direction 7 at the conveyance speed along the conveyance path 6 as the conveyance path facing portion 8 translates. Thus, the waste is placed in the internal image capturing area 14. The waste placed in the internal image capturing area 14 is further conveyed in the conveyance direction 7 at the conveyance speed along the conveyance path 6 as the conveyance path facing portion 8 translates. Thus, the waste is placed in the work area 15. The waste placed in the work area 15 is discharged from the conveyance path 6 as the conveyance path facing portion 8 translates and supplied to a shredder. The shredder shreds the waste supplied from the belt conveyor 3 into small pieces.
The waste may be a battery-containing product having a battery built therein. When a short-circuit occurs in a lithium-ion secondary battery, which is the battery, or when a shock is applied to the lithium-ion secondary battery, a fire may occur. Batteries other than the lithium-ion secondary battery may ignite due to occurrence of a short circuit. Accordingly, when the battery-containing product is supplied from the belt conveyor 3 to the shredder, the battery-containing product is shred and may cause a fire.
A process performed by the object processing apparatus 1 according to the present is described with reference to
The object recognition control apparatus 31 creates integrated information on the basis of the surface image and the internal image (S1103). The object recognition control apparatus 31 determines whether waste indicated by the integrated information is a battery-containing product on the basis of the integrated information, using a learning model for integrated information (S1104).
When the object recognition control apparatus 31 determines that the waste indicated by the integrated information is not a battery-containing product (S1104: No), the process ends.
When the object recognition control apparatus 31 determines that the waste indicated by the integrated information is a battery-containing product (S1104: Yes), the object recognition control apparatus 31 performs image processing on the surface image, to calculate imaging-time position information. The imaging-time position information indicates an imaging time when the battery-containing product was present in the surface image capturing area 12, and indicates a position where the battery-containing product was present at the imaging time. When the object recognition control apparatus 31 determines that the waste indicated by the integrated information is a battery-containing product, the object recognition control apparatus 31 further outputs the calculated imaging-time position information to the projector control apparatus 32 (S1105).
The projector control apparatus 32 creates multiple projection mapping images corresponding to multiple times on the basis of the conveyance speed and the imaging-time position information (S1106). Each of the times is included in a time period during which the battery-containing product is present in the work area 15. A particular projection mapping image corresponding to a certain time among the projection mapping images is created such that an area where the battery-containing product is present in the work area 15 is conspicuously visually recognizable by projecting the particular project mapping image onto the work area 15 at the certain time. Examples of the projection mapping image include an image that illuminates the battery-containing product more brightly than portions in the work area 15 other than the battery-containing product. The projector control apparatus 32 controls the projector 23 to perform projection mapping for projecting the projection mapping images onto the work area 15 at the multiple times (S1107).
The worker 16 searches multiple pieces of waste that is conveyed along the conveyance path 6 and placed in the work area 15 for a battery-containing product. When the worker 16 finds out a battery-containing product, the worker removes the found battery-containing product from the conveyance path 6. The waste sorting apparatus 2 can make it easy for the worker 16 to search for a battery-containing product by the projection mapping that makes the battery-containing product conspicuously visually recognizable, and thus can reduce the frequency with which the worker 16 overlooks the battery-containing product. Since the waste sorting apparatus 2 reduces the frequency with which the worker 16 overlooks the battery-containing product, the waste sorting apparatus 2 can reduce the frequency with which the battery-containing product is supplied to the shredder. Since the waste sorting apparatus 2 reduces the frequency with which the battery-containing product is supplied to the shredder, the waste sorting apparatus 2 can reduce the frequency of ignition in the shredder.
When the object recognition control apparatus 31 makes an erroneous determination such as a determination that waste that is not a battery-containing product is a battery-containing product and a determination that a battery-containing product is not a battery-containing product, the user creates retraining data. The retraining data includes data associating the integrated information used when the waste that is not a battery-containing product is erroneously determined as a battery-containing product with a determination of “being not a battery-containing product.” The retraining data further includes data associating the integrated information used when a battery-containing product is erroneously determined as not a battery-containing product with a determination of “being a battery-containing product.” The user retrains the learning model for integrated information using the retraining data to create a new learning model for integrated information. The learning model for integrated information to be used when the object recognition control apparatus 31 determines whether waste is a battery-containing product is updated to the created new learning model for integrated information according to an operation by a user. The update of the learning model for integrated information enables the object processing apparatus 1 to further reduce the frequency of erroneous determinations.
As illustrated in
As illustrated in
An object processing apparatus according to Comparative Example 1 has the same or substantially the same configuration as the configuration of the object processing apparatus 1 according to Embodiment 1 except that the determination unit 38 of the object processing apparatus 1 according to Embodiment 1 is replaced with another determination unit. The object processing apparatus according to Comparative Example 1 determines whether waste conveyed along the conveyance path 6 is a battery-containing product on the basis of only a surface image without using an internal image. Accordingly, the object processing apparatus of Comparative Example 1 may erroneously determine that a portable electric fan in a bag is not a battery-containing product since the portable electric fan does not appear in the surface image 71. Even when the portable electric fan does not appear in the surface image 71, the object processing apparatus 1 can determine that the portable electric fan in the bag is a battery-containing product on the basis of the integrated information since the portable electric fan appears in the internal image 73. Accordingly, compared to the object processing apparatus according to Comparative Example 1, the object processing apparatus 1 can reduce the frequency of erroneous determination that a bag containing a battery-containing product is not a battery-containing product.
As illustrated in
As illustrated in
An object processing apparatus according to Comparative Example 2 has the same or substantially the same configuration as the configuration of the object processing apparatus 1 according to Embodiment 1 except that the determination unit 38 of the object processing apparatus 1 according to Embodiment 1 is replaced with a still another determination unit. The object processing apparatus according to Comparative Example 2 determines whether waste conveyed along the conveyance path 6 is a battery-containing product on the basis of only an internal image without using a surface image. Accordingly, the object processing apparatus of Comparative Example 2 may erroneously determine that a battery-containing product on a plate that blocks X-rays is not a battery-containing product. Even when the entire portable electric fan does not appear in the internal image 84, the object processing apparatus 1 can determine that the battery-containing product on the plate that blocks X-rays is a battery-containing product on the basis of the integrated information since the entire portable electric fan appears in the surface image 81. Accordingly, compared to the object processing apparatus according to Comparative Example 2, the object processing apparatus 1 can reduce the frequency of erroneous determination that a battery-containing product on a plate that blocks X-rays is not a battery-containing product.
The object processing apparatus 1 according to Embodiment 1 includes the visible line image sensor 21, the X-ray sensor 22, the determination unit 38, and the projector 23. The visible line image sensor 21 captures a surface image in which the surface of waste appears. The X-ray sensor 22 captures an internal image in which an internal object inside the waste appears. The determination unit 38 determines whether the waste is a battery-containing product that includes a built-in battery on the basis of the surface image and the internal image. The projector 23 operates on the basis of the determination result obtained by the determination unit 38. With this configuration, the object processing apparatus 1 according to Embodiment 1 can reduce the frequency of erroneously determining that a battery-containing product is not a battery-containing product or erroneously determining that waste that is not a battery-containing product is a battery-containing product.
Further, the determination unit 38 of the object processing apparatus 1 according to Embodiment 1 determines whether waste is a battery-containing product on the basis of the integrated information obtained by integrating the surface image and the internal image. With this configuration, the number of information dimensions used in the determination by the object processing apparatus 1 according to Embodiment 1 increases. This enhances the determination accuracy, compared to when each of the surface image and the internal image is used alone in the determination.
Further, the object processing apparatus 1 according to Embodiment 1 includes the position calculation unit 39 and the projector control apparatus 32. When the determination unit 38 determines that the waste is a battery-containing product, the position calculation unit 39 calculates a position where the battery-containing product is present on the basis of the surface image. The projector control apparatus 32 controls the projector 23 on the basis of the calculated position to project, in the work area 15, video indicating an area where the battery-containing product is present. With this configuration, the object processing apparatus 1 according to Embodiment 1 makes it easy for the worker 16 to find out a battery-containing product from multiple pieces of waste. The object processing apparatus 1 according to Embodiment 1 does not have to include another means for detecting the position of the battery-containing product in the work area 15 in addition to the visible line image sensor 21. This reduces the number of components and the manufacturing cost.
The description given above is of a case where the internal image used by the object processing apparatus 1 according to Embodiment 1 is generated such that a position where an internal view representing a certain piece of waste appears in the internal image matches a position where a surface view representing the certain piece of waste appears in the surface image. Alternatively, the internal image may be generated such that the position where the internal view representing the certain piece of waste appears in the internal image does not match the position where the surface view representing the certain piece of waste appears in the surface image. In this case, the object recognition control apparatus 31 calculates a position where the waste appears in the surface image and a position where the waste appears in the internal image by image processing, and creates the integrated information on the basis of the calculated positions. With this configuration, even when the internal image is generated in a manner that a position where an internal view representing a certain piece of waste appears in the internal image does not match a position where a surface view representing the certain piece of waste appears in the surface image, the object recognition control apparatus 31 can reduce the frequency of erroneous determination using the integrated information.
As illustrated in
For example, as illustrated in
The battery-containing product candidate determination unit 92 determines whether a battery-containing product candidate appears in a surface image generated by the imaging control unit 37, using a learning model for internal images, the learning model being generated by machine learning. The battery-containing product candidate is a product that possibly includes a built-in battery. Examples of the battery-containing product candidate include a product whose appearance is equivalent to the appearance of a portable electric fan, an electric shaver, or a portable charger. The battery candidate determination unit 93 determines whether a battery candidate appears in a surface image generated by the imaging control unit 37, using a learning model for internal images, the learning model being generated by machine learning. The battery candidate is an object that is possibly a battery. Examples of the battery candidate include an object that is formed of a material that does not transmit X-rays and has a cylindrical shape or a rectangular parallelepiped shape.
When the battery-containing product candidate determination unit 92 determines that a battery-containing product candidate appears in a surface image and when the battery candidate determination unit 93 determines that a battery candidate appears in an internal image, the battery-containing product determination unit 94 determines that the waste appearing in the surface image and the internal image is a battery-containing product. When the battery-containing product candidate determination unit 92 determines that no battery-containing product candidate appears in a surface image or when the battery candidate determination unit 93 determines that no battery candidate appears in an internal image, the battery-containing product determination unit 94 determines that the waste appearing in the surface image and the internal image is not a battery-containing product.
The object processing apparatus according to Comparative Example 1, which determines whether waste conveyed along the conveyance path 6 is a battery-containing product on the basis of only a surface image, may erroneously determine a battery-containing product candidate from which a battery has been removed as a battery-containing product. The object processing apparatus according to Embodiment 2 determines that the waste is not a battery-containing product even when a battery-containing product candidate appears in the surface image, when no battery candidate appears in the internal image. Thus, the object processing apparatus according to Embodiment 2 can reduce the frequency of erroneous determination that a battery-containing product candidate from which a battery has been removed is a battery-containing product.
The object processing apparatus according to Comparative Example 2, which determines whether waste conveyed along the conveyance path 6 is a battery-containing product on the basis of only an internal image, may erroneously determine a product that includes a battery candidate, which is not actually a battery, as a battery-containing product. The object processing apparatus according to Embodiment 2 determines that the waste is not a battery-containing product even when a battery candidate appears in the internal image, when no battery-containing product candidate appears in the surface image. Thus, the object processing apparatus according to Embodiment 2 can reduce the frequency of erroneous determination that waste that is not a battery-containing product is a battery-containing product. In other words, the waste sorting apparatus including the object processing apparatus according to Embodiment 2 can reduce the frequency of erroneously determining waste that is not a battery-containing product as a battery-containing product. This reduces the frequency of misleadingly presenting to the worker 16 that waste that is not a battery-containing product is a battery-containing product.
An object processing apparatus according to Embodiment 3 has the same or substantially the same configuration as the configuration of the object processing apparatus according to Embodiment 2 except that the battery-containing product determination unit 94 of the object processing apparatus according to Embodiment 2 is replaced with another battery-containing product determination unit. When the battery-containing product candidate determination unit 92 determines that a battery-containing product candidate appears in a surface image or when the battery candidate determination unit 93 determines that a battery candidate appears in an internal image, the battery-containing product determination unit determines that the waste appearing in the surface image and the internal image is a battery-containing product. When the battery-containing product candidate determination unit 92 determines that no battery-containing product candidate appears in a surface image and when the battery candidate determination unit 93 determines that no battery candidate appears in an internal image, the battery-containing product determination unit determines that the waste appearing in the surface image and the internal image is not a battery-containing product.
Accordingly, compared to the object processing apparatus according to Embodiment 2, the object processing apparatus according to Embodiment 3 can more reliably reduce the frequency of erroneous determination that a battery-containing product is not a battery-containing product. Thus, compared to the waste sorting apparatus including the object processing apparatus according to Embodiment 2, the waste sorting apparatus including the object processing apparatus according to Embodiment 3 can more reliably prevent a battery-containing product from being supplied to the shredder. In other words, compared to the waste sorting apparatus including the object processing apparatus according to Embodiment 2, the waste sorting apparatus including the object processing apparatus according to Embodiment 3 can more reliably prevent ignition in the shredder.
As illustrated in
For example, as illustrated in
The battery-containing product candidate determination unit 92 determines whether a battery-containing product candidate appears in a surface image generated by the imaging control unit 37, using a learning model for surface images, the learning model being generated by machine learning. When the battery-containing product candidate determination unit 92 determines that no battery-containing product candidate appears in the surface image, the X-ray sensor control unit 96 controls the X-ray sensor 22 not to project X-rays toward the internal image capturing area 14. The area extraction unit 97 extracts a partial image from the internal image generated by the imaging control unit 37. A position where the partial image is present in the internal image matches a position where the surface image representing the battery-containing product candidate is present in the surface image.
The battery-containing product determination unit 98 determines whether a battery candidate appears in the partial image extracted by the area extraction unit 97. When the battery-containing product determination unit 98 determines that a battery candidate appears in the partial image, the determination unit 95 determines that the waste appearing in the surface image and the internal image is a battery-containing product. When the battery-containing product determination unit 98 determines that no battery candidate appears in the partial image, the determination unit 95 determines that the waste appearing in the surface image and the internal image is not a battery-containing product.
In the same or substantially the same manner as the object processing apparatus according to Embodiment 2, the object processing apparatus according to Embodiment 4 can also reduce the frequency of erroneously determining that a battery-containing product is not a battery-containing product or erroneously determining that waste that is not a battery-containing product is a battery-containing product. Further, the X-ray sensor 22 of the object processing apparatus according to Embodiment 4 does not emit X-rays on the basis of the determination that no battery-containing product candidate appears in the surface image. This reduces the operation rate of the X-ray sensor 22, and thus reduces the running cost. Accordingly, the object processing apparatus according to Embodiment 4 can reduce the dose to which objects around the X-ray sensor 22 are exposed to X-rays. Since the information amount of the partial image is smaller than the information amount of the internal image, determining whether the battery candidate appears in the partial image extracted from the internal image is easier and takes less time than determining whether the battery candidate appears in the internal image. Accordingly, the object processing apparatus according to Embodiment 4 can easily determine in a short time whether waste appearing in the surface image and the internal image is a battery-containing product.
When the object recognition control apparatus 31 of the object processing apparatus according to any of Embodiments 2 to 4 makes an erroneous determination such as a determination that waste that is not a battery-containing product is a battery-containing product candidate and a determination that a battery-containing product is not a battery-containing product candidate, a user creates retraining data for surface images. The retraining data for surface images includes data associating the surface image used when the waste that is not a battery-containing product is erroneously determined as a battery-containing product candidate with a determination of “being not a battery-containing product candidate.” The retraining data for surface images further includes data associating the surface image used when a battery-containing product is erroneously determined as not the battery-containing product candidate with a determination of “being a battery-containing product candidate.” The user retrains the learning model for surface images using the retraining data for surface images to create a new learning model for surface images. The learning model for surface images to be used when the object recognition control apparatus 31 determines whether waste is a battery-containing product candidate is updated to the created new learning model for surface images according to an operation by a user. The update of the learning model for surface images enables the object processing apparatus 1 according to any of Embodiments 2 to 4 to further reduce the frequency of erroneously determination as to whether waste appearing in a surface image is a battery-containing product candidate.
When the object recognition control apparatus 31 of the object processing apparatus according to any of Embodiments 2 to 4 makes an erroneous determination such as a determination that an object that is not a battery is a battery candidate and a determination that a battery is not a battery candidate, a user creates retraining data for internal images. The retraining data for internal images includes data associating the internal image used when the object that is not a battery is erroneously determined as a battery candidate with a determination of “being not a battery candidate.” The retraining data for internal images includes data associating the internal image used when the battery is erroneously determined as not a battery candidate with a determination of “being a battery candidate.” The user retrains the learning model for internal images using the retraining data for internal images to create a new learning model for internal images. The learning model for internal images to be used when the object recognition control apparatus 31 determines whether an internal object of waste is a battery candidate is updated to the created new learning model for internal images according to an operation by a user. The update of the learning model for internal images enables the object processing apparatus 1 according to any of Embodiments 2 to 4 to further reduce the frequency of erroneously determination as to whether an internal object of waste appearing in an internal image is a battery candidate.
While the object recognition control apparatus 31 of the object processing apparatus according to any of Embodiments 2 to 4 uses the two learning models of the retraining data for surface image and the retraining data for internal images, the object processing apparatus 1 according to Embodiment 1 uses only one learning model for integrated information. Accordingly, the object processing apparatus 1 according to Embodiment 1 can reduce the frequency of erroneous determination by retraining only one learning model for integrated information, and thus easily reduce the frequency of erroneous determination, compared to the object processing apparatus according to any of Embodiments 2 to 4.
In Embodiments 1 to 4, the description given above is of a case where the object processing apparatus determines whether waste appearing in the surface image and the internal image is a battery-containing product by using the learning model generated by machine learning. Alternatively, the determination may be performed without using the learning model. The object processing apparatus can reduce the frequency of erroneous determination by using the surface image and the internal image even when the learning model is not used.
The description given above is of a case where the internal image capturing area 14 of the object processing apparatus according to Embodiment 4 is disposed downstream from the surface image capturing area 12 in the conveyance direction 7. Alternatively, the internal image capturing area 14 may be disposed upstream from the surface image capturing area 12 in the conveyance direction 7. In this case, the object recognition control apparatus 31 determines whether a battery candidate appears in an internal image. On the basis of the determination result, the object recognition control apparatus 31 controls the visible line image sensor 21 to capture a surface image when the battery candidate appears in the internal image, and controls the visible line image sensor 21 not to capture the surface image when no battery candidate appears in the internal image. Also in this case, the object processing apparatus can reduce the frequency of erroneous determination in the same or substantially the same manner as the object processing apparatus according to Embodiment 4. Further, in such an object processing apparatus, the visible line image sensor 21 does not operate when no battery candidate appears in the internal image. Accordingly, the operation rate of the visible line image sensor 21 is reduced, and thus the running cost is reduced.
The terahertz sensor 103 is disposed above the internal image capturing area 14 such that the internal image capturing area 14 is disposed between the conveyance path facing portion 8 and the terahertz sensor 103. The terahertz sensor 103 visualizes the internal structure of a subject by using absorption, scattering, and diffraction of terahertz waves. The terahertz sensor 103 emits terahertz waves toward the internal image capturing area 14, receives a terahertz wave reflected from the surface or inside of the subject disposed in the internal image capturing area 14 among the emitted terahertz waves, and captures an internal image on the basis of the received terahertz wave. Accordingly, the outer contour of the subject and an internal object in the subject appear in the internal image.
The robot 104 is disposed near the work area 15. The robot 104 can grip an object disposed in the work area 15 and can move and release the gripped object. In other words, the robot 104 can remove the object from the work area.
The object recognition control apparatus 105 is connected to the visible area image sensor 102, the terahertz sensor 103, and the robot control apparatus 106 so as to communicate information with each other. The object recognition control apparatus 105 is a computer in which multiple computer programs that cause the object recognition control apparatus 105 to function as an imaging control unit 111, a determination unit 112, a position calculation unit 113, and an output unit 114 are installed. The computer program to be installed in the object recognition control apparatus 105 may be downloaded from another computer or may be read from a non-transitory tangible storage medium.
The imaging control unit 111 controls the visible area image sensor 102 to capture a surface image in which a subject that is conveyed along the conveyance path 6 and disposed in the surface image capturing area 12 appears. The imaging control unit 111 controls the terahertz sensor 103 to capture an internal image in which the subject appearing in the captured surface image appears. In the same or substantially the same manner as the determination unit 38, the determination unit 112 determines whether a subject conveyed along the conveyance path 6 is a battery-containing product on the basis of the surface image and the internal image captured by the imaging control unit 111. In the same or substantially the same manner as the position calculation unit 39, when the determination unit 112 determines that the subject is a battery-containing product, the position calculation unit 113 performs image processing on the surface image, to generate imaging-time position information on the basis of the surface image. The imaging-time position information indicates an imaging time when the subject was present in the surface image capturing area 12, and indicates a position where the subject was present at the imaging time. When the determination unit 112 determines that the subject is a battery-containing product, the output unit 114 outputs the imaging-time position information generated by the position calculation unit 113 to the robot control apparatus 106.
The robot control apparatus 106 is connected to the robot 104 so as to communicate information with each other. The robot control apparatus 106 controls the robot 104 to remove the subject appearing in the surface image and the internal image captured by the imaging control unit 111 from the conveyance path 6 on the basis of the imaging-time position information that is output from the object recognition control apparatus 105.
The object recognition control apparatus 105 controls the visible area image sensor 102 to capture a surface image in which waste that is conveyed along the conveyance path 6 and disposed in the surface image capturing area 12 appears. The object recognition control apparatus 105 controls the terahertz sensor 103 to capture an internal image in which the outer contour of waste that is conveyed along the conveyance path 6 and disposed in the surface image capturing area 12 and an internal object that is contained in the waste appear. Specifically, the object recognition control apparatus 105 captures the internal image such that a position where an internal view representing certain waste appears in the internal image matches a position where a surface view representing the certain waste appears in the surface image. In other words, the object recognition control apparatus 105 controls the terahertz sensor 103 to capture an internal image at the time that is delayed from the time when the surface image is captured by a delay time that is calculated on the basis of the conveyance speed and the distance between the surface image capturing area 12 and the internal image capturing area 14.
The object recognition control apparatus 105 generates integrated information on the basis of the surface image and the internal image, and determines whether the waste appearing in the surface image and the internal image is a battery-containing product on the basis of the integrated information. With this configuration, in the same or substantially the same manner as the object processing apparatus 1 according to Embodiment 1, the object processing apparatus 101 according to Embodiment 5 can reduce the frequency of erroneously determining that a battery-containing product is not a battery-containing product or erroneously determining that waste that is not a battery-containing product is a battery-containing product.
When the object recognition control apparatus 105 determines that the waste appearing in the surface image and the internal image is a battery-containing product, the object recognition control apparatus 105 performs image processing on the surface image, to calculate imaging-time position information. The imaging-time position information indicates an imaging time when the battery-containing product was present in the surface image capturing area 12, and indicates a position where the battery-containing product was present at the imaging time. When the object recognition control apparatus 105 determines that the waste appearing in the surface image and the internal image is a battery-containing product, the object recognition control apparatus 105 further outputs the calculated imaging-time position information to the robot control apparatus 106.
The robot control apparatus 106 calculates a removal time and a removal position on the basis of the conveyance speed and the imaging-time position information. The removal time indicates the time when the battery-containing product is present in the work area 15. The removal position indicates the position where the battery-containing product is present at the removal time. The robot control apparatus 106 controls the robot 104 on the basis of the removal time and the removal position, to remove the battery-containing product from the conveyance path 6. The waste sorting apparatus including the object processing apparatus 101 according to Embodiment 5 can reduce the frequency of erroneous determination. Thus, in the same or substantially the same manner as the waste sorting apparatus 2, the waste sorting apparatus including the object processing apparatus according to Embodiment 5 reduces the frequency with which the battery-containing product is supplied to the shredder, thus reducing the frequency of ignition in the shredder.
The description given above is of a case where the position calculation unit 39 of the object processing apparatus according to any of Embodiment 1 to 4 and the position calculation unit 113 of the object processing apparatus according to Embodiment 5 calculate the imaging-time position information on the basis of the surface image. Alternatively, the position calculation unit 39 and the position calculation unit 113 may calculate the imaging-time position information on the basis of the internal image. The object processing apparatus can appropriately calculate an area where the battery-containing product is present in the work area 15 even when the imaging-time position information is calculated on the basis of the internal image. The object processing apparatus thus configured does not have to include another means for detecting the position of the battery-containing product in the work area 15 in addition to the X-ray sensor 22 or the terahertz sensor 103. This reduces the number of components and the manufacturing cost.
The X-ray sensor 123 includes a light emitter 127 and a light receiver 128. The light emitter 127 is disposed above the floor 126. The light emitter 127 emits X-rays toward the floor 126. The light receiver 128 is disposed below the floor 126. The light receiver 128 receives transmitted X-rays transmitted through the multiple pieces of garbage on the floor 126 among X-rays emitted from the light emitter 127. On the basis of the transmitted X-rays received by the light receiver 128, the X-ray sensor 123 captures an internal image in which the outer contours of the multiple pieces of waste on the floor 126 and multiple internal objects that are contained in the multiple pieces of waste appear.
The projector 124 is disposed above the floor 126. The projector 124 performs projection mapping to project video onto an object that is present on the floor 126.
The object processing apparatus 121 according to Embodiment 6 is used for waste sorting to find a battery-containing product from multiple pieces of waste scattered on the floor 126. The control apparatus 125 controls the visible area image sensor 122 to capture a surface image in which the surfaces of the multiple pieces of waste on the floor 126 appear. The control apparatus 125 controls the X-ray sensor 123 to capture an internal image in which the outer contours of the multiple pieces of waste on the floor 126 and multiple internal object contained in the multiple pieces of waste appear. In the same or substantially the same manner as the object recognition control apparatus 31 of the object processing apparatus according to any of Embodiment 1 to 4 and the object recognition control apparatus 105 of the object processing apparatus according to Embodiment 5, the control apparatus 125 determines whether each of the multiple pieces of waste appearing in the surface image and the internal image is a battery-containing product on the basis of the surface image and the internal image. With this configuration, in the same or substantially the same manner as the object processing apparatus according to any of the above-described embodiments, the object processing apparatus 121 according to Embodiment 6 can reduce the frequency of erroneously determining that a battery-containing product is not a battery-containing product or erroneously determining that waste that is not a battery-containing product is a battery-containing product.
When multiple battery-containing products are included in the multiple pieces of waste appearing in the surface image and the internal image, the control apparatus 125 performs image processing on the surface image, to calculate multiple pieces of position information corresponding to the multiple battery-containing products. The position information corresponding to a certain battery-containing product among the multiple pieces of position information indicates a position where the certain battery-containing product is present. The control apparatus 125 projects a projection mapping image created on the basis of the multiple pieces of position information onto the multiple pieces of waste on the floor 126. The projection mapping image is created such that an area where the multiple battery-containing products are present among the multiple pieces of waste on the floor 126 is conspicuously visually recognizable.
A worker searches the multiple pieces of waste on the floor 126 for a battery-containing product. When the worker finds out a battery-containing product, the worker removes the found battery-containing product from the floor 126. The object processing apparatus 121 according to Embodiment 6 can make it easy for the worker to search for a battery-containing product by the projection mapping that makes the battery-containing product conspicuously visually recognizable, and thus can reduce the frequency with which the worker overlooks the battery-containing product.
The description given above is of a case where the object processing apparatus 121 according to Embodiment 6 includes the X-ray sensor 123. Alternatively, the object processing apparatus may include a terahertz sensor. In the same or substantially the same manner as the object processing apparatus 121 according to Embodiment 6, the object processing apparatus including the terahertz sensor can also reduce the frequency of erroneous determination. When the terahertz sensor is disposed above the floor 126, the object processing apparatus including the terahertz sensor does not have to include a part of the terahertz sensor below the floor 126. Accordingly, compared to the object processing apparatus 121 according to Embodiment 6, the object processing apparatus including the terahertz sensor can be installed in a simple manner.
The description given above is of a case where the object processing apparatus according to any of Embodiments 1 to 6 includes the projector 23, the projector 124, or the robot 104, which is a processing device according to any of the embodiments. Alternatively, the object processing apparatus may include a different processing device that operates on the basis of the determination result indicating whether waste is a battery-containing product. Examples of the different processing device include an audio device that outputs an alarm sound when the waste is determined as a battery-containing product, and a device (i.e., the object recognition control apparatus 31 itself including the output unit 40 or the object recognition control apparatus 105 itself including the output unit 114) that outputs information indicating that the waste is determined as a battery-containing product to another device.
A waste sorting apparatus according to the related art sometimes make an incorrect determination such as a determination that waste which is an object to be removed is not an object to be removed or a determination that waste which is not an object to be removed is an object to be removed.
An object processing apparatus, an object determination method, and a program according to one or more embodiments of the present disclosure can reduce the frequency of erroneously determining whether an object is a target.
The above-described embodiments are illustrative and do not limit the present disclosure. Thus, numerous additional modifications and variations are possible in light of the above teachings. For example, elements and/or features of different illustrative embodiments may be combined with each other and/or substituted for each other within the scope of the present disclosure. Any one of the above-described operations may be performed in various other ways, for example, in an order different from the one described above.
The functionality of the elements disclosed herein may be implemented using circuitry or processing circuitry which includes general purpose processors, special purpose processors, integrated circuits, application specific integrated circuits (ASICs), digital signal processors (DSPs), field programmable gate arrays (FPGAs), conventional circuitry and/or combinations thereof which are configured or programmed to perform the disclosed functionality. Processors are considered processing circuitry or circuitry as they include transistors and other circuitry therein. In the disclosure, the circuitry, units, or means are hardware that carry out or are programmed to perform the recited functionality. The hardware may be any hardware disclosed herein or otherwise known which is programmed or configured to carry out the recited functionality. When the hardware is a processor which may be considered a type of circuitry, the circuitry, means, or units are a combination of hardware and software, the software being used to configure the hardware and/or processor.
Number | Date | Country | Kind |
---|---|---|---|
2023-012012 | Jan 2023 | JP | national |