The present application claims priority from Japanese Patent Application JP 2023-024093 filed on Feb. 20, 2023, the content of which is hereby incorporated by reference into this application.
The present invention relates to a medical imaging apparatus that captures a medical image of a subject, and particularly, to a technique of specifying the subject from among a plurality of persons in a case where the plurality of persons are included in a camera image to be used to set an imaging position.
A medical imaging apparatus is an apparatus that detects a signal obtained from an imaging position of a subject, for example, X-rays transmitted through the subject, nuclear magnetic resonance signals generated from the subject, or the like, to capture a medical image to be used to diagnose the subject, or the like. In the medical imaging apparatus, the imaging position is set with respect to the subject on a patient table prior to the capturing of the medical image. It is desirable that the setting of the imaging position be automated.
In JP2021-6993A, it is disclosed that camera images are input into two trained independent deep learning models, and an imaging position is automatically set based on a prediction result obtained from each of the two.
However, JP2021-6993A lacks consideration for a case where a plurality of persons are included in the camera image. In a case where a plurality of persons are included in the camera image, it is difficult to set the imaging position with respect to the subject because it is unclear which person is the subject.
In that respect, an object of the present invention is to provide a medical imaging apparatus and a control method of the same for specifying a subject from among a plurality of persons included in a camera image.
In order to achieve the above-described object, according to an aspect of the present invention, there is provided a medical imaging apparatus that captures a medical image of a subject placed on a patient table, the medical imaging apparatus comprising: a camera image acquisition unit that acquires a camera image including the patient table; a person extraction unit that extracts a person from the camera image; and a subject specification unit that specifies the subject from among a plurality of the persons extracted by the person extraction unit, based on a distance between each of the plurality of persons and a reference position set based on the patient table.
In addition, according to another aspect of the present invention, there is provided a control method of a medical imaging apparatus that captures a medical image of a subject placed on a patient table, the control method comprising: a camera image acquisition step of acquiring a camera image including the patient table; a person extraction step of extracting a person from the camera image; and a subject specification step of specifying the subject from among a plurality of the persons extracted in the person extraction step, based on a distance between each of the plurality of persons and a reference position set based on the patient table.
According to the present invention, it is possible to provide a medical imaging apparatus and a control method of the same for specifying a subject from among a plurality of persons included in a camera image.
Hereinafter, examples of a medical imaging apparatus according to the present invention will be described with reference to the accompanying drawings. The medical imaging apparatus is an apparatus that detects a signal obtained from a subject, for example, X-rays transmitted through the subject, nuclear magnetic resonance signals generated from the subject, or the like, to capture a medical image to be used to diagnose the subject, or the like. Hereinafter, as an example of the medical imaging apparatus, an X-ray computed tomography (CT) apparatus that captures a tomographic image of the subject by acquiring X-ray projection images of the subject at various projection angles will be described.
An overall configuration of the X-ray CT apparatus of Example 1 will be described with reference to
The scan gantry unit 100 comprises an X-ray source 101, a rotating plate 102, a collimator 103, an X-ray detector 106, a data collection unit 107, a patient table 105, a rotating plate controller 108, a patient table controller 109, an X-ray controller 110, and a high-voltage generation unit 111. The X-ray source 101 is a device that irradiates a subject 10 placed on the patient table 105 with X-rays and is, for example, an X-ray tube device. The collimator 103 is a device that restricts an irradiation range of X-rays. The rotating plate 102 is provided with an opening portion 104 through which the subject 10 placed on the patient table 105 enters, and is also equipped with the X-ray source 101 and the X-ray detector 106 and rotates the X-ray source 101 and the X-ray detector 106 around the subject 10.
The X-ray detector 106 is a device that is disposed to face the X-ray source 101, that comprises a plurality of detection elements which detect X-rays transmitted through the subject 10, and that detects a spatial distribution of X-rays. The detection elements of the X-ray detector 106 are arranged two-dimensionally in a rotation direction and a rotation axis direction of the rotating plate 102. The data collection unit 107 is a device that collects the spatial distribution of X-rays detected by the X-ray detector 106 as digital data.
The rotating plate controller 108 is a device that controls rotation and inclination of the rotating plate 102. The patient table controller 109 is a device that controls up, down, front, back, left, and right movements of the patient table 105. The high-voltage generation unit 111 is a device that generates a high voltage applied to the X-ray source 101. The X-ray controller 110 is a device that controls an output of the high-voltage generation unit 111. The rotating plate controller 108, the patient table controller 109, and the X-ray controller 110 are, for example, a micro-processing unit (MPU) or the like.
The operation unit 120 comprises an input unit 121, an image generation unit 122, a display unit 125, a storage unit 123, and a system controller 124. The input unit 121 is a device that is used to input examination data such as a name of the subject 10, an examination date and time, and an imaging condition, and is, for example, a keyboard, a pointing device, a touch panel, or the like. The image generation unit 122 is a device that generates the tomographic image by using the digital data collected by the data collection unit 107, and is, for example, an MPU, a graphics processing unit (GPU), or the like. The display unit 125 is a device that displays the tomographic image or the like generated by the image generation unit 122, and is, for example, a liquid crystal display, a touch panel, or the like. The storage unit 123 is a device that stores the digital data collected by the data collection unit 107, the tomographic image generated by the image generation unit 122, a program to be executed by the system controller 124, data to be used by the program, and the like, and is, for example, a hard disk drive (HDD), a solid state drive (SSD), or the like. The system controller 124 is a device that controls each unit such as the rotating plate controller 108, the patient table controller 109, and the X-ray controller 110, and is, for example, a central processing unit (CPU).
The camera 130 is a device that images the subject 10 placed on the patient table 105 together with the patient table 105 from above, and is provided on a ceiling of the imaging room or above the scan gantry unit 100 such that the patient table 105 is located at a substantially center of a field of view of the camera 130. A camera image captured by the camera 130 is displayed on the display unit 125, and is used by an operator in the operation room to confirm a state of the subject 10 or is used to set the imaging position with respect to the subject 10. The camera image may be stored in the storage unit 123.
The high-voltage generation unit 111 generates a tube voltage, which is a high voltage applied to the X-ray source 101, based on the imaging condition set via the input unit 121, whereby X-rays corresponding to the imaging condition are emitted to the subject 10 from the X-ray source 101. The X-ray detector 106 detects the X-rays emitted from the X-ray source 101 and transmitted through the subject 10 with a large number of detection elements and acquires the spatial distribution of the transmitted X-rays. The rotating plate 102 is controlled by the rotating plate controller 108 and rotates based on the imaging condition input through the input unit 121, particularly a rotation speed or the like. The patient table 105 is controlled by the patient table controller 109 and moves relative to the rotating plate 102 to move the imaging position set with respect to the subject 10 to an imaging field of view, which is a range in which the transmitted X-rays are detected.
By repeating the irradiation of X-rays by the X-ray source 101 and the detection of X-rays by the X-ray detector 106 with the rotation of the rotating plate 102, projection data, which is the X-ray projection image of the subject 10, is measured at various projection angles. The projection data is associated with a view representing each projection angle, and a channel (ch) number and a column number which are detection element numbers of the X-ray detector 106. The measured projection data is transmitted to the image generation unit 122. The image generation unit 122 generates the tomographic image by performing back-projection processing on a plurality of pieces of projection data. The generated tomographic image is displayed on the display unit 125 or stored in the storage unit 123 as the medical image.
In order to capture the tomographic image, it is necessary to set the imaging position, which is a position at which the tomographic image is captured, with respect to the subject 10. In a case where a plurality of persons are included in the camera image when the camera image is used to set the imaging position, it is not possible to specify which person is the subject 10, which makes it difficult to set the imaging position with respect to the subject 10. In that respect, in Example 1, the subject 10 is specified from among the plurality of persons extracted from the camera image based on distances between the plurality of persons included in the camera image and a reference position set based on the patient table 105, thereby enabling setting of the imaging position using the camera image.
The functional blocks of Example I will be described with reference to
In Example 1, a camera image acquisition unit 201, a person extraction unit 202, and a subject specification unit 203 are provided. Hereinafter, each unit will be described.
The camera image acquisition unit 201 acquires the camera image captured by the camera 130. The acquired camera image is a digitized image and is a frame image in a video. The camera image is transmitted from the camera 130 or read out from the storage unit 123.
The person extraction unit 202 extracts a person from the camera image acquired by the camera image acquisition unit 201. The camera image acquisition unit 201 may be configured with, for example, a trained deep learning model. In a case where the camera image includes a plurality of persons, the person extraction unit 202 extracts each of the plurality of persons. In addition, the person extraction unit 202 may detect feature points of each person to be extracted, for example, eyes, nose, ears, shoulders, elbows, wrists, waist, knees, ankles, and the like.
The subject specification unit 203 specifies the subject 10 from among the plurality of persons extracted by the person extraction unit 202. For the specification of the subject 10, the distances between the plurality of persons extracted by the person extraction unit 202 and the reference position set based on the patient table 105 are used. That is, since the subject 10 is placed on the patient table 105, for example, a person having a shortest distance to a center of the patient table 105 is specified as the subject 10.
An example of a flow of processing of Example 1 will be described step by step with reference to
The camera image acquisition unit 201 acquires the camera image including the patient table 105. The camera image acquired in S301 is a frame image in a video, and may be a camera image transmitted from the camera 130 or a camera image read out from the storage unit 123.
The person extraction unit 202 determines whether or not a person has been extracted from the camera image acquired in S301. The processing returns to S301 via S303 in a case where no person has been extracted from the camera image, and the processing proceeds to S304 in a case where the person has been extracted.
The camera image acquisition unit 201 updates the frame image in the video to the next frame.
The person extraction unit 202 detects the feature points of the person extracted in S302. The detection of the feature points may be based on pose estimation.
The subject specification unit 203 selects one of the plurality of persons based on the distances between the plurality of persons extracted in S302 and the reference position. In a case where the person extracted in S302 is one person, that person is selected.
An example of the flow of the processing of S305 will be described step by step with reference to
The subject specification unit 203 acquires the reference position. The reference position is set based on the patient table 105. Since the patient table 105 is located at the substantially center of the camera image, a position of a center line or a center point of the camera image may be set as the reference position. For example, as shown in
Further, as shown in
The subject specification unit 203 calculates a distance between the reference position acquired in S601 and each person extracted in S302.
As shown in
Respective distances between the plurality of feature points of each person and the camera image's center line 700 may be calculated to use an average value of the calculated distances as the distance to the reference position of each person. For example, in a case where the coordinates of three feature points of a certain person are (X1,Z1), (X2,Z2), (X3,Z3) and an X coordinate of the camera image's center line 700 is X0, respective distances from the three feature points to the camera image's center line 700 are X1−X0, X2−X0, and X3−X0. Therefore, the distance of the person to the reference position is (X1+X2+X3−3*X0)/3, which is the average value of the three distances.
In addition, each person's center line may be calculated based on a region of interest designated in advance. An examination region may be designated as the region of interest. As shown in
As shown in
As shown in
The subject specification unit 203 selects a person having the shortest distance calculated in S602. By selecting the person having the shortest distance to the reference position, it becomes easier to specify the subject 10 placed on the patient table 105. In any of the examples of
Through the flow of the processing described with reference to
The subject specification unit 203 determines whether or not the person selected in S305 satisfies a condition of the subject 10. The processing returns to S301 via S303 in a case where the person does not satisfy the condition of the subject 10, and the processing proceeds to S307 in a case where the condition is satisfied.
Whether or not the person satisfies the condition of the subject 10 is determined, for example, by whether or not the person is within the region of the patient table 105. That is, determination is made that the condition of the subject 10 is satisfied in a case where the person is within the region of the patient table 105.
In addition, whether or not the condition of the subject 10 is satisfied may be determined based on a motion between frames. That is, while the subject 10 is kept stationary on the patient table 105, the radiologist who assists the subject 10 or prepares for imaging moves around the patient table 105. Therefore, determination is made that the condition of the subject 10 is satisfied in a case where the motion between frames is less than a predetermined threshold value. The motion between frames is calculated based on, for example, a difference value between adjacent frame images in the video.
The subject specification unit 203 specifies the person determined in S306 to satisfy the condition of the subject 10 as the subject 10. The imaging position is set for the person specified as the subject 10 based on, for example, the camera image. In addition, the person specified as the subject 10 may be displayed on the display unit 125. By displaying a specification result on the display unit 125, the operator can confirm whether or not the specified person is appropriate. In a case where the specified person is not appropriate as a result of the confirmation by the operator, the processing may return to S301 via S303.
The camera image acquisition unit 201 determines whether or not the camera image acquired in S301 is a last frame. The processing returns to S301 via S303 in a case where the camera image is not the last frame, and the flow of the processing ends in a case where it is the last frame.
Through the flow of the processing described with reference to
It should be noted that the flow of the processing of Example 1 is not limited to the flow of the processing illustrated in
The plurality of embodiments of the present invention have been described above. The present invention is not limited to the above-described embodiments, and the components can be modified and embodied without departing from the gist of the invention. Additionally, a plurality of components disclosed in the above-described embodiments may be combined as appropriate. Furthermore, some components may be deleted from all the components described in the above-described embodiments.
Number | Date | Country | Kind |
---|---|---|---|
2023-024093 | Feb 2023 | JP | national |