This application claims priority to Japanese Patent Application No. 2023-222807, which was filed on Dec. 28, 2023, and which is hereby incorporated by reference herein in its entirety.
The present disclosure relates to a processing apparatus, an imaging system, a processing method, and a storage medium.
Recently, virtual productions, in which a background image is displayed on an LED display (LED wall) and simultaneously imaged with a real object, have rapidly become popular. Any tiny moirés (interference fringes) on an LED display may not be noticed during imaging but may be noticed in post-production and cause reimaging, so moiré may be detected during imaging. Japanese Patent Laid-Open No. 2008-011334 discloses a method for converting a spatial axis signal representing an image containing moiré into a frequency axis signal, and then converting a frequency axis signal from which a frequency component corresponding to the moiré is removed, into a spatial axis signal.
The method disclosed in Japanese Patent Laid-Open No. 2008-011334 identifies moiré using a plurality of images captured at different imaging angles, and thus is to prepare a camera for moiré detection, which is separate from the camera for imaging.
A processing apparatus according to one aspect of the disclosure is configured to cause a display apparatus to display a first image, and to acquire from an image pickup apparatus a second image made by capturing at an angle of view including a display area of the display apparatus. The processing apparatus includes a processor configured to acquire a distinction result of distinguishing between an area in which an object existing between the display area and the image pickup apparatus has been captured, and the display area in the second image, detect a change in the second image by comparing the first image with an image of the display area in the second image, and perform first control for notifying information regarding a detection result. Alternatively, the processing apparatus includes a processor configured to detect whether or not moiré has occurred in the display area of the second image, and record on a recording medium an image in which the moiré has been corrected and information regarding an area where the moiré has been detected in a case where the moiré has been detected. An imaging system having the above processing apparatus also constitutes another aspect of the disclosure. A processing method corresponding to the above processing apparatus also constitutes another aspect of the disclosure. A storage medium storing a program that causes a computer to execute the above processing method also constitutes another aspect of the disclosure.
Further features of the disclosure will become apparent from the following description of embodiments with reference to the attached drawings.
In the following, the term “unit” may refer to a software context, a hardware context, or a combination of software and hardware contexts. In the software context, the term “unit” refers to a functionality, an application, a software module, a function, a routine, a set of instructions, or a program that can be executed by a programmable processor such as a microprocessor, a central processing unit (CPU), or a specially designed programmable device or controller. A memory contains instructions or programs that, when executed by the CPU, cause the CPU to perform operations corresponding to units or functions. In the hardware context, the term “unit” refers to a hardware element, a circuit, an assembly, a physical structure, a system, a module, or a subsystem. Depending on the specific embodiment, the term “unit” may include mechanical, optical, or electrical components, or any combination of them. The term “unit” may include active (e.g., transistors) or passive (e.g., capacitor) components. The term “unit” may include semiconductor devices having a substrate and other layers of materials having various concentrations of conductivity. It may include a CPU or a programmable processor that can execute a program stored in a memory to perform specified functions. The term “unit” may include logic elements (e.g., AND, OR) implemented by transistor circuits or any other switching circuits. In the combination of software and hardware contexts, the term “unit” or “circuit” refers to any combination of the software and hardware contexts as described above. In addition, the term “element,” “assembly,” “component,” or “device” may also refer to “circuit” with or without integration with packaging materials.
Referring now to the accompanying drawings, a detailed description will be given of embodiments according to the disclosure. Corresponding elements in respective figures will be designated by the same reference numerals, and a duplicate description thereof will be omitted.
A lens unit 106 includes lens units including a zoom lens and a focus lens, an aperture mechanism, and a drive motor. An optical image passing through the lens unit 106 is received by an imaging unit 107. The imaging unit 107 uses a CCD or CMOS sensor, etc., and converts an optical signal into an electrical signal.
A CPU 102 controls each component in the image pickup apparatus 100 using a RAM 104 as a work memory according to a program stored in a ROM 103. The ROM 103 is a nonvolatile recording device, and records a program for operating the CPU 102, various adjustment parameters, etc. The RAM 104 is a volatile memory using a semiconductor device, is generally slower than a frame memory 111, and has a smaller capacity than that of a frame memory 111. The frame memory 111 is a device that temporarily stores an image signal and can read it out if necessary. Since image signals are a huge amount of data, the frame memory 111 is to have a high bandwidth and large capacity. Dual Data Rate 4-Synchronous Dynamic RAM (DDR4-SDRAM) and the like have recently been used as the frame memory 111. Using the frame memory 111 can perform processing such as combining images that differ over time and cut out only a necessary area.
An image processing unit 105 performs various image processing for data from the imaging unit 107 or image data stored in the frame memory 111 or recording medium 112 under the control of the CPU 102. The image processing performed by the image processing unit 105 includes pixel interpolation, coding processing, compression processing, decoding processing, enlargement/reduction processing (resizing), noise reduction processing, color conversion processing, and the like of image data. The image processing unit 105 performs processing such as correction of performance variations of pixels of the imaging unit 107, correction of defective pixels, white balance correction, luminance correction, and correction of distortion and peripheral light loss caused by lens characteristics. The image processing unit 105 performs processing of generating distance information regarding a distance from the image pickup apparatus 100 to the object to be imaged. The image processing unit 105 may be configured as a dedicated circuit block for performing specific image processing. Depending on the type of image processing, the CPU 102 may perform image processing according to a program without using the image processing unit 105.
The CPU 102 controls the lens unit 106 based on the calculation result acquired by the image processing unit 105, and performs optical image enlargement, focal length adjustment, and adjustment of an aperture stop configured to adjust a light amount. The CPU 102 may perform image stabilization by moving a part of the lens unit on a plane perpendicular to the optical axis.
An operation unit 113 is an interface with the external device that accepts user operations. The operation unit 113 can use mechanical elements such as buttons and switches, and includes a power switch, a mode switch, etc.
A display unit 114 serves as a display device that can be visually confirmed by the user. When the display unit 114 displays, for example, images processed by the image processing unit 105 and setting menus, the user can check the operation status of the image pickup apparatus 100. Small and low-power devices such as Liquid Crystal Displays (LCDs) and organic Electroluminescence (EL) have recently been used as the display unit 114. The display unit 114 may include a resistive or capacitive thin film device called a touch panel, and may be used as a substitute for the operation unit 113. The CPU 102 generates character strings for informing the user of the setting status of the image pickup apparatus 100 and a menu for setting the image pickup apparatus 100, and displays them on the display unit 114 by superimposing them on the image processed by the image processing unit 105. In addition to character information, the CPU 102 can further superimpose another imaging assist display such as a histogram, vector scope, waveform monitor, zebra, peaking, and false color.
An image (or video) terminal unit 109 includes a plurality of image terminals. Examples of the image terminals are Serial Digital Interface (SDI), High Definition Multimedia Interface (HDMI) (registered trademark), and DisplayPort (registered trademark). Outputting an image signal via the image terminal unit 109 can display an image in real time on an unillustrated external monitor. The CPU 102 can process an image input via the image terminal unit 109 in the image processing unit 105 and display the processed image on the display unit 114.
A network module 108 is an interface for inputting and outputting image and audio signals. The network module 108 can communicate with external devices via the Internet, etc., and can also send and receive files, commands, and various types of data such as image signals and metadata. The transmission method that is used by the network module 108 can be wireless or wired.
The recording medium 112 can record image data and various setting data, and use a large-capacity storage element. For example, a Hard Disc Drive (HDD) or Solid State Drive (SSD) is used as the recording medium 112, and is attached to a recording medium interface (I/F) 110. In a case where the user presses a predetermined button (hereinafter referred to as a record button) on the operation unit 113, the CPU 102 starts recording the image data processed by the image processing unit 105 into the recording medium 112 via the recording medium I/F 110. In a case where the user presses the record button again, the recording is stopped. During recording, a signal indicating that the image pickup apparatus 100 is recording, is output via the network module 108 and the image terminal unit 109. This configuration can notify an external device that recording is in progress, or record image data output from the image pickup apparatus 100 in the external device in conjunction with a recording operation on the image pickup apparatus 100.
An object detector 115 detects objects using artificial intelligence, such as deep learning using a neural network. In the case of object detection using deep learning, the CPU 102 transmits a program for processing, a network structure, weight parameters, and the like stored in the ROM 103 to the object detector 115. The object detector 115 performs processing to detect an object from the image signal based on various parameters acquired from the CPU 102, and loads the processing result in the RAM 104.
An attitude (or orientation) detector 116 detects the attitude state of the image pickup apparatus 100 using, for example, a gyro sensor or an acceleration sensor. This configuration can detect whether the camera is tilted or shaking.
The image sensor according to this embodiment can output two signals for phase-difference detection (also referred to as A-image signal and B-image signal hereinafter) from each pixel unit. It can also output an imaging signal (A-image signal+B-image signal) acquired by adding the signals of two photodiodes. In a case where the added signal is output, a signal equivalent to the output of the image sensor of the example Bayer array illustrated in
The imaging unit 107 can output a phase-difference detection signal for each pixel unit, but can also output a value acquired by adding and averaging the phase-difference detection signals of a plurality of adjacent pixel units. Outputting the added and averaged value can reduce the time to read signals from the imaging unit 107 and the bandwidth that is used by the internal bus 101. Using such output signals from the imaging unit 107, the CPU 102 performs a correlation calculation of the two image signals and calculates information such as a defocus amount, parallax information, and various types of reliability.
A defocus amount on the image plane is calculated based on a shift between the A-image signal and the B-image signal. The defocus amount has a positive or negative value, and whether the defocus amount is a positive or negative value determines whether the state is a front focus state or a rear focus state. The absolute value of the defocus amount provides information regarding the in-focus degree (degree of focus shift), and a defocus amount of 0 represents an in-focus state. That is, the CPU 102 calculates information on whether the image is in front or rear focus based on the positive or negative sign of the defocus amount. The CPU 102 also calculates in-focus degree information based on the absolute value of the defocus amount. The information on whether the image is in front or rear focus is output in a case where the defocus amount exceeds a predetermined value, and in a case where the absolute value of the defocus amount is within a predetermined value, information indicating that the image is in focus is output.
The CPU 102 controls the lens unit 106 according to a defocus amount for focusing. The CPU 102 also calculates a distance to an object using the principle of triangulation from parallax information and the lens information regarding the lens unit 106.
Although
A description will now be given of the distance information generating processing performed by the image processing unit 105.
In step S301, the image processing unit 105 calculates a B-image signal for phase-difference detection by calculating a difference between two signals, the (A-image signal+B-image signal) for imaging output from the imaging unit 107 and the A-image signal for phase-difference detection. This embodiment uses a method in which the (A-image signal+B-image signal) for imaging and the A-image signal for phase-difference detection are output, but the present disclosure is not limited to this example. The A-image signal and the B-image signal may be output in this state from the imaging unit 107. In this case, the (A-image signal +B-image signal) for imaging can be calculated by adding the A-image signal and the B-image signal. In a case where two sensors are provided, such as a stereoscopic camera, the image signals output from the respective image sensors may be the A-image signal and the B-image signal.
In step S302, the image processing unit 105 corrects shading caused by optical factors for each of the A-image signal for phase-difference detection and the B-image signal for phase-difference detection.
In step S303, the image processing unit 105 performs filtering processing for each of the A-image signal for phase-difference detection and the B-image signal for phase-difference detection. The filtering processing may be performed, for example, by a high-pass filter including a Finite Impulse Response (FIR) filter. In this embodiment, the A-image signal for phase-difference detection and the B-image signal for phase-difference detection are passed through the high-pass filter, but the present disclosure is not limited to this example. Signals may be generated through a band-pass filter or a low-pass filter with different filter coefficients. Then, correlation calculation processing, which will be described later, may be performed using the A-image signal for the generated phase-difference detection and the B-image signal for the generated phase-difference detection.
In step S304, the image processing unit 105 performs correlation calculation using divided fine blocks for the A-image signal for phase-difference detection and the B-image signal for phase-difference detection, for which the filtering processing has been performed in step S303. The size or shape of the fine blocks is not limited, and adjacent blocks may overlap each other.
A description will now be given of the correlation calculation of a pair of images, the A-image and the B-image. A signal sequence of the A-image at the target pixel position is denoted as E(1) to E(m), and a signal sequence of the B-image at the target pixel position is denoted as F(1) to F(m). While the signal sequence F(1) to F(m) of the B-image is shifted from the signal sequence E(1) to E(m) of the A-image, a correlation amount C(k) at a shift amount k between the two signal sequences is calculated using the following equation (1).
In equation (1), the Σ operation means an operation to calculate the sum for n. In the Σ operation, a range of n and n+k is limited to a range from 1 to m. A shift amount k is an integer value, and is a relative pixel shift amount in units of the detection pitch of a pair of data. Hereinafter, k where a discrete correlation amount C(k) is minimum is represented as kj. In an ideal state having no noise,
SLOP in equation (4) represents a minimum and local minimum correlation amount and a slope of a change in the adjacent correlation amount. In
In
In a case where the correlation between the pair of image signal strings is low, a change amount of the correlation amount C(k) is small and a correlation amount curve is generally flat, so a SLOP value is small. Similarly, in a case where an object image has low contrast, the correlation between the pair of image signal strings is low and a correlation amount curve has a flat shape. Based on this property, the reliability of the calculated pixel shift amount x can be determined by the SLOP value. That is, in a case where the SLOP value is large, it can be determined that the correlation between the pair of image signal strings is high, and in a case where the SLOP value is small, it can be determined that no significant correlation has been acquired between the pair of image signal strings. This embodiment uses equation (1) for the correlation calculation, and thus the correlation amount C(k) is smallest and locally minimum at the shift amount where the correlation between the pair of image signal strings is highest. As another method, a correlation calculation method may be used where the correlation amount C(k) is largest and locally maximum at the shift amount where the correlation between the pair of image signal strings is highest.
In step S305, the image processing unit 105 calculates the reliability. The reliability can be defined by C(kj), which indicates the degree of coincidence between the two images calculated in step S304, as described above, and a value of SLOP.
In step S306, the image processing unit 105 performs interpolation processing. The pixel shift amount calculated in step S304 may not be able to be used because the reliability calculated in step S305 is low. In such cases, there is no need to perform interpolation using the pixel shift amount calculated around it. The interpolation method may apply a median filter or reduce pixel shift amount data and then enlarge it again. Color data may be extracted from the (A-image signal+B-image signal) for imaging, and a pixel shift amount may be interpolated using the color data.
In step S307, the image processing unit 105 calculates a defocus amount by referring to a pixel shift amount x calculated in step S304. More specifically, a defocus amount (denoted as DEF) can be acquired by the following equation (5):
In equation (5), P is a conversion coefficient determined by the detection pitch (pixel arrangement pitch) and a distance between projection centers of two left and right viewpoints in a pair of parallax images, and is expressed in mm/pixel.
In step S308, the image processing unit 105 calculates a distance from the defocus amount calculated in step S307. Where Da is a distance to an object, Db is a focus position, and F is a focal length, the following equation (6) approximately holds:
Therefore, the distance Da to the object is expressed by the following equation (7):
In a case where Db is Db0 where DEF-0, the absolute distance Da′ to the object is expressed by the following equation (8):
The relative distance Da-Da′ is expressed by the following equation (9) from equations (7) and (8):
As described above, performing processing according to the flow in
A CPU 702 controls each component of the image signal processing apparatus 700 according to a program stored in a ROM 703, using a RAM 704 as a work memory. The ROM 703 is a nonvolatile storage device that stores a program for operating the CPU 702 and various adjustment parameters. The RAM 704 is a volatile memory that uses a semiconductor device, and is generally slower and has a smaller capacity than a frame memory 709. The frame memory 709 is a device that temporarily stores an image signal and can read it out if necessary. Since image signals are a huge amount of data, the frame memory 709 is to have a high bandwidth and large capacity. Recently, DDR4-SDRAM and the like are used as the frame memory 709. Using the frame memory 709 can perform processing such as combining images that differ over time and cut out only a required area.
An image processing unit 705 performs various image processing for image data stored in the frame memory 709 or the recording medium 712 under control of the CPU 702. The image processing unit 705 may include a dedicated circuit block for performing specific image processing. Depending on the type of image processing, the CPU 702 can perform image processing according to a program without using the image processing unit 705.
The operation unit 710 is an interface with the external device that accepts user operations. The operation unit 710 includes a mouse, a keyboard, a touch panel, or the like.
A display unit 711 serves as a display device that can be visually confirmed by the user. When the display unit 711 displays, for example, images processed by the image processing unit 705, setting menus, and the like, the user can check the operation status of the image signal processing apparatus 700. Small and low-power devices such as LCDs and organic ELs have recently been used as the display unit 711. The display unit 711 may also include a resistive film type or capacitive film element called a touch panel, and may be used as a substitute for the operation unit 710. The CPU 702 generates character strings for informing the user of the setting status of the image signal processing apparatus 700 and a menu for setting the image signal processing apparatus 700, and displays them on the display unit 711 by superimposing them on the image processed by the image processing unit 705.
An image terminal unit 707 includes a plurality of image terminals. Examples of the image terminal include SDI, HDMI, and DisplayPort. Outputting an image signal via the image terminal unit 707 can display an image in real time on an unillustrated external monitor. The CPU 702 can process an image input via the image terminal unit 707 in the image processing unit 705 and display the processed image on the display unit 711.
A network module 706 is an interface for inputting and outputting image signals and audio signals. The network module 706 can communicate with external devices via the Internet etc., and can also send and receive files, commands, and various types of data such as image signals and metadata. The transmission method that is used by the network module 706 may be wireless or wired.
The recording medium 712 can record image data and various setting data, and use a large-capacity storage element. For example, a HDD or SSD is used as the recording medium 712, and is attached to the recording medium I/F 708. In a case where the user presses a predetermined button (referred to as a record button hereinafter) on the operation unit 710, the CPU 702 starts recording the image data processed by the image processing unit 705 or the image data input from the image terminal unit 707 into the recording medium 712 via the recording medium I/F 708. In a case where the user presses the record button again, the recording is stopped. The CPU 702 can also record image data into the recording medium 712 according to a signal indicating that recording is in progress, which is received via the network module 706 or the image terminal unit 707.
The image signal processing apparatus 700 also acquires lens information such as a focal length of the image pickup apparatus 100, information such as exposure, etc. via the network module 706 and the image terminal unit 707. The CPU 702 reads out three-dimensional model data recorded on the recording medium I/F 708. Next, the CPU 702 re-renders the three-dimensional model data in the image processing unit 705 based on the lens information, exposure, etc. of the image pickup apparatus 100, and the position and attitude information regarding the image pickup apparatus 100 acquired from the coordinate detection apparatus 601. Then, the CPU 702 generates a CG image suitable for the angle of view of the image pickup apparatus 100, and outputs it to the display apparatus 300 as a background image.
Referring now to
In step S801, the CPU 702 first reads out three-dimensional model data recorded in the recording medium I/F 708. Next, the CPU 702 re-renders the three-dimensional model data in the image processing unit 705 based on the lens information, information regarding exposure, etc. of the image pickup apparatus 100, and position and attitude information regarding the image pickup apparatus 100 acquired from the coordinate detection apparatus 601. Then, the CPU 702 generates a CG image that matches the angle of view of the image pickup apparatus 100, and outputs it to the display apparatus 300 as a background image (first image). The background image is displayed on the display apparatus 300.
In step S802, the CPU 102 acquires image data (second image) including an object and background image captured by the imaging unit 107 and processed by the image processing unit 105, and distance information, and stores them in the frame memory 111.
In step S803, the CPU 102 first reads out the image data and distance information from the frame memory 111. Next, the CPU 102 determines an object area including the object in the image data and a background area (display area) including the background image, as illustrated in
In this embodiment, the CPU 102 distinguishes between the object area and the background area, but the CPU 702 may acquire image data and distance information from the image pickup apparatus 100 and distinguish between the object area and the background area.
In step S804, the CPU 702 acquires the flags indicating the object area and background area and image data output by the image pickup apparatus 100 via the image terminal unit 707 and the network module 706. That is, the CPU 702 functions as an acquiring unit for acquiring flags indicating the object area and background area as a result of distinguishing between the object area and the background area. In a case where the resolution or imaging angle of the background image output in step S801 differs from that of the image data output by the image pickup apparatus in step S803, the CPU 702 performs resizing processing or angle correction processing for the background image to match the resolution and imaging angle. Then, the CPU 702 compares the image data of the portions corresponding to the background areas. In this embodiment, the CPU 702 functions as a detector configured to compare the image data of the portions corresponding to the background areas of the images and detects a difference (first difference).
In step S805, the CPU 702 determines whether the difference (first difference) between pixels in the areas compared in step S804 is equal to or greater than a predetermined value (first predetermined value or greater). In this embodiment, the CPU 702 determines whether there is a pixel whose difference is equal to or greater than a predetermined value. In a case where the CPU 702 determines that the difference between pixels is equal to or greater than the predetermined value, the CPU 702 executes the processing of step S806. In a case where the CPU 702 determines that the difference between pixels is not equal to or greater than the predetermined value, the CPU 702 ends this flow. This embodiment makes the determination based on whether there is a pixel in which the difference between pixels in the compared areas is equal to or greater than the predetermined value, but the present disclosure is not limited to this example. For example, the processing of step S806 may be executed in a case where an average value of the differences between pixels is equal to or greater than a predetermined value, or in a case where the number of pixels in which the difference between pixels is equal to or greater than a predetermined value is equal to or greater than a predetermined number of pixels.
In step S806, the CPU 702 determines whether the image data is being recorded. In a case where the CPU 702 determines that the image data is being recorded, it executes the processing of step S807, and in a case where the CPU 702 determines that the image data is not being recorded, it executes the processing of step S808.
In step S807, the image processing unit 705 generates a corrected image of the image data input from the image pickup apparatus 100 under the control of the CPU 702. More specifically, the CPU 702 first performs Fast Fourier Transform (FFT) for the background image output in step S801 and for the portion of the image data output by the image pickup apparatus in step S803 that corresponds to the background area. Thereby, a spatial axis signal can be converted into a frequency axis signal. Then, the respective frequency components are compared, and a signal of the frequency component present only in the image data input from the image pickup apparatus 100 is determined to be moiré, and signal processing is performed to remove that frequency component (so that the frequency component is smaller than a first predetermined amount). One method for removing a specific frequency component can use, for example, filtering processing using a notch filter. Then, by performing an inverse FFT for the frequency axis signal from which the frequency component of the moiré has been removed and converting it into a spatial axis signal, image data from which the moiré has been removed can be acquired.
In step S808, the CPU 702 generates an alert (or warning) display notifying the user that moiré or the like has occurred and an unintended background image has been captured, and outputs the alert to the image pickup apparatus 100 and the display apparatus 300 via the image terminal unit 707. Then, an alert is displayed on the display unit 114 of the image pickup apparatus 100, the display apparatus 300, and the display unit 711 of the image signal processing apparatus 700, and is conveyed to the user. The alert display input to the image pickup apparatus 100 may be output from the image terminal unit 109, and the alert may be displayed on an unillustrated external monitor. In this embodiment, the CPU 702 functions as a control unit configured to perform control to notify the user that the difference in image data in the portions corresponding to the background areas is large, moiré has occurred, and an unintended background image has been captured.
Thus, the CPU 702 displays an alert on the display unit in this embodiment, but the present disclosure is not limited to this embodiment as long as it is possible to notify the user that moiré has occurred and an unintended background image has been captured. For example, control to notify the user by vibration or sound may be performed.
In step S809, in a case where the resolution or imaging angle of the background image output in step S801 differs from that of the corrected image generated in step S807, the CPU 702 performs resizing processing and angle correction processing for the background image to match the resolution and imaging angle. Then, the CPU 702 compares the image data of the portions corresponding to the background areas. The CPU 702 determines whether a correction is possible according to a difference between pixels in the compared areas. In a case where the CPU 702 determines that the difference between pixels in the compared areas is equal to or greater than a predetermined value (a second predetermined value or greater), that is, in a case where correction is impossible, the CPU 702 executes the processing of step S810. In a case where the CPU 702 determines that the difference between pixels in the compared areas is less than a predetermined value (less than a second predetermined value), that is, in a case where it determines that correction is possible, it executes the processing of step S811. This embodiment determines whether the difference between pixels in the compared areas is equal to or greater than a predetermined value, but the present disclosure is not limited to this embodiment. For example, the processing of step S810 may be performed in a case where an average value of the differences between pixels is equal to or greater than a predetermined value, or in a case where the number of pixels whose differences between pixels are equal to or greater than a predetermined value is equal to or greater than a predetermined number.
In step S810, the CPU 702 generates an alert display notifying the user that moiré or the like has occurred and an unintended background image has been captured, and outputs the alert to the image pickup apparatus 100 via the image terminal unit 707. Then, an alert is displayed on the display unit 114 of the image pickup apparatus 100 and the display unit 711 of the image signal processing apparatus 700, and the user is informed not to affect the recorded image, and the image data input from the image pickup apparatus 100 is recorded in the recording medium I/F 708. The alert display input to the image pickup apparatus 100 may be output from the image terminal unit 109, and the alert may be displayed on an unillustrated external monitor.
In step S811, the CPU 702 performs processing to record the corrected image generated by the image processing unit 705 in step S807 in the recording medium I/F 708, and to output it via the image terminal unit 707.
In step S812, the CPU 702 performs processing to record data indicating the area corrected in step S807 and a time code indicating the time or frame when the correction was made as metadata in the recording medium I/F 708, and to output it via the image terminal unit 707.
This embodiment determines in step S809 whether a correction is possible, and the corrected image is recorded and output in step S811, but the present disclosure is not limited to this embodiment. It is also possible to perform the processing of step S810 without determining whether correction is possible in step S809, and to correct the background area in a post-processing step after imaging using the recorded captured image, the metadata recorded in step S812, and the background image output in step S801. At this time, the CPU 702 associates the corrected captured image with metadata indicating that moiré has been corrected and records them in the recording medium 712. In addition, the CPU 702 may output and record metadata indicating that moiré has occurred in the display area of the display apparatus 300 together with the captured image into the recording medium 712, without performing correction to remove moiré in step S807, in order to perform correction later.
In this embodiment, a method for removing moiré converts image data into a frequency axis signal, and identifies and removes the frequency components of moiré, but the present disclosure is not limited to this embodiment. For example, moiré may be removed by generating image data by converting the background image output in step S801 according to the angle of view of the image pickup apparatus 100, to match the resolution and imaging angle of the image data output by the image pickup apparatus 100, and by replacing the background area of the image data input from the image pickup apparatus 100.
This embodiment uses distance information calculated using the A-image signal and B-image signal acquired from the imaging unit 107 as a means for separating the object area from the background area, but the present disclosure is not limited to this embodiment. The distance to the object may be acquired by using another measure such as a distance sensor, or the object area and the background area may be separated using an image segmentation technology without using the distance information.
Performing the above processing can detect and remove moiré that occurs during imaging while a background image is displayed on the display unit. In a case where there is a difference of a predetermined value or more between the background image output by the image signal processing apparatus 700 to the display apparatus 300 and the background image displayed on the display apparatus 300 and captured by the image pickup apparatus 100, not only moiré but also changes in luminance, color, distortion, pixel/area defects, etc. can be detected and removed. The method of detecting such changes is not limited to the above method of calculating the difference, and any calculation method can be used as long as it can compare the background image output to the display apparatus 300 and the background image captured by the image pickup apparatus 100 for each area and pixel, such as a ratio.
Moiré may be estimated from optical design information regarding the image pickup apparatus 100 (such as the pixel pitch of the image sensor, focus, zoom, and object distance) and the pixel pitch information regarding the display apparatus 300, in addition to the moiré detecting method in this embodiment.
In the first embodiment, moiré is removed by generating a corrected image of image data input from the image pickup apparatus 100.
Moiré may occur during imaging due to interference with a repetitive pattern contained in the image displayed as the background image. In such a case, moiré can be suppressed by correcting the background image to be displayed. Accordingly, this embodiment removes moiré by correcting and outputting the background image output by the image signal processing apparatus 700.
In step S1001, the image processing unit 705 generates a corrected image of the background image output in step S801 under the control of the CPU 702, and outputs the corrected image to the display apparatus 300. More specifically, for example, an image in which high-frequency components are reduced (high-frequency components are smaller than a second predetermined amount) by applying a low-pass filter to the background image output in step S801 is generated, and output to the display apparatus 300.
In step S1002, in a case where the resolution or imaging angle of the background image output in step S1001 differs from that of the image data captured in step S802, the CPU 702 performs resizing processing or angle correction processing for the background image to match the resolution and imaging angle. Then, the CPU 102 compares the image data of the portions corresponding to the background areas. In a case where the CPU 702 determines that a difference between pixels in the compared areas is equal to or greater than a predetermined value (a third predetermined value or greater), it executes the processing of step S810. In a case where the CPU 702 determines that the difference between pixels in the compared areas is less than a predetermined value (a third predetermined value or less), it executes the processing of step S1003. This embodiment determines whether the difference between pixels in the compared areas is equal to or greater than a predetermined value, but the present disclosure is not limited to this embodiment. For example, the processing of step S810 may be executed in a case where an average value of the differences between pixels is equal to or greater than a predetermined value, or in a case where the number of pixels whose difference between pixels is equal to or greater than a predetermined value is equal to or greater than a predetermined number of pixels.
In step S1003, the CPU 102 captures the background image and object output to the display apparatus 300 in step S1001, and records them in the recording medium I/F 110. The CPU 102 may output the captured image data via the image terminal unit 109 and record it in the recording medium 712 of the image signal processing apparatus 700.
Performing the above processing can detect and remove abnormalities such as moiré that occur during imaging while a background image is displayed on a display unit.
Embodiment(s) of the disclosure can also be realized by a computer of a system or apparatus that reads out and executes computer-executable instructions (e.g., one or more programs) recorded on a storage medium (which may also be referred to more fully as a ‘non-transitory computer-readable storage medium’) to perform the functions of one or more of the above-described embodiment(s) and/or that includes one or more circuits (e.g., application specific integrated circuit (ASIC)) for performing the functions of one or more of the above-described embodiment(s), and by a method performed by the computer of the system or apparatus by, for example, reading out and executing the computer-executable instructions from the storage medium to perform the functions of one or more of the above-described embodiment(s) and/or controlling the one or more circuits to perform the functions of one or more of the above-described embodiment(s). The computer may comprise one or more processors (e.g., central processing unit (CPU), micro processing unit (MPU)) and may include a network of separate computers or separate processors to read out and execute the computer-executable instructions. The computer-executable instructions may be provided to the computer, for example, from a network or the storage medium. The storage medium may include, for example, one or more of a hard disk, a random-access memory (RAM), a read-only memory (ROM), a storage of distributed computing systems, an optical disc (such as a compact disc (CD), digital versatile disc (DVD), or Blu-ray Disc (BD)™), a flash memory device, a memory card, and the like.
While the disclosure has been described with reference to embodiments, it is to be understood that the disclosure is not limited to the disclosed embodiments. The scope of the following claims is to be accorded the broadest interpretation so as to encompass all such modifications and equivalent structures and functions.
Each embodiment can provide a processing apparatus that can detect abnormalities such as moiré that occur during imaging while a background image is displayed on a display unit.
| Number | Date | Country | Kind |
|---|---|---|---|
| 2023-222807 | Dec 2023 | JP | national |