The present disclosure relates to an electronic device and a method of controlling the electronic device.
A synchronous solid-state image sensor that captures image data (frame) in synchronization with a synchronization signal such as a vertical synchronization signal has been used in an electronic device or the like. With this general synchronous solid-state image sensor, image data can be acquired only at every synchronization signal cycle (e.g., 1/60 second). Hence, it is difficult to deal with requests for higher-speed processing in fields such as traffic and robots. Therefore, an asynchronous solid-state image sensor has been proposed in which a detection circuit that detects, for each pixel address, that the light amount of the pixel exceeds a threshold as an address event in real time is provided for each pixel. Such a solid-state image sensor that detects an address event for each pixel is called an event base vision sensor (EVS).
Patent Document 1: WO 2019/087471 A
Patent Document 2: Japanese Patent Application Laid-Open No. 2017-169987
Meanwhile, the state of a front part of an electronic device may be observed by an EVS image based on detection signals generated by EVS pixels. However, the EVS pixels are arranged at an end of a front surface part of the electronic device, and there is a risk that occlusion or the like is likely to occur.
Therefore, the present disclosure provides an electronic device in which EVS pixels can be arranged closer to a central part, and a method of controlling the electronic device.
In order to solve the above problem, according to the present disclosure, there is provided an electronic device including a display unit that has a display region in which display elements are arranged in an array in a first direction and a second direction different from the first direction, and an image sensor that is disposed on a side opposite to a display surface of the display unit so as to overlap the display region in a third direction different from the first direction and the second direction, and includes a plurality of pixels, in which the display unit transmits incident light, and the plurality of pixels outputs an event signal in a case where a change in luminance of light incident via the display unit is larger than a predetermined threshold.
The electronic device may further include a state analysis unit that analyzes a behavior of a user in a contact operation on the display unit using information of the event signal and estimates a user feeling.
The electronic device may further include a contact position analysis unit that estimates a position at which the user has contacted the display unit by using information of the event signal.
The contact position analysis unit may use propagation information of the event signal to distinguish an object that touched the display unit.
The electronic device may further include a control unit that controls the display unit, and may change a display content to be displayed on the display unit according to at least one of the contact position or the touched object.
A display content to be displayed on the display unit may be changed on the basis of a vibration image of a user generated using information of the event signal.
The electronic device may further include a state analysis unit that estimates a user's emotion on the basis of the vibration image of a user generated using information of the event signal.
The electronic device may further include a state processing unit that causes the display unit to display an image according to an estimation result of the state analysis unit.
The state processing unit may cause the display unit to display an image for healthcare according to an estimation result of the state analysis unit.
The state processing unit may cause the display unit to display a content option according to an estimation result of the state analysis unit.
The state processing unit may cause the display unit to display an action proposal to the user according to an estimation result of the state analysis unit.
The action proposal may be based on information of an improvement example of a third party acquired from an external server.
The electronic device may further include a speaker unit that emits a sound, and a sound arrival position analysis unit that estimates a part of the user exposed to the sound emitted from the speaker unit, using information of the event signal.
The sound arrival position analysis unit may determine whether or not an ear of a user is exposed to a sound emitted from the speaker unit.
The electronic device may further include a sound wave direction adjustment unit that controls an orientation of the speaker according to an arrival position of a sound analyzed by the sound arrival position analysis unit.
The electronic device may further include a face shape analysis unit hat records three-dimensional position information of both eyes, both ears, a nose, and a mouth in a three-dimensional image of the user in a recording unit.
The face shape analysis unit may estimate a position of an ear in three images in an oblique direction of the user by using three-dimensional position information of both eyes, both ears, a nose, and a mouth of the user recorded in advance and a rotation angle of the three-dimensional image of the user.
The sound arrival position analysis unit may be able to change an arrival position extracted by analysis according to an audio wavelength, of the speaker.
In a case where the sound arrival position analysis unit determines that a sound reaches the user the basis of the event signal, a depth sensor that captures a three-dimensional image of the user may be activated.
The sound arrival position analysis unit may fuse an image based on the event signal and an image based on the depth sensor, and acquire three-dimensional position information of both eves, both ears, a nose, and a mouth of the user.
The face shape analysis unit may generate a three-dimensional image of the user by skeleton estimation after activation of the depth sensor.
The event signal may be acquired constantly.
The display unit may be caused to emit light so as to satisfy sensitivity of the plurality of pixels.
According to the present disclosure, there is provided a method of controlling an electronic device including a display unit that has a display region in which display elements are arranged in an array in a first direction and a second direction different from the first direction, and an image sensor that is disposed on a side opposite to a display surface or the display unit so as to overlap the display region in a third direction different from the first direction and the second direction, and includes a plurality of pixels, in which the display unit transmits incident light, and the plurality of pixels outputs an event signal in a case where a change in luminance of light incident via the display unit is larger than a predetermined threshold.
Hereinafter, embodiments of an electronic device and a method of controlling the electronic device will be described with reference to the drawings. While main components of the electronic device will be mainly described below, the electronic device can have components and functions that are not illustrated or described. The following description does not exclude components and functions that are not illustrated or described.
The imaging lens 110 collects incident light and guides it to the solid-state image sensor 200. The solid-state image sensor 200 includes an EVS pixel and a gradation pixel. The EVS pixel can detect that the absolute value of the luminance change amount exceeds a threshold as an address event. The address event includes, for example, an on-event indicating that the amount of increase in luminance exceeds the upper limit threshold and an off-event indicating that the amount of decrease in luminance falls below le lower limit threshold less than the upper limit threshold. Then, the solid-state image sensor 200 generates a detection signal indicating the detection result of the address event for each EVS pixel. Each of the detection signals includes an on-event detection signal VCH indicating presence or absence of an on-event and an off-event detection signal CL indicating presence or absence of an off-event. Note that while the solid-state image sensor 200 detects the presence or absence of both the on-event and the off-event, it is also possible to detect only one of the on-event and the off-event. Furthermore, be EVS pixel according to the present embodiment can output an EVS luminance signal in addition to the detection signal. As a result, a first EVS image based on the detection signal of the EVS pixel and a second EVS image based on the luminance signal of the EVS pixel are formed.
On the other hand, the gradation pixel outputs a gradation luminance signal. A gradation image is formed on the basis of the gradation luminance signal output from the gradation pixel. Note that in the present embodiment, an image based on the detection signal of the EVS pixel is referred to as the first EVS image, an image based on the luminance signal of the EVS pixel is referred to as the second EVS image, and an image based on the gradation luminance signal output from the gradation pixel is referred to as a gradation image. The present embodiment has a first mode in which both the gradation pixel and the EVS pixel are driven, and a second mode and a third mode in which only the EVS pixel is driven. The second mode is a mode in which the first EVS image based on the detection signal of the EVS pixel and the second EVS image based on the luminance signal of the EVS pixel are formed. On the other hand, the third mode is a mode in which the first EVS image based on the detection signal of the EVS pixel is formed. Since the gradation pixel and the EVS pixel can be driven independently, the gradation pixel can be imaged at an imaging rate of, for example, 60 fps, whereas the second mode can be imaged at a rate of, for example, 200 fps. Further, in the third mode, since the luminance signal is not read from the EVS pixel, imaging can be performed at an even higher frame rate.
Power consumption is the smallest in the third mode, and is the next smallest in the second mode. Therefore, the EVS pixels are always driven in the third mode, and it is possible to perform state monitoring or the like based on the first EVS image based on the detection signal of the EVS pixels.
The solid-state image sensor 200 performs predetermined signal processing such as image processing on the first EVS image, the second EVS image, and the gradation image, and outputs the processed data to the recording unit 120 via a signal line 209.
The recording unit 120 records the data and the like from the solid-state image sensor 200. The control unit 130 controls the entire electronic device 100. For example, the control unit 130 controls the solid-state image sensor 200 to capture image data.
The analysis unit 140 performs predetermined analysis processing using at least one of the first EVS image, the second EVS image, or the gradation image.
The communication unit 150 performs wireless communication with an external device. As a result, content or the like is received from an external server and recorded in the recording unit 120 via the control unit 130. For example, the control unit 130 causes a display unit 170 to display an image based on the content.
The speaker unit 160 includes a highly directional speaker and can transmit sound information only to the user. The speaker unit 160 can change the direction in which the sound is transmitted.
Note that while the camera module 3 is disposed on the back surface side of a substantially central part of the display screen 1a in
As illustrated in
The polarizing plate 4c and the ¼ wave plate 4b curb incidence of internally reflected light on the camera module 3. In the display and 4, display elements are arranged in an array. The display panel 4 may be, for example, an organic light emitting device (OLED) diode, a liquid crystal display unit, a MicroLED, or a display panel based on other display principles.
The display panel 4 such as an OLED unit includes a plurality of layers. The display panel 4 is often provided with a member having low transmittance such as a color filter layer. A through hole may be formed in the member having a low transmittance in the display panel 4 in accordance with the arrangement place of the camera module 3. If object light passing through the through hole is made incident on the camera module 3, the image quality of the image captured by the camera module 3 can be improved.
The circularly polarizing plate 6 is provided to reduce glare and enhance visibility of the display screen 1a even in a bright environment. A touch sensor is incorporated in the touch panel 5. While there are various types of touch sensors such as a capacitive type and a resistive film type, any type may be used. Furthermore, the touch panel 5 and the display panel 4 may be integrated. The cover glass 7 is provided to protect the display panel 4 and other components.
Here, a configuration of the pixel array unit 30 will be described with reference to
A configuration of the pixel block 30a will be described with reference to
In the first mode and the second mode, the EVS pixel 308b outputs an analog signal of a voltage corresponding to the photocurrent to the EVS AFE 314. Furthermore, the EVS pixel 308b generates an analog signal of a voltage corresponding to the photocurrent as an EVS luminance signal (first luminance signal), and outputs the signal to the AD converter circuit 212b (see
On the other hand, in the third mode, the EVS pixel 308b does not output the EVS luminance signal to the AD converter circuit 212b (see
The EVS analog front end (AFE) 314 generates a detection signal from a voltage signal based on the output of the EVS pixel 308b, and outputs the detection signal to the second signal processing unit 214 (see
Returning to
The second access control circuit 211b controls the plurality of EVS pixels 308b and the plurality of EVS AFEs 314. The second access control circuit 211b according to the present embodiment causes the plurality of EVS AFEs 314 to sequentially detect address events for each row, and causes the second signal processing unit 214 to sequentially output the detection signals for each row.
Furthermore, when an address event is detected, the second access control circuit 211b sequentially outputs the luminance signals of the plurality of EVS pixels 308b to an EVS readout circuit 212b for each row.
A configuration example of the AD converter 212a will be described with reference to
A configuration example of the AD converter 212b for EVS will be described with reference to
Returning to
The timing control circuit 215 controls the timing of each component of the solid-state image sensor 200 on the basis of time stamp information. For example, the timing control circuit 212d controls timings of the first access control circuit 211a and the second access control circuit 211b. As a result, is also possible to synchronize the luminance signal of the gradation pixel 308a read by the AD converter 212a with the EVS luminance signal of the EVS pixel 308b read by the EVS readout circuit 212b.
Returning to
The second signal processing unit 214 performs predetermined signal processing on the detection signals from the plurality of EVS AFEs 314. The second signal processing unit 214 generates a first EVS image by, for example, arranging detection signals as pixel signals in a two-dimensional lattice pattern.
As illustrated in
Here, a detailed configuration example and a control operation example of the gradation pixel 308a will be described with reference to
N-type metal-oxide-semiconductor (MOS) transistors are used as the reset transistor 321, the amplification transistor 322, the selection transistor 323, and a transfer transistor 3310, for example. Furthermore, a photoelectric conversion element 311 is disposed on the light receiving chip 201. All the elements other than the photoelectric conversion element 311 are arranged on the detection chip 202.
The photoelectric conversion element 311 photoelectrically converts incident light to generate charge.
The charge photoelectrically converted by the photoelectric conversion element 311 is supplied from the photoelectric conversion element 311 to the floating diffusion layer 324 by the transfer transistor 3310. The charge supplied from the photoelectric conversion element 311 is accumulated in the floating diffusion layer 324. The floating diffusion layer 324 generates a voltage signal having a voltage value corresponding to the amount of accumulated charges.
The amplification transistor 322 is connected in series with the selection transistor 323 between the power line of a power supply voltage VDD and the vertical signal line VSL1. The amplification transistor 322 amplifies the voltage signal subjected to charge-voltage conversion by the floating diffusion layer 324.
A selection signal SEL is supplied from the first access control circuit 211a to the gate electrode of the selection transistor 323. In response to the selection signal SEL, the selection transistor 323 outputs the voltage signal amplified by the amplification transistor 322 to the AD converter 212a (see
Here, a detailed configuration example of the EVS pixel 308b will be described with reference to
In the EVS pixel 308b having the above configuration, the light receiving unit 31 includes a light receiving element (photoelectric conversion element) 311, a transfer transistor 312, and an OFG (Over Flow Gate) transistor 313. N-type metal oxide semiconductor (MOS) transistors are used as the transfer transistor 312 and the OFG transistor 313, for example. The transfer transistor 312 and the OFG transistor 313 are connected in series to each other.
The light receiving element 311 is connected between a common connection node N1 of the transfer transistor 312 and the OFG transistor 313 and the ground, and photoelectrically converts incident light to generate charge of a charge amount corresponding to the amount of the incident light.
A transfer signal TRG is supplied from the second access control circuit 21 illustrated in
A control signal OFG is supplied from the second access control circuit 211b to the gate electrode of the OFG transistor 313. In response to the control signal OFG, the OFG transistor 313 supplies the electric signal generated by the light receiving element 311 to the EVS AFE 314. The electric signal supplied to the EVS ATE 314 is a photocurrent including charges.
The pixel signal generation unit 32 includes a reset transistor 321, an amplification transistor 322, a selection transistor 323, and a floating diffusion layer 324. N-type MOS transistors are used as the reset transistor 321, the amplification transistor 322, and the selection transistor 323, for example.
The charge photoelectrically converted by the light receiving element 311 is supplied from the light receiving unit 31 to the pixel signal generation unit 32 by the transfer transistor 312. The charge supplied from the light receiving unit 31 is accumulated in the floating diffusion layer 324. The floating diffusion layer 324 generates a voltage signal having a voltage value corresponding to the amount of accumulated charges. That is, the floating diffusion layer 324 converts charge into voltage.
The reset transistor 321 is connected between the power line of the power supply voltage VDD and the floating diffusion layer 324. A reset signal RST is supplied from the second access control circuit 211b to the gate electrode of the reset transistor 321. The reset transistor 321 initializes (resets) the charge amount of the floating diffusion layer 324 in response to the reset signal RST.
The amplification transistor 322 is connected in series with the selection transistor 323 between the power line of the power supply voltage VDD and the vertical signal line VSL. The amplification transistor 322 amplifies the voltage signal subjected to charge-voltage conversion by the floating diffusion layer 324.
A selection signal SEL is supplied from the second access control circuit 211b to the gate electrode of the selection transistor 323. In response to the selection signal SEL, the selection transistor 323 outputs, the voltage signal amplified by the amplification transistor 322 to the EVS readout circuit 212b (see
In the electronic device 100 including the pixel array unit 30 in which the EVS pixels 308b having the above-described configuration are two-dimensionally arranged, when the control unit 130 illustrated in
Then, when an address event is detected in a certain EVS pixel 308b, the second access control circuit 211b turns off the OFG transistor 313 of the EVS pixel 308b to stop the supply of photocurrent to the EVS AFE 314. Next, the second access control circuit 211b supplies a transfer signal TRG to the transfer transistor 312 to drive the transfer transistor 312, and transfers the charge photoelectrically converted by the light receiving element 311 to the floating diffusion layer 324.
In this manner, the electronic device 100 including the pixel array unit 30 in which the EVS pixels 308b having the above-described configuration are two-dimensionally arranged outputs only the pixel signal of the EVS pixel 308b in which the address event is detected to the EVS readout circuit 212b. As a result, regardless of the presence or absence of an address event, power consumption of the electronic device 100 and the processing amount of image processing can be reduced as compared with the case of outputting the pixel signals of all the pixels.
Note that the configuration of the EVS pixel 308b exemplified here is an example, and the EVS pixel 308b is not limited to this configuration example. For example, the pixel configuration may omit the pixel signal generation unit 32. In the case of this pixel configuration, it is only required that the OFG transistor 313 be omitted from the light receiving unit 31, and the transfer transistor 312 have the function of the OFG transistor 313.
The current-voltage conversion unit 331 converts the photocurrent from the light receiving unit 31 of the gradation pixel 308a into a logarithmic voltage signal. The current-voltage conversion unit 331 supplies the converted voltage signal to the buffer 332. The buffer 332 buffers the voltage signal supplied from the current-voltage conversion unit 331 and supplies the voltage signal to the subtractor 333.
A row drive signal is supplied from the second access control circuit 211b to the subtractor 333. The subtractor 333 lowers the level of the voltage signal supplied from the buffer 332 in accordance with the row drive signal. Then, the subtractor 333 supplies the voltage signal with lowered level to the quantizer 334. The quantizer 334 quantizes the voltage signal supplied from the subtractor 333 into a digital signal and outputs the digital signal to the transfer unit 335 as an address event detection signal.
The transfer unit 335 transfers the address event detection signal supplied from the quantizer 334 to the second signal processing unit 214 and other components. When an address event is detected, the transfer unit 335 supplies an address event detection signal to the second signal processing unit 214 and the second access control circuit 211b.
Next, configuration examples of the current-voltage conversion unit 331, the subtractor 333, and the quantizer 334 in the EVS AFE 314 will be described.
The N-type transistor 3311 is connected between the power line of the power supply voltage VDD and a signal input line 3314. The P-type transistor 3312 and the N-type transistor 3313 are connected in series between the power line of the power supply voltage VDD and the ground. Then, a common connection node N2 of the P-type transistor 3312 and the N-type transistor 3313 is connected to the gate electrode of the N-type transistor 3311 and the input terminal of the buffer 332 illustrated in
A predetermined bias voltage Vbias is applied to the gate electrode of the P-type transistor 3312. With this configuration, the P-type transistor 3312 supplies a constant current to the N-type transistor 3313. A photocurrent is input from the light receiving unit 31 to the gate electrode of the N-type transistor 3313 through the signal input line 3314.
The drain electrodes of the N-type transistor 3311 and the N-type transistor 3313 are connected to the power supply side, and such a circuit is called a source follower. These two source followers connected in a loop convert the photocurrent from the light receiving unit 31 into a logarithmic voltage signal.
The subtractor 333 according to the present example includes a capacitive element 3331, an inverter circuit 3332, a capacitive element 3333, and a switch element 3334.
One end of the capacitive element 3331 is connected to the output terminal of the buffer 332 illustrated in
In the subtractor 333 having the above configuration, when the switch element 3334 is turned on (closed), a voltage signal Vinit is input to the terminal of the capacitive element 3331 on the buffer 332 side, and the terminal on the opposite side becomes a virtual ground terminal. The potential of the virtual ground terminal is set to zero for convenience. At this time, when the capacitance value of the capacitive element 3331 is C1, charge Qinit accumulated in the capacitive element 3331 is expressed by the following formula (1). On the other hand, since both ends of the capacitive element 3333 are short-circuited, the accumulated charge is 0.
Qinit=C1×Vinit (1)
Next, considering a case where the switch element 3334 is turned off (open) and the voltage of the terminal of the capacitive element 3331 on the buffer 332 side changes to Vafter, charge Qafter accumulated in the capacitive element 3331 is expressed by t e following formula (2).
Qafter=C1×Vafter (2)
On the other hand, when the capacitance value of the capacitive element 3333 is C2 and the output voltage is Vout, charge Q2 accumulated in the capacitive element 3333 is expressed by the following formula (3).
Q2=−C2×Vout (3)
At this time, since the total charge amount of the capacitive element 3331 and the capacitive element 3333 does not change, the following formula (4) is established.
Qinit=Qafter+Q2 (4)
When the formulae (1) to (3) are substituted into the formula (4) and deformed, the following formula (5) is obtained.
Vout=−(C1/C2)×(Vafter−Vinit) (5)
The formula (5) represents the subtraction operation of the voltage signal, and the gain of the subtraction result is C1/C2. Since it is usually desired to maximize the gain, it is preferable to design C1 to be large and C2 to be small. On the other hand, when C2 is too small, kTC noise increases, and noise characteristics may deteriorate. Therefore, capacity reduction of C2 is limited to a range in which noise can be tolerated. Furthermore, since the EVS AFE 314 including the subtractor 333 is mounted for each EVS pixel 308b, the capacitive element 3331 and the capacitive element 3333 have area restrictions. In consideration of these, the capacitance values C1 and C2 of the capacitive elements 3331 and 3333 are determined.
In
The storage unit 336 is provided between the quantizer 334 and the transfer unit 335, and accumulates the output of the quantizer 334, that is, the comparison result of the comparator 3341 on the basis of a sample signal supplied from the control unit 337. The storage unit 336 may be a sampling circuit such as a switch, plastic, or a capacitor, or may be a digital memory circuit such as a latch or a flip-flop.
The control unit 337 supplies a predetermined threshold voltage Vth to the inverting (−) input terminal of the comparator 3341. The threshold voltage Vth supplied from the control unit 337 to the comparator 3341 may have different voltage values in a time division manner. For example, the control unit 337 supplies a threshold voltage Vth1 corresponding to an on-event indicating that the change amount of the photocurrent exceeds an upper limit threshold and a threshold voltage Vth2 corresponding to an off-event indicating that the change amount falls below a lower limit threshold at different timings, so that one comparator 3341 can detect a plurality of types of address events.
For example, the storage unit 336 may accumulate the comparison result of the comparator 3341 using the threshold voltage Vth1 corresponding to the on-event in a period in which the threshold voltage Vth2 corresponding to the off-event is supplied from the control unit 337 to the inverting (−) input terminal of the comparator 3341. Note that the storage unit 336 may be inside the EVS pixel 308b or outside the EVS pixel 308b. Furthermore, the storage unit 336 is not an essential component of the EVS AFE 314. That is, the storage unit 336 may be omitted.
As described above, according to the present embodiment, as described above, according to the present embodiment, the individual image sensor 200 including a plurality of EVS pixels 308b is arranged on the opposite side of the display surface of the display unit 170. With this configuration, according to the luminance signals of the plurality of EVS pixels 308b, an event signal can be output in a case where the change in the luminance of the light incident via the display unit 170 is larger than a predetermined threshold. Furthermore, by providing the individual image sensor 200 including the EVS pixel 308b below the display unit 170, occlusion can be curbed.
Furthermore, a wide-angle lens having a predetermined thickness such as a fisheye lens can be arranged in the optical system 110.
An electronic device 100 according to a second embodiment is different from the electronic device 100 according to the first embodiment in further including a function capable of estimating the user's emotional state. Hereinafter, differences from the electronic device 100 according to the first embodiment will be described.
The recognition processing unit 1400 recognizes the observation target on the basis of, for example, the first EVS image. The recognition target according to the present embodiment is, for example, the fingertip. A general processing algorithm can be used for the recognition processing. For example, an occurrence region of an address event in the first EVS image is labeled, and if the area in the region labeled in a U shaped or a ring shape is within a predetermined range, it is recognized as a fingertip. In the first EVS image, a region corresponding to the edge part of the observation target is an occurrence region of an address event. Therefore, the case of the fingertip, for example, the occurrence region of an address event is labeled in a U shape or a ring shape. Then, the recognition processing unit 1400 sequentially outputs recognition signals including information indicating that the observation target is the finger and information indicating the barycentric coordinates of the fingertip region f16 to the state analysis unit 1402.
The state analysis unit 1402 estimates a user feeling on the basis of a behavior (such as hesitation) of the user in a touch panel operation on the display unit 160.
Here, (a) of
As illustrated in (a) of
Therefore, the state analysis unit 1402 generates an evaluation value based on the time until the target position g16 is touched and the oscillation state, and evaluates the mental state on the basis of the evaluation value. For example, the evaluation value generated by the state analysis unit 1402 becomes larger as the time until the target position g16 is touched becomes longer, and becomes larger as the oscillation number increases. With this configuration, the state analysis unit 1402 estimates that the mental state is stable when the evaluation value is a first threshold or less, unstable when the evaluation value is a second threshold or more, and normal when the evaluation value is greater than the first threshold and less than the second threshold. In this manner, by estimating the emotional state until the target position g16 is touched, it is possible to give feedback to improve operability. For example, in a case where hesitation is estimated, it is possible to improve the display mode such as the size and display color of the target position g16.
Furthermore, in a case where the target position g16 is a content selection button related to e-commerce, it is possible to give feedback regarding the user's psychological state and reflect the user's psychological state in a customer attraction method, an advertisement method, or the like. For example, in a case where hesitation is estimated, it is possible to improve the customer attraction method and advertisement method to reduce the hesitation.
On the other hand, if the luminance of the external environment is appropriate (Y in step S100), the control unit 130 proceeds to the third mode and repeats imaging of only the first EVS image (step S104). Subsequently, the recognition processing unit 1400 recognizes the observation target on the basis of, for example, the first EVS image (step S106).
Next, the state analysis unit 1402 determines whether or not the recognition processing unit 1400 has recognized the finger (step S108). If it is determined that the finger is recognized (Y in step S108), the state analysis unit 1402 captures only the first EVS image (step s106), and records the position coordinates and time of the fingertip until the user's finger touches the display unit 170 in the recording unit 120.
Next, the state analysis unit 1402 determines whether or not the user's finger has touched the display unit 170 on the basis of a signal from the touch panel 5 (see
As described above, the state analysis unit 1402 records the behavior of the user regarding the touch panel operation of the display unit 170, and estimates a reason state using the oscillation of the user's finger with respect to the target position g160 and the time until the touch as evaluation values. As a result, the psychological state of the user can be objectively estimated. Furthermore, the relationship between the operation and the psychological state makes it possible to give feedback to improve operability. Moreover, in a case where the target position g16 is a content selection button related to e-commerce, it is possible to feed back the psychological state of the user at the time of selecting the content and reflect the psychological state in the customer attraction method, the advertisement method, or the like.
An electronic device 100 according to a third embodiment is different from the electronic device 100 according to the second embodiment in that a function capable of estimating a touch position of the user by a first EVS image is further mounted. Hereinafter, differences from the electronic device 100 according the second embodiment will be described.
The contact position analysis unit 1404 determines whether or not the finger has touched the cover glass 7 of the display unit 170 from the form of the spread of the address event values of the first EVS images captured in time series. Then, when determining that there has been a touch, the contact position analysis unit 1404 analyzes the coordinates of the touch center. For example, barycentric coordinates of an address event value spreading in a ripple ring shape is set as the touch center.
Furthermore, the contact position analysis unit 1404 performs two-dimensional Fourier analysis on the superimposed images of the time-series first EVS images. As a result, the amplitude component for each frequency is analyzed, and it is determined whether the touch is made with the ball of the finger or with the fingertip such as a nail. For example, in a case where the touch is made with the ball of the finger, the ratio between the value of the amplitude component near the low frequency and the value of the amplitude component near the high frequency is larger than that in a case where the touch is made with the fingertip such as a nail. In this manner, the contact position analysis unit 1404 determines whether the touch is made with the ball of the finger or with the fingertip such as the nail on the basis of the ratio between the value of the amplitude component near the low frequency and the value of the amplitude component near the high frequency. Then, a control unit 130 (see
As described above, the contact position analysis unit 1404 analyzes the coordinates of the touch center from the form of the spread of the address event values of the first EVS images captured in time series. As a result, even when the electronic device 100 does not have a touch panel, the touch position on the cover glass 7 (see
An electronic device 100 according to a fourth embodiment is different from the electronic device 100 according to the third embodiment in further including a function capable of estimating an interaction between the user and a nearby person by a first EVS image. Hereinafter, differences from the electronic device 100 according to the second embodiment will be described.
For example, a region below 0 of the line L240 in
The interaction analysis unit 1406 estimates the interaction between the subject and the nearby persons by the form of the temporal change of the position of the lower jaw of the first EVS images captured in time series. For example, when the movement of the jaw of the subject is observed and the movement of the jaw of the person around is observed in conjunction with the movement, it is estimated that the degree of agreement is high. On the other hand, in a case where the linkage of the movement of the jaw of the person around is not observed, it is estimated that the degree of agreement is low.
More specifically, the interaction analysis unit 1406 records position coordinates such as the vertical coordinates of the address event value corresponding to the position of the lower jaw for each of the face regions a230, a232, and a234 in the recording unit 120 (see
As described above, the interaction analysis unit 1406 analyzes the temporal change of the position of the lower jaw in the first EVS images captured in time series. As a result, it is possible to estimate that the degree of agreement is high when the movement of the jaw of the person around is observed in conjunction with the position of the lower jaw of the subject, and it is possible to estimate that the degree of agreement is low when the movement of the jaw of the person around is not observed.
An electronic device 100 according to a fifth embodiment is different from the electronic device 100 according to the fourth embodiment in further including a function of estimating a psychological state by analyzing the vibration of the user from a first EVS image. Hereinafter, differences from the electronic device 100 according to the fourth embodiment will be described.
Furthermore, in a case where the largest number of people show instability at a certain point of time, an improvement example for stabilizing emotions of the people is acquired as knowledge. This improvement example is also stored in association with the emotion information. As the improvement example, action examples include showing a relaxed content, promoting a relaxing action such as deep breathing, stretching, or the like.
The vibration image generation unit 1408 generates a vibration image of the user on the basis of a cycle of the address event for each pixel acquired in time series.
A state analysis unit 1402 according to the fifth embodiment estimates a psychological state of the user, such as an emotion, on the basis of a vibration image generated by the vibration image generation unit 1408. For this estimation method, for example, a technology disclosed in Patent Document 2 can be used. The state analysis unit 1402 according to the fifth embodiment is different from the technology disclosed in Patent Document 2 in that a vibration image generated by the vibration image generation unit 1408 is used.
The state processing unit 1500 displays, on the display unit 170, an image in a display form according to the estimation result of the state analysis unit 1402.
On the other hand, as illustrated in (b) of
As illustrated in
On the other hand, if the images are acquired (Y in step S202), the vibration image generation unit 1408 generates a vibration image (step S204).
Next, the state analysis unit 1402 estimates the psychological state of the user using the vibration image generated by the vibration image generation unit 1408. Subsequently, the state analysis unit 1402 determines whether or not to end the processing (step S208), and if the processing is not to be ended (N in step S208), repeats the processing from step S200. On the other hand, if the processing is to be ended (Y in step S208), the entire processing is ended.
Next, the state processing unit 1500 determines whether or not the psychological state of the user acquired from the state analysis unit 1402 is unstable (step S304). If the state is not unstable (N in step S304), the processing from step S300 is repeated.
On the other hand, if the state is unstable (Y in step S304), information on an improvement example associated with the content being displayed is acquired from the content accumulation unit 1000a of the server 1000 via the communication unit 150 (see
Next, the state processing unit 1500 determines whether or not to end the entire processing (step S310) If it is determined not, to end the processing (N in step S310), the processing from step S300 is repeated. On the other hand, if it is determined to end the processing (Y in step S310), the overall processing is ended.
As described above, the state analysis unit 1402 according to the present embodiment estimates the psychological state of the user using the vibration image of the user generated by the vibration image generation unit 1408. As a result, the psychological state of the user can be objectively estimated. Furthermore, since the psychological state of the user who is displaying the content can be estimated, options for the next content can be changed according to the psychological state of the user.
Furthermore, in a case where a psychological state of the user who is displaying the content is unstable, contents corresponding to an improvement measure associated with the content is displayed on the display unit 170 as options. As a result, it is possible to allow the user to select content having a proven record of improvement.
An electronic device 100 according to a sixth embodiment is different from the electronic device 100 according to the fifth embodiment in further including a function of changing an arrival region of a sound emitted by a speaker unit 160 by analyzing the arrival region of the sound emitted by the speaker unit 160. Hereinafter, differences from the electronic device 100 according to the fifth embodiment will be described.
A processing example of the face shape analysis unit 1502 will be described with reference to
As illustrated in
As illustrated in
As illustrated in
Processing examples of the sound arrival position analysis unit 1504 and the sound wave direction adjustment unit 1506 will be described with reference to
As illustrated in
As illustrated in
Furthermore, the sound arrival position analysis unit 1504 can also perform frequency analysis of a region exposed to sound by two-dimensional Fourier transform. In this case, the region corresponding to the frequency of the sound emitted from the speaker unit 160 alone can be estimated as the region of the ear E39. Therefore, in a case where there are a large number of sound sources, the estimation accuracy can be further improved.
Furthermore, the sound wave direction adjustment unit 1506 can combine the wavefronts of the sound waves Sa and Sb of the first speaker 160a and the second speaker 160b to generate a sound field specialized for the user B37. More specifically, the sound wave direction adjustment unit 1506 adjusts the orientation of the first speaker 160a and the second speaker 160b and the overlap of the wavefronts of the waves Sa and Sb, and generates a sound field that reaches the region of the ear E39 more intensively. Furthermore, the sound wave direction adjustment unit 1506 can determine whether or not there is a person nearby by a sensor on the basis of the processing result of the recognition processing unit 1400, and can change the intensity and range of the sound to be transmitted when there is a person nearby.
Next, the face shape analysis unit 1502 generates a three-dimensional image of the user B37 on the basis of the depth image of the depth sensor 2000 (step S404). Subsequently, the face shape analysis unit 1502 rotates the front three-dimensional face image recorded in advance, and determines the direction of the face of the user B37 on the basis of the angle that matches the three-dimensional face image generated in step S404 (step S406).
Next, the face shape analysis unit 1502 estimates the position of the ear in the first EVS image using three-dimensional position information of both eyes, both ears, the nose, and the mouth recorded in advance and the rotation angle, that is, information of the direction of the face (step S408).
Next, the sound arrival position analysis unit 1504 estimates a region exposed to sound waves (step S410). Subsequently, the sound wave direction adjustment unit 1506 determines whether or not the position of the ear estimated by the face shape analysis unit 1502 matches the region exposed to sound waves estimated by the sound arrival position analysis unit 1504 (step S412). Then, if the regions match (Y in step S412), the sound wave direction adjustment unit 1506 repeats the processing from step S402.
On the other hand, if the regions do not match (N in step S412), the sound wave direction adjustment unit 1506 adjusts angles θ1 and θ2 with respect to the first speaker 160a and the second speaker 160b such that the position of the ear estimated by the face shape analysis unit 1502 matches the area exposed to sound waves estimated by the sound arrival position analysis unit 1504 (step S414). Next, the sound wave direction adjustment unit 1506 determines whether or not to end the entire processing (step S416), and if it is determined not to end the entire processing (step S416), repeats the processing from step S402.
As described above, the sound arrival position analysis unit 1504 according to the present embodiment estimates the region exposed to sound waves using the first EVS image, and the sound wave direction adjustment unit 1506 adjusts angles θ1 and θ2 with respect to the first speaker 160a and the second speaker 160b such that the position of the ear estimated by the face shape analysis unit 1502 matches the region exposed to sound waves estimated by the sound arrival position analysis unit 1504. As a result, even if the user B37 moves, the region of the ear E39 of the user B37 can be constantly exposed to sound waves.
Note that the present technology can also be configured in the following manner.
(1) An electronic device including
(2) The electronic device according to (1), further including a state analysis unit that analyzes a behavior of a user in a contact operation on the display unit using information of the event signal and estimates a user feeling.
(3) The electronic device according to (1) or (2), further including a contact position analysis unit that estimates a position at which the user has contacted the display unit by using information of the event signal.
(4) The electronic device according to (3), in which the contact position analysis unit uses propagation information of the event signal to distinguish an object that touched the display unit.
(5) The electronic device according to further including a control unit that controls the display unit, in which
(6) The electronic device according to (1), in which
(7) The electronic device according to (6), further including a state analysis unit that estimates a user's emotion on the basis of the vibration image of a user generated using information of the event signal.
(8) The electronic device according to (7), further including a state processing unit that causes the display unit to display an image according to an estimation result of the state analysis unit.
(9) The electronic device according to (7), in which
(10) The electronic device according to (8), in which
(11) The electronic device according to (8), in which
(12) The electronic device according to (11), in which
(13) The electronic device according to (1), further including a speaker unit that emits a sound, and
(14) The electronic device according to (13), in which
(15) The electronic device according to (14), further including a sound wave direction adjustment unit that controls an orientation of the speaker according to an arrival position of a sound analyzed by the sound arrival position analysis unit.
(16) The electronic device according to (15), in which
(17) The electronic device according to (16), further including a face shape analysis unit that records three-dimensional position information of both eyes, both ears, a nose, and a mouth in a three-dimensional image of the user in a recording unit.
(18) The electronic device according to (17), in which
(19) The electronic device according to (18), in which
(20) The electronic device according to (19), in which
(21) The electronic device according to (20), in which
(22) The electronic device according to (21), in which
(23) The electronic device according to (1), in which the event signal is acquired constantly.
(24) The electronic device according to (1), in which
(25) A method of control a g an electronic device including
Aspects of the present disclosure are not limited to the above-described individual embodiments, but include various modifications conceivable by those skilled in the art, and the effects of the present disclosure are not limited to the above-described contents. That is, various additions, modifications, and partial deletions can be made lout departing from the conceptual idea and gist of the present disclosure derived from the contents defined in the claims and equivalents thereof.
Number | Date | Country | Kind |
---|---|---|---|
2020-162530 | Sep 2020 | JP | national |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/JP2021/032981 | 9/8/2021 | WO |