This application claims priority to Japanese Patent Application No. 2021-009401 filed on Jan. 25, 2021, incorporated herein by reference in its entirety.
The present disclosure relates to a method of specifying a generation point of abnormal sound and an application program.
For example, Japanese Unexamined Patent Application Publication No. 9-196825 discloses a device that detects abnormality based on output when image data is input into a neural network and output when waveform data obtained by analyzing the frequency of an acoustic signal is input into the neural network.
In the case of objects that generate sound at normal times, such as vehicles and the like, the sound pressure level of an abnormal sound generated when abnormality occurs does not necessarily exceed the sound pressure level of the sound generated at the normal times. Therefore, in order to make it possible to specify the generation point of abnormality by using a neural network as described above, the neural network has much information to identify, which often makes it hard for the neural network to learn. Furthermore, it is also necessary to identify the relationship between an image represented by image data and sound represented by an acoustic signal. However, it can be difficult for a person to identify the generation point of sound relating to abnormality. Therefore, to specify the relationship, a high level information processing ability is necessary.
Hereinafter, measures to solve the stated problems and the operation effects thereof will be described.
1. A method of specifying a generation point of abnormal sound includes: a reproduction step; an indication reception step; an association step; a selection step; and a specification step. The reproduction step is for reproducing sound of each of a plurality of frequency components by a reproduction device. The indication reception step is for receiving, after the reproduction step, an indication indicating which component, out of the frequency components, is close to a component of an abnormal sound of which a generation point is to be specified, via a human interface. The association step is for associating sound data with image data and storing the sound data and the image data in a storage device, the sound data being recorded by a recording device that is arranged in proximity to each of a plurality of regions that are obtained by dividing a target object, the image data being a part of image data on the target object imaged by an imaging device, the part being related to the target object that is in proximity to the recording device at a time of recording. The selection step is for selecting, out of the sound data associated with the image data in the association step, the image data associated with the data that is highest in sound pressure level of an indication frequency component that is the frequency component indicated in the indication reception step. The specification step is for specifying the generation point of the abnormal sound based on the image data selected in the selection step.
The method enables a person to indicate the frequency component that is characteristic of a sound that the person feels abnormal by reproducing sound of each of a plurality of frequency components. Therefore, regardless of whether the sound pressure level of the sound that the person feels abnormal is the maximum sound pressure level of the sound generated from the target object, the abnormal sound can properly be grasped. The sound data recorded by the recording device that is arranged in proximity to each of a plurality of regions that are obtained by dividing a target object is stored in association with a part of the image data related to the target object that is in proximity to the recording device. By selecting the image data associated with the data highest in sound pressure level of the indication frequency component, the selected image data is highly likely to represent the image data close to the generation point of the abnormal sound. Therefore, it becomes possible to specify the generation point of the abnormal sound detected by the person.
2. The method according to the first aspect may include: a recording step of recording the sound generated from the target object with the recording device and storing the sound as sound data in the storage device; and an extraction step of extracting the frequency components of the sound indicated by the sound data. In the method according to the first aspect, the reproduction step may be a step of reproducing sound of each of the frequency components extracted in the extraction step.
In the method, the frequency components of the sound data are extracted and reproduced. This enables a person to specify the component that the person feels abnormal based on the sound that the person actually hear.
3. In the method according to the second aspect, the target object may include a rotary machine, and the frequency components may be predetermined components that are proportional to a rotational frequency of the rotary machine.
When the target object includes a rotary machine, the rotary machine or a rotor coupled to the rotary machine may generate abnormal sound with rotation. In that case, the frequency of the abnormal sound is proportional to the rotation frequency of the rotary machine. In the above method, the frequency components to be extracted are the predetermined components that are proportional to the rotational frequency of the rotary machine. This makes it possible to accurately extract the components of the abnormal sound generated with rotation of the rotary machine.
4. The method according to the third aspect may include a measurement result acquisition step of acquiring a measurement result by a device that measures a variable value indicating rotation speed of the rotary machine. The extraction step may have a step of extracting the frequency components based on the rotational frequency indicated by the measurement result.
In the method, frequency components are extracted based on the rotational frequency indicated by the measurement result. Accordingly, target components can accurately be extracted without having to fix the rotation speed of the rotary machine to predetermined values.
5. The method according to the third or fourth aspect may include an inquiry step of inquiring whether or not the component of the abnormal sound of which the generation point is to be specified is a component dependent on the rotational frequency of the rotary machine. The extraction step may include: a step of extracting a plurality of predetermined components independently of the rotational frequency of the rotary machine when a response as a result of the inquiry is that the component is not dependent, and a step of extracting predetermined components proportional to the rotational frequency of the rotary machine when the response as the result of the inquiry is that the component is dependent.
In the method, it is possible to obtain useful information to determine the type of abnormal sound by inquiring whether the component of the abnormal sound is dependent on the rotation frequency of the rotary machine.
6. In the method according to any one of the first to fifth aspects, the specification step may include a first probability calculation step of using the image data as input to calculate probability that one or more candidates of the generation point of the abnormal sound are points indicated by the image that is indicated by the above image data, a second probability calculation step of calculating probability that the one or more candidates of the generation point of the abnormal sound are actual generation points of the abnormal sound based on the sound data associated with the image data selected in the selection step, and a comprehensive probability calculation step of calculating comprehensive probability that the one or more candidates of the generation point of the abnormal sound are actual generation points of the abnormal sound based on a result of probability calculation in the first probability calculation step and a result of probability calculation in the second probability calculation.
In the above method, the generation point of the abnormal sound is specified based on the position information on the target object indicated by the image data when the sound pressure level of the specified frequency component is the highest, as well as on the sound data at that time. This makes it possible to specify the generation point of the abnormal sound with high accuracy as compared with the case of specifying based on only the position information.
7. In the method according to the sixth aspect, the second probability calculation step may be a step of calculating probability corresponding to one or more candidates of the generation point of the abnormal sound, based on similarity between the sound data associated with the image data selected in the selection step and waveform pattern data representing relationship between frequencies that are characteristic when the candidates of the generation point of the abnormal sound generate sound and sound pressure levels.
In the above method, the probability corresponding to candidates of the generation point is calculated based on the similarity between the sound data on the candidates of the generation point and waveform pattern data on the sound pressure levels relative to the frequencies. This makes it possible to specify the generation point of the abnormal sound by using specific frequency that is characteristic of the abnormal sound, as well as the sound pressure waveform when the abnormal sound is generated.
8. In the method according to any one of the first to seventh aspects, the first probability calculation step may be a step of using the image data as input to calculate probability corresponding to the candidates of the generation point indicated by the image data. The second probability calculation step may be a step of calculating probability corresponding to the candidates of the generation point of the abnormal sound, based on similarity between the sound data associated with the image data selected in the selection step and waveform pattern data representing relationship between frequencies that are characteristic when the candidates of the generation point generate sound and sound pressure levels.
In the method, the recording device as well as the imaging device are arranged in proximity to each of a plurality of regions obtained by dividing a target object. This makes it easy to associate the recorded sound with the position information on the target object that faces the recording device.
9. An application program that causes a computer to execute the reproduction step, the indication reception step, the association step, and the selection step in the method of specifying a generation point of abnormal sound according to any one of the first to eighth aspects.
10. The application program according to the ninth aspect may cause the computer to execute: a transmission step of transmitting the image data selected in the selection step and the sound data associated with the image data; and a reception step of receiving a signal indicating information regarding the generation point of the abnormal sound specified in the specification step based on the image data and the sound data transmitted in the transmission step.
The above method allows the computer to execute the transmission step and reception step. Accordingly, another computer other than the computer can execute the specification step. Therefore, a computational load of the computer executing the reproduction step and other steps can be reduced as compared with the case where the computer also executes the specification step.
Features, advantages, and technical and industrial significance of exemplary embodiments of the present disclosure will be described below with reference to the accompanying drawings, in which like signs denote like elements, and wherein:
Hereinafter, an embodiment of the present disclosure will be described with reference to the drawings.
A vehicle 10 is a vehicle brought into a repair shop by a user who detects abnormal sound. The vehicle 10 is mounted with an internal combustion engine 12 and a control device 14 of the internal combustion engine 12. The repair shop may also be a dealer shop.
A mobile terminal 20 is a portable terminal possessed by an operator at the repair shop. The mobile terminal 20 includes a CPU 22, a storage device 24, a display 26, a touchpanel 28, a speaker 30, a microphone 32, a camera 34 and a communication device 36, which can communicate with each other via a communication line 38. Here, the storage device 24 is an electrically rewritable, non-volatile memory. The display 26 is constituted of, for example, an LCD or an LED. The touchpanel 28 is arranged so as to be superimposed on the display 26. The communication device 36 can communicate with the control device 14 of the vehicle 10. The communication device 36 can also communicate with an analyzer 50 possessed by a vehicle manufacturer via a network 40.
The analyzer 50 includes a CPU 52, a storage device 54, and a communication device 56, which can communicate via a communication line 58. The storage device 54 is an electrically rewritable, non-volatile memory.
The storage device 24 of the mobile terminal 20 stores an application program 24a. The storage device 54 of the analyzer 50 stores a specification program 54a. The CPU 22 executes instructions defined by the application program 24a, and the CPU 52 executes instructions defined by the specification program 54a so as to perform processing to specify abnormal sound generated in the vehicle 10. Hereinafter, the processing will be described in detail.
In a series of processing shown in
With reference to
The CPU 22 then determines whether input operation performed on the touchpanel 28 is input operation indicating the rotational synchronization or input operation indicating no synchronization (S14). When it is determined that the input operation indicates synchronization (S14: YES), the CPU 22 communicates with the control device 14 to receive data about rotation speed NE of a crank shaft that is a rotation shaft of the internal combustion engine 12 (S16). The CPU 22 then sets the frequency band of n-th order rotation as a prescribed frequency component that is characteristic as a candidate of the abnormal sound, based on the rotation speed NE (S18). In this process, the CPU 22 refers to frequency component data 24b stored in the storage device 24 shown in
As shown in
The process of S18 is to set frequency bands that are defined based on the line segments f1, f2, . . . and the rotation speed NE. In other words, the number of the frequency bands set as prescribed frequency bands here is equal to the number of the line segments f1, f2, . . . .
With reference to
When the processes of S18 and S20 are completed, the CPU 22 operates the speaker 30 to reproduce sounds of a plurality of frequency bands set in the process of S18 or the process of S20 (S22). This process can be achieved by band-pass filter processing applied to the sounds recorded in the process of S10 and thereby generating data on sounds in the frequency bands set by the process of S18 or process of S20.
The CPU 22 then operates the display 26 to display visual information for urging a user to indicate the sound closest to the abnormal sound among the reproduced sounds (S24). The CPU 22 continues the processes of S22 and S24 until the frequency band closest to the abnormal sound is indicated (S26: NO) by input operation performed on the touchpanel 28. Specifically, the processes of S22 to S26 may be performed by the CPU 22 to display the number of buttons, corresponding to the number of the frequency bands to be reproduced, on the display 26, with the button currently selected for reproduction being displayed in color different from other buttons. In other words, when a specific button is indicated via the touchpanel 28, the CPU 22 may determine that the sound closest to the abnormal sound is indicated.
When it is determined that there is input operation indicating the sound closest to the abnormal sound (S26: YES), the CPU 22 determines the indicated band as the indication frequency band (S28).
Then, the CPU 22 operates the display 26 as shown in
For proximity photographing of a moving image, it is desirable that a distance between the target object and the mobile terminal 20, and the speed of moving the mobile terminal 20 are set in advance and the operator performs the processing based on the set distance and speed. This can be achieved by, for example, giving an instruction regarding the distance and speed by operating the display 26 or the speaker 30 in the process of S30 shown in
When proximity photographing of the moving image is started by operation of the operator, the CPU 22 stores image data from the camera 34 and sound data from the microphone 32 in the storage device 24 for each region obtained by dividing the target object (S32).
Dividing the target object into regions can be achieved when, for example, the CPU 22 determines the position of the image indicated by the proximity video image in an overall image stored by the process of S10. Instead of determining the position, the CPU 22 may display on the display 26 the current position on the moving path shown in
With reference to
The CPU 22 performs processes of S32 to S36 for all the regions of the target object. In other words, when it is determined that processing for all the regions are not yet complete (S38: NO), the CPU 22 updates the label variable i (S40) and returns to the process of S32. Contrary to this, when it is determined that processing for all the regions are complete (S38: YES), the CPU 22 selects the image data and the sound data corresponding to the maximum sound pressure level L(i), among the image data and the sound data stored by the process of S32 (S42). The CPU 22 then operates the communication device 36 to transmit the selected image data and sound data to the analyzer 50 (S44).
In this connection, as shown in
The mapping defined by the mapping data 54b is a learned model that is learned using the image data and training data that sets the value of a node corresponding to the object indicated by the image data to “1” and the value of other nodes to “0”.
The CPU 52 then calculates a second probability b(j) regarding the candidates j of the generation point of the abnormal sound based on the waveform of the sound pressure level corresponding to the frequency indicated by the received sound data (S54). The process is based on the similarity between a waveform pattern indicated by waveform pattern data 54c for each of the candidates of the generation point of the abnormal sound stored in the storage device 54 shown in
When, among the waveform patterns indicated by the waveform pattern data 54c, some waveform patterns are below a prescribed similarity to the waveform of the sound pressure level corresponding to the frequency indicated by the received sound data, the CPU 52 may set the second probability b that the candidates corresponding to the patterns are the generation point of the abnormal sound to “0”.
With reference to
The CPU 52 then operates the communication device 56 to transmit the result of specifying the generation point of the abnormal sound to a transmission source of the data received by the process of S50 (S58). Here, the CPU 52 transmits information about three candidates of the generation point higher in the comprehensive probability. The CPU 52 temporarily ends a series of the processes shown in
In contrast, as shown in
The CPU 22 temporarily ends a series of processing shown in
Now, the functions and effects of the present embodiment will be described.
When an abnormal sound of which generation point is expected to be specified, is proportional to the rotation frequency of the crank shaft, the CPU 22 reproduces sounds of frequencies corresponding to candidates of the generation point, which are component parts that can generate proportional abnormal sounds. When the operator indicates the frequency closest to the abnormal sound, out of the reproduced sounds, the CPU 22 sets the frequency as an indication frequency. The CPU 22 then photographs images and records sounds when mobile terminal 20 is arranged in proximity to a plurality of regions obtained by dividing the target object. Then, the CPU 22 calculates the sound pressure level of the indication frequency component among the recorded sounds, and stores the calculated sound pressure level. The CPU 22 selects the region, out of the regions, highest in recorded sound pressure level, and transmits the image data and sound data on the region to the analyzer 50.
The CPU 52 of the analyzer 50 calculates the first probability a that the object, indicated by the transmitted image data, is each of the candidates of the generation point. The CPU 52 also calculates the second probability b that each of the candidates of the generation point is actually the generation point of the abnormal sound, based on the similarity between the waveform indicated by the transmitted sound data and the waveform indicated by the waveform pattern data 54c. Then, the CPU 52 multiplies values of the first probability a and the second probability b of the same candidate to obtain the comprehensive probability of the candidate, and transmits the obtained comprehensive probability to the CPU 22. The CPU 22 displays the transmitted result of specifying the generation point of the abnormal sound.
This makes it possible to calculate the generation point of the abnormal sound with high accuracy. Specifically, for example, since the frequency of inverse number of the time interval at a compression top dead center of the internal combustion engine 12 is the frequency of inverse number of the period of a combustion phase, generated sound has a frequency with a high sound pressure level. However, the sound sensed as abnormal sound is not necessarily the sound of this frequency. Therefore, it is difficult to grasp which sound the user recognizes as abnormal sound based only on the frequency analysis of the sound data. On the contrary, in the present embodiment, sounds of frequencies that are characteristic as the candidates of the generation point of abnormal sound are reproduced, and a person indicates which sound of frequency is close to the abnormal sound. This makes it possible to correctly understand the abnormal sound and then specify the generation point thereof. As a result, it is possible to grasp the generation point of the abnormal sound with high accuracy.
According to the present embodiment described in the foregoing, the functions and effects as described below are further achieved.
(1) An indication frequency component is extracted from the sounds recorded when the mobile terminal 20 is arranged in proximity to each of a plurality of regions obtained by dividing a target object, and the point with a highest sound pressure level is specified. Accordingly, even in the case where an operator who heard abnormal sound has an incorrect preconception regarding the generation point of the abnormal sound, it is possible to specify the place highly likely to generate the sound of the indication frequency regardless of the preconception.
(2) In the case of abnormal sound proportional to the rotation frequency of the crank shaft, the rotation speed NE is obtained from the control device 14. Based on the rotation speed NE, frequency components, characteristic of the sound generated by each of the candidates of the generation point of the abnormal sound, are extracted. Accordingly, a target component can accurately be extracted without having to fix the rotation speed NE of the crank shaft to predetermined values.
(3) An inquiry is made to a person regarding whether the abnormal sound is proportional to the rotation frequency of the crank shaft. As a result, it is possible to obtain useful information to determine the type of abnormal sound.
(4) The CPU 52 calculates the first probability a that each of the candidates is the generation point, from the image data when the indication frequency is at its maximum, and calculates the second probability b that each of the candidates is the generation point, based on the similarity between the waveform indicated by the sound data with the frequency at its maximum and the waveform indicated by the waveform pattern data 54c. This makes it possible to specify the generation point of the abnormal sound by using the indication frequency that is a specific frequency characteristic of the abnormal sound as well as the sound pressure waveform when the abnormal sound is generated. Accordingly, even when the frequencies generated by a plurality of generation candidates are the same-order frequencies as the rotation frequency of the crank shaft, it becomes possible to specify which of these candidate is the generation point.
(5) The analyzer 50 executes the processes of S52 and S54. This makes it possible to reduce the computational load on the mobile terminal 20 as compared with the case where the processes of S52 and S54 are executed by the mobile terminal 20.
Correspondence Relation
The correspondence relation between the matters in the embodiments and the matters described in the column “Solution to Problem” is as follows. In the followings, the correspondence relation is shown for every number of the solutions stated in the column “Solution to Problem”. [1, 8] The target object corresponds to the object housed in the engine compartment 16 shown in
The present embodiment can be implemented with modifications as shown below. The present embodiment and following modifications can be implemented in combination with each other without departing from the range of technically consistency.
About Rotary Machine
About Measurement Result Acquisition Step
About Inquiry Step
About Reproduction Step and Indication Reception Step
About Association Step
About First Probability Calculation Step
About Second Probability Calculation Step
About Comprehensive Probability
About Display of Specification Result
About Computer that Executes Processing to Specify Generation Point
About Computer
The computer is not limited to those executing software processing. The computer may include a dedicated hardware circuit for hardware processing, such as ASICs. The computer may further be a combination of a software execution device that executes programs and the dedicated hardware processing circuit. Specifically, the computer may have any one of the configurations (a) to (c) below: (a) including a software execution device that executes all the processing based on programs; (b) including a software execution device that executes some of the processing based on programs, and a dedicated hardware circuit that executes the remaining processing; and (c) including a dedicated hardware circuit that executes all the processing. Here, the number of the software execution devices, or the number of the dedicated hardware circuits may be two or more.
About Target Object as Specification Target of Generation Point of Abnormal Sound
Number | Date | Country | Kind |
---|---|---|---|
2021-009401 | Jan 2021 | JP | national |