The entire disclosure of Japanese Patent Application No. 2023-150743 filed on Sep. 19, 2023 is incorporated herein by reference in its entirety.
The present disclosure relates to a sound diagnostic system, and more specifically, to a sound diagnostic system using the sound pressure.
In a case where an apparatus such as a home appliance used at home or an image forming apparatus used in an office breaks down, the apparatus may produce an abnormal sound. A serviceperson can identify a faulty component from the abnormal sound and replace the faulty component. However, in a case where an apparatus including a plurality of motors or the like breaks down, it is difficult for a serviceperson to specify which component of the apparatus is producing the abnormal sound. In this case, the serviceperson cannot specify which component should be repaired or replaced. Therefore, there is a need for a sound diagnostic technology for easily identifying a component producing an abnormal sound in an apparatus.
Regarding a sound diagnosis, for example, Japanese Laid-Open Patent Publication No. 2017-138151 discloses a “diagnostic apparatus” that supports a task of identifying a cause of an abnormal sound. The diagnostic apparatus identifies the frequency of the abnormal sound in frequency spectrum waveform data. The diagnostic apparatus performs a fast Fourier transform on a frequency component of the abnormal sound, in the time axis direction, to extract information on the period and the frequency of the abnormal sound. The extracted information is used for identifying the cause of the abnormal sound.
Japanese Laid-Open Patent Publication No. H04-167699 discloses a “rattling sound detection apparatus.” The apparatus obtains an autocorrelation function of a time waveform of a sound. The apparatus determines whether an abnormality is present or not, based on a result of comparison between a peak value of the autocorrelation function and a predetermined value.
An apparatus having a complicated mechanism, such as an image forming apparatus, includes a plurality of components that generate a drive sound. The plurality of components may include a component that may generate an abnormal sound having a plurality of frequency components. Alternatively, the plurality of components may include a component that may generate an abnormal sound multiple times or an abnormal sound having a certain time width, during one rotation. The apparatus disclosed in Japanese Laid-Open Patent Publication No. 2017-138151 is not able to accurately detect the period of the abnormal sound of such a component. As a result, the apparatus disclosed in Japanese Laid-Open Patent Publication No. 2017-138151 is not able to accurately identify a causative component that causes the abnormal sound.
The apparatus disclosed in Japanese Laid-Open Patent Publication No. H04-167699 merely determines whether an abnormality is present or not, and does not identify a causative component that causes the abnormal sound.
An object in one aspect of the present disclosure is to provide a technology that enables a causative component causing an abnormal sound to be identified among a plurality of components.
To achieve at least one of the abovementioned objects, according to an aspect of the present invention, a sound diagnostic system reflecting one aspect of the present invention comprises: one or more hardware processors; and one or more memories storing instructions that, when executed by the one or more hardware processors, cause the one or more hardware processors to perform operations. The operations include: calculating, from sound information on a sound produced by an apparatus, a sound pressure of each of a plurality of frequencies at each timing; calculating an autocorrelation function for a change, with time, of a representative value of the sound pressures of respective frequencies included in a target frequency band, among the plurality of frequencies; extracting one or more peak timings of the autocorrelation function; detecting a period of each of one or more abnormal sounds, based on the one or more peak timings; and identifying, based on the period, a causative component for each of the one or more abnormal sounds, among a plurality of components included in the apparatus.
The advantages and features provided by one or more embodiments of the invention will become more fully understood from the detailed description given hereinbelow and the appended drawings which are given by way of illustration only, and thus are not intended as a definition of the limits of the present invention.
Hereinafter, one or more embodiments of the present invention will be described with reference to the drawings. However, the scope of the invention is not limited to the disclosed embodiments.
Hereinafter, embodiments of the technical idea according to the present disclosure are described with reference to the drawings. In the following description, the same components are denoted by the same reference characters. Their names and functions are the same. Therefore, a detailed description thereof is not herein repeated. Furthermore, each embodiment and each modification or the like may be selectively combined as appropriate.
(a. Configuration of a System to which the Technology of the Present Disclosure is Applicable)
The sound diagnostic system 100 includes, by way of example, a diagnostic apparatus 120 and a server 110. According to an embodiment, the diagnostic apparatus 120 may have the function of the server 110. An apparatus 130 is a subject on which a sound diagnosis is to be conducted. The apparatus 130 is any equipment that includes a plurality of driven components. Examples of the apparatus 130 include office equipment such as an image forming apparatus, home appliances, equipment in a factory, research equipment, and the like. The diagnostic apparatus 120 is communicably connected to the server 110 via a wireless local area network (LAN) terminal such as a Wi-Fi (registered trademark) router, or via an Internet communication network.
The diagnostic apparatus 120 collects sound information on a sound produced by the apparatus 130, and displays a result of an analysis of the sound information. A serviceperson carries the diagnostic apparatus 120 in order to perform maintenance management, repair, and the like of an apparatus to be maintained, which is used by an end user. The diagnostic apparatus 120 acquires at least an abnormal sound produced by the apparatus to be maintained, and displays a result of an analysis of the acquired abnormal sound. The result of the analysis of the abnormal sound indicates a causative component that causes the abnormal sound, among a plurality of components included in the apparatus 130.
According to an embodiment, the diagnostic apparatus 120 transmits sampled sound information to the server 110. Then, the diagnostic apparatus 120 receives a result of an analysis from the server 110. Further, the diagnostic apparatus 120 displays the result of the analysis on a display. The result of the analysis indicates at least a causative component causing the abnormal sound. The diagnostic apparatus 120 and the server 110 are connected to each other via a wireless or wired network so as to be able to transmit and receive data to and from each other.
According to another embodiment, the diagnostic apparatus 120 analyzes the sampled sound information. Then, the diagnostic apparatus 120 displays the result of the analysis on the display. The result of the analysis indicates at least a causative component causing the abnormal sound. In this case, the sound diagnostic system 100 may not include the server 110.
According to still another embodiment, the diagnostic apparatus 120 is a tablet, a smartphone, a personal computer, smart glasses, or any other information processing apparatus. The diagnostic apparatus 120 may contain a normal microphone, a directional microphone, or the like. The diagnostic apparatus 120 may be configured in such a manner that a normal microphone or a directional microphone is connectable to the diagnostic apparatus.
The server 110 receives the sound information from the diagnostic apparatus 120 via the network. The server 110 analyzes the sound information and returns a result of the analysis to the diagnostic apparatus 120. The result of the analysis indicates a causative component causing the abnormal sound. A procedure for analyzing the sound information is described in detail later herein. According to an embodiment, the server 110 may be implemented as one or more apparatuses, a virtual machine built on a cloud environment, or a combination thereof.
The apparatus 130 is an apparatus to be diagnosed. The apparatus 130 includes a plurality of driven components. The driven component is a component driven by a motor, for example. Sounds produced by respective driven components may have respective frequencies different from each other.
(b. Problems that May Occur in Sound Diagnosis)
As described above, the apparatus 130 includes a plurality of driven components. For example, an image forming apparatus used in an office or the like includes numerous driven components. Therefore, even when the diagnostic apparatus 120 collects an abnormal sound, it is not easy to identify a driven component producing the abnormal sound. In order to solve this problem, the sound diagnostic system 100 calculates a sound pressure of each of the plurality of frequencies, from sound information on the sound produced by the apparatus 130. The sound diagnostic system 100 calculates an autocorrelation function for a change, with time, of a representative value of the sound pressures of respective frequencies included in a target frequency band. The sound diagnostic system 100 extracts one or more peak timings of the autocorrelation function. The sound diagnostic system 100 detects a period of each of one or more abnormal sounds, based on the one or more peak timings. The sound diagnostic system 100 identifies, based on the detected period, a causative component for each of one or more abnormal sounds, among a plurality of driven components included in the apparatus 130.
(c. Terms Used in the Present Specification)
Next, terms used herein to describe the technology of the present disclosure are described.
“System” in the present specification encompasses a configuration formed of one or multiple apparatuses. The system may also encompass a virtual machine or a container built in a cloud environment, or a configuration formed of at least a part of them.
“Apparatus” in the present specification may be any information processing apparatus such as a personal computer, a workstation, a server apparatus, a tablet, or a smartphone. The apparatus may also be a combination of them.
According to an embodiment, the sound diagnostic system 100 may be connected to input/output devices such as a display and a keyboard, and used by a user. According to another embodiment, the sound diagnostic system 100 may provide various functions to a user via a network, as a service or web application. In this case, the user may use the functions of the sound diagnostic system 100 via a browser or client software installed in a terminal of the user.
“Apparatus to be diagnosed” in the present specification encompasses an apparatus installed in any place such as a home, an office, a factory, a school, or a laboratory. By way of example, the apparatus to be diagnosed is any apparatus such as a refrigerator, an air conditioner, an image forming apparatus, a manufacturing apparatus placed in a factory, or a robot. The apparatus to be diagnosed includes a plurality of driven components. In the present specification, the apparatus to be diagnosed may be referred to simply as apparatus.
“Component (driven component)” in the present specification is a component included in an apparatus to be diagnosed. The component is typically a rotationally driven component. The component may be a motor or a combination of a motor and a power transmission component. The component may also be a combination of a motor, a power transmission component, and a component located downstream thereof. Alternatively, the component may be a component driven by a motor.
“Sound information” in the present specification relates to a sound produced by an apparatus to be diagnosed. The sound information may also be data picked up by the diagnostic apparatus 120 through a microphone, that is, sampled sound data. The sampled sound data is digital data represented at any resolution.
“Sound pressure” in the present specification is a deviation from the atmospheric pressure caused by a sound. The sound pressure is defined as an effective value of the maximum amplitude.
“Overall sound pressure (overall sound intensity)” in the present specification is a representative value of sound pressures of respective frequency components included in a target frequency band, at each timing of a spectrogram. An overall sound pressure (a representative value of sound pressures) is, for example, a sum or an average value. The spectrogram includes a plurality of frequency components for each timing. For example, it is assumed that frequency components X, Y, and Z are included in the target frequency band, at timing A of the spectrogram. In this case, the overall sound pressure (overall sound intensity) at timing A is calculated by an equation: “overall sound pressure at timing A=a sound pressure of frequency component X+a sound pressure of frequency component Y+a sound pressure of frequency component Z.” Alternatively, the overall sound pressure (overall sound intensity) at timing A may be an average value of the sound pressures of respective frequency components X, Y, and Z.
“Autocorrelation function” in the present specification indicates a relationship between a time shift amount τ and an autocorrelation between a time waveform x(t) of an overall sound pressure (a representative value of sound pressures) and a waveform x(t+τ) obtained by shifting x(t) by the time shift amount τ. “Autocorrelation” represents the degree of similarity between the waveform x(t) and the waveform x(t+τ). “Autocorrelation” is also referred to as “autocorrelation coefficient.”
When x(t) is a discrete signal, the autocorrelation function R(τ) is typically expressed by the following Expression (1). Note that the autocorrelation function R(τ) may be normalized using the mean and the variance of x(t) so as to have a value of −1 to 1.
“Peak timing” in the present specification is a timing having an autocorrelation higher than an autocorrelation at a preceding timing and an autocorrelation at a subsequent timing, by a certain extent or more, in a autocorrelation graph.
(a. Apparatus Configuration)
The sound diagnostic system 100 includes a microphone (mic) 202, a sound collection unit 204, a sound pressure calculation unit 205, an autocorrelation calculation unit 206, a peak extraction unit 208, and a detection unit 210. Furthermore, the sound diagnostic system 100 includes a storage unit 212, an identification unit 222, and an output unit 224. The storage unit 212 stores period information 214 and a user interface (UI) program 218.
According to an aspect, the diagnostic apparatus 120 includes the microphone 202, the sound collection unit 204, and the output unit 224. The server 110 includes the sound pressure calculation unit 205, the autocorrelation calculation unit 206, the peak extraction unit 208, the detection unit 210, the storage unit 212, and the identification unit 222.
According to another aspect, the diagnostic apparatus 120 may have the function of the server 110 in addition to the microphone (mic) 202, the sound collection unit 204, and the output unit 224. In this case, the diagnostic apparatus 120 includes the sound pressure calculation unit 205, the autocorrelation calculation unit 206, the peak extraction unit 208, the detection unit 210, the storage unit 212, and the identification unit 222.
The microphone 202 and the sound collection unit 204 operate as a collection unit that collects sound information on a sound produced by the apparatus.
The microphone 202 converts a sound into an electrical signal and outputs the electrical signal to the sound collection unit 204. The apparatus 130 includes a plurality of sound sources (driven components). Therefore, the microphone 202 picks up a synthesized sound produced by a plurality of sound sources.
The sound collection unit 204 converts the electrical signal output from the microphone 202 into a digital signal to generate sound information (a sampling result). The sound collection unit 204 converts an analog signal into sound information that is a digital signal at any resolution. The sound collection unit 204 outputs the generated sound information.
The sound pressure calculation unit 205 is an example of “first calculation unit” in the present disclosure. The sound pressure calculation unit 205 calculates a sound pressure of each of a plurality of frequencies included in each timing of sound information. More specifically, the sound pressure calculation unit 205 performs a short-time Fourier transform (STFT) on the sound information (sound data). The sound pressure calculation unit 205 performs the STFT on the sound information to thereby obtain a frequency spectrum waveform (spectrogram). The spectrogram represents a change, with time, in sound pressure distribution for each frequency.
The sound pressure calculation unit 205 calculates, for each timing, an overall sound pressure (a representative value of sound pressures) of frequencies included in the target frequency band, among the plurality of frequencies. The overall sound pressure is, for example, the sum or the average value of the sound pressures of respective frequencies. The target frequency band may include all of the plurality of frequencies at which the sound pressure is calculated by the sound pressure calculation unit 205, or may include only a part of the plurality of frequencies.
The autocorrelation calculation unit 206 is an example of “second calculation unit” in the present disclosure. The autocorrelation calculation unit 206 calculates an autocorrelation function for a change, with time, in overall sound pressure (a representative value of sound pressures) of frequencies included in a target frequency band. That is, the autocorrelation calculation unit 206 calculates an autocorrelation function for the time waveform x(t) of the overall sound pressure of the frequencies included in the target frequency band. The autocorrelation calculation unit 206 calculates an autocorrelation between the time waveform x(t) and the time waveform x(t+τ) obtained by shifting the time waveform x(t) by the time shift amount τ. The autocorrelation calculation unit 206 calculates an autocorrelation function R(τ) representing a relationship between the time shift amount τ and the autocorrelation, while changing the time shift amount τ within a predetermined range. The autocorrelation calculation unit 206 is described in detail later herein with reference to
The peak extraction unit 208 extracts one or more peak timings of the autocorrelation function R(τ). More specifically, the peak extraction unit 208 extracts, as a peak timing, a time shift amount τ at which an autocorrelation differs from adjacent autocorrelations at respective timings by a preset value or more. Alternatively, the peak extraction unit 208 may extract the peak timing based on a prominence value (a degree of projection) of the autocorrelation. The peak extraction unit 208 is described in detail later herein with reference to
The detection unit 210 detects the period of each of one or more abnormal sounds, based on one or more peak timings. For example, the detection unit 210 detects, as a period, a peak timing at which the height of the peak (prominence value) is largest among the extracted one or more peak timings. Alternatively, when a plurality of peak timings are extracted, the detection unit 210 may detect the period of an abnormal sound based on the interval between peak timings. For example, when the peak timing 1.5 sec and the peak timing 3.0 sec are extracted, the detection unit 210 may determine that the period is 3.0−1.5=1.5 sec.
The detection unit 210 may detect respective periods of a plurality of abnormal sounds, from a plurality of peak timings. For example, the detection unit 210 detects, as a period of a first abnormal sound, the shortest peak timing among the plurality of peak timings. Then, the detection unit 210 extracts, from the plurality of peak timings, one or more remaining peak timings at which the difference from an integer multiple of the period of the first abnormal sound exceeds a threshold value. The detection unit 210 detects, as a period of the second abnormal sound, the shortest peak timing among the extracted one or more remaining peak timings. The detection unit 210 may detect the period of the k-th abnormal sound by repeating the same process. k is an integer of 3 or more.
The storage unit 212 stores data required for processing by the identification unit 222 and data of a UI to be provided to a user. The storage unit 212 stores period information 214. The period information 214 indicates the period of a sound produced by each driven component. The storage unit 212 stores the UI program 218. The UI program 218 is data or a program of a UI to be provided to a user. The UI program 218 includes UI design data, an event program associated with each UI component, and the like.
The output unit 224 outputs a result of an analysis of the sound information. The result of the analysis indicates at least a causative component causing the abnormal sound. By way of example, the output unit 224 displays the screens illustrated in
(b. Procedure of Sound Diagnosis)
In step S301, the diagnostic apparatus 120 picks up a sound of the apparatus 130 to be diagnosed through a microphone (microphone 202). More specifically, the microphone 202 of the diagnostic apparatus 120 picks up a sound produced by the apparatus 130, in response to a user's (e.g., a serviceperson's) operation. In addition, the sound collection unit 204 converts the sound (an analog signal) into sampled sound information (digital data). The diagnostic apparatus 120 transmits the sound information to the server 110. In connection with the following steps, it is assumed that the sound information is to be transmitted to the server 110. Note that in one aspect in which the diagnostic apparatus 120 additionally has the function of the server 110, the diagnostic apparatus 120 does not transmit the sound information to the server 110. By way of example, the diagnostic apparatus 120 acquires sound information 400 as illustrated in
In step S302, the server 110 performs STFT on the acquired sound information. Thus, the server 110 calculates a spectrogram representing a change, with time, in sound pressure distribution for each frequency. By way of example, the server 110 acquires a spectrogram 500 as illustrated in
In step S303, the server 110 calculates, for each timing, an overall sound pressure (a representative value of sound pressures) of all of a plurality of frequencies included in the spectrogram. Typically, the server 110 calculates, as the overall sound pressure, the sum of the sound pressures of respective frequencies for each timing in the spectrogram. Alternatively, the server 110 calculates, as the overall sound pressure, the average value of the sound pressures of respective frequencies for each timing in the spectrogram. By way of example, the server 110 acquires data 600 as illustrated in
In step S304, the server 110 calculates an autocorrelation function for a change, with time, in overall sound pressure. By way of example, the server 110 calculates an autocorrelation function 700 as illustrated in
In step S305, the server 110 extracts one or more peak timings of the autocorrelation function. The peak timing is a timing at which the autocorrelation is larger than those of the preceding and following timings by a certain degree or more, in the graph of the autocorrelation. In other words, it is a portion protruding to a greater extent in the graph of the autocorrelation. By way of example, the server 110 acquires peak timing information 800 indicating peak timings as illustrated in
In step S306, the server 110 determines whether one or more peak timings have been extracted.
When one or more peak timings have not been extracted (NO in step S306), the process proceeds to step S307. In step S307, the server 110 generates result information indicating that the causative component of the abnormal sound cannot be identified.
When one or more peak timings have been extracted (NO in step S306), the process proceeds to step S308. In step S308, the server 110 detects a period of each of one or more abnormal sounds, on the basis of one or more peak timings indicated by the peak timing information 800.
In step S309, the server 110 determines whether a period effective for identifying a causative component of the abnormal sound has been detected. For example, the server 110 determines that a period having a difference from any period indicated by the period information 214 that is less than a predetermined threshold value is an effective period.
When the period effective for identifying the causative component of the abnormal sound is not detected (NO in step S309), the process proceeds to step S307.
When the period effective for identifying the causative component of the abnormal sound is detected (YES in step S309), the process proceeds to step S310. In step S310, the server 110 identifies, as a causative component, a driven component that produces a sound with a period different from the detected period by a predetermined threshold value or less, based on the period information 214. The server 110 generates result information indicating the identified causing component.
After step S307 or step S310, the process proceeds to step S311. In step S311, the server 110 transmits the result information to the diagnostic apparatus 120. The diagnostic apparatus 120 displays the result indicated by the received result information on a display. According to an aspect, the diagnostic apparatus 120 may transfer the result information to another apparatus (such as a PC or a smartphone).
(c. Details of Each Procedure)
Next, details of each operation of the flow in
For example, the time waveform x(t) is expressed as a discrete function corresponding to a time resolution Δt of the spectrogram calculated by the sound pressure calculation unit 205. Δt is, for example, 0.005 sec. In this case, the autocorrelation calculation unit 206 calculates an autocorrelation between the time waveform x(t) and the time waveform x(t+n×Δt). n is a natural number and is also referred to as a lag. The time shift amount τ is represented by n×Δt. The autocorrelation calculation unit 206 calculates the autocorrelation function 700 by changing the lag n within a predetermined range.
Note that it is also possible to calculate an autocorrelation function for the sound information 400 illustrated in
Furthermore, in order to further reduce the calculation load for calculating the autocorrelation function 700, the range of the time shift amount τ may be determined depending on the maximum rotation period that can be taken by a plurality of components of the apparatus 130. For example, when the maximum period indicated by the period information 214 is three seconds, three seconds or a time slightly longer than three seconds is determined as the upper limit value of the range of the time shift amount τ. The server 110 calculates the autocorrelation function 700 by changing the time shift amount τ in a range from 0 to the upper limit value.
When a periodic abnormal sound is generated by a driven component, the overall sound pressure for each time has periodicity. In this case, as illustrated in
It is preferable that the server 110 removes a trend component from the change with time indicated by the data 600, in order to more accurately determine the period of the abnormal sound. The trend component represents a tendency of data over a long period of time, such as a certain increasing tendency or decreasing tendency, which occurs with passage of time. Therefore, the server 110 removes the trend component, for example, by applying a high-pass filter to the data 600. The server 110 may calculate an autocorrelation function for a change, with time, in the overall sound pressure from which the trend component has been removed.
The server 110 (peak extraction unit 208) changes the time shift amount τ from 0 to the upper limit value successively, and determines whether a target time shift amount is a peak timing.
Specifically, the server 110 (peak extraction unit 208) compares an autocorrelation for the target time shift amount with an autocorrelation for an adjacent time shift amount. It is assumed that the autocorrelation for the target time shift amount is higher than the autocorrelation for the adjacent time shift amount by a predetermined threshold value or more. In this case, the server 110 (peak extraction unit 208) determines that the target time shift amount is a peak timing. The server 110 acquires the autocorrelation at that time as a peak value. In
Note that the server 110 may exclude a peak representing an autocorrelation equal to or less than a preset threshold value (for example, 0.1). The server 110 extracts, as a peak timing, a time shift amount τ for the peak is left without being excluded.
A specific method of calculating the prominence is as follows. First, the server 110 focuses on a certain peak. Next, the server 110 extends rightward and leftward the horizontal line of the peak, until the horizontal line of the peak intersects with a higher peak or reaches the right or left end (graph end) of the signal. Next, the server 110 finds signals of which autocorrelations are a minimum value, on the right side and the left side of a section where the horizontal line is extended. Next, the server 110 sets a larger one of the minimum value of the right section and the minimum value of the left section, as a reference value. Next, the server 110 sets a difference between the autocorrelation of the peak and the reference value, as a prominence value. The prominence value indicates the degree of projection from the surroundings (the height of the peak relative to the surroundings). The server 110 excludes peaks equal to or smaller than a predetermined prominence value (e.g., 0.1). The server 110 extracts, as the peak timing, the time shift amount τ of the peak that is left without being excluded.
Note that the server 110 may impose a limitation on the length of left and right extensions of the horizontal line in order to calculate a more accurate prominence value.
According to an aspect, after detecting the target time shift amount as the peak timing, the server 110 may not detect the time shift amount in a certain period from the peak timing as a peak timing. Thus, successive detection of peaks due to noise is suppressed.
According to an aspect, the server 110 may restrict the one or more peak timings based on peak information. The peak information indicates the shape of the peak. The peak information indicates, for example, the height of the peak (the value of the autocorrelation or prominence value), the kurtosis of the peak, or the difference or ratio between the autocorrelation of the peak and the maximum value of the autocorrelation. For example, the server 110 excludes, from the peak timing, a time shift amount having an autocorrelation of which difference from the maximum value of the autocorrelation is less than or equal to a predetermined threshold value (for example, a value half the maximum value). Thus, the time shift amount having a smaller autocorrelation is not erroneously detected as the peak timing. Alternatively, the server 110 may exclude, from the peak timing, a time shift amount by which the height of the peak is equal to or smaller than a predetermined threshold value. Alternatively, the server 110 may exclude, from the peak timing, a time shift amount by which the kurtosis of the peak is equal to or smaller than a predetermined threshold.
The server 110 (the detection unit 210) detects the period of each of one or more abnormal sounds on the basis of one or more peak timings.
The identification unit 222 checks the detected period with the period information 214 to identify a causative component causing the abnormal sound.
For example, the result information 1100 illustrated in
According to an aspect, the identification unit 222 may select a plurality of components as candidates for the causative component of the abnormal sound having the detected period. For example, as illustrated in
By way of example, the identification unit 222 may determine that a component p is a candidate for a causative component of an abnormal sound having a detected period, based on the following Expressions (2) to (5).
The first screen 1200 is a screen before sound diagnosis is conducted. The first screen 1200 includes buttons or the like for parameter settings necessary for analysis, and for execution of recording and analysis. On the first screen 1200, a user (serviceperson or the like) may select a sound file (information on collected sound) and press an analysis button to acquire an analysis result. When the analysis button is pressed, the first screen 1200 transitions to a second screen 1300.
The second screen 1300 is a screen showing a result of an analysis of the sound file (information on collected sound). The second screen 1300 displays the period of an abnormal sound detected based on the peak timing of the autocorrelation function, and the name of a component that produces a sound having a period close to the period of the abnormal sound. The second screen 1300 may display a plurality of component names when there are a plurality of components that generate a sound having a period close to the period of the detected abnormal sound.
As described above, the target frequency band for which the overall sound pressure is calculated may include all of a plurality of frequencies included in the spectrogram 500. Alternatively, the target frequency band for which the overall sound pressure is calculated may include only a part of the plurality of frequencies included in the spectrogram 500.
For example, an image forming apparatus includes a plurality of rollers. An abnormal sound such as a rubbing sound produced by the rollers has only some frequency components and has a low sound pressure. Therefore, when the overall sound pressure is calculated for all frequency components, the abnormal sound having only some of the frequency components is buried in environmental sounds having other frequency components. Thus, as illustrated in
As illustrated in
The sound diagnostic system according to Embodiment 2 presets a plurality of frequency bands. The plurality of frequency bands include a first frequency band and one or more second frequency bands. The first frequency band includes all of a plurality of frequencies included in the spectrogram. Each of the one or more second frequency bands includes only some of a plurality of frequencies included in the spectrogram. When the one or more second frequency bands include a plurality of second frequency bands, the plurality of second frequency bands are different from each other. Each of the one or more second frequency bands is a part of the first frequency band.
The one or more second frequency bands include, for example, frequency bands of “0 to 861 Hz,” “861 to 1722 Hz” and “1722 to 2583 Hz.”
Note that a certain second frequency band may partially overlap another second frequency band. For example, the one or more second frequency bands may include frequency bands of “0 to 861 Hz,” “761 to 1722 Hz” and “1622 to 2583 Hz.”
In step S1901, the server 110 (autocorrelation calculation unit 206) selects, as a target frequency band, an unselected one of the one or more second frequency bands. The second frequency band selected as the target frequency band corresponds to “second target frequency band” in the present disclosure.
In step S1902, the server 110 calculates, for each timing, an overall sound pressures (a representative values of sound pressures) of frequencies included in the target frequency band.
In step S1903, the server 110 calculates an autocorrelation function for a change, with time, of the overall sound pressure.
In step S1904, the server 110 (peak extraction unit 208) extracts one or more peak timings in the autocorrelation function.
In step S1905, the server 110 (detection unit 210) detects the period of each of one or more abnormal sounds on the basis of one or more peak timings.
In step S1906, the server 110 registers, in period candidate information illustrated in
Note that when no peak timing is extracted in step S1904, steps S1905 and S1906 are skipped.
In step S1907, the server 110 determines whether all of the one or more second frequency bands have been selected each as a target frequency band.
When all of the one or more second frequency bands have not been selected each as a target frequency band (NO in step S1907), the process returns to step S1901.
When all of the one or more second frequency bands have been selected each as a target frequency band (YES in step S1907), the process proceeds to step S1908. In step S1908, the server 110 (identification unit 222) selects a period corresponding to the largest peak height from period candidate information.
After step S1908, the process proceeds to step S310. In step S310 subsequent to step S1908, the server 110 (identification unit 222) identifies a causative component. Specifically, the server 110 identifies, as a causative component, a driven component that produces a sound having a period different from the period selected in step S1908 by less than a predetermined threshold value.
Thus, the server 110 according to Embodiment 2 first calculates a first autocorrelation function corresponding to a frequency band including all of a plurality of frequencies included in the spectrogram. Then, in response to the fact that no causative component is identified based on the first autocorrelation function, the server 110 performs steps S1901 to S1908. Specifically, the server 110 calculates a second autocorrelation function corresponding to each of the one or more second frequency bands which are a part of the first frequency band. Then, the server 110 identifies a causative component based on the second autocorrelation function.
Although the configuration in which the diagnostic apparatus 120 includes the microphone 202 is described herein, application examples of the technology of the present disclosure are not limited thereto. According to an aspect, the apparatus 130 to be diagnosed may include a microphone. In this case, the apparatus 130 transmits the sound information to the diagnostic apparatus 120 or the server 110.
The sound information may be analog sound data. In this case, the detection unit 210 may have a function of encoding analog sound data. The sound information may also be digital sound data. By way of example, the sound collection unit 204 may perform encoding or the like of a sound captured by the microphone 202. The sound information including a plurality of frequencies produced by the apparatus 130 may include a sampling result (sampling information) of a recorded sound. A predetermined period for extracting information on a plurality of sound pressures may be a period of any length set in advance in the sound diagnostic system 100. For example, the predetermined period may include a period of any length such as a few milliseconds, a few seconds, and a few minutes. The information on the sound pressures may include information such as the overall sound pressure of sounds of respective frequencies, a relative sound pressure of a sound of each frequency, and a peak timing included in the overall sound pressure.
In an aspect, the apparatus 130 may be an image forming apparatus including a plurality of drive sources. In this case, the plurality of drive sources may drive any of a fixing belt, a sheet ejection roller, a sheet feed roller, and a fixing gear.
In the present specification, timing may include not only an instant of time in a series of time but also a time (period) of any length. By way of example, the sound diagnostic system 100 may analyze sound information delimited every 1 microsecond. In this case, a certain timing means a period of one microsecond. Thus, the timing may be any period of time, such as one microsecond, one millisecond, or the like. The sound pressure within a certain period (e.g., a period of one microsecond) may for example be the average value or the median value of sound pressures within the period.
All or some of the sound pressure calculation unit 205, the autocorrelation calculation unit 206, the peak extraction unit 208, the detection unit 210, the storage unit 212, the identification unit 222, and the output unit 224 may be implemented by a computer.
The processor 1 includes, for example, a CPU (central processing unit), an MPU (micro processing unit), or the like, and executes a program 7 stored in the storage device 3.
The RAM 2 stores, in a volatile manner, data generated through execution of the program 7 by the processor 1, data input via the input device 5, data externally received via the communication interface 4, and the like.
The storage device 3 stores data in a nonvolatile manner. The storage device 3 is implemented by a hard disk device, an SSD (solid state drive), or the like.
The input device 5 includes, for example, a touch panel, a keyboard, a mouse, and the like. The display 6 is, for example, a liquid crystal display or an organic EL (Electro Luminescence) display.
Processing in the computer 2100 is implemented through cooperation of the program 7 executed by each hardware device and the processor 1. The program 7 is stored in the storage device 3 in advance. Alternatively, the program 7 may be stored in a CD-ROM (Compact Disc-Read Only Memory) or another recording medium and distributed. Alternatively, the program 7 may also be provided as a downloadable application program by an information provider connected to the so-called Internet. Such a program 7 is read from the recording medium by an optical disc drive (not shown) or another reading device, or is downloaded via the communication interface 4, and then stored temporarily in the storage device 3.
Each component of the computer 2100 illustrated in
Note that the recording medium is not limited to a CD-ROM, an FD (Flexible Disk), and a hard disk, but may be a magnetic tape, a cassette tape, an optical disc (MO (Magnetic Optical Disc)/MD (Mini Disc)/DVD (Digital Versatile Disc)), an IC (Integrated Circuit) card (including a memory card), an optical card, or a medium carrying the program in a stationary manner like a semiconductor memory such as a mask ROM, an EPROM (Electronically Programmable Read-Only Memory), an EEPROM (Electronically Erasable Programmable Read-Only Memory), or a flash ROM.
The program 7 includes not only a program directly executable by the processor 1 but also a program in the form of a source program, a compressed program, an encrypted program, and the like.
As described above, the present embodiment includes the following disclosure.
A sound diagnostic system including:
The sound diagnostic system according to configuration 1, wherein the target frequency band includes all of the plurality of frequencies or only a part of the plurality of frequencies.
The sound diagnostic system according to configuration 1, wherein
The sound diagnostic system according to any one of configurations 1 to 3, wherein the second calculation unit determines a range of a time shift amount of the autocorrelation function, in accordance with a maximum rotation period that the plurality of components can take.
The sound diagnostic system according to any one of configurations 1 to 4, wherein
The sound diagnostic system according to any one of configurations 1 to 5, wherein the extraction unit extracts a target time shift amount as one of the one or more peak timings, in response to a fact that an autocorrelation at the target time shift amount differs from respective autocorrelations preceding and subsequent to the target time shift amount by more than or equal to a predetermined threshold value.
The sound diagnostic system according to any one of configurations 1 to 6, wherein the extraction unit restricts the one or more peak timings based on peak information indicating a shape of a peak.
The sound diagnostic system according to any one of configurations 1 to 7, wherein the identification unit identifies the causative component, using period information that associates each of the plurality of components with a rotation period.
A sound diagnostic method including:
A sound diagnostic program that causes a computer to perform a sound diagnostic method, the sound diagnostic method includes:
It should be understood that the embodiments disclosed herein are illustrative and non-restrictive in every respect. The scope of the present disclosure is defined not by the above description but by the appended claims, and all changes that come within the meaning and range of equivalency of the claims are intended to be embraced therein. In addition, it is intended that the disclosed contents described in the embodiment and the modifications can be implemented alone or in combination as much as possible.
Although embodiments of the present invention have been described and illustrated in detail, the disclosed embodiments are made for purposes of illustration and example only and not limitation. The scope of the present invention should be interpreted by terms of the appended claims.
Number | Date | Country | Kind |
---|---|---|---|
2023-150743 | Sep 2023 | JP | national |