The present disclosure relates to a data processing method, in particular, to a method and device of processing data conforming to statistical distribution.
Spectroscopy such as angle-resolved photoelectron spectroscopy (ARPES) and momentum resolved photoemission electron microscopy (k-PEEM) is used for the research of microscopic structure, for example, electronic structure of a wide range of materials. As the technology advances, its coverage increases to two and three dimensions. Accordingly, the time for acquiring data with given signal-to-noise ratio (SNR) may increase and insufficient data acquisition time may increase noise.
An example of the methods of reducing the noise of data is Gaussian smoothing, but it may cause data blurring that may cause damage on the essential information.
The embodiments of the present disclosure is provided to reduce the data acquisition time and noise.
A data processing method according to an embodiment of the present disclosure includes: training a neural network; receiving input data from an external source, and converting the input data by the trained neural network, wherein the training comprises: generating one or more generated data from an original data; converting, by the neural network, the generated data into an output data; estimating the output data based on the original data; and optimizing the neural network based on result of the estimation, wherein the original data and the generated data conform to a statistical distribution, and wherein the original data and the output data have signal-to-noise ratio higher than the generated data.
The generating may include generating the generated data at random.
The statistical distribution may include Poisson distribution.
The input data and the original data may include spectroscopy data.
The input data and the original data may include angle-resolved photoelectron spectroscopy (ARPES) count data.
The number of the generated data may be equal to or greater than two.
The neural network may include a deep neural network.
The neural network may include a deep convolutional neural network.
The deep convolutional neural network may include equal to or less than 20 layers.
The estimating may be performed using a loss function; and the loss function may include weighted sum of mean absolute error and multiscale structural similarity index.
A data processing device according to an embodiment of the present disclosure includes: a processor converting input data, wherein the processor comprises: a receiver receiving an original data; a generator generating one or more generated data at random from the original data; a neural network converting the generated data into an output data; and an estimator estimating the output data based on the original data, wherein the original data and the generated data conform to a statistical distribution, wherein the original data and the output data have signal-to-noise ratio higher than the generated data, and wherein the neural network is optimized according to output of the estimator.
The statistical distribution may include Poisson distribution.
The input data and the original data may include angle-resolved photoelectron spectroscopy (ARPES) count data.
The number of the generated data may be equal to or greater than two.
The neural network may include a deep convolutional neural network having 20 layers or less.
The estimator may estimate using a loss function; and the loss function may include weighted sum of mean absolute error and multiscale structural similarity index.
The embodiments of the present disclosure can reduce the data acquisition time with further reduced noise.
The embodiments of the present disclosure will be described in detail with reference to accompanying drawings such that those skill in the art can easily implement the present invention. However, the embodiments of the present invention may not be limited thereto and achieved in various forms.
Referring to
The input data IN may include spectroscopy data, for example, the data measured using angle-resolved photoelectron spectroscopy (ARPES), momentum resolved photoemission electron microscopy (k-PEEM), etc. An example of the input data IN may include ARPES count data.
The processor 1 may convert the input data IN to produce the output data OUT, which has signal-to-noise ratio (SNR) higher than that of the input data IN. For example, when the input data IN is count data, the output data OUT from the processor 1 may be similar to those higher than the input data IN.
Referring to
The neural network 10 may include, for example, deep neural network or deep convolutional neural network, and may be trained using training data.
The training data may include a plurality of data pairs of original data Di and generated data Dg.
the receiver 20 is connected to the generator 30 and the estimator 40, and may receive the original data Di from an external source such as ARPES electronic analyzer (not shown), and transmit the original data Di to the generator 30 and the estimator 40.
The generator 30, connected to the receiver 20 and the neural network 10, may generate the generated data Dg from the original data Di received from the receiver 20 using the statistical distribution to which the original data Di conforms and may transmit the generated data Dg to the neural network 10. The original data Di and the generated data Dg may conform to given statistical distribution such as Poisson distribution, but the embodiment is not limited thereto. The generated data Dg may be randomly generated from the original data Di by the generator 30, and two or more generated data Dg may be generated from one original datum Di. The generation of the generated data Dg may reduce overfitting of the neural network 10 during the training. For data augmentation, the data set may be randomly rotated or flipped.
According to an embodiment of the present disclosure, the generated data Dg has a lower SNR than the original data Di. For example, the original data Di may be ARPES high count data while the generated data Dg may be low count data. In
Referring to
The estimator 40 may estimate the quality of the output data Do with reference to the original data Di. The estimation of the quality of the output data Do may use loss function or cost function, and the loss function may be defined, for example, as weighted sum of mean absolute error and multiscale structural similarity index. The estimation result may be fed back to the neural network 10. The neural network 10 may perform optimization, for example, adjusting inner parameters, and may repeat the generation of the output data Do, thereby proceeding the training. An example of optimization may include Adam optimization.
During the training, the intensity of data may be randomly adjusted, and overfitting may be carefully checked using a validation data set which is not used in the training.
Some experimental examples of a data processing device according to embodiments of the present disclosure will be described in detail with reference to
A deep network with 20 convolution layers was used in these experiments. Each layer of the convolutional neural network has a filter number of 64 and a kernel size of 3. An Adam optimizer was adopted to train the network for 150 epochs. The learning rate was initially set to be 5×10−4 and multiplied by 0.1 after every 50 epochs. For the training data set, 50 different original high-count ARPES data were used, and 50 low-count data for each original data were randomly generated, resulting in a total of 2500 low-count data.
Referring to
Referring to
Referring to
The reason why the data processing, for example, ARPES data processing is effective for ARPES data as described above may be the correlation of the data among the neighboring pixels. Two major factors contribute to the correlation. First, the dimension for ARPES features is larger than the pixel size, leading to occupation of several pixels for an ARPES feature. Thus, if the value at a pixel is large, it is likely for neighboring pixels to have a large value. Second, the length scale over which the band structure changes is larger than the data pixel size. Hence, the band structure does not change abruptly in a pixel. This means that the band structure has an approximate translational symmetry in a short length scale. Even if the information at a pixel is corrupted with noise, the value at the pixel can be recovered from the most statistically probable value inferred from adjacent pixel values. Therefore, a data set carries more information than just the pixel-wise sum of information. With the additional information, the seemingly imperfect information of the noisy data can be recovered.
As such, since the information at a pixel is compensated by the information from adjacent pixel values, better performance is expected if there are more neighboring pixels. Therefore, the data processing according to embodiments of the present disclosure is more effective for the higher-dimensional data than for the lower dimensional data. The data processing according to embodiments of the present disclosure is more useful in higher dimension because the data acquisition time for higher-dimensional data is longer.
Referring to
The embodiments of the present disclosure can reduce the data acquisition time with further reduced noise.
The embodiments of the present disclosure can reduce the data acquisition time with further reduced noise.
| Number | Date | Country | Kind |
|---|---|---|---|
| 10-2021-0041786 | Mar 2021 | KR | national |
| Filing Document | Filing Date | Country | Kind |
|---|---|---|---|
| PCT/KR2022/004518 | 3/30/2022 | WO |