OPTICAL 4f SYSTEM PERFORMING EXTENDED CONVOLUTION OPERATION AND OPERATING METHOD THEREOF

Information

  • Patent Application
  • 20230206620
  • Publication Number
    20230206620
  • Date Filed
    December 21, 2022
    a year ago
  • Date Published
    June 29, 2023
    a year ago
  • Inventors
  • Original Assignees
    • ELECTRONICS AND TELECOMMUNICATIONS RESAERCH INSTITUTE
Abstract
Disclosed is an optical system which includes a first lens that receives input data from an object, a kernel that performs a first Fourier transform on the input data and generates learning data by performing calculation on a result of the first Fourier transform and pattern data, and a second lens that generates result data by performing a second Fourier transform on the learning data, and the input data, the learning data, and the result data include both a positive number and a negative number.
Description
CROSS-REFERENCE TO RELATED APPLICATIONS

This application claims priority under 35 U.S.C. § 119 to Korean Patent Application No. 10-2021-0185587 filed on Dec. 23, 2021 and No. 10-2022-0173746 filed on Dec. 13, 2022, in the Korean Intellectual Property Office, the disclosures of which are incorporated by reference herein in their entireties.


BACKGROUND
1. Field of the Invention

Embodiments of the present disclosure described herein relate to a convolution operation, and more particularly, relate to an optical 4f system performing extended convolution operations and an operating method thereof.


2. Description of Related Art

A convolutional neural network (CNN) is a type of deep neural network that includes one or more convolution layers. The CNN provides good performance in image recognition, video recognition, image classification, medical image analysis, natural language processing, and time-series analysis, and fields to which the CNN is applied continue to expand.


An optical 4f system refers a system that has a structure in which two lenses are paired and transfers an image from a first lens of one side to a second lens of an opposite side; when the first and second lenses have the same focal length of “f”, a distance from an object plane (or an input plane) to the first lens is “f”, a distance from the first lens to the second lens is 2f, and a distance from the second lens to an image plane (or an output plane) is “f”. That is, a total length from the object plane to the image plane is 4f.


Nowadays, the convolution operation using the optical 4f system is being actively researched. However, because an input and an output of the optical 4f system are proportional to the intensity of light but a value used in the convolution operation is based on the amplitude of light, it is necessary to correct the input and the output. Also, because the amplitude of light is expressed only by a positive number, when the convolution operation is performed only by using the optical 4f system without a separate auxiliary device, it is impossible to receive a value including a negative number.


SUMMARY

Embodiments of the present disclosure provide a method for extending a convolution operation of an optical 4f system such that a convolution operation is performed on an input including a negative number in the optical 4f system, an optical 4f system performing the extended convolution operation, and an operating method of the optical 4f system.


According to an embodiment, an optical system includes a first lens that receives input data from an object, a kernel that performs a first Fourier transform on the input data and generates learning data by performing calculation on a result of the first Fourier transform and pattern data, and a second lens that generates result data by performing a second Fourier transform on the learning data, and the input data, the learning data, and the result data include both a positive number and a negative number.


According to an embodiment, an operating method of an optical system which includes a first lens, a kernel, and a second lens includes receiving input data from an object through the first lens, performing a first Fourier transform on the input data through the kernel, generating learning data by performing calculation on a result of the first Fourier transform and pattern data, and generating result data by performing a second Fourier transform on the learning data, and the input data, the learning data, and the result data include both a positive number and a negative number.





BRIEF DESCRIPTION OF THE DRAWINGS

The above and other objects and features of the present disclosure will become apparent by describing in detail embodiments thereof with reference to the accompanying drawings.



FIG. 1 is a block diagram of an optical 4f system according to an embodiment of the present disclosure.



FIG. 2 is a diagram illustrating a result of expressing values associated with an operation performed in an optical 4f system of FIG. 1 through a light amplitude and a light intensity.



FIG. 3 illustrates operation rules of Fourier transform and operation rules of inverse Fourier transform, which are used to extend a convolution operation.



FIG. 4 illustrates theorems used for extension of a convolution operation.



FIG. 5 illustrates theorems used for extension of an extended convolution operation.



FIG. 6 is a flowchart illustrating an example of an operating method of an optical 4f system according to an embodiment of the present disclosure.





DETAILED DESCRIPTION

Below, embodiments of the present disclosure will be described in detail and clearly to such an extent that one skilled in the art easily carries out the present disclosure.


In the detailed description, components described with reference to the terms “unit”, “module”, “block”, “˜er or ˜or”, etc. and function blocks illustrated in drawings will be implemented with software, hardware, or a combination thereof. For example, the software may be a machine code, firmware, an embedded code, and application software. For example, the hardware may include an electrical circuit, an electronic circuit, a processor, a computer, an integrated circuit, integrated circuit cores, a pressure sensor, an inertial sensor, a microelectromechanical system (MEMS), a passive element, or a combination thereof.



FIG. 1 is a diagram illustrating an example of an optical 4f system 100 according to an embodiment of the present disclosure. Referring to FIG. 1, the optical 4f system 100 may include a first lens 110, a kernel 120, and a second lens 130. Assuming that a focal length of each of the first lens 110 and the second lens 130 is “f”, a distance between an object 10 and the first lens 110, a distance between the first lens 110 and the second lens 130, and a distance between the second lens 130 and an image 20 are “f”, 2f, and “f”; in this case, a total of length may be 4f.


The first lens 110 may receive input data f(x,y) from the object 10 and may perform a first Fourier transform at a location of the kernel 120, and the second lens 130 may perform a second Fourier transform. For example, the kernel 120 may serve as a filter between the first Fourier transform and the second Fourier transform and may include pattern data G(u,v) associated with learning. The convolution operation of the optical 4f system 100 may be implemented by performing the Fourier transform continuously two times. That is, the optical 4f system 100 may output the image 20 corresponding to a result of the convolution operation of the kernel 120 with the object 10 through the lenses 110 and 130. Herein, because the Fourier transform is performed two times, the object 10 and the image 20 are inverted (or reverse) images.


In detail, f(x,y) may correspond to the first input image or an input value of a previous CNN layer, G(u,v) may correspond to an image pattern trained in the CNN, and custom-character−1{custom-character{f(x,y)}(u,v)·G(u,v)}(x,y) may correspond to a value (i.e., the output image 20) generated by the optical 4f system 100. {f(x,y)}(u,v) may indicate the first Fourier transform while passing through the first lens 110, which corresponds to the transformation from the x-y plane to the u-v plane. {f(x,y)}(u,v)·G(u,v) may correspond to an image pattern learning result in the form of multiplying the result of the first Fourier transform and a learning image pattern G(u,v) together. Also, custom-character−1{custom-character{f(x,y)}(u,v)·G(u,v)}(x,y) may indicate the second Fourier transform performed while {f(x,y)}(u,v)·G(u,v) obtained by multiplying the result of the first Fourier transform and G(u,v) together passes through the second lens 130 and may be expressed in the form of inverse Fourier transform for the convenience of equation expansion in the specification.



FIG. 2 is a diagram illustrating a result of expressing values associated with an operation performed in the optical 4f system 100 of FIG. 1 through a light amplitude and a light intensity. All f(x,y), G(u,v), and custom-character−1{custom-character{f(x,y)}(u,v)·G(u,v)}(x,y) that are used in the convolution operation using optical 4f system 100 described above are equations based on the light amplitude, and an actual input value and actual output values of the optical 4f system 100 are proportional to the light intensity. For example, the actual input value and the actual output values may be expressed in form of |f(x,y)|2, |G(u,v)|2, and |custom-character−1{custom-character{f(x,y)}(u,v)·G(u,v)}(x,y)|2. Accordingly, these values should be corrected to a form that could be used in the convolution operation. Because the light intensity at an arbitrary location is proportional to a square of the light amplitude, a value of the light amplitude may be calculated by performing a square root operation on both the input value and the output values in the convolution operation using the optical 4f system 100, and the convolution operation may then be performed.


That is, because the input value and the output values of two-dimensional data used in the convolution operation using the optical 4f system 100 are based on the light amplitude value, limitations f(x,y)≥0, G(u,v)≥0, and custom-character−1{custom-character{f(x,y)}(u,v)·G(u,v)}(x,y)≥0 that input and output values could be only a positive number exist. As such, it may be impossible to use an input value including a negative number and to train an image pattern including a negative number, and only a positive number may be generated as a result; in this case, some nonlinear functions may not operate, resulting in the reduction of CNN performance. Below, a convolution operation extending method for overcoming the limitations will be described.



FIG. 3 illustrates operation rules FT1 and FT2 of Fourier transform and operation rules IFT1 and IFT2 of inverse Fourier transform, which are used to extend a convolution operation. In FIG. 3, an operator “·” represents a multiplication operation, and an operator “*” represents a convolution operation. Operation rules of the Fourier transform include the homogeneity FT1 indicating that an output magnitude is proportional to an input magnitude, the additivity FT2 indicating that a total result when plural input signals are added is equal to a result of adding results of individual input signals, and the Fourier transform FT3 for convolution indicating that the convolution operation in the x-y plane is expressed by a multiplication operation in the u-v plane. Operation rules of the inverse Fourier transform include the inverse property IFT1 indicating that an original input value is obtained by performing the inverse Fourier transform on a result of the Fourier transform and the distributive property for addition IFT2.



FIG. 4 illustrates theorems used for extension of a convolution operation. With regard to f(x,y) in the x-y plane, a first theorem indicates that an input signal is expressed by a difference between signals f1(x,y)≥0 and f2(x,y)≥0 being two arbitrary different positive numbers; with regard to G(u,v) in the u-v plane, a second theorem indicates that an input signal is expressed by a difference between signals G1(u,v)≥0 and G2(u,v)≥0 being two arbitrary different positive numbers.


First, the way to remove the limitation associated with f(x,y) will be described. Assuming an arbitrary f(x,y) in which there is not limitation condition of a positive number or a negative number, according to the first theorem of FIG. 4, f1(x,y) and f2(x,y) that satisfy f(x,y)=f1(x,y)−f2(x,y) and in which there is a limitation condition of a positive number may certainly exist (f1(x,y)≥0 and f2(x,y)≥0). Also, because the additivity FT2 of FIG. 3 is applied to the Fourier transform, {custom-character(x,y)}(u,v)·G(u,v) may be expressed by Equation 1 below.












{

f

(

x
,
y

)

}




(

u
,
v

)

·

G

(

u
,
v

)



=


{



f
1

(

x
,
y

)

-


f
2

(

x
,
y

)


}




(

u
,
v

)

·

G

(

u
,
v

)







=


(



{


f
1

(

x
,
y

)

}



(

u
,
v

)


-


{


f
2

(

x
,
y

)

}



(

u
,
v

)



)

·

G

(

u
,
v

)






=



{


f
1

(

x
,
y

)

}




(

u
,
v

)

·

G

(

u
,
v

)



-


{


f
2

(

x
,
y

)

}




(

u
,
v

)

·

G

(

u
,
v

)









[

Equation


1

]







That is, custom-character{f(x,y)}(u,v)·G(u,v) may be expressed by a difference between custom-character{f1(x,y)}(u,v)·G(u,v) and custom-character{f2(x,y)}(u,v)·G(u,v). Now, when the inverse property IFT1 of FIG. 3 is applied, Equation 1 above may be expressed like Equation 2 below through the inverse Fourier transform.













-
1



{


{

f

(

x
,
y

)

}




(

u
,
v

)

·

G

(

u
,
v

)



}



(

x
,
y

)


=



-
1



{



{


f
1

(

x
,
y

)

}




(

u
,
v

)

·

G

(

u
,
v

)



-


{


f
2

(

x
,
y

)

}




(

u
,
v

)

·

G

(

u
,
v

)




}



(

x
,
y

)






=




-
1



{


{


f
1

(

x
,
y

)

}




(

u
,
v

)

·

G

(

u
,
v

)



}



(

x
,
y

)


-



-
1



{


{


f
2

(

x
,
y

)

}




(

u
,
v

)

·

G

(

u
,
v

)



}



(

x
,
y

)








[

Equation


2

]







Herein, as described above, in the case of f(x,y), because it is assumed that there is no special limitation condition, custom-character−1{custom-character{f(x,y)}(u,v)·G(u,v)}(x,y) may be calculated through the optical 4f system 100 of FIG. 1. However, because f1(x,y) and f2(x,y) have a positive number limitation condition, both custom-character−1{custom-character{f1(x,y)}(u,v)·G(u,v)}(x,y) and custom-character−1{custom-character{f2(x,y)}(u,v)·G(u,v)}(x,y) of Equation 2 above may be calculated through the optical 4f system 100. Accordingly, when an arbitrary f(x,y) and G(u,v) being a positive number are given, a value of custom-character−1{custom-character{f(x,y)}(u,v)·G(u,v)}(x,y) may be obtained through a difference between values corresponding to results of respectively calculating custom-character−1{custom-character{f1(x,y)}(u,v)·G(u,v)}(x,y) and custom-character−1{custom-character{f2(x,y)}(u,v)·G(u,v)}(x,y) through the optical 4f system 100.


Next, the way to remove the limitation associated with G(u,v) will be described. Assuming an arbitrary G(u,v) in which there is no limitation condition of a positive number or a negative number, according to the second theorem of FIG. 4, G1(u,v) and G2(u,v) that satisfy G(u,v)=G1(u,v)−G2(u,v) and have a positive number limitation condition may certainly exist (G1(u,v)≥0 and G2(u,v)≥0). Also, because the additivity FT2 of FIG. 3 is applied to the Fourier transform, custom-character{f(x,y)}(u,v)·G(u,v) may be expressed by Equation 3 below.











{

f

(

x
,
y

)

}




(

u
,
v

)

·

G

(

u
,
v

)






=


{

f

(

x
,
y

)

}




(

u
,
v

)

·

(



G
1

(

u
,
v

)

-


G
2

(

u
,
v

)


)








=



{

f

(

x
,
y

)

}




(

u
,
v

)

·


G
1

(

u
,
v

)



-


{

f

(

x
,
y

)

}




(

u
,
v

)

·


G
2

(

u
,
v

)









[

Equation


3

]







That is, custom-character{f(x,y)}(u,v)·G(u,v) may be expressed by a difference between custom-character{f(x,y)}(u,v)·G1(u,v) and {f(x,y)}(u,v)·G2(u,v). Now, when the inverse property IFT1 of FIG. 3 is applied, Equation 3 above may be expressed like Equation 4 below through the inverse Fourier transform.













-
1



{


{

f

(

x
,
y

)

}




(

u
,
v

)

·

G

(

u
,
v

)



}



(

x
,
y

)


=



-
1



{



{


f
1

(

x
,
y

)

}




(

u
,
v

)

·


G
1

(

u
,
v

)



-


{

f

(

x
,
y

)

}




(

u
,
v

)

·


G
2

(

u
,
v

)




}



(

x
,
y

)






=




-
1



{


{

f

(

x
,
y

)

}




(

u
,
v

)

·


G
1

(

u
,
v

)



}



(

x
,
y

)


-



-
1



{


{

f

(

x
,
y

)

}




(

u
,
v

)

·


G
2

(

u
,
v

)



}



(

x
,
y

)








[

Equation


4

]







According, when f(x,y) being a positive number and an arbitrary G(u,v) are given, a value of custom-character−1{custom-character{f(x,y)}(u,v)·G(u,v)}(x,y) may be obtained through a difference between values corresponding to results of respectively calculating custom-character−1{custom-character{f(x,y)}(u,v)·G1(u,v)}(x,y) and custom-character−1{custom-character{f(x,y)}(u,v)·G2(u,v)}(x,y) through the optical 4f system 100.


Now, the way to remove all the limitations associated with f(x,y) and G(u,v) will be described. According to the first theorem and the second theorem of FIG. 4, an arbitrary f(x,y) in which there is no limitation condition may satisfy f(x,y)=f1(x,y)−f2(x,y) (f1(x,y)≥0 and f2(x,y)≥0), and an arbitrary G(u,v) in which there is no limitation condition may satisfy G(u,v)=G1(u,v)−G2(u,v) (G1(u,v)≥0 and G2(u,v)≥0); in this case, when the additivity FT2 of FIG. 3 is applied, custom-character{f(x,y)}(u,v)·G(u,v) may be expressed by Equation 5 below.












{

f

(

x
,
y

)

}




(

u
,
v

)

·

G

(

u
,
v

)



=


{



f
1

(

x
,
y

)

-


f
2

(

x
,
y

)


}




(

u
,
v

)

·

G

(

u
,
v

)







=


(



{


f
1

(

x
,
y

)

}



(

u
,
v

)


-


{


f
2

(

x
,
y

)

}



(

u
,
v

)



)

·

G

(

u
,
v

)






=


(



{


f
1

(

x
,
y

)

}



(

u
,
v

)


-


{


f
2

(

x
,
y

)

}



(

u
,
v

)



)

·

(



G
1

(

u
,
v

)

-


G
2

(

u
,
v

)


)






=



{


f
1

(

x
,
y

)

}




(

u
,
v

)

·


G
1

(

u
,
v

)



-


{


f
2

(

x
,
y

)

}




(

u
,
v

)

·


G
1

(

u
,
v

)



-


{


f
1

(

x
,
y

)

}




(

u
,
v

)

·


G
2

(

u
,
v

)



+


{


f
2

(

x
,
y

)

}




(

u
,
v

)

·


G
2

(

u
,
v

)









[

Equation


5

]







Now, when the inverse property IFT1 of FIG. 3 is applied, a result of Equation 5 above may be expressed by Equation 6 below through the inverse Fourier transform.













-
1



{


{

f

(

x
,
y

)

}




(

u
,
v

)

·

G

(

u
,
v

)



}



(

x
,
y

)


=



-
1



{



{


f
1

(

x
,
y

)

}




(

u
,
v

)

·


G
1

(

u
,
v

)



-


{


f
2

(

x
,
y

)

}




(

u
,
v

)

·


G
1

(

u
,
v

)



-


{


f
1

(

x
,
y

)

}




(

u
,
v

)

·


G
2

(

u
,
v

)



+


{


f
2

(

x
,
y

)

}




(

u
,
v

)

·


G
2

(

u
,
v

)




}



(

x
,
y

)






=




-
1



{


{


f
1

(

x
,
y

)

}




(

u
,
v

)

·


G
1

(

u
,
v

)



}



(

x
,
y

)


-



-
1



{


{


f
2

(

x
,
y

)

}




(

u
,
v

)

·


G
1

(

u
,
v

)



}



(

x
,
y

)


-



-
1



{


{


f
1

(

x
,
y

)

}




(

u
,
v

)

·


G
2

(

u
,
v

)



}



(

x
,
y

)


+



-
1



{


{


f
2

(

x
,
y

)

}




(

u
,
v

)

·


G
2

(

u
,
v

)



}



(

x
,
y

)








[

Equation


6

]







Accordingly, custom-character{f(x,y)}(u,v)·G(u,v) including f(x,y) and G(u,v) in which there is no positive number limitation condition may be incapable of being calculated directly through the optical 4f system 100 but may be expressed through f1(x,y)≥0, f2(x,y)≥0, G1(u,v)≥0, and G2(u,v)≥0 like Equation 6 above. Therefore, custom-character{f(x,y)}(u,v)·G(u,v) for arbitrary f(x,y) and G(u,v) may be calculated by respectively calculating custom-character−1{custom-character{f1(x,y)}(u,v)·G1(u,v)}(x,y), custom-character−1{custom-character{f2(x,y)}(u,v)·G1(u,v)}(x,y), custom-character−1{custom-character{f1(x,y)}(u,v)·G2(u,v)}(x,y), and custom-character−1{custom-character{f1(x,y)}(u,v)·G2(u,v)}(x,y) by using the optical 4f system 100 four times and then adding differences thereof. As such, the limitations associated with f(x,y) and G(u,v) may be removed, and the convolution operation using the optical 4f system 100 may be extended.


Until now, a method in which the convolution operation of the optical 4f system 100 of FIG. 1 is extended by removing positive number limitations associated with f(x,y) and G(u,v) is described. However, referring to Equation 6 above, the convolution operation using the optical 4f system 100 should be performed four times to draw an operation result which is not limited to only a positive number. Below, the way to optimize an extended convolution operation by using the number of necessary convolution operations will be described.



FIG. 5 illustrates theorems used for extension of an extended convolution operation. With regard to the x-y plane f(x,y), a third theorem indicates that an input signal for G(u,v) is expressed by a difference between a signal (x,y)≥0 being an arbitrary positive number and a positive constant (a≥0 and b≥0); with regard to the u-v plane, a fourth theorem indicates that an input signal for G(u,v) is expressed by a difference between a signal G′(u,v)≥0 being an arbitrary positive number and a positive constant (a≥0 and b≥0).


First, the way to optimize a convolution operation for an arbitrary G(u,v) through the fourth theorem will be described. In detail, f(x,y) may be optimized similarly by applying the third theorem, and thus, additional description will be omitted to avoid redundancy. After the optimization of G(u,v) is described, the way to optimize both f(x,y) and G(u,v) will be described.


According to the fourth theorem, because G(u,v) is expressed like G′(u,v)−b (G′(u,v)≥0, b≥0), custom-character{f(x,y)}(u,v)·G(u,v) may be expressed by Equation 7 below by applying the additivity FT2 of FIG. 3.











{

f

(

x
,
y

)

}




(

u
,
v

)

·

G

(

u
,
v

)



=



{

f

(

x
,
y

)

}




(

u
,
v

)

·

(



G


(

u
,
v

)

-
b

)



=




{

f

(

x
,
y

)

}




(

u
,
v

)

·


G


(

u
,
v

)



-


{

f

(

x
,
y

)

}




(

u
,
v

)

·
b



=




{

f

(

x
,
y

)

}




(

u
,
v

)

·


G


(

u
,
v

)



-


b
·



{

f

(

x
,
y

)

}



(

u
,
v

)



=



{

f

(

x
,
y

)

}




(

u
,
v

)

·


G


(

u
,
v

)



-


{

b
·

f

(

x
,
y

)


}



(

u
,
v

)










[

Equation


7

]







That is, custom-character{f(x,y)}(u,v)·G(u,v) may be expressed by a difference between custom-character{f(x,y)}(u,v)·G′(u,v) and custom-character{b f(x,y)}(u,v). Now, when both the inverse property IFT1 and the distributive property for addition IFT2 are applied, custom-character{f(x,y)}(u,v)·G(u,v) may be expressed by Equation 8 below through the inverse Fourier transform.












-
1



{


{

f

(

x
,
y

)

}




(

u
,
v

)

·

G

(

u
,
v

)



}



(

x
,
y

)


=




-
1



{



{

f

(

x
,
y

)

}




(

u
,
v

)

·


G


(

u
,
v

)



-


{

b
·

f

(

x
,
y

)


}



(

u
,
v

)



}



(

x
,
y

)


=





-
1



{


{

f

(

x
,
y

)

}




(

u
,
v

)

·


G


(

u
,
v

)



}



(

x
,
y

)


-



-
1



{


{

b
·

f

(

x
,
y

)


}



(

u
,
v

)


}



(

x
,
y

)



=




-
1



{


{

f

(

x
,
y

)

}




(

u
,
v

)

·


G


(

u
,
v

)



}



(

x
,
y

)


-

b
·

f

(

x
,
y

)









[

Equation


8

]







Because G′(u,v) is a positive number, the operation using the optical 4f system 100 may be possible; because b·f(x,y) is capable of being calculated by directly multiplying an input f(x,y) and a constant “b” together without using the optical 4f system 100, unlike Equation 4 that requires the operation using the optical 4f system 100 two times, the operation (custom-character{{f(x,y)}(u,v)·G′(u,v)}(x,y)) using the optical 4f system 100 may be required only once, and thus, the convolution operation may be optimized.


Finally, how to optimize arbitrary f(x,y) and G(u,v) will be described. According to the fourth theorem, because f(x,y) is capable of being expressed like f′(x,y)−a(f′(x,y)≥0, a≥0) and G(u,v) is capable of being expressed like G′(u,v)−b(G′(u,v)≥0, b≥0), custom-character{f(x,y)}(u,v)·G(u,v) may be expressed by Equation 9 below by applying the homogeneity FT1 and the additivity FT2 of FIG. 3 together.











{

f

(

x
,
y

)

}




(

u
,
v

)

·

G

(

u
,
v

)



=



{

f

(

x
,
y

)

}




(

u
,
v

)

·

(



G


(

u
,
v

)

-
b

)



=




{

f

(

x
,
y

)

}




(

u
,
v

)

·


G


(

u
,
v

)



-


{

f

(

x
,
y

)

}




(

u
,
v

)

·
b



=




{

f

(

x
,
y

)

}




(

u
,
v

)

·


G


(

u
,
v

)



-


b
·



{

f

(

x
,
y

)

}



(

u
,
v

)



=




{

f

(

x
,
y

)

}




(

u
,
v

)

·


G


(

u
,
v

)



-


{

b
·

f

(

x
,
y

)


}



(

u
,
v

)



=




{



f


(

x
,
y

)

-
a

}




(

u
,
v

)

·


G


(

u
,
v

)



-


{

b
·

f

(

x
,
y

)


}



(

u
,
v

)



=




(



{


f


(

x
,
y

)

}



(

u
,
v

)


-


{
a
}



(

u
,
v

)



)

·


G


(

u
,
v

)


-


{

b
·

f

(

x
,
y

)


}



(

u
,
v

)



=



{


f


(

x
,
y

)

}




(

u
,
v

)

·


G


(

u
,
v

)



-


{
a
}




f

(

u
,
v

)

·


G


(

u
,
v

)



-


{

b
·

f

(

x
,
y

)


}



(

u
,
v

)













[

Equation


9

]







That is, custom-character{f(x,y)}(u,v)·G(u,v) may be calculated by using a difference between custom-character{f(x,y)}(u,v)·G′(u,v), custom-character{a}(u,v)·G′(u,v), and custom-character{b f(x,y)}(u,v). Now, when both the inverse property IFT1 and the distributive property for addition IFT2 of FIG. 3 are applied, custom-character{f(x,y)}(u,v)·G(u,v) may be expressed by Equation 10 below through the inverse Fourier transform.















-
1



{


{

f

(

x
,
y

)

}




(

u
,
v

)

·

G

(

u
,
v

)



}



(

x
,
y

)


=




-
1



{



{


f


(

x
,
y

)

}




(

u
,
v

)

·


G


(

u
,
v

)



-


{
a
}




(

u
,
v

)

·


G


(

u
,
v

)



-


{

b
·

f

(

x
,
y

)


}



(

u
,
v

)



}



(

x
,
y

)


=




-
1



{


{


f


(

x
,
y

)

}




(

u
,
v

)

·


G


(

u
,
v

)



}



(

x
,
y

)


-



-
1



{


{
a
}




(

u
,
v

)

·


G


(

u
,
v

)



}



(

x
,
y

)





}

-



-
1



{


{

b
·

f

(

x
,
y

)


}



(

u
,
v

)


}



(

x
,
y

)



=



-
1



{


{




f


(

x
,
y

)



{


(

u
,
v

)

·


G


(

u
,
v

)


}



(

x
,
y

)


-



-
1



{


{
a
}




(

u
,
v

)

·


G


(

u
,
v

)



}



(

x
,
y

)


-

b
·

f

(

x
,
y

)











[

Equation


10

]







Because arbitrary f(x,y) and G(u,v) are not limited to a positive number, custom-character−1{custom-character{f(x,y)}(u,v)·G(u,v)}(x,y) may be incapable of being calculated directly through the optical 4f system 100; however, because f′(x,y)≥0 and G′(u,v)≥0, custom-character−1{custom-character{f(x,y)}(u,v)·G′(u,v)}(x,y) and custom-character−1{custom-character{f(a)}(u,v)·G′(u,v)}(x,y) may be calculated through the optical 4f system 100, and b·f(x,y) may be calculated by directly multiplying an input f(x,y) and a constant “b” without using the optical 4f system 100. Therefore, unlike Equation 6 that requires the operation using the optical 4f system 100 four times, because only two operations (custom-character−1{custom-character{f(x,y)}(u,v)·G′(u,v)}(x,y) and custom-character−1{custom-character{f(a)}(u,v)·G′(u,v)}(x,y)) using the optical 4f system 100 are required, the convolution operation may be extended to a negative number input as well as a positive number input, and the number of operations may be optimized.



FIG. 6 is a flowchart illustrating an example of an operating method of the optical 4f system 100 according to an embodiment of the present disclosure. Below, the description will be given with reference to FIG. 1 together with FIG. 6.


In operation S110, the first lens 110 may receive input data from the object 10. In operation S120, the kernel 120 may perform the first Fourier transform on the input data, and in operation S130, the kernel 120 may generate learning data by performing calculation (e.g., a multiplication operation) on a result of the first Fourier transform and pattern data. In operation S140, the second lens 130 may generate result data by performing the second Fourier transform (e.g., inverse Fourier transform) on the learning data. Herein, the kernel 120 may correspond to a filter of the CNN, and an image indicated by the input data and an image indicated by the result data may be inverted images.


In conclusion, through embodiments of the present disclosure, the convolution operation using the optical 4f system may be performed on f(x,y) and G(u,v) including both a positive number and a negative number, and a result of the convolution operation may also include both a positive number and a negative number. As such, the following are obtained.


As described with reference to FIG. 1, f(x,y) is the first input image of the CNN or an input value of a previous CNN layer. Because there is a limitation condition that f(x,y) must always be a positive number, it is difficult to use the conventional optical 4f system in the artificial intelligence and machine learning fields such as CNN; however, according to an embodiment of the present disclosure, because the limitation associated with f(x,y) is removed, the optical 4f system may be applied to the CNN including a plurality of convolution layers.


G(u,v) is an image pattern that is trained in the CNN. When there are provided a variety of patterns that is capable of being trained in the machine learning field, a learning result of the CNN may be excellent. Because there was a constraint condition that G(u,v) must always be a positive number, when a conventional optical 4f system is applied to the CNN, a CNN learning result may be degraded; however, according to an embodiment of the present disclosure, because the limitation associated with G(u,v) is removed, learning and inference results of the CNN may be improved when a variety of patterns are trained.


Also, because a value generated in the conventional optical 4f system is always a positive number, some nonlinear functions may not operate properly, resulting in the reduction of CNN performance. For example, because a nonlinear function such as a rectified linear unit (ReLU) acts as an identity function when an input value is positive, it is impossible to express a characteristic of the nonlinear function. In this case, because a multi-layer neural network operates only as a linear function, the multi-layer neural network performs only a role of a single-layer neural network. According to an embodiment of the present disclosure, because the limitations associated with a convolution result are removed, an image generated by an optical 4f system according to an embodiment of the present disclosure may be input to the ReLU function to perform CNN learning and inference.


The optical 4f system according to embodiments of the present disclosure is not limited to being used for the convolution operation and may be used for a variety of alternative purposes. In addition, the optical 4f system according to an embodiment of the present disclosure may be used in combination with many existing computing systems and devices including a processor. For example, the processor may include a special purpose processing device such as an application-specific integrated circuit (ASIC), a programmable array logic (PAL), a programmable logic array (PLA), a programmable logic device (PLO), or a field programmable gate array (FPGA), or any other custom and/or programmable device. A computing device may also include storage devices such as a static RAM, a dynamic RAM, a ROM, a CD-ROM, a magnetic tape, an optical disk, a flash memory, a non-transitory computer readable medium, or any other machine readable medium.


According to an embodiment of the present invention, as an optical 4f system may be applied to a CNN including a plurality of convolution layers, more diverse image patterns may be trained, and thus, learning and reasoning results of the CNN may be improved.


Also, according to an embodiment of the present invention, a convolution result value may also be output as various values including negative numbers, and thus, CNN learning and inference may be performed in conjunction with a ReLU-based nonlinear function.


While the present disclosure has been described with reference to embodiments thereof, it will be apparent to those of ordinary skill in the art that various changes and modifications may be made thereto without departing from the spirit and scope of the present disclosure as set forth in the following claims.

Claims
  • 1. An optical system comprising: a first lens configured to receive input data from an object;a kernel configured to perform a first Fourier transform on the input data and to generate learning data by performing calculation on a result of the first Fourier transform and pattern data; anda second lens configured to generate result data by performing a second Fourier transform on the learning data,wherein the input data, the learning data, and the result data include both a positive number and a negative number.
  • 2. The optical system of claim 1, wherein phases of an image represented by the input data and an image represented by the result data are opposite.
  • 3. The optical system of claim 1, wherein the kernel corresponds to a filter of a convolutional neural network (CNN).
  • 4. The optical system of claim 1, wherein the learning data are calculated by Equation 1:
  • 5. The optical system of claim 4, wherein the result data are calculated by Equation 2: −1{{f(x,y)}(u,v)·G(u,v)}(x,y)=−1{{f1(x,y)}(u,v)·G1(u,v)}(x,y)−−1{{f2(x,y)}(u,v)·G1(u,v)}(x,y)−−1{{f1(x,y)}(u,v)·G2(u,v)}(x,y)+−1{{f1(x,y)}(u,v)·G2(u,v)}(x,y).
  • 6. The optical system of claim 1, wherein the learning data are calculated by Equation 3:
  • 7. The optical system of claim 6, wherein the result data are calculated by Equation 4: −1{{f(x,y)}(u,v)·G(u,v)}(x,y)=−1{{f(x,y)}(u,v)·G′(u,v)}(x,y)−−1{{f(a)}(u,v)·G′(u,v)}(x,y)−b·f(x,y).
  • 8. The optical system of claim 1, wherein the result data are input to a ReLU (Rectified Linear Unit) function.
  • 9. An operating method of an optical system which includes a first lens, a kernel, and a second lens, the operating method comprising: receiving input data from an object through the first lens;performing a first Fourier transform on the input data through the kernel;generating learning data by performing calculation on a result of the first Fourier transform and pattern data; andgenerating result data by performing a second Fourier transform on the learning data,wherein the input data, the learning data, and the result data include both a positive number and a negative number.
  • 10. The operating method of claim 9, wherein phases of an image represented by the input data and an image represented by the result data are opposite.
  • 11. The operating method of claim 9, wherein the kernel corresponds to a filter of a convolutional neural network (CNN).
  • 12. The operating method of claim 9, wherein the generating of the learning data includes generating the learning data based on Equation 1:
  • 13. The operating method of claim 12, wherein the generating of the result data includes generating the result data based on Equation 2: −1{{f(x,y)}(u,v)·G(u,v)}(x,y)=−1{{f1(x,y)}(u,v)·G1(u,v)}(x,y)−−1{{f2(x,y)}(u,v)·G1(u,v)}(x,y)−−1{{f1(x,y)}(u,v)·G2(u,v)}(x,y)+−1{{f1(x,y)}(u,v)·G2(u,v)}(x,y).
  • 14. The operating method of claim 9, wherein the generating of the learning data includes generating the learning data based on Equation 3:
  • 15. The operating method of claim 14, wherein the generating of the result data includes generating the result data based on Equation 4: −1{{f(x,y)}(u,v)·G(u,v)}(x,y)=−1{{f(x,y)}(u,v)·G′(u,v)}(x,y)−−1{{f(a)}(u,v)·G′(u,v)}(x,y)−b·f(x,y).
  • 16. The operating method of claim 9, further comprising: inputting the result data to a ReLU (Rectified Linear Unit) function.
Priority Claims (2)
Number Date Country Kind
10-2021-0185587 Dec 2021 KR national
10-2022-0173746 Dec 2022 KR national