The present invention relates to a method for processing data associated with a neural network.
The present invention furthermore relates to an apparatus for processing data associated with a neural network.
Exemplary embodiments of the present invention relate to a method, for example a computer-implemented method, for processing data associated with a, for example artificial, for example deep, neural network, for example convolutional neural network, CNN, comprising: representing at least one filter of the neural network based on at least one filter dictionary, and, optionally, processing input data, and/or data that can be derived or are derived from input data, by using the at least one filter. In further exemplary embodiments, the use of the at least one filter dictionary or of the filter that can be represented thereby may increase a quality of training or of processing of data by the neural network (inference) and may, for example, decrease a need for computing time resources and/or memory resources, for example for the training and/or the inference.
In further exemplary embodiments of the present invention, it is provided that the at least one filter dictionary at least partially characterizes, for example spans, a linear space, wherein the at least one filter dictionary may, for example, be characterized by :={g(1), . . . , g(N)}⊂K1×K2, wherein g(i) characterizes an i-th element of the at least one filter dictionary, for example an i-th filter, for example filter kernel, where i=1, . . . , N, wherein K1 characterizes a size of the filters of the at least one filter dictionary (FD) in a first dimension, wherein K2 characterizes a size of the filters of the at least one filter dictionary in a second dimension, wherein, for example, K1=K2=K applies, wherein span{} characterizes the linear space that the at least one filter dictionary at least partially characterizes.
In further exemplary embodiments of the present invention, at least one filter or filter kernel may also have more than two dimensions, for example three or more, or one dimension, wherein the principle according to the embodiments is also applicable to such configurations, without limiting generality.
In further exemplary embodiments of the present invention, at least one filter or filter kernel may be square, for example, where K1=K2, wherein K1< >K2 is also possible in further exemplary embodiments.
In further exemplary embodiments of the present invention, more than one filter dictionary may also be provided. For example, in the case of a plurality of filter dictionaries, at least a first filter dictionary with filters of a first size (e.g., K1×K2) may be provided, and at least a second filter dictionary with filters of a second size (e.g., K1′×K2′, wherein K1′=K2′ is also possible in further exemplary embodiments) may be provided.
In further exemplary embodiments of the present invention, it is provided that a) the at least one filter dictionary does not completely span a space, for example K1×K2, for example is undercomplete, or that b) at least some elements of the at least one filter dictionary are linearly dependent on one another, wherein the at least one filter dictionary is, for example, overcomplete.
In further exemplary embodiments of the present invention, it is provided that the at least one filter dictionary is different from a standard basis , for example according to :={e(n): n=1, . . . , K2}, wherein e(n) characterizes an n-th unit vector associated with the standard basis . In further exemplary embodiments, further degrees of freedom for representing at least one filter, for example in the form of a linear combination of a plurality of elements of the filter dictionary, are thus given, for example.
In further exemplary embodiments of the present invention, it is provided that the representing of the at least one filter of the neural network based on the at least one filter dictionary can be characterized by the following equation and/or is performed based on the following equation: h=Σn=1Nλn·g(n), wherein h characterizes the at least one filter, wherein g(n) characterizes an n-th element, for example an n-th filter, of the at least one filter dictionary, wherein λn characterizes a coefficient associated with the n-th element, for example n-th filter, of the at least one filter dictionary, and wherein n is an index variable that characterizes one of the N elements, for example one of the N filters, of the at least one filter dictionary.
In further exemplary embodiments of the present invention, representing a plurality of filters h(α,β), associated with, for example, a layer of the neural network, based on the at least one filter dictionary can be characterized by the following equation and/or is performed based on the following equation: h(α,β)=Σn=1Nλn(α,β)·g(n), wherein α characterizes an index variable associated with a number of output channels of the layer, wherein β characterizes an index variable associated with a number of input channels of the layer, wherein λn(α,β) characterizes a coefficient, associated with the n-th element, for example n-th filter, of the at least one filter dictionary, for the output channel α and the input channel β of the layer.
In further exemplary embodiments of the present invention, it is provided that the processing of the input data, and/or of the data that can be derived or are derived from the input data (e.g., data that are output by an inner layer (“hidden layer”) of the neural network), by using the at least one filter can be characterized by the following equation and/or is performed based on the following equation:
wherein X characterizes the input data, or the data that can be derived or are derived from the input data, for example an input feature map for one or the layer of the neural network, wherein α characterizes an index variable associated with a number of output channels of the layer, wherein β characterizes an index variable associated with a number of input channels of the layer, wherein λn(α,β) characterizes a coefficient, associated with the n-th element, for example n-th filter, of the at least one filter dictionary, for the output channel α and the input channel β of the layer, wherein cin characterizes a number of the input channels of the layer, wherein * characterizes a convolution operation.
In further exemplary embodiments of the present invention, it is provided that the method comprises: initializing the at least one filter dictionary, for example prior to representing the at least one filter and/or processing input data, for example, wherein initializing, for example, comprises at least one of the following elements: a) random-based initializing, for example by assigning random numbers or pseudorandom numbers to at least some filter coefficients gi,j(n) of at least some elements or filters of the at least one filter dictionary (for example, an n-th filter or filter kernel of the at least one filter dictionary has, e.g., 3×3 filter coefficients: g1,1(n), g1,2(n), g1,3(n), g2,1(n), . . . , g3,3(n)), b) random-based initializing such that a or the linear space span{} that can be characterized by the at least one filter dictionary characterizes a orthonormal basis, for example comprising b1) initializing at least some, for example all, filter coefficients gi,j(n)—of at least some, for example all, elements or filters of the at least one filter dictionary with filter coefficient values, for example independently equally distributed filter coefficient values, b2) applying the Gram-Schmidt orthogonalization method to the elements or filters of the at least one filter dictionary, c) random-based initializing by means of c1) initializing at least some, for example all, filter coefficients gi,j(n) of at least some, for example all, elements or filters of the at least one filter dictionary with filter coefficient values, for example independently equally distributed filter coefficient values, c2) scaling, or rescaling, the at least one filter dictionary based on at least one statistical quantity, for example a mean and/or a standard deviation.
In further exemplary embodiments of the present invention, it is provided that the method comprises: initializing coefficients of, for example, some, for example all, elements or filters of the at least one filter dictionary, comprising at least one of the following aspects: a) random-based or pseudorandom-based initializing of the coefficients, b) initializing the coefficients based on the at least one filter dictionary.
In further exemplary embodiments of the present invention, it is provided that the method comprises: reducing, for example thinning out, for example pruning, at least one component of the at least one filter dictionary, wherein reducing comprises at least one of the following elements: a) reducing at least one element, for example filter, of the at least one filter dictionary, for example by zeroing at least one filter coefficient, for example a plurality of filter coefficients, of the at least one element, for example filter, of the at least one filter dictionary, b) removing or deleting at least one element, for example filter, of the at least one filter dictionary, c) removing or deleting at least one coefficient associated with the at least one filter dictionary.
In further exemplary embodiments of the present invention, it is provided that the method comprises at least one of the following elements: a) performing the reducing after an or the initializing of the at least one filter dictionary, b) performing the reducing after an or the initializing of coefficients or of the coefficients of, for example, some, for example all, elements or filters of the at least one filter dictionary, c) performing the reducing during a training of the neural network, d) performing the reducing after a or the training of the neural network.
In further exemplary embodiments of the present invention, the reducing may occur, e.g., in an event-driven manner, for example based on an occurrence of particular data values, e.g. of the output data that can be determined by means of the neural network, and/or in a time-controlled manner, for example repeatedly, for example periodically. Combinations thereof are also possible in further exemplary embodiments.
In further exemplary embodiments of the present invention, it is provided that the method comprises at least one of the following elements: a) using the at least one, for example the same, filter dictionary for a plurality of layers, for example all layers, of the neural network, b) using the at least one, for example the same, filter dictionary for a plurality, for example all, layers of the neural network that are associated with the same spatial size of data to be processed, for example feature maps, c) using the at least one, for example the same, filter dictionary for a respective residual block, for example in the case of a residual neural network, for example ResNet, d) using the at least one, for example the same, filter dictionary for a layer of the neural network.
In further exemplary embodiments of the present invention, the neural network may also comprise, in addition to one or more layers which respectively perform filtering by using the at least one filter dictionary or by using filters that can be represented by means of the at least one filter dictionary (i.e., layers which, for example, perform two-dimensional convolution operations of corresponding input data for the respective layer, e.g., input feature map, with the respective filter mask), one or more further components, such as other functional layers, for example pooling layers, such as max-pooling layers, fully connected layers, for example in terms of a multi-layer perceptron (MLP), at least one, for example non-linear, activation function, etc.
In further exemplary embodiments of the present invention, it is provided that the method comprises: training the neural network, for example based on training data, wherein a trained neural network is, for example, obtained, and, optionally, using the, for example trained, neural network, for example for processing the input data.
Further exemplary embodiments of the present invention relate to a method, for example a computer-implemented method, for training a, for example artificial, for example deep, neural network, for example convolutional neural network, CNN, wherein at least one filter of the neural network can be represented and/or is represented based on at least one filter dictionary, wherein the method comprises: training at least one component of the at least one filter dictionary, wherein the training of the at least one component of the at least one filter dictionary is, for example, performed at least temporarily simultaneously and/or together with a training of at least one other component of the neural network.
In further exemplary embodiments of the present invention, it is provided that the training comprises a training of one, for example only one or at least one, element of the at least one filter dictionary.
In further exemplary embodiments of the present invention, it is provided that the method comprises: providing a filter dictionary characterizing a standard basis, wherein the standard basis can, for example, be characterized according to :={e(n): n=1, . . . , K2}, wherein e(n) characterizes an n-th unit vector associated with the standard basis , changing the filter dictionary, characterizing the standard basis, based on the training. Thus, in further exemplary embodiments, flexibility with regard to the representation of filters for the neural network is increased in comparison to using the standard basis.
In further exemplary embodiments of the present invention, it is provided that the method comprises: providing a filter dictionary not characterizing a standard basis, changing the filter dictionary, not characterizing a standard basis, based on the training.
In further exemplary embodiments of the present invention, it is provided that the method comprises: providing a pre-trained neural network or performing a first training, for example pre-training, for the neural network, optionally performing a reducing, for example the reducing according to exemplary embodiments, on the pre-trained neural network and, optionally, performing a further training.
In further exemplary embodiments of the present invention, it is provided that the training comprises: training the at least one filter dictionary together with at least one coefficient associated with the at least one filter dictionary.
In further exemplary embodiments of the present invention, it is provided that the processing of the input data comprises at least one of the following elements: a) processing multi-dimensional data, b) processing image data, c) processing audio data, for example voice data and/or operating noises from technical equipment systems, such as machines, d) processing video data or parts of video data, e) processing sensor data, wherein the processing of the input data comprises, for example, an analysis, for example a classification, of the input data.
In further exemplary embodiments of the present invention, it is provided that the method comprises: using output data obtained based on the processing of the input data to influence, for example control and/or regulate, at least one component of a technical system, for example cyber-physical system.
In further exemplary embodiments of the present invention, it is provided that the method comprises at least one of the following elements: a) initializing the at least one filter dictionary, b) initializing coefficients associated with the at least one filter dictionary, c) reducing, for example thinning out, for example pruning, at least one component of the at least one filter dictionary, d) training the neural network, for example the at least one filter dictionary, for example together with at least one further component of the neural network, for example based on a gradient-based optimization method, for example a stochastic gradient-based optimization method.
Further exemplary embodiments of the present invention relate to an apparatus for performing the method according to the embodiments.
Further exemplary embodiments of the present invention relate to a computer-readable storage medium comprising instructions that, when executed by a computer, cause the computer to perform the method according to the embodiments.
Further exemplary embodiments of the present invention relate to a computer program comprising instructions that, when the program is executed by a computer, cause the computer to perform the method according to the embodiments.
Further exemplary embodiments of the present invention relate to a data carrier signal that transmits and/or characterizes the computer program according to the embodiments.
Further exemplary embodiments of the present invention relate to a use of the method according to the embodiments and/or of the apparatus according to the embodiments and/or of the computer-readable storage medium according to the embodiments and/or of the computer program according to the embodiments and/or of the data carrier signal according to the embodiments for at least one of the following elements: a) representing at least one filter of the neural network based on the at least one filter dictionary, b) processing input data, and/or data that can be derived or are derived from input data, by using the at least one filter, c) increasing flexibility with regard to the representation of the at least one filter, d) adapting dynamically, i.e., adapting can be performed, for example, during a performance of the method, the at least one filter, for example during a training in which at least one further component of the neural network is also trained, e) decreasing a complexity of the neural network, f) improving a generalization by the neural network, for example in the sense that a behavior of the neural network during a training becomes more similar to a behavior of the neural network outside of the training, for example when evaluating input data other than training data, g) reducing or decreasing an overfitting, for example “memorizing” the training data, h) saving storage resources and/or computing time resources required for a representation and/or an evaluation of the neural network, i) decreasing a training duration, j) enabling use of existing reduction methods or pruning methods for neural networks, for example structured and/or unstructured pruning methods, for example also for reducing at least one component of the at least one filter dictionary, k) increasing flexibility with regard to initializing the at least one filter dictionary, l) enabling flexible use of the at least one filter dictionary, for example selectively, for at least one component, for example a layer, of the neural network, for example a flexible sharing of the at least one filter dictionary between different components of the neural network, m) increasing a quality of a training and/or an evaluation, for example inference, of the neural network.
Further features, possible applications and advantages of the present invention emerge from the description below of exemplary embodiments of the present invention, which are illustrated in the figures. All described or depicted features by themselves or in any combination constitute the subject matter of the present invention, regardless of their formulation or representation in the description or in the figures.
Exemplary embodiments of the present invention, cf.
In further exemplary embodiments, the use of the at least one filter dictionary FD or of the filter FILT-1 that can be represented thereby may increase a quality of training or of processing of data by the neural network (inference) and may, for example, reduce a need for computing time resources and/or memory resources, for example for the training and/or the inference.
In further exemplary embodiments, it is provided that the at least one filter dictionary FD at least partially characterizes a linear space, wherein, for example, the at least one filter dictionary FD can be characterized by :={g(1), . . . , g(N)}⊂K1×K2 wherein g(i) characterizes an i-th element of the at least one filter dictionary FD, for example an i-th filter, for example filter kernel, where i=1, . . . , N, wherein K1 characterizes a size of the filters of the at least one filter dictionary FD in a first dimension, wherein K2 characterizes a size of the filters of the at least one filter dictionary FD in a second dimension, wherein, for example, K1=K2=K applies, wherein characterizes the linear space that the at least one filter dictionary FD at least partially characterizes.
In further exemplary embodiments, at least one filter or filter kernel may also have more than two dimensions, for example three or more, wherein the principle according to the embodiments is also applicable to such configurations, without limiting generality.
In further exemplary embodiments, at least one filter or filter kernel may be square, for example, where K1=K2, wherein K1< >K2 is also possible in further exemplary embodiments.
In further exemplary embodiments, more than one filter dictionary FD may also be provided. For example, in the case of a plurality of filter dictionaries, at least a first filter dictionary with filters of a first size (e.g., K1×K2) may be provided, and at least a second filter dictionary with filters of a second size (e.g., K1′×K2′, wherein K1′=K2′ is also possible in further exemplary embodiments) may be provided.
In further exemplary embodiments, it is provided that a) the at least one filter dictionary FD does not completely span a space, for example K1×K2, for example is undercomplete, or that b) at least some elements of the at least one filter dictionary FD are linearly dependent on one another, wherein the at least one filter dictionary FD is, for example, overcomplete.
In further exemplary embodiments, it is provided that the at least one filter dictionary FD, which can, for example, be characterized according to :={g(1), . . . , g(N)}⊂K1×K2 or :={g(1), . . . , g(N)}⊂K×K, is different from a standard basis , for example according to :={e(n): n=1, . . . , K2}, wherein e(n) characterizes an n-th unit vector associated with the standard basis . In further exemplary embodiments, further degrees of freedom for representing 100 at least one filter, for example in the form of a linear combination of a plurality of elements of the filter dictionary FD, are thus given, for example.
In further exemplary embodiments, it is provided that the representing 100 (
In further exemplary embodiments, representing 100 a plurality of filters h(α,β), associated with, for example, a layer L1 of the neural network NN, based on the at least one filter dictionary FD can be characterized by the following equation and/or is performed based on the following equation: h(α,β)=Σn=1Nλn(α,β)·g(n), wherein a characterizes an index variable associated with a number of output channels of the layer L1, wherein β characterizes an index variable associated with a number of input channels of the layer L1, wherein λn(α,β) characterizes a coefficient, associated with the n-th element, for example n-th filter, of the at least one filter dictionary FD, for the output channel α and the input channel β of the layer L1.
In further exemplary embodiments, it is provided that the processing of the input data ED, and/or of the data ED′, ED′ that can be derived or are derived from the input data ED (e.g., data that are output by an inner layer (“hidden layer”) L2 of the neural network NN), by using the at least one filter FILT-1 can be characterized by the following equation and/or is performed based on the following equation:
wherein X characterizes the input data, or the data that can be derived or are derived from the input data, for example an input feature map for one or the layer L1, L2 of the neural network NN, wherein a characterizes an index variable associated with a number of output channels of the layer L1, wherein β characterizes an index variable associated with a number of input channels of the layer L1, wherein λn(α,β) characterizes a coefficient, associated with the n-th element, for example n-th filter, of the at least one filter dictionary FD, for the output channel α and the input channel β of the layer L1, wherein cin characterizes a number of the input channels of the layer L1, wherein * characterizes a convolution operation.
In further exemplary embodiments,
In further exemplary embodiments,
Initializing 110, 110a, 110b, 110c results in at least one initialized filter dictionary FD′ which can be used for representing 100 according to
In further exemplary embodiments, the random-based initializing 110b such that a or the linear space span{} that can be characterized by the at least one filter dictionary is spanned by an orthonormal basis, may, for example, comprise at least one of the aspects mentioned by way of example below:
1) Initializing at least some, for example all, filter coefficients g(1), . . . , g(K
2) Applying the Gram-Schmidt orthogonalization method to the basis {g(1), . . . , q(K
3) Optionally, for an initialization of the coefficients λ, μh←0 (average of the spatial (filter) coefficients,
4)
Variance of the spatial coefficients, for example according to a Kaiming normal initialization, wherein cin characterizes a number of input channels. In further exemplary embodiments, other values for the mean or the variance may also be selected.
5) Initializing the spatial coordinates φn(α,β)˜(μh,σh2) in a manner independently equally distributed for all α∈{1, . . . , cout}, β∈{1, . . . , cin}, n∈{1, . . . , K2},
6) Calculating a basis transformation matrix W, for example according to
7) Determining the coefficients λ(α,β)←ψT·φ(α,β) with respect to the at least one filter dictionary,
8) Providing the initialized filter dictionary ={g(1), . . . , g(N)} and associated coefficients λ=(λn(α,β))α,β,n.
In further exemplary embodiments, the random-based initializing 110c by means of c1) initializing 110c-1 at least some, for example all, filter coefficients gi,j(n) of at least some, for example all, elements or filters of the at least one filter dictionary with filter coefficient values, for example independently equally distributed filter coefficient values, c2) scaling 110c-2, or rescaling, the at least one filter dictionary based on at least one statistical quantity, for example a mean and/or a standard deviation, may, for example, comprise at least one of the aspects mentioned by way of example below:
10) Initializing at least some, for example all, filter coefficients g(1), . . . , g(N)∈K×K with independently equally distributed gi,j(n)˜(0,1),
11) For example, for each spatial component i, j of the elements of the at least one element, for example filter, of the at least one filter dictionary, a sample mean μi,j or a sample variance σi,j, is, for example, determined over the entire filter dictionary, e.g., according to
12) Scaling, or rescaling, the filter dictionary, for example according to
13) Optionally, for an initialization of the coefficients λ, μh←0 (average of the spatial (filter) coefficients,
14)
Variance of the spatial coefficients, for example according to a Kaiming normal initialization, wherein cin characterizes a number of input channels. In further exemplary embodiments, other values for the mean or the variance may also be selected.
15) Initializing the coordinates according to λn(α,β)˜(μh,σh2), in a manner independently equally distributed for all α∈{1, . . . , cout}, β∈{1, . . . , cin}, n∈{1, . . . , N}
16) Providing the initialized filter dictionary ={{tilde over (g)}(1), . . . , {tilde over (g)}(N)} and associated coefficients λ=(λn(α,β))α,β,n.
In further exemplary embodiments,
In further exemplary embodiments,
In further exemplary embodiments,
In further exemplary embodiments, the reducing 130 may occur, e.g., in an event-driven manner, for example based on an occurrence of particular data values, e.g. of the output data AD that can be determined by means of the neural network, and/or in a time-controlled manner, for example repeatedly, for example periodically. Combinations thereof are also possible in further exemplary embodiments.
In further exemplary embodiments,
In further exemplary embodiments,
Further exemplary embodiments,
In further exemplary embodiments, the training may also comprise, for example only, a training of the at least one filter dictionary, for example without training coefficients associated with the at least one filter dictionary in the process.
The optional block 163 symbolizes a use of the trained neural network.
In further exemplary embodiments,
In further exemplary embodiments,
In further exemplary embodiments,
In further exemplary embodiments, it is provided that the training 150, 160 comprises: training the at least one filter dictionary FD together with at least one coefficient associated with the at least one filter dictionary FD.
In further exemplary embodiments, the training 150, 160 may also comprise, for example only, a training of the at least one filter dictionary, for example without training coefficients associated with the at least one filter dictionary in the process.
In further exemplary embodiments, the training 150, 160 may also comprise, for example only, a training of at least one coefficient associated with the at least one filter dictionary.
In further exemplary embodiments,
In further exemplary embodiments,
In further exemplary embodiments,
In further exemplary embodiments, the following sequence may be provided in order to provide a trained neural network NN′ comprising (e.g., trainable) filters that can be represented by means of the at least one filter dictionary FD:
1) optionally: initializing k filter dictionaries 0(1), . . . , 0(1) (for example according to
1a) optionally: sharing at least some of the filter dictionaries 0(1), . . . , 0(1) initialized according to step 1), i.e., using, for example, the at least some of the filter dictionaries 0(1), . . . , 0(1) initialized according to step 1), for example for other layers of the neural network NN,
2a) assigning a respective filter dictionary 0(J
2b) initializing the coefficients λ0(l) for the L layers, for example according to
3a) optionally: determining a, for example global, pruning mask p for the reducing, for example according to
3b) optionally: reducing, for example pruning, the coefficients λ0(l) for the filter dictionaries, for example by means of the pruning mask p, for example according to λ0⊙μ, wherein λ0=(λ0(1), . . . , λ0(L)) characterizes the (e.g., global) filter coefficients, and wherein ⊙ characterizes the Hadamard product or element-wise product. This operation may also be referred to as “interspace pruning” in further exemplary embodiments because the optional pruning can at least partially be applied to the interspace that can be characterized by the filter dictionaries or to the coefficients associated with the filter dictionaries.
4) For example, for T many training steps, t∈{1, . . . , T},
4a) performing a forward pass, for example based on the filter dictionaries t-1(1), . . . , t-1(k) and based on the coefficients λt-1⊙μ (e.g., pruned or reduced by means of pruning mask μ), for example according to
4b) performing a backward pass, for example based on the filter dictionaries t-1(1), . . . , t-1(k) and based on the coefficients λt-1⊙μ (e.g., pruned or reduced by means of pruning mask μ), for example according to
if a sharing of filter dictionaries occurs in the forward pass 4a), this may also be performed in the backward pass 4b) in further exemplary embodiments,
4c) applying a, for example stochastic, gradient-based optimization to the filter dictionaries t-1(1), . . . , t-1(k) and the coefficients λt-1⊙μ based on the backward pass according to previous step 4b),
wherein, for example, after the T training steps 4a), 4b), 4c), trained filter dictionaries T(1), . . . , T(k), for example with sparsely populated coefficients λT⊙μ, are obtained, by means of which, for example, a trained neural network NN′ can be provided.
In further exemplary embodiments, the optional pruning 3a), 3b) may, for example, also be omitted or be performed during the training 4) or after the training 4).
In further exemplary embodiments, an infinite number of training steps t are also possible, which, for example, corresponds to continuous training.
In further exemplary embodiments, different pruning masks μ may also be used for at least two different training steps t1, t2.
In further exemplary embodiments, in addition to the aspects described above with reference to steps 4a), 4b) 4c), further parameters or hyperparameters of the neural network NN may also be trained, for example weights of fully connected layers NN-K1, etc.
Further exemplary embodiments,
In further exemplary embodiments, it is provided that the apparatus 200 comprises: a computing device (“computer”) 202 comprising, for example, one or more, in the present case, for example, two, computing cores 202a, 202b; a memory device 204 assigned to the computing device 202 for at least temporarily storing at least one of the following elements: a) data DAT (e.g., input data ED and/or training data TD and/or data for an operation of the neural network NN (e.g., weights and/or filter coefficients, data of the at least one filter dictionary FD), b) computer program PRG, in particular for performing a method according to the embodiments.
In further exemplary embodiments, the memory device 204 comprises a volatile memory 204a (e.g., random access memory (RAM)) and/or a non-volatile memory 204b (e.g., flash EEPROM).
In further exemplary embodiments, the computing device 202 comprises at least one of the following elements or is designed as at least one of these elements: microprocessor (μP), microcontroller (μC), application-specific integrated circuit (ASIC), system on chip (SoC), programmable logic module (e.g., FPGA, field programmable gate array), hardware circuitry, graphics processor, tensor processor, or any combinations thereof.
Further exemplary embodiments relate to a computer-readable storage medium SM comprising instructions PRG that, when executed by a computer 202, cause the latter to perform the method according to the embodiments.
Further exemplary embodiments relate to a computer program PRG comprising instructions that, when the program is executed by a computer 202, cause the latter to perform the method according to the embodiments.
Further exemplary embodiments relate to a data carrier signal DCS that characterizes and/or transmits the computer program PRG according to the embodiments. For example, the data carrier signal DCS can be received via an optional data interface 206 of the apparatus 200, via which, for example, at least some of the following data can also be exchanged (sent and/or received): DAT, ED, ED′, AD.
Further exemplary embodiments,
Further exemplary embodiments provide an adaptivity of the at least one filter dictionary so that the neural network can, for example, be better represented with comparatively few parameters than in a conventional spatial representation of the filter coefficients.
Number | Date | Country | Kind |
---|---|---|---|
10 2021 210 607.2 | Sep 2021 | DE | national |