This application claims priority to and the benefit of Korean Patent Application No. 10-2023-0156761, filed on Nov. 13, 2023, the disclosure of which is incorporated herein by reference in its entirety.
The present disclosure relates to a data processing technology using audio feature encoding.
Closed-circuit televisions (CCTVs) are increasingly installed for video surveillance and security, and the number of CCTV systems equipped with Internet Protocol (IP) cameras and network video recorder (NVR) servers or video management system (VMS) servers are increasing. A CCTV system includes at least one camera that captures a video, a digital video recorder (DVR) or an NVR that is connected to at least one camera with unshielded twisted pair (UTP) cables to compress and store video data from each camera, and a storage device that is connected to the DVR or the NVR, and the CCTV system transmits a video from each camera channel to a monitoring server or an administrator terminal through a network.
Video data captured by CCTV cameras may be used in various cases, such as the case of committing a criminal act (e.g., threatening a passerby with a knife or shooting a gun, etc.), the case where more than a certain number of people enter a specific area, the case where defects in production equipment or defects in manufactured products are detected in production plants, etc. However, in the case where it is difficult to identify an object using video data at night or the case where it is not possible to know from video alone whether an event has occurred (e.g., when an explosion occurred nearby but the explosion was not captured in the captured video, etc.), it is difficult to detect the occurrence of an event using only video data.
Further, when video data captured by a CCTV camera is transmitted together with audio data through a network, large amounts of transmission data are generated. As the resolution of a CCTV camera increases, the size of transmission data increases, and broadband transmission bandwidth is required to transmit such large amounts of video and audio data. Therefore, it is necessary to introduce a technology for reducing transmission bandwidth of audio and video data for security and surveillance according to machine processing.
The present disclosure is directed to providing a data processing technology for improving accuracy of event detection compared to the related art.
The present disclosure is also directed to providing a technology for reducing transmission bandwidth when transmitting video data and audio data according to machine processing.
According to an aspect of the present disclosure, there is provided a method. The method may include obtaining video data and extracting a video feature from the obtained video data, obtaining audio data related to the video data and extracting an audio feature from the obtained audio data, detecting a preset event on the basis of one or more of the video feature and the audio feature, when the preset event is detected, performing conversion processing and encoding processing on the video feature and the audio feature, and generating transmission data including the video feature and the audio feature on which the conversion processing and the encoding processing have been performed.
In detecting the preset event, when a condition related to the preset event is satisfied in both the video feature and the audio feature, it may be determined that the preset event has been detected. Alternatively, in detecting the preset event, when a condition related to the preset event is satisfied in any one of the video feature and the audio feature, it may be determined that the preset event has been detected.
The performing of the conversion processing and the encoding processing may include converting the video feature and encoding the converted video feature to generate video feature encoded data, and converting the audio feature and encoding the converted audio feature to generate audio feature encoded data. Here, the transmission data may include the video feature encoded data and the audio feature encoded data.
The performing of the conversion processing and encoding processing may include converting the video feature, converting the audio feature, and fusing the converted video feature and the converted audio feature to generate fusion feature data. Here, the transmission data may include fusion-encoded data generated by encoding the fusion feature data.
The transmission data may include metadata related to the detected event.
The transmission data may include time stamp information for synchronization of the video feature and the audio feature.
According to another aspect of the present disclosure, there is provided a method. The method may include obtaining audio data and extracting an audio feature from the obtained audio data, detecting a preset event on the basis of the audio feature, when the preset event is detected, extracting a video feature from video data related to the detected event, performing conversion processing and encoding processing on the video feature and the audio feature, and generating transmission data including the video feature and the audio feature on which the conversion processing and the encoding processing have been performed.
The performing of the conversion processing and the encoding processing may include converting the video feature and encoding the converted video feature to generate video feature encoded data, and converting the audio feature and encoding the converted audio feature to generate audio feature encoded data. Here, the transmission data may include the video feature encoded data and the audio feature encoded data.
The performing of the conversion processing and encoding processing may include converting the video feature, converting the audio feature, and fusing the converted video feature and the converted audio feature to generate fusion feature data. Here, the transmission data may include fusion-encoded data obtained by encoding the generated fusion feature data.
The transmission data may include metadata related to the detected event. The transmission data may include time stamp information for synchronization of the video feature and the audio feature.
According to still another aspect of the present disclosure, there is provided a program stored in a recording medium to allow a computer to perform the method.
According to yet another aspect of the present disclosure, there is provided a data processing device. The data processing device may include a memory configured to store input data, and a processor coupled to the memory. The processor may be configured to perform operations including obtaining video data and extracting a video feature from the obtained video data, obtaining audio data related to the video data and extracting an audio feature from the obtained audio data, detecting a preset event on the basis of one or more of the video feature and the audio feature, when the preset event is detected, performing conversion processing and encoding processing on the video feature and the audio feature, and generating transmission data including the video feature and the audio feature on which the conversion processing and the encoding processing have been performed.
According to yet another aspect of the present disclosure, there is provided a data processing device. The data processing device may include a memory configured to store input data, and a processor coupled to the memory. The processor may be configured to perform operations including obtaining audio data and extracting an audio feature from the obtained audio data, detecting a preset event on the basis of the audio feature, when the preset event is detected, extracting a video feature from video data related to the detected event, performing conversion processing and encoding processing on the video feature and the audio feature, and generating transmission data including the video feature and the audio feature on which the conversion processing and the encoding processing have been performed.
The above and other objects, features and advantages of the present disclosure will become more apparent to those of ordinary skill in the art by describing examples thereof in detail with reference to the accompanying drawings, in which:
Specific structural or step-by-step descriptions the present disclosure are merely for describing examples according to the concept of the present disclosure. Therefore, the concept of the present disclosure may be implemented in various forms. The present disclosure should not be construed as limited to the examples of the present disclosure.
While the present disclosure is open to various modifications and alternative forms, specific examples in the accompanying drawings will herein be described in detail. However, there is no intent to limit the present disclosure to the particular forms disclosed. On the contrary, it should be understood that the present disclosure is to cover all modifications, equivalents, and alternatives falling within the spirit and scope of the present disclosure.
It will be understood that, although the terms “first,” “second,” etc., may be used herein to describe various components, these components should not be limited by these terms. These terms are used only to distinguish one component from another component. For example, a first component could be termed a second component, and similarly, a second component could be termed a first component without departing from the scope of the present disclosure.
It should be understood that when a first component is referred to as being “connected” or “coupled” to a second component, the first component may be directly connected or coupled to the second component. However, it should be understood that other components may be present among a plurality of components. In contrast, when a component is referred to as being “directly connected” or “directly coupled” to another component, there are no intervening components present. Other words used to describe the relationship between components should be interpreted in a like fashion (i.e., “between” versus “directly between,” “adjacent” versus “directly adjacent,” and the like).
The terminology used herein is for the purpose of describing particular examples only and is not intended to be limiting to the present disclosure. As used herein, the singular forms “a” and “an” are intended to also include the plural forms, unless the context clearly indicates otherwise. It should be further understood that the terms “comprise,” “comprising,” “include,” and/or “including” used herein specify the presence of stated features, integers, steps, operations, components, parts, or combinations thereof, but do not preclude the presence or addition of one or more other features, integers, steps, operations, components, parts, or combinations thereof.
Unless otherwise defined, all terms including technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which the present disclosure belongs. It should be further understood that terms, such as those defined in commonly used dictionaries, should be interpreted as having a meaning that is consistent with their meaning in the context of the relevant art and are not to be interpreted in an idealized or overly formal sense unless expressly so defined herein.
In describing examples, descriptions of technical content that is well known in the technical field to which the present disclosure belongs and that are not directly related to the present disclosure will be omitted. This is to convey the gist of the present disclosure more clearly without obscuring it by omitting unnecessary explanation.
Hereinafter, the present disclosure will be described in detail by describing examples of the present disclosure with reference to the accompanying drawings.
Referring to
The data processing device 10 may be connected to a monitoring server 30 through a network. The data processing device 10 may process the video and/or audio data generated by the camera 20 to generate transmission data, and transmit the generated transmission data to the monitoring server 30. The transmission data may include a video feature and an audio feature. Further, the transmission data may include a fusion feature in which the video feature and the audio feature are fused.
The data processing device 10 may process the video and/or audio data generated by the camera 20 to generate transmission data using a typical video codec or audio codec, and transmit the generated transmission data to the monitoring server 30. The transmission data may include video data or audio data encoded using a typical video codec or audio codec.
The monitoring server 30 may restore video and/or audio feature data from the transmission data. The monitoring server 30 may restore video and/or audio feature information from the transmission data, and use the restored feature information for processing by a machine.
The monitoring server 30 may restore original video and/or audio data from the transmission data. The monitoring server 30 may restore the video and/or audio data from the transmission data, and transmit the restored video and/or audio data to a terminal 40.
The terminal 40 used by a user or administrator may be connected to the monitoring server 30 through a network. The terminal 40 may receive the video data and the audio data that are generated by the camera 20 and information related to the video and/or audio data from the monitoring server 30, or receive a result of processing, by the machine, the audio data and the video data.
The camera 20 may photograph a monitoring target region to obtain video data for the monitoring target region. The camera 20 may photograph the monitoring target region in real time for surveillance or security purposes. The camera 20 may be a pan-tilt-zoom (PTZ) camera capable of panning and tilting and with an adjustable zoom ratio of a lens thereof. The camera 20 may be provided as a plurality of cameras.
The cameras 20 may communicate with network devices using various communication methods such as a wired or wireless local area network (LAN), Wi-Fi, ZigBee, Bluetooth, near-field communication, etc. For example, the cameras 20 may communicate with each other according to a low-power wireless communication protocol using radio frequencies in industrial-scientific-medical (ISM) bands.
The data processing device 10 may store video data and audio data that are received from the camera 20, and provide transmission data obtained by processing the stored video and audio data to the monitoring server 30 through a network. The monitoring server 30 may restore original video and audio features by decoding the video and audio features included in the transmission data. The monitoring server 30 may use the restored video and audio features for machine processing. Through machine processing, a specific object may be recognized, an event may be detected, or information on an object related to the event may be recognized through metadata without an administrator.
The network may include a wired network or a wireless network. The wireless network may be a 2nd generation (2G) or 3rd generation (3G) cellular communication system, a 3rd Generation Partnership Project (3GPP), a 4th generation (4G) communication system, Long-Term Evolution (LTE), World Interoperability for Microwave Access (WiMAX), etc.
The terminal 40 may include a terminal used by an administrator or user of the monitoring server 30. The terminal 40 may access the monitoring server 30 through the network and receive and process monitoring data provided by the monitoring server 30.
The data processing device 10 may be a digital video recorder, a network video recorder, etc., but the present disclosure is not limited thereto. The data processing device 10 may be implemented as a single physical device or implemented by organically combining a plurality of physical devices. The data processing device 10 may be configured as a device integrated with the camera 20.
Referring to
The communication interface 11 may receive video data and audio data from the plurality of cameras 20. The communication interface 11 may be configured to transmit transmission data generated by the data processing device 10 to the monitoring server 30 through a network.
The processor 12 may obtain video data and extract a video feature from the obtained video data. The processor 12 may obtain audio data related to the video data and extract an audio feature from the obtained audio data. The processor 12 may detect a preset event on the basis of one or more of the video feature and the audio feature. When the preset event is detected, the processor 12 may perform conversion processing and encoding processing on the video feature and the audio feature. The processor 12 may generate transmission data including the video feature and the audio feature on which the conversion processing and the encoding processing have been performed.
Further, the processor 12 may obtain audio data, extract an audio feature from the obtained audio data, and detect a preset event on the basis of the extracted audio feature. When the preset event is detected, the processor 12 may extract a video feature from video data related to the detected event and perform conversion processing and encoding processing on the extracted video feature and audio feature. The processor 12 may generate transmission data including the video feature and the audio feature on which the conversion processing and the encoding processing have been performed.
The memory 13 may store input data including the video data and audio data obtained from the camera 20. The memory 13 may store data that are generated while the processor 12 processes the video data and the audio data, and may store the video feature and the audio feature that are generated by the processor 12. The memory 13 may store video feature on which the conversion processing and the encoding processing have been performed, the audio feature on which the conversion processing and the encoding processing have been performed, and metadata related to the video data and the audio data, etc.
Referring to
The video feature extractor 111 may extract a video feature from obtained video data. The audio feature extractor 121 may obtain audio data related to the video data and extract an audio feature from the obtained audio data.
The video feature extractor 111 may extract feature information or a feature map from input video data. The feature information may be called a feature map (or sparse map), a feature vector, a latent vector, or the like. The feature information may be in any possible form extracted from the video feature extractor 111, and may be defined according to at least one of a user, a device, a network type, and an environment.
When the video feature extractor 111 extracts features from input video data, various conventional image processing-based feature extraction techniques may be applied. Further, the video feature extractor 111 may extract features using one or more of deep learning or machine learning-based feature extraction techniques.
The audio feature extractor 121 may extract feature information or a feature map from input audio data. The feature information may be called a feature map (or sparse map), a feature vector, a latent vector, or the like. The feature information may be in any possible form extracted from the audio feature extractor 121, and may be defined according to at least one of a user, a device, a network type, and an environment.
When the audio feature extractor 121 extracts features from input audio data, various conventional sound processing-based feature extracting methods may be applied. Further, the audio feature extractor 121 may extract features using one or more of deep learning or machine learning-based feature extracting techniques.
There is no particular limitation on the method of expressing the feature information extracted by the video feature extractor 111 or the audio feature extractor 121, and the extracted feature information may be expressed in various ways. For example, the expression of the extracted feature information may vary depending on at least one of the type and size of the data, the type and size of the network, and the type and size of the network layer. Further, the extracted feature information may consist of data having at least one characteristic among highly correlated data such as general images, sparse data, and dense data.
The event detector 131 may detect a preset event on the basis of one or more of the video feature and the audio feature. The event detector 131 may generate metadata related to the detected event. The metadata may include information such as the type of event (dangerous situation, situation requiring rescue, situation in which a failure occurs, etc.), information on an object type related to the event (vehicle, person, animal, etc.), and detailed information on an object related to the event (when the object is a person, the person's sex, clothing color, etc.), etc.
The event may include vehicle crashes, firearm firing, graffiti, assaults, robberies, etc. For example, car crashes have sounds of glass breaking, rubber slipping, steel plates crumpling, etc. Firearm firing has an explosion sound due to a gun being cocked or a sound of a bullet hitting at supersonic speed, etc. Graffiti has sounds such as those made by spray paint being released. Vehicles breaking entry barriers, robberies, etc. are accompanied by sounds resulting from glass breakage or similar impacts. The event detector 131 may determine whether an event has occurred in consideration of video data captured at the same time as such sounds.
Further, the event may include an equipment failure event in a product production plant and a defective event in a manufactured product. The equipment failure event in the product production plant may be detected using a unique signature sound that occurs when specific equipment fails. The defective event in the manufactured product may be detected using a unique signature sound that occurs when a specific manufactured product is defective. The event detector 131 may determine whether an event has occurred in consideration of video data captured at the same time for such unique signature sounds.
When a preset event is detected by the event detector 131, the video feature converter 112 may perform conversion processing on the video feature, and the video feature encoder 113 may encode the video feature on which the conversion processing has been performed, to generate video feature encoded data.
The video feature converter 112 performs predetermined conversion processing on the extracted video feature. The video feature converter 112 may convert the extracted video feature information as necessary before subsequent encoding processing. The video feature converter 112 may convert the video feature information into a form suitable for compressing and decompressing the video feature information.
The video feature converter 112 may convert the extracted video feature information into another form using various methods. For example, the video feature information may be converted by applying one or more methods of normalization, scaling, rearrangement, representation bit reduction, quantization, and filtering.
The video feature encoder 113 may perform encoding on the converted video feature information. There is no particular limitation on the encoding technique performed by the video feature encoder 113, and encoding techniques based on deep learning and machine learning as well as encoding techniques in conventional video compression standards may be applied.
In one example, deep learning-based video feature encoding may include one or more convolutional layers and a fully connected layer. In this case, the type of filter for the convolutional layer may change depending on one or more of the type, learning type, and size of the feature.
The video feature encoder 113 may perform encoding on the video feature in units of at least one of sample, line, block, and frame. In this case, encoding may be performed in units of at least one of sample, line, block, and frame, depending on at least one of the shape, size, and dimension of input video feature. Further, the video feature encoder 113 may perform prediction-based encoding, binarization-based encoding, entropy-based encoding, or transformation-based encoding.
Further, when a preset event is detected by the event detector 131, the audio feature converter 122 may perform conversion processing on the audio feature, and the audio feature encoder 123 may encode the audio feature on which the conversion processing has been performed, to generate audio feature encoded data.
The audio feature converter 122 performs predetermined conversion processing on the extracted audio feature. The audio feature converter 122 may convert the extracted audio feature information as necessary before subsequent encoding processing. The audio feature converter 122 may convert the audio feature information into a form suitable for compressing and decompressing the audio feature information.
The audio feature converter 122 may convert the extracted video feature information into another form using various methods. For example, the audio feature information may be converted by applying one or more methods of normalization, scaling, rearrangement, representation bit reduction, quantization, and filtering.
The audio feature encoder 123 may perform encoding on the converted audio feature information. There is no particular limitation on the encoding technique performed by the audio feature encoder 123, and encoding techniques based on deep learning and machine learning as well as encoding techniques in conventional audio compression standards may be applied.
In one example, deep learning-based audio feature encoding may include one or more convolutional layers and a fully connected layer. In this case, the type of filter for the convolutional layer may change depending on one or more of the type, learning type, and size of the feature.
The data processing device may generate transmission data including the video feature encoded data and the audio feature encoded data. The transmission data may be transmitted to the monitoring server 30 through a communication network. The transmission data may include metadata related to the detected event.
The transmission data may include a time stamp for synchronization of the video feature and the audio feature. The time stamp is information indicating a time at which a video and an audio are generated. The video data, the video feature, the converted video feature, and the video feature encoded data may each include a time stamp indicating a generation time, and the video and the audio may be synchronized by the time stamps.
When a condition related to the preset event is satisfied in both the video feature and the audio feature, the event detector 131 may determine that the preset event has been detected. In another example, when the condition related to the preset event is satisfied in any one of the video feature and the audio feature, the event detector 131 may determine that the preset event has been detected.
Referring to
The descriptions of the video feature extractor 111, the video feature converter 112, the audio feature extractor 121, the audio feature converter 122, and the event detector 131 regarding the data processing device are the same as illustrated in
The feature fusion device 141 may receive the converted video feature and the converted audio feature and fuse the converted video feature and the converted audio feature to generate fusion feature data. Since the video feature and the audio feature have feature vectors of different dimensions, a process of converting the video feature and the audio feature into video features and audio features of the same dimension is required. For example, when a feature vector of the video feature is a p-dimensional feature vector and a feature vector of the audio feature is a q-dimensional feature vector, fusion feature data obtained by fusing the two feature vectors may have a d-dimensional feature vector.
A single-layer neural network may be used to match feature vectors of two different dimensions to the same d dimension. A rectified linear unit (ReLu) may be used as an activation function of a single-layer neural network. In consideration of differences in distribution of feature vectors, normalization to a unit vector may be performed on the feature vectors. A fused feature vector may be generated by vector synthesis from the feature vectors of the video feature and the audio feature matched to the d dimension.
The fusion feature encoder 142 may perform encoding on the fusion feature data and generate fusion-encoded data. The fusion-encoded data may be included in the transmission data and transmitted to the monitoring server 30.
There is no particular limitation on the encoding technique performed by the fusion feature encoder 142, and encoding techniques based on deep learning and machine learning as well as encoding techniques in conventional video and/or audio compression standards may be applied.
In one example, deep learning-based feature encoding may include one or more convolutional layers and a fully connected layer. In this case, the type of filter for the convolutional layer may change depending on one or more of the type, learning type, and size of the feature.
The data processing device may include an audio feature extractor 121, an audio feature converter 122, an audio feature encoder 123, and an event detector 131. Regarding the audio feature extractor 121, the descriptions of the audio feature extractor 121, the audio feature converter 122, the audio feature encoder 123, and the event detector 131 are the same as illustrated in
The audio feature extractor 121 may obtain audio data and extract an audio feature from the obtained audio data.
The event detector 131 may detect a preset event on the basis of the audio feature. When the preset event is detected, the audio feature converter 122 may perform conversion processing on the audio feature. The audio feature encoder 123 may encode the audio feature on which the conversion processing has been performed, to generate audio feature encoded data. The generated audio feature encoded data may be included in transmission data and transmitted to the monitoring server 30.
Referring to
Referring to
Among the video feature extractor 111, the video feature converter 112, the video feature encoder 113, the audio feature extractor 121, the audio feature converter 122, the audio feature encoder 123, and the event detector 131 of the data processing device illustrated in
Referring to
When the preset event is detected, the video feature extractor 111 may extract a video feature from video data related to the detected event. When the preset event is detected, the event detector 131 transmits information on a time stamp (recording time information) at which the event is detected to the video feature extractor 111, and thus the video feature extractor 111 may extract the video feature from the video data corresponding to time information of the event detected using the audio data.
The video feature converter 112 may perform conversion processing on the extracted audio feature. The video feature encoder 113 may perform encoding processing on the audio feature on which the conversion processing has been performed.
When the preset event is detected, the audio feature converter 122 may perform conversion processing on the extracted audio feature. The audio feature encoder 123 may perform encoding processing on the audio feature on which the conversion processing has been performed.
Referring to
Referring to
Referring to
When the preset event is detected, the video feature extractor 111 may extract a video feature from video data related to the detected event. The video feature converter 112 may perform conversion processing on the extracted video feature. Further, when the preset event is detected, the audio feature converter 122 may perform conversion processing on the extracted audio feature.
The feature fusion device 141 may perform fusing processing on the video feature generated by the video feature converter 112 and the audio feature generated by the audio feature converter 122. The fusion feature encoder 142 may perform encoding processing on the features on which the fusing process has been performed.
The feature fusion device 141 and the fusion feature encoder 142 of
The audio feature extractor 121 may use one method or combine two or more methods among algorithms for extracting features of audio data, such as linear predictive coding (LPC), mel-frequency cepstral coefficients (MFCCs), mel spectrogram, etc., to extract features from audio data.
In
The audio feature extractor 121 may pre-process an audio signal in a time domain and boost high-frequency energy (a). An input audio signal goes through a pre-emphasis filter with high-pass characteristics to model the frequency characteristics of the human outer/middle ear. The audio signal that has gone through the pre-emphasis filter is converted into a digital signal through sampling, quantization, and encoding.
The audio feature extractor 121 may divide the pre-processed audio signal into a plurality of sections and obtain a spectrum in a frequency domain by performing Fourier transform (FFT) (b). Framing processing may be performed by dividing the pre-processed audio signal into sections of 20 ms each. Discontinuous points at both ends of a frame may contain unwanted information when converted into the frequency domain, and accordingly, each frame is multiplied by a window coefficient to minimize discontinuities at start and end points. Hann window, Hamming window, Blackman window, Kaiser window, etc. may be used as a window function.
After framing processing, the frequency characteristics may be determined by converting the signal from the time domain to the frequency domain using a Fast Fourier Transform (FFT) algorithm. Discrete and aperiodic audio signals in the time domain are converted into periodic and continuous signals in the frequency domain through Fourier transform.
The audio feature extractor 121 may obtain the sum of sizes in each band by matching a triangular filter bank tailored to the mel scale to the obtained spectrum (c). By passing a result value obtained by the FFT algorithm through the filter bank, the frequency band may be divided into several filter banks and energy in each bank may be obtained.
The audio feature extractor 121 may take the logarithm of the output value of the filter bank (d) and then perform a discrete cosine transform (DCT) (e) to obtain an MFCC feature vector (f). Since humans feel the loudness of sound as a logarithmic function, the logarithm of the output energy value of each filter bank is taken, and then when a DCT is performed, the energy value is converted into a sinusoidal coefficient of frequency in mel scale. Through this process, a feature vector may be obtained from the audio signal.
Referring to
A preset event may be detected based on one or more of the obtained video data and the obtained audio feature (operation S130). When the preset event is detected, conversion processing and encoding processing may be performed on the video feature and the audio feature (operation S140).
Thereafter, transmission data including the video feature and the audio feature on which the conversion processing and the encoding processing have been performed may be generated (operation S150). The generated transmission data may be transmitted to the monitoring server 30. The detailed descriptions of the data processing device of
Referring to
When the preset event is detected, a video feature may be extracted from video data related to the detected event (operation S230). Conversion processing and encoding processing may be performed on the video feature and the audio feature (operation S240).
Thereafter, transmission data including the video feature and the audio feature on which the conversion processing and the encoding processing haven been performed may be generated (operation S250). The generated transmission data may be transmitted to the monitoring server 30. The detailed descriptions of the data processing device of
According to the present disclosure, it is possible to improve accuracy of event detection compared to the related art.
Further, according to the present disclosure, it is possible to reduce transmission bandwidth when transmitting video data and audio data according to machine processing.
The present disclosure disclosed in this specification and drawings are only examples to aid understanding of the technical content of the present disclosure and not intended to limit the scope of the present disclosure. It is clear to those skilled in the art that various modifications based on the technological scope of the present disclosure in addition to the examples disclosed herein can be made.
Number | Date | Country | Kind |
---|---|---|---|
10-2023-0156761 | Nov 2023 | KR | national |