This application claims priority to Spanish Patent Application No. P202230148, filed 23 Feb. 2022 and U.S. provisional application 63/431,590, filed 9 Dec. 2022, all of which are incorporated herein by reference in their entirety.
The present application relates to audio processing and machine learning.
The approaches described in this section are approaches that could be pursued, but not necessarily approaches that have been previously conceived or pursued. Therefore, unless otherwise indicated, it should not be assumed that any of the approaches described in this section qualify as prior art merely by virtue of their inclusion in this section.
Speech coding found its first major application in secure communication, and later enabled low-cost mobile and internet communication. With the continuously decreasing bitrate of speech codecs, decoded speech suffers from various quality issues. It would be helpful to address such quality issues.
A computer-implemented method of restoring clean speech from coded audio data. The method comprises obtaining coded audio data comprising a first set of frames. The method further comprises extracting a set of feature vectors from the coded audio data using a self-supervised deep learning model including a neural network, where the set of feature vectors is respectively extracted from the first set of frames. The method additionally comprises generating enhanced speech data comprising a second set of frames from the set of feature vectors using a generative deep learning model including a neural network, where the enhanced speech data corresponds to clean speech in the coded audio data.
Techniques described in this specification can be advantageous over conventional audio processing techniques. For example, the method enables the generation of improved speech data from real-life coded audio data containing speech that is often contaminated. The generation is conditioned on robust features obtained from self-supervised learning of speech properties at various levels, from low-level raw audio properties to high-level speaker-related properties, which helps achieve a high quality in the improved speech data despite the distortion and coding. The improved speech data is improved with efficiency using relatively compact deep learning models that may run in real time as raw speech signals are generated. The improved speech data leads to better perception of the audio and better user enjoyment of the audio.
The example embodiment(s) of the present invention are illustrated by way of example, and not in way by limitation, in the figures of the accompanying drawings and in which like reference numerals refer to similar elements and in which:
In the following description, for the purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding of the example embodiment(s) the present invention. It will be apparent, however, that the example embodiment(s) may be practiced without these specific details. In other instances, well-known structures and devices are shown in block diagram form in order to avoid unnecessarily obscuring the example embodiment(s).
Embodiments are described in sections below according to the following outline:
A system for generating enhanced speech data using robust audio features is disclosed. In some embodiments, a system is programmed to use a self-supervised deep learning model to generate a set of feature vectors from given audio data that contains contaminated speech and is coded. The system is further programmed to use a generative deep learning model to create improved audio data corresponding to clean speech from the set of feature vectors.
In some embodiments, the system is programmed to train the self-supervised deep learning model. The training includes using specific audio features of clean speech signals to compute the weights of an encoder that is configured to generate a set of feature vectors from contaminated, coded speech signals. The audio features can include low-level audio features, such as energy or pitch, or high-level audio features, such as speaker accent. A speech signal refers to a digitalized audio waveform that includes speech. The digitalization results in one or more frames, each frame including one or more samples corresponding to a particular sampling rate. The distortion can come from extra speech or non-speech audio in the environment, such as noise or reverberation, or limitations of the recording equipment, such as clipping, for example. The coding can result from application of a compression algorithm with a target bitrate. The trained self-supervised deep learning model is expected to generated a set of robust feature vectors, for each frame of a contaminated, coded speech signal, that characterize a frame of the corresponding clean speech signal.
In some embodiments, while the clean speech signals are of a certain sampling rate, the system is programmed to down-sample coded speech signals, which can be derived from the clean speech signals, to form the training dataset for the self-supervised deep learning model. For example, speech signals of a sampling rate of 48 kHz can be down-sampled to speech signals of a sampling rate of 16 kHz. The down-sampled speech signals are expected to still capture main audio features that can be used to restore 48 kHz clean signals.
In some embodiments, the system is programmed to then train the generative deep learning model. The training includes using clean speech signals to compute the weights of a recurrent neural network that is configured to generate improved speech signals from sets of feature vectors, which are generated by the self-supervised deep learning model from contaminated, coded speech signals. The trained generative deep learning model is expected to generate improved speech signals of a predetermined sampling rate, such as 48 kHz, even if the sets of feature vectors are generated from speech signals of a lower sampling rate, such as 16 kHz.
In some embodiments, given a new speech signal that has been coded using any coding algorithm for any bitrate, the system is configured to execute the trained self-supervised deep learning model on the new speech signal to generate a set of feature vectors. The system is configured to further execute the trained generative deep learning model on the set of feature vectors to generate an improved speech signal that corresponds to decoded, clean speech signal.
The system produces technical benefits. With the continuously decreasing bitrate of speech codecs, decoded speech suffers more from lost bandwidths, spectral holes, or other quality issues. The system addresses the technical problem of unpredictable quality of decoded speech signals that result from lowering bitrates of the coding process. The system enables the generation of improved speech data from real-life coded audio data containing speech that is often contaminated. The generation is conditioned on robust features obtained from self-supervised learning of speech properties at various levels, from low-level raw audio properties to high-level speaker-related properties, which helps achieve a high quality in the improved speech data despite the distortion and coding. The improved speech data is improved with efficiency using relatively compact deep learning models that may run in real time as raw speech signals are generated. The improved speech data leads to better perception of the audio and better user enjoyment of the audio.
In some embodiments, the networked computer system comprises an audio management server computer 102 (“server”), one or more sensors 104 or input devices, and one or more output devices 110, which are communicatively coupled through direct physical connections or via one or more networks 118.
In some embodiments, the server 102 broadly represents one or more computers, virtual computing instances, and/or instances of an application that is programmed or configured with data structures and/or database records that are arranged to host or execute functions related to restoring clean speech signals from contaminated, coded speech signals. The server 102 can comprise a server farm, a cloud computing platform, a parallel computer, or any other computing facility with sufficient computing power in data processing, data storage, and network communication for the above-described functions.
In some embodiments, each of the one or more sensors 104 can include a microphone or another digital recording device that converts sounds into electric signals. Each sensor is configured to transmit raw or processed audio data to the server 102. Each sensor may include a processor or may be integrated into a typical client device, such as a desktop computer, laptop computer, tablet computer, smartphone, or wearable device, where a processor coupled with the sensor can perform initial processing of the audio data.
In some embodiments, each of the one or more output devices 110 can include a speaker or another digital playing device that converts electrical signals back to sounds. Each output device is programmed to play audio data received from the server 102. Similar to a sensor, an output device may include a processor or may be integrated into a typical client device, such as a desktop computer, laptop computer, tablet computer, smartphone, or wearable device, where a processor coupled with the sensor can perform subsequent processing of the audio data.
The one or more networks 118 may be implemented by any medium or mechanism that provides for the exchange of data between the various elements of
In some embodiments, the server 102 is programmed to receive input audio data corresponding to sounds in a given environment from the one or more sensors 104. The input audio data may comprise a plurality of frames over time. The sever 102 is programmed to next process the input audio data, which is coded and typically has a mixture of speech and noise, to generate improved audio data that corresponds to clean speech. The server can also be programmed to send the output audio data to the one or more output devices.
In some embodiments, the server 102 comprises feature extraction model training instructions 202, speech enhancement model training instructions 206, model execution instructions 208, and communication interface instructions 210. The server 102 also comprises a database 220.
In some embodiments, the feature extraction model training instructions 202 enable training deep learning models for extracting audio features corresponding to clean speech signals from coded, contaminated speech signals. The deep learning models include various neural networks or other self-supervised models that can learn features from clean speech signals.
In some embodiments, the speech enhancement model training instructions 206 enable training deep learning models for generating improved speech signals from sets of audio features generated by the feature extraction model. The deep learning models include various neural networks or other generative models that can generate improved speech signals from audio features corresponding to clean speech signals.
In some embodiments, the model execution instructions 208 enable executing deep learning models for generating improved speech data from contaminated, coded speech data. The execution can include executing the feature extraction model on coded audio data containing a mixture of speech and non-speech to generate a set of feature vectors. The execution can further include executing the speech enhancement model on the set of feature vectors to generate improved audio data that corresponds to clean speech.
In some embodiments, the communication interface instructions 210 enable communication with other systems or devices through computer networks. The communication can include receiving audio data or trained deep learning models from audio sources or other systems. The communication can also include transmitting enhanced speech data to other processing devices or output devices.
In some embodiments, the database 220 is programmed or configured to manage storage of and access to relevant data, such as audio data, including clean speech signals, contaminated speech signals, coded speech signals, or down-sampled speech signals, digital models, features extracted from the audio data, or results of executing the digital models.
Speech signals, namely digital audio data containing speech and typically represented as waveforms in the time domain, encompass a rich set of acoustic and linguistic properties, ranging from individual lexical units, such as phonemes or words, to speaker characteristics, such as their intent or emotional status. However, such properties of speech are poorly captured by low-level, signal-processing features, such as amplitudes of a wave signal, log Mel spectrograms, or Mel frequency cepstral coefficients (MFCCs). Coded speech is already missing information. Restoring original speech using only low-level, signal-processing features extracted from the coded speech. which do not fully characterize the speech signals, could be especially difficult.
In some embodiments, the server therefore identifies properties or extract features from coded speech using a feature extraction model, which is a self-supervised deep learning model. The extracted features are then used to build a speech enhancement model, which is a generative deep learning model.
Speech signals are generally distorted by various contaminations caused by the environment or the recording apparatus, such as reverberation, additive noise, overlapping speech, temporal or frequency masking, or clipping. Speech encoding introduces yet another level of distortion.
A sampling rate of 48 kHz for digital speech is common in today's multimedia communication. Based on a reasonable assumption that acoustic and linguistic content, such as phoneme or pitch, mainly resides in a low-frequency region, a lower sampling rate can be used for training the feature extraction model. Using a lower sampling rate can reduce the amount of data needed and also increase the case of finding needed data. The assumption further includes that the low-frequency region generally lies below 8 kHz.
Therefore, in some embodiments, the server 102 chooses as the lower sampling rate one that is between 8 kHz and 48 kHz, such as 16 kHz. The server 102 then includes 16 kHz speech signals in the first dataset and uses the first dataset as a first training set. Alternatively, the server 102 down-samples 48 kHz speech signals in the first dataset to obtain a first training set of the lower sampling rate that is much bigger than the first dataset. While the feature extraction model is built mainly from training data of the lower sampling rate, the feature extraction model is expected to work with and thus is used for extracting features from actual data of the original, higher sampling rate, as further discussed below. Experiments show that a feature extraction model built from the first training set of a sampling rate of 16 kHz is capable of restoring 48 kHz clean speech signals with high quality.
In some embodiments, the server 102 builds a feature extraction model, such as 320, from the first training set. The server 102 can build a PASE+ model, as described in the paper by Ravanelli et al. PASE+ uses a self-supervised learning approach to extract speech information at various levels, such as phonemes or speaker emotions. PASE+ is a deep neural network that takes as input a distorted speech signal as a waveform and generates as output a set of high-dimensional feature vectors, which is expected to characterize a corresponding clean speech signal, where each feature is for one frame of the distorted speech signal. For example, each frame may be 10 ms long with 480 samples for a sampling rate of 48 kHz. Therefore, the server 102 uses each distorted speech signal in the first training set for building the PASE+ model with the goal to generate a set of feature vectors, such as 308, that characterizes corresponding clean speech without distortions.
Specifically, PASE+ comprises an encoder (including a quasi-recurrent neural network to learn long-term dependencies across time steps) that extracts a group of features as a feature vector from a frame of a distorted speech signal. PASE+ also comprises a number of workers. Each worker is a small feed-forward neural network that accomplishes a self-supervised signal-transformation task corresponding to a known speech property. For example, when the signal transformation task corresponds to the MFCC, the worker constructs the “target” MFCC values from a clean speech signal corresponding to the distorted speech signal (thus the flow from 304 to 320, for example) and determines the loss of the current feature vector from the target MFCC values. The signal-transformation task is considered self-supervised rather than supervised in that while no classification or label of the ground truth (e.g., the clean speech signal) is separately given, properties of the ground truth can be and are automatically computed from the ground truth. The total loss of the current feature vector is computed based on the losses determined by all the workers.
Experimental results show that distortions introduced into clean speech signals tend to augment the training data with more variety and further force denoising the training data to learn distortion-invariant, robust features. Using the first training set of distorted signals, the workers are trained with the encoder to determine the weights for the encoder. Given an actual speech signal, only the trained encoder needs to be executed to obtain a corresponding group of robust features, as further discussed in Section 4.3.
In some embodiments, the workers' tasks can be related to signal processing features, such as filter banks, prosody (interpolated logarithm of the fundamental frequency, voiced/unvoiced probability, zero-crossing rate, and energy), log Mel spectrograms, or MFCCs, or sampling strategies used to capture speaker characteristics, such as local info max or global info max, as also discussed in the paper by Ravanelli et al. In this manner, PASE+ provides a deep and compact representation of various levels of speech abstraction, from low-level spectral information to high-level speaker information.
In some embodiments, the server 102 builds an alternative feature extraction model instead of the PASE+ model. To utilize the first training set, the alternative feature extraction model would also adopt a self-supervised or unsupervised learning approach to extract speech information at various levels. An example of the alternative feature extraction model is contrastive predictive coding (CPC), which uses a contrastive loss function that relies on encodings of positive samples.
In some embodiments, the server 102 starts with a sampling rate common in today's multimedia communication. The server 102 prepares a second dataset of speech signals, such as 334, that are contaminated to various extents for a speech enhancement model. The speech signals in the second dataset can also be subject to different codecs with different bitrates. The server 102 then down-samples the speech signals in the second dataset to obtain a down-sampled dataset of a lower sampling rate, such as 16 kHz. Next, the server runs the down-sampled dataset through the trained feature extraction model, and obtains a second training set of sets of feature vectors, such as 338, for the speech enhancement model.
In some embodiments, the server 102 builds a speech enhancement model, such as 350, from the second training set. The server 102 can build a WaveRNN model, as described in the paper titled “Efficient Neural Audio Synthesis” by Kalchbrenner et al. WaveRNN uses a sequential process to generate high-quality samples efficiently. WaveRNN is a deep neural network that takes as input a set of feature vectors, one for each frame of the original speech signal, which is expected to characterize a corresponding clean speech signal, and generates as output a corresponding improved speech signal, which is expected to correspond to the clean speech signal. Therefore, the server 102 can use sets of feature vectors generated from distorted speech signals by the feature extraction model for building the WaveRNN model with the goal to obtain, for each distorted speech signal, an improved speech signal that is free from any distortions present in the distorted speech signal, such as 340.
Specifically, WaveRNN includes a conditional network and a recurrent network. The conditional network consists of a pair of convolutional networks: a residual network and an upsampling network that work in parallel. The residual network with growing dilations through blocks of learned transforms can map a set of feature vectors to a latent representation, which is then split into multiple parts that are used as input to the subsequent recurrent network. At the same time, the set of feature vectors goes through the upsampling network to generate a second set of feature vectors that match the temporal size of the original speech signal. The outputs of these two convolutional networks are concatenated to form a set of output feature vectors, which is then fed into the recurrent network.
The recurrent network, which contains fully connected (FC) layers, unidirectional gated recurrent units (GRU), and a softmax function, generates the improved speech signal one sample at a time. As each output feature vector generally corresponds to a frame of multiple samples, the same output feature vector is used in generating the corresponding samples in the improved speech signal. For each current sample of the improved speech signal, the output feature vector and the previous samples already generated are concatenated before going through blocks of learned transforms ending with the softmax activation to compute the probability of each possible value for the current sample, which is then used to calculate the cross-entropy loss of the clean speech samples.
As discussed in the previous section, while the sets of feature vectors can be generated from distorted speech signals, the sets of feature vectors are expected to characterize the clean speech signals. In some embodiments, to increase the accuracy of the resulting WaveRNN model, for each current sample of the improved speech signal, the output feature vector and the previous samples of the known clean speech signal corresponding to the previous samples of the improved speech signal are concatenated instead (thus the flow from 332 to 350, for example) before going through the blocks of learned transforms ending with the softmax activation to compute the probability of each possible value for current sample. The probability is then used to calculate the cross-entropy loss of the clean speech samples and update the model parameter values.
In some embodiments, the original speech signal is of a sampling rate of 48 kHz, the set of feature vectors are generated by the feature extraction model from down-sampled data of a sampling rate of 16 kHz, and the speech enhancement network generates the improved speech signal that is again of a sampling rate of 48 kHz.
In some embodiments, the server 102 builds an alternative speech enhancement model instead of the WaveRNN model. To utilize the second training set, the alternative speech enhancement model would also be a generative model to restore the clean speech signal. Examples of the alternative speech enhancement model include LPCNet, WaveNet, or SampleRNN.
In some embodiments, instead of training the feature extraction model and the speech enhancement model separately, the server 102 trains the two models together. Then, instead of potentially using separate datasets, such as 306 and 336, to train the two models, the output 308 of the feature extraction model 320 could flow directly into the generative deep model 350 during training, and parameters of both models can be tuned during the same training process. In this manner, the extracted features can be further refined to better adapt themselves to the application of interest, which in this case is restoring clean speech signals.
In some embodiments, given a new contaminated, coded speech signal, the server 102 generates a set of feature vectors using the PASE+ model. When the new speech signal is of a sampling rate common in today's multimedia communication, such as 48 kHz, the server 102 down-samples the new speech signal to the lower sampling rate, such as 16 kHz, used to train the feature extraction model before executing the feature extraction model. From the set of feature vectors, the server 102 then generates an improved signal using the WaveRNN model. In this case, the samples being used in generating the current sample of the improved speech signal would be the previous samples generated for the improved speech signal. The generation process is autoregressive by randomly sampling from a predicted distribution of the current sample based on the previously generated samples.
In some embodiments, the PASE+ model can be replaced by an alternative feature extraction model, or the WaveRNN model can be replaced by an alternative speech enhancement model in generating the improved speech signal for the given new contaminated, coded speech signal, as discussed above.
In some embodiments, the server 102 sends the improved speech signal to an output device, such as an audio player or another computing device on which the improved speech signal can be further processed.
In step 402, the server 102 is programmed to obtain coded audio data comprising a first set of frames.
In some embodiments, the server 102 is programmed to receive original coded data and down-sample the original coded data to obtain the coded audio data. The original coded data can correspond to a sampling rate of 48 kHz, and the coded audio data can correspond to a sampling rate of 16 kHz. The coded audio data can contain noise or reverbs.
In step 404, the server 102 is programmed to extract a set of feature vectors from the coded audio data using a self-supervised deep learning model including a neural network, the set of feature vectors being respectively extracted from the first set of frames.
In some embodiments, the self-supervised deep learning model includes an encoder and a plurality of workers. Each worker of the plurality of workers performs a self-supervised task related to a distinct speech property. One worker of the plurality of workers can perform the self-supervised task related to a pre-defined sampling strategy that draws anchor, positive, and negative samples from a pool of representations generated by the encoder.
In step 406, the server 102 is programmed to generate enhanced speech data comprising a second set of frames from the set of feature vectors using a generative deep learning model including a neural network, the enhanced speech data corresponding to clean speech in the coded audio data.
In some embodiments, the generative deep learning model includes a conditional network and a recurrent network. The conditional network converts the set of feature vectors into a set of output feature vectors by considering multiple frames each time. The recurrent network generates the enhanced speech data from the set of output feature vectors one sample at a time, where each frame of the second set of frames comprises a plurality of samples. The recurrent network can generate a new sample of each frame the enhanced speech data using a corresponding feature vector of the set of feature vectors and samples of the enhanced speech data generated previously.
In some embodiments, the server 102 is programmed to obtain a training set of distorted, down-sampled speech signals relative to a predetermined sampling rate. The server is further configured to build the self-supervised deep learning model using the training set of distorted, down-sampled speech signals.
In some embodiments, the server 102 is programmed to obtain a second dataset of clean speech signals of the predetermined sampling rate corresponding to the coded, down-sampled speech signals. The server is configured to also distort a copy of the second dataset with one or more artifacts caused by a recording environment, a recording equipment, or a coding algorithm, to obtain the first training set. The server is further configured to build the self-supervised deep learning model using also the second dataset.
In some embodiments, the server 102 is programmed to obtain a dataset of coded, down-sampled speech signals relative to a predetermined sampling rate. The server 102 is further configured to generate a training set of sets of feature vectors from the dataset using the self-supervised deep learning model. In addition, the server 102 is configured to build the generative deep learning model using the training set of sets of feature vectors.
In some embodiments, the server 102 is programmed to obtain a third dataset of clean speech signals of the predetermined sampling rate corresponding to the coded. down-sampled speech signals. The server 102 is further configured to build the generative deep learning model also using the third dataset.
In some embodiments, the server 102 is programmed to obtain a dataset of distorted, down-sampled speech signals relative to a predetermined sampling rate and train a combined model comprising the self-supervised deep learning model connected with the generative deep learning model using the dataset.
Various aspects of the disclosed embodiments may be appreciated from the following enumerated example embodiments (EEEs):
According to one embodiment, the techniques described herein are implemented by at least one computing device. The techniques may be implemented in whole or in part using a combination of at least one server computer and/or other computing devices that are coupled using a network, such as a packet data network. The computing devices may be hard-wired to perform the techniques, or may include digital electronic devices such as at least one application-specific integrated circuit (ASIC) or field programmable gate array (FPGA) that is persistently programmed to perform the techniques, or may include at least one general purpose hardware processor programmed to perform the techniques pursuant to program instructions in firmware, memory, other storage, or a combination. Such computing devices may also combine custom hard-wired logic, ASICs, or FPGAs with custom programming to accomplish the described techniques. The computing devices may be server computers, workstations, personal computers, portable computer systems, handheld devices, mobile computing devices, wearable devices, body mounted or implantable devices, smartphones, smart appliances, internetworking devices, autonomous or semi-autonomous devices such as robots or unmanned ground or aerial vehicles, any other electronic device that incorporates hard-wired and/or program logic to implement the described techniques, one or more virtual computing machines or instances in a data center, and/or a network of server computers and/or personal computers.
Computer system 500 includes an input/output (I/O) subsystem 502 which may include a bus and/or other communication mechanism(s) for communicating information and/or instructions between the components of the computer system 500 over electronic signal paths. The I/O subsystem 502 may include an I/O controller, a memory controller and at least one I/O port. The electronic signal paths are represented schematically in the drawings, for example as lines, unidirectional arrows, or bidirectional arrows.
At least one hardware processor 504 is coupled to I/O subsystem 502 for processing information and instructions. Hardware processor 504 may include, for example, a general-purpose microprocessor or microcontroller and/or a special-purpose microprocessor such as an embedded system or a graphics processing unit (GPU) or a digital signal processor or ARM processor. Processor 504 may comprise an integrated arithmetic logic unit (ALU) or may be coupled to a separate ALU.
Computer system 500 includes one or more units of memory 506, such as a main memory, which is coupled to I/O subsystem 502 for electronically digitally storing data and instructions to be executed by processor 504. Memory 506 may include volatile memory such as various forms of random-access memory (RAM) or other dynamic storage device. Memory 506 also may be used for storing temporary variables or other intermediate information during execution of instructions to be executed by processor 504. Such instructions, when stored in non-transitory computer-readable storage media accessible to processor 504, can render computer system 500 into a special-purpose machine that is customized to perform the operations specified in the instructions.
Computer system 500 further includes non-volatile memory such as read only memory (ROM) 508 or other static storage device coupled to I/O subsystem 502 for storing information and instructions for processor 504. The ROM 508 may include various forms of programmable ROM (PROM) such as erasable PROM (EPROM) or electrically erasable PROM (EEPROM). A unit of persistent storage 510 may include various forms of non-volatile RAM (NVRAM), such as FLASH memory, or solid-state storage, magnetic disk or optical disk such as CD-ROM or DVD-ROM, and may be coupled to I/O subsystem 502 for storing information and instructions. Storage 510 is an example of a non-transitory computer-readable medium that may be used to store instructions and data which when executed by the processor 504 cause performing computer-implemented methods to execute the techniques herein.
The instructions in memory 506, ROM 508 or storage 510 may comprise one or more sets of instructions that are organized as modules, methods, objects, functions, routines, or calls. The instructions may be organized as one or more computer programs, operating system services, or application programs including mobile apps. The instructions may comprise an operating system and/or system software; one or more libraries to support multimedia, programming or other functions; data protocol instructions or stacks to implement TCP/IP, HTTP or other communication protocols; file processing instructions to interpret and render files coded using HTML, XML, JPEG, MPEG or PNG; user interface instructions to render or interpret commands for a graphical user interface (GUI), command-line interface or text user interface; application software such as an office suite, internet access applications, design and manufacturing applications, graphics applications, audio applications, software engineering applications, educational applications, games or miscellaneous applications. The instructions may implement a web server, web application server or web client. The instructions may be organized as a presentation layer, application layer and data storage layer such as a relational database system using structured query language (SQL) or NoSQL, an object store, a graph database, a flat file system or other data storage.
Computer system 500 may be coupled via I/O subsystem 502 to at least one output device 512. In one embodiment, output device 512 is a digital computer display. Examples of a display that may be used in various embodiments include a touch screen display or a light-emitting diode (LED) display or a liquid crystal display (LCD) or an e-paper display. Computer system 500 may include other type(s) of output devices 512, alternatively or in addition to a display device. Examples of other output devices 512 include printers, ticket printers, plotters, projectors, sound cards or video cards, speakers, buzzers or piezoelectric devices or other audible devices, lamps or LED or LCD indicators, haptic devices, actuators or servos.
At least one input device 514 is coupled to I/O subsystem 502 for communicating signals, data, command selections or gestures to processor 504. Examples of input devices 514 include touch screens, microphones, still and video digital cameras, alphanumeric and other keys, keypads, keyboards, graphics tablets, image scanners, joysticks, clocks, switches, buttons, dials, slides, and/or various types of sensors such as force sensors, motion sensors, heat sensors, accelerometers, gyroscopes, and inertial measurement unit (IMU) sensors and/or various types of transceivers such as wireless, such as cellular or Wi-Fi, radio frequency (RF) or infrared (IR) transceivers and Global Positioning System (GPS) transceivers.
Another type of input device is a control device 516, which may perform cursor control or other automated control functions such as navigation in a graphical interface on a display screen, alternatively or in addition to input functions. Control device 516 may be a touchpad, a mouse, a trackball, or cursor direction keys for communicating direction information and command selections to processor 504 and for controlling cursor movement on display 512. The input device may have at least two degrees of freedom in two axes, a first axis (e.g., x) and a second axis (e.g., y), that allows the device to specify positions in a plane. Another type of input device is a wired, wireless, or optical control device such as a joystick, wand, console, steering wheel, pedal, gearshift mechanism or other type of control device. An input device 514 may include a combination of multiple different input devices, such as a video camera and a depth sensor.
In another embodiment, computer system 500 may comprise an internet of things (IoT) device in which one or more of the output device 512, input device 514, and control device 516 are omitted. Or, in such an embodiment, the input device 514 may comprise one or more cameras, motion detectors, thermometers, microphones, seismic detectors, other sensors or detectors, measurement devices or encoders and the output device 512 may comprise a special-purpose display such as a single-line LED or LCD display, one or more indicators, a display panel, a meter, a valve, a solenoid, an actuator or a servo.
When computer system 500 is a mobile computing device, input device 514 may comprise a global positioning system (GPS) receiver coupled to a GPS module that is capable of triangulating to a plurality of GPS satellites, determining and generating geo-location or position data such as latitude-longitude values for a geophysical location of the computer system 500. Output device 512 may include hardware, software, firmware and interfaces for generating position reporting packets, notifications, pulse or heartbeat signals, or other recurring data transmissions that specify a position of the computer system 500, alone or in combination with other application-specific data, directed toward host 524 or server 530.
Computer system 500 may implement the techniques described herein using customized hard-wired logic, at least one ASIC or FPGA, firmware and/or program instructions or logic which when loaded and used or executed in combination with the computer system causes or programs the computer system to operate as a special-purpose machine. According to one embodiment, the techniques herein are performed by computer system 500 in response to processor 504 executing at least one sequence of at least one instruction contained in main memory 506. Such instructions may be read into main memory 506 from another storage medium, such as storage 510. Execution of the sequences of instructions contained in main memory 506 causes processor 504 to perform the process steps described herein. In alternative embodiments, hard-wired circuitry may be used in place of or in combination with software instructions.
The term “storage media” as used herein refers to any non-transitory media that store data and/or instructions that cause a machine to operation in a specific fashion. Such storage media may comprise non-volatile media and/or volatile media. Non-volatile media includes, for example, optical or magnetic disks, such as storage 510. Volatile media includes dynamic memory, such as memory 506. Common forms of storage media include, for example, a hard disk, solid state drive, flash drive, magnetic data storage medium, any optical or physical data storage medium, memory chip, or the like.
Storage media is distinct from but may be used in conjunction with transmission media. Transmission media participates in transferring information between storage media. For example, transmission media includes coaxial cables, copper wire and fiber optics, including the wires that comprise a bus of I/O subsystem 502. Transmission media can also take the form of acoustic or light waves, such as those generated during radio-wave and infra-red data communications.
Various forms of media may be involved in carrying at least one sequence of at least one instruction to processor 504 for execution. For example, the instructions may initially be carried on a magnetic disk or solid-state drive of a remote computer. The remote computer can load the instructions into its dynamic memory and send the instructions over a communication link such as a fiber optic or coaxial cable or telephone line using a modem. A modem or router local to computer system 500 can receive the data on the communication link and convert the data to be read by computer system 500. For instance, a receiver such as a radio frequency antenna or an infrared detector can receive the data carried in a wireless or optical signal and appropriate circuitry can provide the data to I/O subsystem 502 such as place the data on a bus. I/O subsystem 502 carries the data to memory 506, from which processor 504 retrieves and executes the instructions. The instructions received by memory 506 may optionally be stored on storage 510 either before or after execution by processor 504.
Computer system 500 also includes a communication interface 518 coupled to bus 502. Communication interface 518 provides a two-way data communication coupling to network link(s) 520 that are directly or indirectly connected to at least one communication networks, such as a network 522 or a public or private cloud on the Internet. For example, communication interface 518 may be an Ethernet networking interface, integrated-services digital network (ISDN) card, cable modem, satellite modem, or a modem to provide a data communication connection to a corresponding type of communications line, for example an Ethernet cable or a metal cable of any kind or a fiber-optic line or a telephone line. Network 522 broadly represents a local area network (LAN), wide-area network (WAN), campus network, internetwork or any combination thereof. Communication interface 518 may comprise a LAN card to provide a data communication connection to a compatible LAN, or a cellular radiotelephone interface that is wired to send or receive cellular data according to cellular radiotelephone wireless networking standards, or a satellite radio interface that is wired to send or receive digital data according to satellite wireless networking standards. In any such implementation, communication interface 518 sends and receives electrical, electromagnetic or optical signals over signal paths that carry digital data streams representing various types of information.
Network link 520 typically provides electrical, electromagnetic, or optical data communication directly or through at least one network to other data devices, using, for example, satellite, cellular, Wi-Fi, or BLUETOOTH technology. For example, network link 520 may provide a connection through a network 522 to a host computer 524.
Furthermore, network link 520 may provide a connection through network 522 or to other computing devices via internetworking devices and/or computers that are operated by an Internet Service Provider (ISP) 526. ISP 526 provides data communication services through a world-wide packet data communication network represented as internet 528. A server computer 530 may be coupled to internet 528. Server 530 broadly represents any computer, data center, virtual machine or virtual computing instance with or without a hypervisor, or computer executing a containerized program system such as DOCKER or KUBERNETES. Server 530 may represent an electronic digital service that is implemented using more than one computer or instance and that is accessed and used by transmitting web services requests, uniform resource locator (URL) strings with parameters in HTTP payloads, application programming interface (API) calls, app services calls, or other service calls. Computer system 500 and server 530 may form elements of a distributed computing system that includes other computers, a processing cluster, server farm or other organization of computers that cooperate to perform tasks or execute applications or services. Server 530 may comprise one or more sets of instructions that are organized as modules, methods, objects, functions, routines, or calls. The instructions may be organized as one or more computer programs, operating system services, or application programs including mobile apps. The instructions may comprise an operating system and/or system software; one or more libraries to support multimedia, programming or other functions; data protocol instructions or stacks to implement TCP/IP, HTTP or other communication protocols; file format processing instructions to interpret or render files coded using HTML, XML, JPEG, MPEG or PNG; user interface instructions to render or interpret commands for a graphical user interface (GUI), command-line interface or text user interface; application software such as an office suite, internet access applications, design and manufacturing applications, graphics applications, audio applications, software engineering applications, educational applications, games or miscellaneous applications. Server 530 may comprise a web application server that hosts a presentation layer, application layer and data storage layer such as a relational database system using structured query language (SQL) or NoSQL, an object store, a graph database, a flat file system or other data storage.
Computer system 500 can send messages and receive data and instructions, including program code, through the network(s), network link 520 and communication interface 518. In the Internet example, a server 530 might transmit a requested code for an application program through Internet 528, ISP 526, local network 522 and communication interface 518. The received code may be executed by processor 504 as it is received, and/or stored in storage 510, or other non-volatile storage for later execution.
The execution of instructions as described in this section may implement a process in the form of an instance of a computer program that is being executed, and consisting of program code and its current activity. Depending on the operating system (OS), a process may be made up of multiple threads of execution that execute instructions concurrently. In this context, a computer program is a passive collection of instructions, while a process may be the actual execution of those instructions. Several processes may be associated with the same program; for example, opening up several instances of the same program often means more than one process is being executed. Multitasking may be implemented to allow multiple processes to share processor 504. While each processor 504 or core of the processor executes a single task at a time, computer system 500 may be programmed to implement multitasking to allow each processor to switch between tasks that are being executed without having to wait for each task to finish. In an embodiment, switches may be performed when tasks perform input/output operations, when a task indicates that it can be switched, or on hardware interrupts. Time-sharing may be implemented to allow fast response for interactive user applications by rapidly performing context switches to provide the appearance of concurrent execution of multiple processes simultaneously. In an embodiment, for security and reliability, an operating system may prevent direct communication between independent processes, providing strictly mediated and controlled inter-process communication functionality.
In the foregoing specification, embodiments of the disclosure have been described with reference to numerous specific details that may vary from implementation to implementation. The specification and drawings are, accordingly, to be regarded in an illustrative rather than a restrictive sense. The sole and exclusive indicator of the scope of the disclosure, and what is intended by the applicants to be the scope of the disclosure, is the literal and equivalent scope of the set of claims that issue from this application, in the specific form in which such claims issue, including any subsequent correction.
| Number | Date | Country | Kind |
|---|---|---|---|
| P202230148 | Feb 2022 | ES | national |
| Filing Document | Filing Date | Country | Kind |
|---|---|---|---|
| PCT/US2023/062682 | 2/15/2023 | WO |
| Number | Date | Country | |
|---|---|---|---|
| 63431590 | Dec 2022 | US |