This application claims priority to Chinese Patent Application No. 202110216633.6, filed on Feb. 26, 2021, which is hereby incorporated by reference in its entirety.
The present application relates to the field of smart devices, in particular, to an information prompt method, apparatus, medium, glasses, and program product.
With the development of competitive sports, when various sports competitions (such as the Olympics, Winter Olympics, and Basketball Association) are held, more and more sports fans will choose to go to the stadium to watch the games.
At present, when watching a game in the stadium, the spectator can only acquire the performance of athletes through the naked eyes, while can only acquire the specific performance parameters of the athletes through the live broadcast or large screen on-site.
Upon the traditional way of watching the game in the stadium, the types of information that the spectator can acquire are relatively simple, and the timeliness of acquiring specific data about the game is poor.
The embodiment of the present application provides an information prompt method, apparatus, medium, glasses, and program product to solve the technical problem that upon the traditional way of watching the game in the stadium, the types of information that the spectator can acquire are relatively simple, and the timeliness is poor.
In a first aspect, an embodiment of the present application provides an information prompt method, which is applied to a pair of smart glasses, and the method includes:
acquiring a scene video of an external scene, where the scene video includes a behavior feature of a target object;
matching prompt information corresponding to the target object according to the behavior feature, where the prompt information is associated with the behavior feature; and
displaying the prompt information on lenses of the pair of smart glasses.
In a possible design, the matching prompt information corresponding to the target object according to the behavior feature includes:
acquiring face information of candidate objects in the scene video, and determining the target object from the candidate objects according to preset face information; and
determining, according to an action feature of the target object in the scene video, technical parameters of sports, where the behavior feature includes the action feature, and the prompt information includes the technical parameters of sports.
In a possible design, the determining, according to an action feature of the target object in the scene video, technical parameters of sports includes:
determining, according to the action feature of the target object, a behavior type of the target object; and
determining, according to both state change information of a target device in the scene video and the behavior type, the technical parameters of sports.
In a possible design, after the determining the target object from the candidate objects according to preset face information, the method further includes:
sending object information of the target object to other devices, where the other devices include a server and/or another pair of smart glasses;
receiving another scene video and another piece of prompt information sent by the other devices, where the another scene video differs from the scene video in at least a shooting angle of the external scene, and the another piece of prompt information is information associated with the behavior feature of the target object in the another scene video; and
displaying the another scene video and/or the another piece of prompt information on lenses of the pair of smart glasses.
In a possible design, the information prompt method further includes:
selecting, in response to a selection instruction, the target object from the candidate objects.
In a possible design, the matching prompt information corresponding to the target object according to the behavior feature includes:
determining, according to a sound feature in audio information, the prompt information of target audio, where the target object includes the target audio, the scene video includes the audio information, and the behavior feature includes the sound feature.
In a second aspect, an embodiment of the present application provides an information prompt apparatus, including:
a video acquisition module, configured to acquire a scene video of an external scene, where the scene video includes a behavior feature of a target object;
an information processing module, configured to match prompt information corresponding to the target object according to the behavior feature, where the prompt information is associated with the behavior feature; and an information display module, configured to display the prompt information.
In a possible design, the information prompt apparatus further includes:
a face acquisition module, configured to acquire face information of candidate objects in the scene video, and determine the target object from the candidate objects according to preset face information;
the information processing module is further configured to determine, according to an action feature of the target object in the scene video, technical parameters of sports, where the behavior feature includes the action feature, and the prompt information includes the technical parameters of sports.
In a possible design, the information processing module is specifically configured to:
determine, according to the action feature of the target object, a behavior type of the target object; and
determine, according to both state change information of a target device in the scene video and the behavior type, the technical parameters of sports.
In a possible design, the information prompt apparatus further includes:
an information sending module, configured to send object information of the target object to other devices, where the other devices include a server and/or another pair of smart glasses; and
an information receiving module, configured to receive another scene video and another piece of prompt information sent by the other devices, where the another scene video differs from the scene video in at least a shooting angle of the external scene, and the another piece of prompt information is information associated with the behavior feature of the target object in the another scene video;
the information display module is further configured to display both the another scene video and the another piece of prompt information.
In a possible design, the information prompt apparatus further includes:
an object selection module, configured to select, in response to a selection instruction, the target object from the candidate objects.
In a possible design, the information processing module is specifically configured to:
determine, according to a sound feature in audio information, the prompt information of target audio, where the target object includes the target audio, the scene video includes the audio information, and the behavior feature includes the sound feature.
In a third aspect, an embodiment of the present application further provides a pair of smart glasses, including: a processor, a memory, lenses, and an image sensor, where the processor is connected to the memory, the lenses, and the image sensor, respectively;
the image sensor is configured to acquire a scene video of an external scene;
the memory is configured to store a computer program of the processor;
where the processor is configured to implement any one of the information prompt methods in the first aspect by executing the computer program;
the lenses are configured to process generated prompt information through the information prompt method.
In a fourth aspect, an embodiment of the present application also provides a storage medium on which a computer program is stored, where, when the computer program is executed by a processor, any one of the information prompt methods in the first aspect is implemented.
In a fifth aspect, an embodiment of the present application also provides a computer program product, including a computer program, where, when the computer program is executed by a processor, any one of the information prompt methods in the first aspect is implemented.
The embodiments of the present application provide an information prompt method, apparatus, medium, glasses, and program product, which acquire the scene video of the external scene in real time by wearing a pair of smart glasses, and then automatically match, combined with the behavior feature of the target objects, the corresponding related information, and provide real-time feedback to the spectator by direct displaying on the pair of smart glasses, thereby enriching the types of information that the spectator can acquire on-site, and at the same time greatly improving the timeliness of acquiring the related information by the spectator.
In order to more clearly illustrate embodiments of the present application or the technical solution in the prior art, accompanying drawings required for describing the embodiments or the prior art will be briefly described below. Obviously, the accompanying drawings in the following description are some of the embodiments of the present application, and other drawings can be acquired by those skilled in the art based on these accompanying drawings without any creative effort.
In order to make the purpose, technical solution, and advantages of the embodiments of the present application clearer, the technical solution in the embodiments of the present application will be described clearly and completely together with the accompanying drawings in the embodiments of the present application. Obviously, the described embodiments are part of the embodiments of the present application, and not all of them. Based on the embodiments of the present application, all other embodiments acquired by those of ordinary skill in the art without creative labor shall fall within the scope of protection of the present application.
The terms “first”, “second”, “third”, “fourth”, etc. (if any) in the specification, in claims of the present application and in the above accompanying drawings are used to distinguish similar objects, and are not necessarily used to describe a particular order or sequence. It should be understood that the data thus used may be interchanged in some cases, so that the embodiments of the present application described herein can be implemented in an order other than those illustrated or described herein. Furthermore, the terms “include” and “have” and any variations thereof are intended to cover non-exclusive inclusion, for example, a process, method, system, product or device that include a series of steps or units are not necessarily limited to those explicitly listed, but may include other steps or units that are not explicitly listed or inherent to the process, method, product or device.
At present, when watching a game in the stadium, the spectator can only acquire the performance of athletes through the naked eyes, while can only acquire the specific performance parameters of the athletes through the live broadcast or large screen on-site. For example, when watching a football game in the stadium, the spectator can only acquire the scores of both sides and the duration of the current game through the large screen on-site, and cannot acquire the specific performance data of each player in time, which can only be acquired after the game. In addition, when watching a track and field event (for example: sprinting) in the stadium, the spectator can only know the sequence of current athletes on the runway, but cannot acquire the current running parameters of the athletes. Similarly, they can only get the sports performance parameters of each athlete through the large screen after the game.
The purpose of the present application is to provide an information prompt method, so that the user can acquire a scene video of the external scene in real time by wearing a pair of smart glasses when watching a game in the stadium; and then corresponding related information can be automatically matched based on behavior features of various target objects; the corresponding related information is directly displayed on the pair of smart glasses to feed back to the user in real time, further enriching the types of information that the spectator can acquire on-site, and at the same time greatly improving the timeliness of acquiring the related information by the spectator. In addition, by directly displaying the prompt information on lenses of the pair of smart glasses, the connection between the target object and the prompt information can be effectively established in the spectator's field of vision, thereby enhancing the pertinence of the prompt information display.
In addition, the information prompt method provided in this embodiment can also be used when watching a concert. When watching a concert, the spectator can acquire the live audio in real time, and then the information corresponding to the audio is matched through the audio technology, for example, the song's title, lyrics, composition, etc. The acquired audio-related information is displayed on the lenses of the pair of smart glasses as prompt information, so that the spectator can acquire the current song information in real time while watching the concert, thereby greatly improving the watching experience of the concert.
Step 101, Acquire a scene video of an external scene, where the scene video includes a behavior feature of a target object.
In this step, when watching a game or a performance, the spectator can wear a pair of smart glasses. Thereby, the scene video of the current external scene is acquired through an image sensor on the pair of smart glasses, and the scene video includes the behavior feature of the target object. Worthy of explanation, the target object in this step can be a person object, a device object, and other related objects that are required to be monitored, and the behavioral feature can be the action feature, voice feature, and other related features that are required to be monitored.
Step 102, Match prompt information corresponding to the target object according to the behavior feature.
After the pair of smart glasses acquire the scene video, the behavior feature in the scene video can be extracted by calculation upon a processing function of the pair of smart glasses itself or by calculation upon uploading to the relevant server, and the prompt information, which is associated with the behavior feature, corresponding to the target object can be matched according to the extracted behavior feature.
For example, speed information can be matched with a running action of the runner as the current prompt information, the current score, assist, foul and other pieces of information can be matched with players' shooting, dribbling or passing actions as the current prompt information, or the song's title, lyrics, composition and other pieces of information can be matched with the currently playing song as the current prompt information.
Step 103, Display the prompt information on lenses of the pair of smart glasses.
After matching the prompt information according to the behavior feature of the target object in the scene video, the generated prompt information can be displayed on the lenses of the pair of smart glasses.
In this embodiment, the scene video of the external scene is acquired in real time by wearing a pair of smart glasses, and then corresponding related information is automatically matched in combination with the behavior feature of the target objects, and real-time feedback is provided to the spectator by direct displaying on the pair of smart glasses, thereby enriching the types of information that the spectator can acquire on-site, and at the same time greatly improving the timeliness of acquiring the related information by the spectator.
Step 201, Acquire a scene video of an external scene, where the scene video includes a behavior feature of a target object.
In this step, when watching a game or a performance, the spectator can wear a pair of smart glasses. Thereby, the scene video of the current external scene is acquired through an image sensor on the pair of smart glasses, and the scene video includes the behavior feature of the target object. Worthy of explanation, the target object in this step can be a person object, a device object, and other related objects that are required to be monitored, and the behavioral feature can be the action feature, voice feature, and other related features that are required to be monitored.
Step 202, Acquire face information of candidate objects in the scene video, and determine the target object from the candidate objects according to preset face information.
After acquiring the scene video through the pair of smart glasses, since there may be multiple elements in the scene video, for example, there are multiple athletes in a game, and the spectator does not pay more attention to all athletes. If the information prompt for all athletes are required to be realized in the process of watching the game, the displayed information will be too complicated, which actually causes information interference to the spectator.
Therefore, after acquiring the scene video, face recognition may be first performed on each candidate object therein, and then the target object can be determined based on the screening condition, which can be preset by the user or can be input by the user in the process of watching the game, for example, it can be input by voice or through a third-party device.
All the screening conditions that are preset or set by input can be used as preset face information required in the following process of object determination. In the specific determination process, the face information of each candidate object in the scene video can be first acquired, for example, face information of four athletes A, B, C, and D in the scene video can be acquired. At this point, since the preset person information is the face information of the athlete B, the athlete B can be selected from the four athletes as the target object, so as to perform the following monitoring and corresponding information prompt for the determined target object.
Step 203, Determine, according to an action feature of the target object in the scene video, technical parameters of sports.
In this step, the technical parameters of sports can be determined according to the action feature of the target object in the scene video, the behavioral feature includes the action feature, and the prompt information to be displayed later includes the technical parameters of sports.
Specifically, in a sprint event, the starting speed can be determined according to the starting action of the target athlete in the scene video, and the current speed can be determined according to the athlete's running action. For example, the image position of the target athlete on the runway at each moment can be acquired, the distance moved by the target athlete in each time interval can be then calculated based on the image ranging technology, and the running speed of the athlete at each moment can be calculated finally.
In another possible case, which can be a basketball game, the current technical parameters of sports (for example, scores, assists, rebounds, fouls, etc.) of the target player can be calculated according to various action features of the target player in the scene video. For such competitions with many types of actions, the behavior type of the target object can be determined first according to the behavior feature of the target object, and then the technical parameters of sports can be determined according to both state change information of the target device in the scene video and the behavior type. For example, after acquiring the shooting action of the target player, it can be determined that the current behavior type of the target player is shooting; the flight track of basketball within a preset time period is monitored then, and it can be determined whether the score is finally obtained based on the change in the positional relationship between the basketball and the basket; finally, scores from the beginning to the current moment is counted by continuously monitoring the motion of the target player, so that it can be displayed as prompt information later.
Step 204, Display the prompt information on lenses of the pair of smart glasses.
After matching the prompt information according to the behavior feature of the target object in the scene video, the generated prompt information can be displayed on the lenses of the pair of smart glasses.
In the process of watching the game, the spectator may pay attention to multiple target objects. Therefore, in the process of watching the game, the spectator can select the target object from all the candidate objects in the scene video by inputting a selection instruction, or can switch the target object. The selection instruction can be manually input by the spectator, or it is selected or switched by capturing the eyeball position of the spectator, so as to meet the requirement of the spectator to select or switch the target object in the process of watching the game.
Step 301, Acquire a scene video of an external scene, where the scene video includes a behavior feature of a target object.
In this step, when watching a game or a performance, the spectator can wear a pair of smart glasses. Thereby, the scene video of the current external scene is acquired through an image sensor on the pair of smart glasses, and the scene video includes the behavior feature of the target object. Worthy of explanation, the target object in this step can be a person object, a device object, and other related objects that are required to be monitored, and the behavioral feature can be the action feature, voice feature, and other related features that are required to be monitored.
Step 302, Acquire face information of candidate objects in the scene video, and determine the target object from the candidate objects according to preset face information.
After acquiring the scene video through the pair of smart glasses, since there may be multiple elements in the scene video, for example, there are multiple athletes in a game, and the spectator does not pay more attention to all athletes. If the information prompt for all athletes are required to be realized in the process of watching the game, the displayed information will be too complicated, which actually causes information interference to the spectator.
Therefore, after acquiring the scene video, face recognition may be first performed on each candidate object therein, and then the target object can be determined based on the screening condition, which can be preset by the user or can be input by the user in the process of watching the game, for example, it can be input by voice or through a third-party device.
All the screening conditions that are preset or set by input can be used as preset face information required in the following process of object determination. In the specific determination process, the face information of each candidate object in the scene video can be first acquired, for example, face information of four athletes A, B, C, and D in the scene video can be acquired. At this point, since the preset person information is the face information of the athlete B, the athlete B can be selected from the four athletes as the target object, so as to perform the following monitoring and corresponding information prompt for the determined target object.
Step 303, Send object information of the target object to other devices.
Step 304, Receive another scene video and another piece of prompt information sent by the other devices.
After determining the target object, in order to acquire more information about the target object, the object information of the target object can be sent to other devices first, for example, the other devices can be a related server or another pair of smart glasses.
Step 305, Display the another scene video and/or the another piece of prompt information on the lenses of the pair of smart glasses.
After acquiring the another scene video and the another piece of prompt information sent by the other devices, the another scene video and/or the another piece of prompt information can be displayed on the lenses of the pair of smart glasses.
In this embodiment, information sharing is achieved by the spectator through the communication between the pair of smart glasses worn by the spectator and other devices, so as to further enrich the prompt information of the target object, which greatly improves the watching experience of the spectator.
a video acquisition module 401, configured to acquire a scene video of an external scene, where the scene video includes a behavior feature of a target object;
an information processing module 402, configured to match prompt information corresponding to the target object according to the behavior feature, where the prompt information is associated with the behavior feature; and
an information display module 403, configured to display the prompt information.
Based on the embodiment shown in
a face acquisition module 404, configured to acquire face information of candidate objects in the scene video, and determine the target object from the candidate objects according to preset face information;
the information processing module 402 is further configured to determine, according to an action feature of the target object in the scene video, technical parameters of sports, where the behavior feature includes the action feature, and the prompt information includes the technical parameters of sports.
In a possible design, the information processing module 402 is specifically configured to:
determine, according to the action feature of the target object, a behavior type of the target object; and
determine, according to both state change information of a target device in the scene video and the behavior type, the technical parameters of sports.
In a possible design, the information prompt apparatus 400 further includes:
an information sending module 405, configured to send object information of the target object to other devices, where the other devices include a server and/or another pair of smart glasses; and
an information receiving module 406, configured to receive another scene video and another piece of prompt information sent by the other devices, where the another scene video differs from the scene video in at least a shooting angle of the external scene, and the another piece of prompt information is information associated with the behavior feature of the target object in the another scene video;
the information display module 403 is further configured to display both the another scene video and the another piece of prompt information.
In a possible design, the information prompt apparatus 400 further includes:
an object selection module 407, configured to select, in response to a selection instruction, the target object from the candidate objects.
In a possible design, the information processing module 402 is specifically configured to:
determine, according to a sound feature in audio information, the prompt information of target audio, where the target object includes the target audio, the scene video includes the audio information, and the behavior feature includes the sound feature.
The processing component 1202 generally controls the overall operation of the pair of smart glasses 1200, such as the operation associated with display, data communication, multimedia operation and recording operation. The processing component 1202 may include one or more processors 1220 to execute the instruction, so as to implement all or part of the steps of the foregoing method. In addition, the processing component 1202 may include one or more modules to facilitate interaction between the processing component 1202 and other components. For example, the processing component 1202 may include a multimedia module to facilitate the interaction between the multimedia component 1208 and the processing component 1202.
The memory 1204 is configured to store various types of data to support the operation of the pair of smart glasses 1200. Examples of such data include instructions for any application or method operated on the pair of smart glasses 1200, various kinds of data, messages, pictures, videos, etc. The memory 1204 may be implemented by any type of volatile or nonvolatile storage device or a combination thereof, such as static random access memory (SRAM), electrically erasable programmable read-only memory (EEPROM), erasable programmable read-only memory (EPROM), programmable read-only memory (PROM), read-only memory (ROM), magnetic memory, flash memory, magnetic disk or optical disk.
The power supply component 1206 provides power for various components of the pair of smart glasses 1200. The power supply component 1206 may include a power management system, one or more power supplies, and other components, which are configured to generate, manage, and distribute power for the pair of smart glasses 1200.
The multimedia component 1208 includes lenses with a display function provided on the pair of smart glasses 1200.
The audio component 1210 is configured to output and/or input audio signals. For example, the audio component 1210 includes a microphone (MIC) configured to receive external audio signals when the pair of smart glasses 1200 are in an operation mode, such as the call mode, recording mode and voice recognition mode. The received audio signal may be further stored in the memory 1204 or transmitted via the communication component 1216. In some embodiments, the audio component 1210 further includes a speaker for outputting audio signals.
The I/O interface 1212 provides an interface between the processing component 1202 and the peripheral interface module, which may be the keyboard, click wheel, button, etc.
The sensor component 1214 includes one or more sensors configured to provide the pair of smart glasses 1200 with various aspects of state evaluation. For example, the sensor component 1214 may include an image sensor, and the image sensor is configured to acquire the environmental image or the video of the external environment of the pair of smart glasses 1200. In addition, the sensor component 1214 can also detect the ON/OFF state of the pair of smart glasses 1200 and the relative positioning of the component, for example, the components are lenses of the pair of smart glasses 1200. The sensor 1214 can also detect the position change of the pair of smart glasses 1200 or a component of the pair of smart glasses 1200, whether the user is in contact with the pair of smart glasses 1200, the orientation or acceleration/deceleration of the pair of smart glasses 1200, and the temperature change of the pair of smart glasses 1200. The sensor component 1214 may include a proximity sensor configured to detect whether there is an object nearby when no physical contact is available. The sensor component 1214 may also include a light sensor, such as a complementary metal-oxide-semiconductor transistor (CMOS) or charge coupled device (CCD) image sensor, used in imaging applications. In some embodiments, the sensor component 1214 may also include an acceleration sensor, a gyroscope sensor, a magnetic sensor, a pressure sensor, or a temperature sensor.
The communication component 1216 is configured to facilitate wired or wireless communication between the pair of smart glasses 1200 and other devices. The pair of smart glasses 1200 can access a wireless network based on a communication standard, such as wireless fidelity (Wi-Fi), 2nd generation telecommunication (2G), or 3rd generation telecommunication (3G), or a combination thereof. In an exemplary embodiment, the communication component 1216 receives broadcast signals or broadcast-related information from an external broadcast management system via a broadcast channel. In an exemplary embodiment, the communication component 1216 also includes a near field communication (NFC) module to facilitate short-range communication. For example, in the NFC module, implementation may be based on the radio frequency identification (RFID) technology, infrared data association (IrDA) technology, ultra-wideband (UWB) technology, Bluetooth (BT) technology and other technologies.
In an exemplary embodiment, the pair of smart glasses 1200 can be implemented by one or more application-specific integrated circuits (ASIC), digital signal processors (DSP), digital signal processing devices (DSPD), programmable logic devices (PLD), field programmable gate arrays (FPGA), controllers, microcontrollers, microprocessors, or other electronic components, so as to perform the information prompt method, including:
acquiring a scene video of an external scene, where the scene video includes a behavior feature of a target object;
matching prompt information corresponding to the target object according to the behavior feature, where the prompt information is associated with the behavior feature; and
displaying the prompt information on lenses of the pair of smart glasses.
In an exemplary embodiment, there is also provided a non-transitory computer-readable storage medium including the instruction, such as the memory 1204 including the instruction, and the instruction can be executed by the processor 1220 of the pair of smart glasses 1200 to implement the foregoing method. For example, the non-transitory computer-readable storage medium may be the ROM, random access memory (RAM), compact disc read-only memory (CD-ROM), magnetic tape, floppy disk, optical data storage device, etc.
This embodiment also provides a program product, which includes a computer program stored in a readable storage medium. At least one processor of an electronic device can read the computer program from a readable storage medium, and execute the computer program to enable the electronic device to perform steps of the foregoing methods.
Those skilled in the art can understand that all or part of the steps of the foregoing method embodiments can be implemented by hardware related to program instructions. The program can be stored in a computer readable storage medium. When the program is executed, the steps of the foregoing method embodiments are performed. The aforementioned storage media include various media that can store program codes: ROM, RAM, magnetic disks, or optical disks.
Finally, it should be noted that: the above embodiments are only used to illustrate the technical solutions of the present application, not to constitute a limitation on it. Although the present application has been described in detail with reference to the foregoing embodiments, those skilled in the art should understand that it is still possible to modify the technical solutions recorded in the above embodiments, or equivalently replace some or all of the technical features; and these modifications or replacements do not deviate the essence of the corresponding technical solutions from the range of technical solutions of the embodiments of the present application.
Number | Date | Country | Kind |
---|---|---|---|
202110216633.6 | Feb 2021 | CN | national |