This application claims the priority benefit of Taiwanese application no. 110123328, filed on Jun. 25, 2021. The entirety of the above-mentioned patent application is hereby incorporated by reference herein and made a part of this specification.
The disclosure relates to a video system. Particularly, the disclosure relates to an augmented reality (AR) system and an operation method thereof.
Various audio-visual streaming services are becoming more and more popular. Common audio-visual streaming services include video conferencing. In a video conference, a user A may show something to a remote user B through a communication network. For example, a mobile phone held by the user A is displaying an interesting digital content (an image or a three-dimensional digital object), and the user A may intend to show the digital content to the remote user B through the video conference. Therefore, the user A sets a video conferencing device to capture an image of the mobile phone. Subject to various environmental factors (e.g., resolution, color shift, etc.), the user B may not be able to clearly see the content displayed by the mobile phone of the user A.
The disclosure provides an augmented reality (AR) system and an operation method thereof for performing an augmented reality application.
In an embodiment of the disclosure, the augmented reality system includes a target device and an augmented reality device. The target device is configured to sense an attitude of the target device to generate attitude information. The target device provides a digital content and the attitude information to the augmented reality device. The augmented reality device is configured to capture the target device to generate an image. The augmented reality device tracks a target location of the target device in the image to perform an augmented reality application. In the augmented reality application, the augmented reality device overlays the digital content on the target location in the image. The AR device correspondingly adjusts an attitude of the digital content in the image based on the attitude information of the target device.
In an embodiment of the disclosure, the operation method includes the following. An attitude of a target device is sensed by the target device to generate attitude information. A digital content and the attitude information is provided by the target device to an augmented reality device. The target device is captured by the augmented reality device to generate an image. A target location of the target device in the image is tracked by the augmented reality device to perform an augmented reality application. The digital content is overlaid on the target location in the image by the augmented reality device in the augmented reality application. An attitude of the digital content in the image is correspondingly adjusted by the augmented reality device based on the attitude information of the target device.
Based on the foregoing, in the embodiments of the disclosure, the augmented reality device may capture the target device to generate an image to perform an augmented reality application. The target device may sense the attitude of the target device, and provide the attitude information and the digital content to the augmented reality device. During the process of performing the augmented reality application, the augmented reality device may overlay the digital content provided by the target device on the target location of the target device in the image, and correspondingly adjust the attitude of the digital content in the image based on the attitude information of the target device. Since the digital content is not fixedly stored in the augmented reality device, the augmented reality device may present augmented reality effects more flexibly.
To make the aforementioned more comprehensible, several embodiments accompanied with drawings are described in detail as follows.
The accompanying drawings are included to provide a further understanding of the disclosure, and are incorporated in and constitute a part of this specification. The drawings illustrate exemplary embodiments of the disclosure and, together with the description, serve to explain the principles of the disclosure.
The term “couple (or connect)” used throughout the whole description of the disclosure (including the claims) may refer to any direct or indirect connection means. For example, if the disclosure describes that a first device is coupled (or connected) to a second device, it should be interpreted that the first device may be directly connected to the second device, or that the first device may be indirectly connected to the second device through other devices or certain connection means. Terms such as “first” and “second” mentioned throughout the whole description of the disclosure (including the claims) are used for naming elements or differentiating between different embodiments or ranges, instead of limiting an upper bound or lower bound of the number of elements or the sequence of elements. Moreover, wherever possible, elements/members/steps using the same reference numerals in the drawings and the embodiments represent the same or similar parts. Cross-reference may be made to related descriptions of elements/members/steps using the same reference numerals or using the same terms in different embodiments.
In step S230, the AR device 120 may capture the target device 110 to generate an image (or an image stream). In step S240, the AR device 120 may track a target location of the target device 110 in the image to perform an AR application. Depending on the actual design, the AR application may include a game application, an educational application, a video conferencing application, and/or other applications. During the process of performing the AR application, the AR device 120 may overlay the digital content DC provided by the target device 110 on the target location in the image (step S250). Therefore, in some application scenarios, the digital content DC may replace the target device 110 in the image. In step S260, the AR device 120 may correspondingly adjust an attitude of the digital content DC in the image based on the attitude information A_inf of the target device 110. For example, the digital content DC may include a three-dimensional digital object (e.g., a car, an animal, or other three-dimensional objects), and the AR device 120 may correspondingly adjust an attitude of the three-dimensional digital object in the image based on the attitude information A_inf.
As the target device 110 moves, the location of the digital content DC in the image changes accordingly. The user may rotate the target device 110. As the attitude of the target device 110 changes, the attitude of the digital content DC in the image changes accordingly. Depending on the actual application scenario, the target location in the image may be the location of the target device 110, or the target location in the image may be different from the location of the target device 110. For example, when the target device 110 is in the image, the target location may be the location of the target device 110 in the image. When the target device 110 is removed from the image, the AR device 120 takes an effective location of the target device 110 before the target device 110 is removed from the image as the target location. Depending on the actual design, under the premise that the digital content DC may be fully presented, the effective location may be the final location of the target device 110 before being removed from the image.
In some other embodiments, the target device 110 may display a marker MRK (not shown in
In still some other embodiments, the target device 110 and/or the AR device 120 may have a human-machine interface for the user to operate. When the user triggers the human-machine interface, the AR device 120 takes a current location of the target device 110 constantly as the target location. That is, after the human-machine interface is triggered, the movement of the target device 110 does not affect the target location (the location of the digital content DC).
In the video conference shown in
Therefore, in the video conference (AR application), the target device 110 may provide the digital content DC and the attitude information A_inf that are being displayed to the AR device 120. The AR device 120 may capture the target device 110 and the user A to generate an image (herein referred to as a conference image). The AR device 120 may overlay the digital content DC on the target device 110 in the conference image to generate an AR conference image. In addition, the AR device 120 may correspondingly adjust the attitude of the digital content DC in the AR conference image (e.g., rotate the direction of the digital content DC) based on the attitude information A_inf of the target device 110. The AR device 120 may transmit the AR conference image to the remote device 300 through the communication network to perform the video conference. The remote device 300 may display the AR conference image to the user B. The digital content that is being displayed by the target device 110 to the user B is not the captured result of the AR device 120. Accordingly, the digital content does not have problems such as resolution, color shift, etc. In addition, the user A may rotate the direction of the target device 110 to present the user B with the digital content DC in different view angles.
For example, based on the actual design, the digital content provided by the target device 110 to the AR device 120 may include a three-dimensional digital object, and the target device 110 has at least one attitude sensor (not shown in
Depending on different design requirements, the application processor 111 and/or the image processor 121 may be realized by hardware, firmware, software (i.e., programs), or a combination of more than one of the above three forms. In terms of the hardware form, the application processor 111 and/or the image processor 121 may be implemented in a logic circuit on an integrated circuit. The relevant functions of the application processor 111 and/or the image processor 121 may be implemented as hardware by utilizing hardware description languages (e.g., Verilog HDL or VHDL) or other suitable programming languages. For example, the relevant functions of the application processor 111 and/or the image processor 121 may be implemented in various logic blocks, modules, and circuits in one or more controllers, microcontrollers, microprocessors, application-specific integrated circuits (ASIC), digital signal processors (DSP), field programmable gate arrays (FPGA), and/or other processing units.
In terms of the software form and/or firmware form, the relevant functions of the application processor 111 and/or the image processor 121 may be implemented as programming codes. For example, the application processor 111 and/or the image processor 121 may be implement by utilizing general programming languages (e.g., C, C++, or assembly language) or other suitable programming languages. The programming codes may be recorded/stored in a “non-transitory computer readable medium”. In some embodiments, the non-transitory computer readable medium includes, for example, read only memory (ROM), a tape, a disk, a card, semiconductor memory, a programmable logic circuit, and/or a storage device. The storage device includes a hard disk drive (HDD), a solid-state drive (SSD), or other storage devices. A computer, a central processing unit (CPU), a controller, a microcontroller, or a microprocessor may read and execute the programming codes from the non-transitory computer readable medium, thereby realizing the relevant functions of the application processor 111 and/or the image processor 121. Moreover, the programming codes may also be provided to the computer (or CPU) through any transmission medium (a communication network, a radio wave, or the like). The communication network is, for example, the Internet, a wired communication network, a wireless communication network, or other communication media.
In summary of the foregoing, in the foregoing embodiments, the AR device 120 may 5/4 capture the target device 110 to generate an image to perform an AR application. The target device 110 may sense the attitude of itself, and provide the attitude information A_inf and the digital content DC to the AR device 120. During the process of performing the AR application, the AR device 120 may overlay the digital content DC provided by the target device 110 on the target location of the target device 110 in the image IMG, and correspondingly adjust the attitude of the digital content DC in the image IMG′ (e.g., rotate the direction of the digital content DC) based on the attitude information A_inf of the target device 110. Since the digital content DC is not fixedly stored in the AR device 120, the AR device 120 may present AR effects more flexibly.
It will be apparent to those skilled in the art that various modifications and variations can be made to the disclosed embodiments without departing from the scope or spirit of the disclosure. In view of the foregoing, it is intended that the disclosure covers modifications and variations provided that they fall within the scope of the following claims and their equivalents.
Number | Date | Country | Kind |
---|---|---|---|
110123328 | Jun 2021 | TW | national |