Virtual reality head-mounted apparatus

Information

  • Patent Grant
  • 11523039
  • Patent Number
    11,523,039
  • Date Filed
    Wednesday, July 31, 2019
    5 years ago
  • Date Issued
    Tuesday, December 6, 2022
    2 years ago
Abstract
Methods, systems, and apparatus, including computer programs encoded on computer storage media, for a virtual reality head-mounted apparatus are provided. One of the apparatus includes: an apparatus body being provided with a convex lens and a camera, the camera being located on a user side of the convex lens, and a lens of the camera facing an eye of the user to acquire eye pattern features of the user. The virtual reality head-mounted apparatus may acquire eye pattern features of the user without interrupting the displaying of virtual reality contents, and quickly and accurately perform identity verification on the user based on the eye pattern features.
Description
TECHNICAL FIELD

This application relates generally to virtual reality (VR) technology, and more specifically, to a virtual reality head-mounted apparatus.


BACKGROUND

Virtual reality (VR) technology provides three-dimensional immersive scenes for a user by utilizing a computer graphics system and various control interfaces to generate an interactive environment on a computer. Existing techniques may create VR experience through a virtual reality head-mounted apparatus, such as VR glasses or a VR helmet.


However, due to unique characteristics of VR technology, techniques that work on traditional electronic devices, such as mobile phones or PCs, may not work well on a VR device. For example, to conduct identity verification for a user, conventional methods that based on password or gesture recognition may take too long to complete, if it can be completed at all, in a VR environment. Therefore, a VR device that can quickly and accurately perform identify verification without interrupting the user's immersive VR experience is desirable.


SUMMARY

In view of the limitations of exiting techniques described above, this application provides a virtual reality head-mounted apparatus. This apparatus may acquire eye pattern features of a user without interrupting virtual reality content, and may quickly and accurately perform identity verification on the user wearing the virtual reality head-mounted apparatus based on the acquired eye pattern features.


This application is directed to a virtual reality head-mounted apparatus. The apparatus may comprise: an apparatus body comprising a convex lens and a camera. The camera may be on a user side of the convex lens, and a lens of the camera may face an eye of a user for acquiring eye pattern features of the user.


In some embodiments, the camera may be located below the convex lens, and tilt upward by a first angle, so that the lens of the camera may face the eye of the user. In some embodiments, the camera may be located above the convex lens, and tilt downward by a second angle, so that the lens of the camera may face the eye of the user.


In some embodiments, the apparatus body may comprise two convex lenses, and the camera may be located on the user side of one of the two convex lenses for acquiring eye pattern features of the user corresponding to the one of the two convex lenses.


In some embodiments, the apparatus body may comprise two cameras and two convex lenses, the two cameras may be respectively located at the user side of the two convex lenses for acquiring eye pattern features of the user corresponding to each of the two convex lenses.


In some embodiments, the camera may be located outside a visible area of the convex lens with respect to the user.


In some embodiments, a position of the camera may be in contact with the convex lens.


In some embodiments, the camera may be an RGB camera or an RGB-IR camera.


In some embodiments, the apparatus body may further comprise: an apparatus interface; and an electronic device for displaying virtual reality convent. The apparatus interface may be electrically connected to the electronic device, and the camera may be connected to the apparatus interface through a data line. Upon receiving a switch control instruction from the electronic device through the apparatus interface and the data line, the camera may perform a status switching operation in response to the switch control instruction, and transmit the acquired eye pattern features to the electronic device through the apparatus interface and the data line.


In some embodiments, the apparatus may further comprise an adjustment component for performing angular adjustment for the camera, such that the lens of the camera may face the eye of the user.


In the virtual reality head-mounted apparatus disclosed in this application, the apparatus body may comprise a camera, with a lens of the camera facing an eye of a user for acquiring eye pattern features of the user. The acquired eye pattern features may be used for identify verification of the user. Thus, without interrupting the user's viewing of virtual reality contents, identity information of the user may be quickly and accurately verified by comparing and analyzing the eye pattern features. This apparatus provides improved user experience and security for virtual reality applications.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 is a side cross-sectional view of a VR helmet provided by a first embodiment of this application.



FIG. 2 is a side cross-sectional view of a VR helmet provided by a second embodiment of this application.



FIG. 3 is a structural diagram of a VR helmet provided by an embodiment of this application, viewed from an wearer's direction.



FIG. 4 is a side cross-sectional view of a VR helmet provided by a third embodiment of this application.



FIG. 5 is a side cross-sectional view of a VR helmet provided by a fourth embodiment of this application.





DETAILED DESCRIPTION OF THE INVENTION

Specific, non-limiting embodiments of the present invention will be described with reference to the drawings. It should be understood that particular features and aspects of any embodiment disclosed herein may be used and/or combined with particular features and aspects of any other embodiment disclosed herein. It should also be understood that such embodiments are by way of example and are merely illustrative of a small number of embodiments within the scope of the present invention. Various changes and modifications obvious to one skilled in the art to which the present invention pertains are deemed to be within the spirit, scope and contemplation of the present invention as further defined in the appended claims.


In order to further illustrate this application, several embodiments are provided, using a VR helmet as an example, to introduce the structures of a virtual reality head-mounted apparatus of this application.



FIG. 1 is a side cross-sectional view of a VR helmet provided by an embodiment of this application. As shown in FIG. 1, the virtual reality head-mounted apparatus may comprise an apparatus body 1. The apparatus body 1 may comprise a convex lens 2 and a VR playing component 3. The convex lens 2 may be located between a user (an eye 4 of the user is shown in FIG. 1) and the VR playing component 3 in the apparatus body 1, so that VR display content played by the VR playing component 3 may, in a form of visible light S1, go through the convex lens 2 and reach the eye 4 of the user. The eye 4 of the user may receive the visible light S1 for viewing the VR display content.


Additionally, the apparatus body 1 may comprise a camera 5. The camera 5 may be located on a user side (i.e., left side of the convex lens 2 as shown in FIG. 1) of the convex lens 2, and a lens of the camera 5 may face an eye 4 of a user for acquiring eye pattern features of the user. The camera 5 may be an infrared radiation (IR) camera or an integrated red-green-blue (RGB)-IR camera, which is not limited in this application. Identity verification on a user wearing the VR helmet may be performed by comparing and analyzing the acquired eye pattern features of the user. Compared to conventional identity verification methods, such as those based on password or gesture recognition, the identify verification in the apparatus of this application is performed based on eye pattern features of users that are acquired by the camera 5 installed in the VR helmet. Therefore, it does not require any complicated structures and equipment being installed on the VR apparatus, or any separate and additional operations from the users. That simplifies the operations and improves the efficiency of identity verification.


In some embodiments, to avoid blocking the VR display content played by the VR playing component 3 (i.e., to avoid blocking the propagation of the visible light S1 from the VR playing component 3), the camera 5 may need to be placed outside of a visible area of the convex lens 2 with respect to the eye 4 of the user (e.g., the visible area may be an area enclosed by an upper boundary T1 and a lower boundary T2 as shown in FIG. 1). The camera 5 may be located at the bottom of the apparatus body 1, that is, the camera 5 may be located below the convex lens 2, and may tilt upward by a first angle, so that the lens of the camera 5 may face the eye 4 of the user.


While using the VR helmet, a user may maintain his/her eye 4 in a straight-looking state as shown in FIG. 1, thus the light S2 from the eye 4 may propagate substantially horizontally, and the line from the camera 5 to the eye 4 may form an angle α with the light S2. A large angle α may lead to severe deformation of the acquired eye pattern features, therefore the camera 5 may be disposed in contact with an edge of the convex lens 2 (in term of at least one of a horizontal distance and a vertical distance) to minimize the angle α without blocking the convex lens 2.


Additionally, as shown in FIG. 2, the VR helmet may include an adjustment component 6 that is electrically connected to the camera 5 and may perform angle adjustment for the camera 5, so that when different users use the same VR helmet, the adjustment component 6 may adjust the angle of the camera 5 to ensure that the lens of the camera 5 faces the eye 4 of the user, thereby reducing, if not eliminating, the distortion of the acquired eye pattern features.


In the aforementioned embodiment, as shown in FIG. 1, in order to avoid blocking the visible light S1, the camera 5 may need to be positioned outside the visible area and thus cannot be at the same height as the eye 4 in a vertical direction. But in a horizontal direction (a direction perpendicular to the paper in FIG. 1), the camera 5's position may be chosen to minimize a horizontal angle between the lens of the camera 5 and the light S2. For example, since the eye 4 of the user is usually located at a center position of the convex lens 2 in the horizontal direction, the camera 5 may be located at the center position of the convex lens 2 in the horizontal direction as shown in FIG. 3. That reduces the deformation or distortion of the eye pattern features acquired by the camera 5.


In the aforementioned embodiment, the VR helmet may comprise two convex lenses 2 corresponding to the eyes of the user. For example, FIG. 3 shows a right convex lens 21 corresponding to a right eye of the user, and a left convex lens 22 corresponding to a left eye of the user. Depending on actual requirements, the apparatus of this application may comprise one or more cameras 5. For example, as shown in FIG. 3, the apparatus may comprise only one camera 5 disposed at the right convex lens 21 to acquire eye pattern features of the right eye of the user. Similarly, the apparatus may comprise only one camera 5 disposed at the left convex lens 22 to acquire eye pattern features of the left eye of the user. Certainly, the apparatus may comprise two cameras 5 disposed at the right convex lens 21 and the left convex lens 22 to respectively acquire eye pattern features of the right eye and the left eye of the user simultaneously or separately.


Although in the embodiments shown in FIGS. 1-3, the camera 5 is located at the bottom of the apparatus body 1 and below the convex lens 2. The camera 5 may be located at other positions as well, which is not limited in this application. For example, as shown in FIG. 4, the camera 5 may be disposed at the top of the apparatus body 1 and above the convex lens 2, so that the camera 5 may tilt downward by a second angle, and the lens of the camera 5 may face the eye 4 of the user for the acquisition of the eye pattern features. Similar to the embodiment shown in FIG. 2, the camera 5 in the embodiment shown in FIG. 4 may also be electrically connected to the adjustment component 6 for performing angular adjustment of the camera 5. Similar to the embodiment shown in FIG. 3, the camera 5 in the embodiment shown in FIG. 4 may also be disposed at a center position of the corresponding convex lens 2 in the horizontal direction (i.e., a direction perpendicular to the paper in FIG. 4) to reduce the deformation or distortion of the eye pattern features acquired by the camera 5. Detail implementation will not be repeated here.



FIG. 5 is a side cross-sectional view of a VR helmet provided by an embodiment of this application. As shown in FIG. 5, the VR helmet may be a split VR head-mounted apparatus. The VR helmet may comprise an apparatus body 1 which comprises an apparatus interface 7 that is electrically connected to an electronic device such as a mobile phone or a tablet installed in the apparatus body 1. The electronic device may, by using a processor and a graphic card for image rendering, and a screen component for displaying VR display content, work as the VR playing component 3 in the apparatus body 1.


Additionally, the camera 5 in the apparatus body 1 may be connected to the apparatus interface 7 through a data line 8, so that when the electronic device connected to the apparatus interface 7 issues a switch control instruction, the camera 5 may receive the switch control instruction through the apparatus interface 7 and the data line 8, and perform a state switching operation in response to the switch control instruction. In other words, by controlling the electronic device or an application program running on the electronic device, the electronic device may send a switch control instruction to the camera 5, thereby controlling the camera 5 to acquire the eye pattern features of the eye 4. That improves the controllability of the eye pattern features acquisition.


Additionally, after the camera 5 completes the acquisition of the eye pattern features, if the VR helmet comprises a processing module, the acquired eye pattern features may be transmitted to the processing module of the VR helmet for processing of the acquired eye pattern features. Alternatively, the camera 5 may transmit, through the apparatus interface 7 and the data line 8, the acquired eye pattern features to the aforementioned electronic device for processing by the electronic device.


Other than a split VR head-mounted apparatus paired with an electronic device such as a mobile phone, the VR helmet of this application may include other forms of VR head-mounted apparatuses. For example, for the split VR head-mounted apparatus, the VR helmet may be accompanied by a PC host, a game console or another external apparatus. The VR playing component 3 may be a display component built into the VR helmet, etc., and the external apparatus may be used for rendering of VR display content. In some embodiments, the VR helmet may be an integrated VR head-mounted apparatus, that is, the VR helmet may be able to play VR display content without resorting to an external displaying apparatus, and the VR playing component 3 may be built-in in the VR helmet. The VR playing component 3 may have playing functions such as rendering and displaying of VR display content.


It should also be noted that the terms “include”, “comprise” and any other variants mean to cover the non-exclusive inclusion. Thereby, the process, method, article, or device which include a series of elements not only include those elements, but also include other elements which are not clearly listed, or include the inherent elements of the process, method, article and device. Without further limitation, the element defined by a phrase “include one . . . ” does not exclude other same elements in the process, method, article or device which include the element.


Reference will be made in detail to exemplary embodiments, examples of which are illustrated in the accompanying drawings. The above description refers to the accompanying drawings in which the same numbers in different drawings represent the same or similar elements unless otherwise represented. The implementations set forth in the following description of exemplary embodiments do not represent all implementations consistent with this application. Instead, they are merely examples of apparatuses and methods consistent with aspects related to this application as recited in the appended claims.


The terms used in this application are merely for the purpose of describing specific embodiments, and are not intended to limit this application. The terms “a”, “said” and “the” of singular forms used in this application and the appended claims are also intended to include plural forms, unless otherwise specified in the context clearly. It should also be understood that, the term “and/or” used herein indicates and includes any or all possible combinations of one or more associated listed items.


It should be understood that although the terms such as first, second, and third may be used herein to describe various information, such information should not be limited to these terms. These terms are merely used for distinguishing information of the same type from each other. For example, within the scope of this application, first information may also be referred to as second information, and similarly, second information may also be referred to as first information. Depending on the context, the term “if” as used herein may be interpreted as “when . . . ” or “upon . . . ” or “in response to determining.”


The foregoing descriptions are merely exemplary embodiments of this application, but are not intended to limit this application. Any modification, equivalent replacement, or improvement made without departing from the spirit and principle of this application should fall within the protection scope of this application.

Claims
  • 1. A virtual reality head-mounted apparatus, comprising: an apparatus body comprising two convex lenses, an electronic device, and two cameras, the two cameras being respectively located at a user side of the two convex lenses for acquiring eye pattern features of a user corresponding to each of the two convex lenses, and a lens of a first camera of the two cameras directly facing an eye of the user for acquiring the eye pattern features of the user, the electronic device located on a side of the two convex lenses away from the user and configured for displaying virtual reality content and receiving the eye pattern features from the cameras,wherein the apparatus further comprises an apparatus interface provided on the apparatus body on the side of the convex lens away from the user, the apparatus interface being electrically connected to the electronic device and the two cameras,and the two cameras are configured to, upon receiving a control instruction transmitted by the electronic device through the apparatus interface, acquire, without relying on an optical surface of the electronic device, the eye pattern features, and send the acquired eye pattern features to the electronic device, andwherein the electronic device is further configured to conduct, based on the eye pattern features, identity verification on the user without interrupting the displaying of the virtual reality content.
  • 2. The apparatus of claim 1, wherein the first camera is located below a first convex lens of the two convex lenses, and tilts upward by a first angle, so that the lens of the first camera faces the eye of the user.
  • 3. The apparatus of claim 1, wherein the first camera is located above a first convex lens of the two convex lenses, and tilts downward by a second angle, so that the lens of the first camera faces the eye of the user.
  • 4. The apparatus of claim 1, wherein the first camera is located outside a visible area of a first convex lens of the two convex lenses with respect to the user.
  • 5. The apparatus of claim 1, wherein the first camera is in contact with an edge of a first convex lens of the two convex lenses.
  • 6. The apparatus of claim 1, wherein the first camera is an RGB camera or an RGB-IR camera.
  • 7. A virtual reality head-mounted apparatus, comprising: an apparatus body comprising a convex lens and a camera, the camera being on a user side of the convex lens, and a lens of the camera directly facing an eye of a user for acquiring eye pattern features of the user;an apparatus interface; andan electronic device for displaying virtual reality content and receiving the eye pattern features from the camera, and located on a side of the convex lens away from the user, the apparatus interface being electrically connected to the electronic device and located on the side of the convex lens away from the user,wherein the camera is connected to the apparatus interface through a data line, and, is configured to, upon receiving a switch control instruction from the electronic device through the apparatus interface and the data line, acquire, in response to the switch control instruction and without relying on an optical surface of the electronic device, eye pattern features, and send the acquired eye pattern features to the electronic device,wherein the electronic device is further configured to conduct, based on the eye pattern features, identity verification on the user without interrupting the displaying of the virtual reality content.
  • 8. The apparatus of claim 1, further comprising: an adjustment component for performing angular adjustment for the first camera, such that the lens of the first camera faces the eye of the user.
  • 9. The apparatus of claim 7, wherein the camera is located below the convex lens, and tilts upward by a first angle, so that the lens of the camera faces the eye of the user.
  • 10. The apparatus of claim 7, wherein the camera is located above the convex lens, and tilts downward by a second angle, so that the lens of the camera faces the eye of the user.
  • 11. The apparatus of claim 7, wherein the camera is located outside a visible area of the convex lens with respect to the user.
  • 12. The apparatus of claim 7, wherein the camera is in contact with an edge of the convex lens.
  • 13. The apparatus of claim 7, wherein the camera is an RGB camera or an RGB-IR camera.
  • 14. The apparatus of claim 7, further comprising: an adjustment component for performing angular adjustment for the camera, such that the lens of the camera faces the eye of the user.
Priority Claims (1)
Number Date Country Kind
201710109091.6 Feb 2017 CN national
CROSS-REFERENCE TO RELATED APPLICATION

This application is a continuation application of International Patent Application No. PCT/CN2018/077280, filed on Feb. 26, 2018, which is based on and claims priority of the Chinese Patent Application No. 201710109091.6, filed on Feb. 27, 2017 and entitled “VIRTUAL REALITY HEAD-MOUNTED APPARATUS.” The above-referenced applications are incorporated herein by reference in their entirety.

US Referenced Citations (47)
Number Name Date Kind
5993000 Kobayashi et al. Nov 1999 A
9329387 Border May 2016 B2
9723117 Buckley Aug 2017 B2
9788714 Krueger Oct 2017 B2
9811908 Raghoebardajal Nov 2017 B2
9829707 Border Nov 2017 B2
9857595 Costa Jan 2018 B2
10013055 Perek et al. Jul 2018 B2
10055887 Gil Aug 2018 B1
10220181 Giap et al. Mar 2019 B2
10488917 Ollila Nov 2019 B2
10573080 Gromotka Feb 2020 B2
10642569 Lin May 2020 B2
20060017654 Romo Jan 2006 A1
20120056717 Maharbiz et al. Mar 2012 A1
20120206588 Kishigami et al. Aug 2012 A1
20130235169 Kato et al. Sep 2013 A1
20140055746 Nistico et al. Feb 2014 A1
20140266990 Makino et al. Sep 2014 A1
20140354953 Chen et al. Dec 2014 A1
20150009574 Liesecke et al. Jan 2015 A1
20150049013 Rahman Feb 2015 A1
20150054734 Raghoebardajal Feb 2015 A1
20150070489 Hudman Mar 2015 A1
20150185475 Saarikko Jul 2015 A1
20150309263 Abovitz et al. Oct 2015 A2
20160063767 Lee et al. Mar 2016 A1
20160080720 Fullam Mar 2016 A1
20160085302 Publicover et al. Mar 2016 A1
20160180591 Shiu et al. Jun 2016 A1
20160259986 Yun et al. Sep 2016 A1
20160267712 Nartker Sep 2016 A1
20160358181 Bradski Dec 2016 A1
20170082858 Klug et al. Mar 2017 A1
20170124309 Tang May 2017 A1
20170124392 Gu May 2017 A1
20170140224 Wilson et al. May 2017 A1
20170235931 Publicover et al. Aug 2017 A1
20180032133 Cho Feb 2018 A1
20180053051 Chen et al. Feb 2018 A1
20180096503 Kaehler et al. Apr 2018 A1
20180203505 Trail Jul 2018 A1
20180218211 Ishioka et al. Aug 2018 A1
20180239423 Mardanbegi Aug 2018 A1
20190113968 Huang et al. Apr 2019 A1
20190258314 Ollila Aug 2019 A1
20200050257 Lee Feb 2020 A1
Foreign Referenced Citations (36)
Number Date Country
102693437 Sep 2012 CN
103429143 Dec 2013 CN
104793741 Jul 2015 CN
104834852 Aug 2015 CN
105068249 Nov 2015 CN
105212418 Jan 2016 CN
105653227 Jun 2016 CN
105718046 Jun 2016 CN
105955491 Sep 2016 CN
205594581 Sep 2016 CN
106203410 Dec 2016 CN
106250749 Dec 2016 CN
205844631 Dec 2016 CN
106339087 Jan 2017 CN
106406509 Feb 2017 CN
205942608 Feb 2017 CN
205942609 Feb 2017 CN
106873159 Jun 2017 CN
206249347 Jun 2017 CN
206584119 Oct 2017 CN
107562179 Jan 2018 CN
202016104179 Aug 2016 DE
202016104179 Aug 2016 DE
H11-44862 Feb 1999 JP
2011-069978 Apr 2011 JP
2014-191386 Oct 2014 JP
2016-105555 Jun 2016 JP
201445212 Dec 2014 TW
2012172719 Dec 2012 WO
2013076994 May 2013 WO
2015198502 Dec 2015 WO
2016018488 Feb 2016 WO
2016191709 Dec 2016 WO
2017026371 Feb 2017 WO
2017026371 Feb 2017 WO
2018067357 Apr 2018 WO
Non-Patent Literature Citations (23)
Entry
Search Report for European Application No. 18757692.1 dated Nov. 19, 2019 (4 pages).
Examination Report for European Application No. 18757692.1 dated Nov. 29, 2019 (7 pages).
First Office Action for Chinese Application No. 201710109091.6 dated Sep. 3, 2018 with English machine translation (13 pages).
Supplementary Search for Chinese Application No. 201710109091.6 dated Mar. 28, 2019 (1 page).
Second Office Action for Chinese Application No. 201710109091.6 dated Apr. 3, 2019 with English machine translation (12 pages).
Third Office Action for Chinese Application No. 201710109091.6 dated Sep. 17, 2019 with English machine translation (15 pages).
Non-final rejection and Search Report for Taiwanese Application No. 106140067 dated Oct. 15, 2018 (7 pages).
Written Opinion of the International Searching Authority and International Search Report for PCT Application No. PCT/CN2018/077280 dated May 25, 2018 (15 pages).
International Preliminary Report on Patentability Chapter 1 for PCT Application No. PCT/CN2018/077280 dated Sep. 6, 2019 (11 pages).
First Search for Chinese Application No. 201710109091.6 dated Aug. 27, 2018 (1 page).
Written Opinion for Singaporean Application No. 11201907256Q dated Apr. 7, 2020.
Communication pursuant to Article 94(3) EPC for European Application No. 18757692.1 dated May 20, 2021.
Scott Hayden, “Hands On: SMIs Gear VR Eye Tracking is Accurate, Fast and Lightweight”, Feb. 24, 2016.
Scott W. Greenwald et al., “Eye gaze tracking with google cardboard using purkinje images”, Proceedings of the 22nd ACM Conference on Virtual Reality Software and Technology, Nov. 2, 2016.
Written Opinion for Singapore Patent Application No. 11201907256Q dated Mar. 2, 2022.
Office Action for Korean Application No. 10-2019-7024490 dated Nov. 19, 2020.
Office Action for Japanese Application No. 2019-546342 dated Nov. 17, 2020.
Communication pursuant to Article 94(3) EPC for European Application No. 18757692.1 dated Aug. 1, 2022.
Pupil Dev Team, “Blog—HTC Vive Eye Tracking Add On—Pupil Labs,” Retrieved from Internet: https://pupil-labs.com/blog/news/htc-vive-eye-tracking-add-on/. Aug. 9, 2016.
Pupil Labs, “Unbox the Pupil Labs add-on,” https://www.youtube.com/watch?v=HGMjJLnK2_4, Oct. 14, 2016.
Pupil Labs, “Take HTC Vive apart a bit more so we can show a close up view,” https://www.youtube.com/watch?v=nlzuwHaglXQ, Oct. 14, 2016.
Pupil Labs, “Closeup of the add-on engagement with the lens holder,” https://www.youtube.com/watch?v=zswmKmlBrss, Oct. 14, 2016.
Anthes et al., “State of the art of virtual reality technology,” 2016 IEEE Aerospace Conference, IEEE, Mar. 5, 2016.
Related Publications (1)
Number Date Country
20190354758 A1 Nov 2019 US
Continuations (1)
Number Date Country
Parent PCT/CN2018/077280 Feb 2018 US
Child 16527978 US