The present disclosure relates, generally, to patient support apparatuses and, more specifically, to patient support apparatus user interfaces.
Patient support apparatuses, such as hospital beds, stretchers, cots, tables, wheelchairs, and chairs are used to help caregivers facilitate care of patients in a health care setting. Conventional patient support apparatuses generally comprise a base and a patient support surface upon which the patient is supported. Often, these patient support apparatuses have one or more powered devices with motors to perform one or more functions, such as lifting and lowering the patient support surface, articulating one or more deck sections, raising a patient from a slouched position, turning a patient, centering a patient, extending a length or width of the patient support apparatus, and the like. Furthermore, these patient support apparatuses typically employ one or more sensors arranged to detect patient movement, monitor patient vital signs, and the like.
When a caregiver wishes to perform an operational function, such as operating a powered device that adjusts the patient support surface relative to the base, the caregiver actuates an input device of a user interface, often in the form of a touchscreen or a button on a control panel. Here, the user interface may also employ a screen to display visual content to the caregiver, such as patient data and operating or status conditions of the patient support apparatus. The visual content may further comprise various graphical menus, buttons, indicators, and the like, which may be navigated via the input device. Certain operational functions or features of the patient support apparatus may also be accessible to and adjustable by the patient. Here, the user interface may allow the patient to adjust the patient support surface between various positions or configurations, view and navigate visual content displayed on a screen (for example, a television program), adjust audio output (for example, volume), and the like.
As the number and complexity of functions integrated into conventional patient support apparatuses has increased, the associated user interfaces have also become more complex and expensive to manufacture. While conventional patient support apparatuses have generally performed well for their intended purpose, there remains a need in the art for a patient support apparatus which overcomes the disadvantages in the prior art and which affords caregivers and patients with improved usability and functionality in a number of different operating conditions.
Referring to
A support structure 32 provides support for the patient P. In the representative embodiment illustrated herein, the support structure 32 comprises a base 34, an intermediate frame 36, and a patient support deck 38. The intermediate frame 36 and the patient support deck 38 are spaced above the base 34 in
As is best depicted in
A mattress 52 is disposed on the patient support deck 38 during use. The mattress 52 comprises a secondary patient support surface 42 upon which the patient P is supported. The base 34, the intermediate frame 36, and the patient support deck 38 each have a head end and a foot end corresponding to designated placement of the patient's P head and feet on the patient support apparatus 30. It will be appreciated that the specific configuration of the support structure 32 may take on any known or conventional design, and is not limited to that specifically illustrated and described herein. In addition, the mattress 52 may be omitted in certain embodiments, such that the patient P can rest directly on the patient support surface 42 defined by the deck sections 40 of the patient support deck 38.
Side rails 54, 56, 58, 60 are coupled to the support structure 32 and are supported by the base 34. A first side rail 54 is positioned at a right head end of the intermediate frame 36. A second side rail 56 is positioned at a right foot end of the intermediate frame 36. A third side rail 58 is positioned at a left head end of the intermediate frame 36. A fourth side rail 60 is positioned at a left foot end of the intermediate frame 36. The side rails 54, 56, 58, 60 are advantageously movable between a raised position in which they block ingress and egress into and out of the patient support apparatus 30, one or more intermediate positions, and a lowered position in which they are not an obstacle to such ingress and egress. It will be appreciated that there may be fewer side rails for certain embodiments, such as where the patient support apparatus 30 is realized as a stretcher or a cot. Moreover, it will be appreciated that in certain configurations the patient support apparatus 30 may not include any side rails. Similarly, it will be appreciated that side rails may be attached to any suitable component or structure of the patient support apparatus 30. Furthermore, in certain embodiments the first and third side rails 54, 58 are coupled to a deck section 40 for concurrent movement between section positions 40A, 40B (for example, see
As shown in
One or more caregiver interfaces 66, such as handles, are shown in
Wheels 68 are coupled to the base 34 to facilitate transportation over floor surfaces. The wheels 68 are arranged in each of four quadrants of the base 34, adjacent to corners of the base 34. In the embodiments shown, the wheels 68 are caster wheels able to rotate and swivel relative to the support structure 32 during transport. Here, each of the wheels 68 forms part of a caster assembly 70 mounted to the base 34. It should be understood that various configurations of the caster assemblies 70 are contemplated. In addition, in some embodiments, the wheels 68 may not be caster wheels. Moreover, it will be appreciated that the wheels 68 may be non-steerable, steerable, non-powered, powered, or combinations thereof. While the representative embodiment of the patient support apparatus 30 illustrated herein employs four wheels 68, additional wheels are also contemplated. For example, the patient support apparatus 30 may comprise four non-powered, non-steerable wheels, along with one or more additional powered wheels. In some cases, the patient support apparatus 30 may not include any wheels. In other embodiments, one or more auxiliary wheels (powered or non-powered), which are movable between stowed positions and deployed positions, may be coupled to the support structure 32. In some cases, when auxiliary wheels are located between caster assemblies 70 and contact the floor surface in the deployed position, they may cause two of the caster assemblies 70 to be lifted off the floor surface, thereby shortening a wheel base of the patient support apparatus 30. A fifth wheel may also be arranged substantially in a center of the base 34. Other configurations are contemplated.
The patient support apparatus 30 further comprises a lift mechanism, generally indicated at 72, which operates to lift and lower the intermediate frame 36 relative to the base 34 which, in turn, moves the patient support deck 38 between a first vertical configuration 38A (for example, a “lowered” vertical position as depicted in
As noted above, the patient support deck 38 is operatively attached to the intermediate frame 36, and the deck section 40 is arranged for movement between a first section position 40A (see
Those having ordinary skill in the art will appreciate that the patient support apparatus 30 could employ any suitable number of deck actuators 80, of any suitable type or configuration sufficient to effect selective movement of the deck section 40 relative to the support structure 32. By way of non-limiting example, the deck actuator 80 could be a linear actuator or one or more rotary actuators driven electronically and/or hydraulically, and/or controlled or driven in any suitable way. Moreover, the deck actuator 80 could be mounted, secured, coupled, or otherwise operatively attached to the intermediate frame 36 and to the deck section 40, either directly or indirectly, in any suitable way. In addition, one or more of the deck actuators 80 could be omitted for certain applications.
Referring now to
As noted above, the controller 84 is best depicted schematically
In the representative embodiment illustrated in
In the representative embodiments illustrated herein, the user interface 86 is realized as a touchscreen 92 comprising a screen 94 and a touch sensor 96. As is described in greater detail below, the screen 94 is configured to display visual content VC to the user, and may be of any suitable size, shape, and/or orientation sufficient to display visual content VC. By way of non-limiting example, the screen 94 could be realized as a curved LCD panel extending along the length or width of the patient support apparatus 30. The touch sensor 96 is operatively attached to the screen 94, and defines an input surface 98 arranged adjacent to the screen 94 that is adapted for interaction with the user (e.g., the caregiver C and/or the patient P). Thus, the screen 94 generally forms part of one or more of the user interfaces 86 for operating the patient support apparatus 30, such as where activation or manipulation of the input device 90 (for example, a touch sensor 96 operatively attached to the screen 94) generates the input signal IS used by the controller 84 to facilitate navigation and/or activation of the visual content VC (e.g., by navigating menus and/or actuating virtual buttons represented as visual content VC on the screen 94 by engaging the touch sensor 96).
In the illustrated embodiment, the screen 94 is operatively attached to the patient support apparatus 30 for concurrent movement. More specifically, the screen 94 is coupled to the footboard 64 for concurrent movement with the patient support deck 38 between the vertical configurations 38A, 38B via the lift mechanism 72, as noted above (compare
In some embodiments, the patient support apparatus 30 further comprises a lift sensor, generally indicated at 100, to determine movement of the patient support deck 38 between the vertical configurations 38A, 38B via the lift mechanism 72 (compare
Those having ordinary skill in the art will appreciate that the lift sensor 100 could be realized in a number of different ways. By way of non-limiting example, the lift sensor 100 could be realized as one or more discrete components, such as a linear potentiometer, a range sensor, a hall-effect sensor, a limit switch, an accelerometer, a gyroscope, and the like generally configured or arranged to measure position, height, and/or movement. Further, the lift sensor 100 could be an encoder, a current sensor, and the like coupled to or in communication with one of the lift actuators 78. Moreover, the functionality afforded by the lift sensor 100 could be entirely or partially realized with software or code for certain applications. Other configurations are contemplated.
In some embodiments, the patient support apparatus 30 further comprises a proximity sensor, generally indicated at 102, to determine movement occurring within an envelope 104 defined adjacent to a caregiver-accessible user interface 86 coupled to the footboard 64 of the patient support apparatus 30 (see
Those having ordinary skill in the art will appreciate that the proximity sensor 102 could be realized in a number of different ways. By way of non-limiting example, the proximity sensor 102 could be realized as one or more discrete components, such as a photoelectric emitter/sensor, a photodetector sensor, a laser rangefinder, a passive charge-coupled device (e.g., a digital camera), a passive thermal infrared sensor (e.g., a forward-looking infrared camera), a radar transmitter/sensor, a sonar transmitter/sensor, and the like generally configured or arranged to detect changes in heat, air pressure, and/or position and/or movement. The proximity sensor 102 could also be realized as a detector configured to respond to the presence or absence of a token, tracker, badge, portable electronic device, and the like carried by the caregiver (e.g., via radio-frequency identification, near-field communication, global positioning satellites, Bluetooth®, Wi-Fi™, and the like). Further, the proximity sensor 102 could be realized as a part of the touchscreen 92 in some embodiments, such as based on electrostatic fields generated with a capacitive-type touch sensor 96. Here too, the functionality afforded by the proximity sensor 102 could be entirely or partially realized with software or code for certain applications. Other configurations are contemplated.
As noted above, in the representative embodiment illustrated in
It will be appreciated that multiple lift sensors 100 could be employed by the controller 84 in certain embodiments, such as to facilitate differentiating the respective heights of the head-end and the foot-end of the patient support deck 38. This differentiation may be used to adjust, optimize, or otherwise change how visual content VC is presented on the screen 94 in some embodiments. Here, it will be appreciated that the specific position and/or orientation of the screen 94 may change relative to the caregiver C based on how the patient support deck 38 is orientated because the screen 94 is coupled to the footboard 64 in the illustrated embodiment. Thus, in embodiments of the patient support apparatus 30 where the head end lift member 74 and the foot end lift member 76 can be driven or otherwise actuated independently (e.g., to place the patient support deck 38 in a Trendelenburg position), the controller 84 could be configured to display visual content VC in different ways based on the orientation and/or position of the screen 94 relative to the base 34 (e.g., using one or more accelerometers, gyroscopes, inertial sensors, and the like). To this end, and by way of non-limiting example, the controller 84 could accommodate changes in the orientation of the screen 94 by presenting, rendering, or otherwise displaying different types of visual content VC, by scaling visual content VC, and/or by otherwise modifying visual content VC as the lift members 74, 76 move the patient support deck 38 to and between different configurations. Other embodiments are contemplated.
As will be appreciated from the subsequent description of the control system 82 in connection with
With continued reference to the embodiment illustrated in
Those having ordinary skill in the art will appreciate that the visual content VC can be delineated in a number of different ways, and may comprise any suitable number of content portions CP1, CP2 which may be distinguishable from each other (e.g., discrete icons, menus, graphics, symbols, buttons, and the like). Thus, in some embodiments, discrete portions of the same content portion may change in different ways. By way of non-limiting example, and as is described in greater detail below in connection with the embodiment depicted in
As will be appreciated from the subsequent description below, the first and/or the second content layouts CL1, CL2 can be configured in a number of different ways so as to optimize the functionality and visibility of the visual content VC based on changes between the vertical configurations 38A, 38B (and/or based on other parameters associated with utilization of the patient support apparatus 30 as described in greater detail below). By way of non-limiting example, the second content layout CL2 shown in
Furthermore, it will be appreciated that the visual content VC could dynamically change between the content layouts CL1, CL2 as the patient support deck 38 moves between the vertical configurations 38A, 38B. By way of non-limiting example, the controller 84 could linearly scale the first content portion CP1 of the visual content VC, such as by using the scaling factors X1, X2 as respective end-points between the vertical configurations 38A, 38B, to display visual content VC at an intermediate content layout CL1.5 with an intermediate scaling factor X1.5 when the patient support deck 38 is between the first and second vertical configurations 38A, 38B (not shown). In addition, it will be appreciated that the controller 84 can display visual content VC in different ways, based such as on which direction the patient support deck 38 is moving. By way of non-limiting example, when the patient support deck 38 is moved towards the floor as the caregiver C actuates a “lower bed” button, the content layout could dynamically change to increase the size of the “lower bed” button as the patient support deck 38 moves closer to the floor. In some embodiments, one or more content portions CP1, CP2 of the visual content VC may change concurrently or separately in ways other than by resizing graphics and text TX based on the scaling factors X1, X2. By way of non-limiting example, as the “lower bed” button associated with the first content portion CP1 is scaled up in size in response to movement toward the first vertical configuration 38A, other buttons, controls, or information (e.g., associated with the second content portion CP2 or another content portion) could be hidden, moved off-screen, and the like. In some embodiments, once the patient support deck 38 has been positioned as close to the floor as possible in the first vertical configuration 38A, the controller 84 could hide everything associated with the first content portion CP1 except for a large “bed up” button used to subsequently move the patient support deck 38 away from the floor. Here in this embodiment, after the “bed up” button has actuated by the caregiver C, the “bed up” button could remain unchanged in size until released, and then the controller 84 could subsequently display different visual content VC (e.g., the first content portion CP1 depicted in
Referring now to
In the representative embodiment depicted in
Those having ordinary skill in the art will appreciate that this configuration affords the caregiver C with the ability to view visual content VC from a distance and without necessarily approaching the patient support apparatus 30. This can be advantageous in situations where the caregiver C wants to observe certain types of relevant visual content VC displayed on the screen 94 without disturbing the patient P, such as when the patient P is resting or asleep. In addition, it will be appreciated that this configuration also affords the caregiver C with the ability to automatically transition to visual content VC which is more relevant when viewed closer to the screen 94.
By way of non-limiting example, in the representative embodiment illustrated in
As shown in
With continued reference to
As noted above, when visual content VC is displayed in the first content layout CL1, the sub-portions of the first content portion CP1 are scaled larger than they are scaled when in the second content layout CL2, whereas the sub-portions of the second content portion CP2 are scaled smaller (see
Because the controller 84 is configured to display visual content VC on the screen 94 in the first content layout CL1 during an absence of movement occurring within the envelope 104 in this embodiment, the caregiver C is able to observe visual content VC on the screen 94 which reflects the patient's P heartrate in a larger size than virtual “buttons” of the user interface 86 while they are positioned away from the patient support apparatus 30. This allows the caregiver C to view the patient's P heartrate from a distance and without necessarily requiring that the caregiver C come into close proximity with the patient support apparatus 30 (e.g., with a quick glance into the patient's P room while making rounds so as not to disturb to the patient P). Moreover, while it is contemplated that the second content portion CP2 of the visual content VC could be displayed or otherwise adjusted in ways other than by scaling as noted above (e.g., by hiding one or more sub-portions of the second content portion CP2), it may be advantageous in certain applications for one or more virtual “buttons” to remain displayed on the screen 94 in a smaller size alongside the larger heartrate text TX. Here, the continued presence of the second content portion CP2 in the first content layout CL1, even with a smaller size when viewed from a distance, may communicate useful status information about the patient support apparatus 30. By way of example, status information communicated by the smaller sized second content portion CP2 could assure the caregiver that certain features of the patient support apparatus 30 remained “locked,” that one or more of the deck sections 40 haven't been adjusted recently, that the patient P hasn't attempted to exit the patient support apparatus 30 without assistance, and the like. Other configurations are contemplated.
Furthermore, because the controller 84 is configured to display visual content VC on the screen 94 in the second content layout CL2 in response to movement occurring within the envelope 104 in this embodiment, the caregiver C is able to observe visual content VC on the screen 94 which reflects the patient's P heartrate in a smaller size than virtual “buttons” of the user interface 86 while they are positioned nearby the patient support apparatus 30. This allows the caregiver C to continue viewing the patient's P heartrate on the screen 94 in a smaller size as they approach the patient support apparatus 30, while simultaneously improving their ability to view the virtual “buttons” of the user interface 86 by scaling the second content portion CP2 to a larger size. Put differently, because the embodiments of the present disclosure allow visual content VC to be presented to the caregiver C in different ways based on their proximity to the patient P and/or to the patient support apparatus 30, visual content VC which is relevant when the caregiver C is standing next to the patient support apparatus 30 (e.g., virtual “buttons” of the user interface 86 that raise or lower the patient support deck 38) can automatically be presented on the screen 94 more prominently than when they the caregiver C is further away from the patient support apparatus 30, where different visual content VC be more relevant to the caregiver C (e.g., large text TX representing the patient's P current heartrate).
Like the embodiment described above in connection with
In this way, the embodiments of the present disclosure afford significant opportunities for enhancing the functionality and operation of user interfaces 86 employed by patient support apparatuses 30. Specifically, visual content VC can be displayed and viewed in a number of different ways which contribute to improved usability of the patient support apparatus 30 without necessitating the use of overtly expensive hardware. Moreover, visual content can be displayed and in ways that provide caregivers C with convenient, easy-to-use, and intuitive features. Thus, the patient support apparatus 30 can be manufactured in a cost-effective manner while, at the same time, affording opportunities for improved functionality, features, and usability.
As noted above, the subject patent application is related to U.S. Provisional Patent Application No. 62/525,373 filed on Jun. 27, 2017. In addition, the subject patent application is also related to: U.S. Provisional Patent Application No. 62/525,353 filed on Jun. 27, 2017 and its corresponding Non-Provisional patent application Ser. No. 16/020,068 filed on Jun. 27, 2018, now U.S. Pat. No. 11,337,872; U.S. Provisional Patent Application No. 62/525,359 filed on Jun. 27, 2017 and its corresponding Non-Provisional patent application Ser. No. 16/020,052 filed on Jun. 27, 2018, now U.S. Pat. No. 11,382,812; U.S. Provisional Patent Application No. 62/525,363 filed on Jun. 27, 2017 and its corresponding Non-Provisional patent application Ser. No. 16/020,085 filed on Jun. 27, 2018; U.S. Provisional Patent Application No. 62/525,368 filed on Jun. 27, 2017 and its corresponding Non-Provisional patent application Ser. No. 16/019,973 filed on Jun. 27, 2018, now U.S. Pat. No. 11,096,850; and U.S. Provisional Patent Application No. 62/525,377 filed on Jun. 27, 2017 and its corresponding Non-Provisional patent application Ser. No. 16/019,986 filed on Jun. 27, 2018, now U.S. Pat. No. 10,811,136. The disclosures of each of the above-identified Provisional Patent Applications and corresponding Non-Provisional patent applications are each hereby incorporated by reference in their entirety.
It will be further appreciated that the terms “include,” “includes,” and “including” have the same meaning as the terms “comprise,” “comprises,” and “comprising.” Moreover, it will be appreciated that terms such as “first,” “second,” “third,” and the like are used herein to differentiate certain structural features and components for the non-limiting, illustrative purposes of clarity and consistency.
Several configurations have been discussed in the foregoing description. However, the configurations discussed herein are not intended to be exhaustive or limit the invention to any particular form. The terminology which has been used is intended to be in the nature of words of description rather than of limitation. Many modifications and variations are possible in light of the above teachings and the invention may be practiced otherwise than as specifically described.
The invention is intended to be defined in the independent claims, with specific features laid out in the dependent claims, wherein the subject-matter of a claim dependent from one independent claim can also be implemented in connection with another independent claim.
This application is a Continuation of U.S. patent application Ser. No. 16/020,003, filed on Jun. 27, 2018, now U.S. Pat. No. 11,202,729, which claims priority to and all the benefits of U.S. Provisional Patent Application No. 62/525,373 filed on Jun. 27, 2017, the disclosures of each of which are hereby incorporated by reference in their entirety.
Number | Name | Date | Kind |
---|---|---|---|
5113214 | Nagata et al. | May 1992 | A |
5276432 | Travis | Jan 1994 | A |
5434621 | Yu | Jul 1995 | A |
5640953 | Bishop et al. | Jun 1997 | A |
5645667 | Kusen | Jul 1997 | A |
5664270 | Bell et al. | Sep 1997 | A |
5971913 | Newkirk et al. | Oct 1999 | A |
6320510 | Menkedick et al. | Nov 2001 | B2 |
6340977 | Lui et al. | Jan 2002 | B1 |
6351678 | Borders | Feb 2002 | B1 |
6362725 | Ulrich et al. | Mar 2002 | B1 |
6396224 | Luff et al. | May 2002 | B1 |
6560492 | Borders | May 2003 | B2 |
6702314 | Crose | Mar 2004 | B1 |
6876303 | Reeder et al. | Apr 2005 | B2 |
6948592 | Kavounas | Sep 2005 | B2 |
7017208 | Weismiller et al. | Mar 2006 | B2 |
7038588 | Boone et al. | May 2006 | B2 |
7154397 | Zerhusen et al. | Dec 2006 | B2 |
7296312 | Menkedick et al. | Nov 2007 | B2 |
7319386 | Collins, Jr. et al. | Jan 2008 | B2 |
7336187 | Hubbard, Jr. et al. | Feb 2008 | B2 |
7389552 | Reed et al. | Jun 2008 | B1 |
7443302 | Reeder et al. | Oct 2008 | B2 |
7472439 | Lemire et al. | Jan 2009 | B2 |
7487562 | Frondorf et al. | Feb 2009 | B2 |
7490021 | Holland et al. | Feb 2009 | B2 |
7557718 | Petrosenko et al. | Jul 2009 | B2 |
7570152 | Smith et al. | Aug 2009 | B2 |
7690059 | Lemire et al. | Apr 2010 | B2 |
7702481 | Dionne et al. | Apr 2010 | B2 |
7747644 | Reihl et al. | Jun 2010 | B1 |
7888901 | Larson et al. | Feb 2011 | B2 |
7895519 | Allegrezza et al. | Feb 2011 | B1 |
7911349 | Zerhusen et al. | Mar 2011 | B2 |
7945452 | Fathallah et al. | May 2011 | B2 |
7973666 | Petrosenko et al. | Jul 2011 | B2 |
8069157 | Jam | Nov 2011 | B2 |
8117701 | Bobey et al. | Feb 2012 | B2 |
8121856 | Huster et al. | Feb 2012 | B2 |
8143846 | Herman et al. | Mar 2012 | B2 |
8165908 | Bolle et al. | Apr 2012 | B2 |
8209608 | Linyard et al. | Jun 2012 | B1 |
8266742 | Andrienko | Sep 2012 | B2 |
8308237 | Kunou | Nov 2012 | B2 |
8319633 | Becker et al. | Nov 2012 | B2 |
8334779 | Zerhusen et al. | Dec 2012 | B2 |
8341777 | Hensley et al. | Jan 2013 | B2 |
8344860 | Collins, Jr. et al. | Jan 2013 | B2 |
8410943 | Metz et al. | Apr 2013 | B2 |
8413270 | Turner et al. | Apr 2013 | B2 |
8413271 | Blanchard et al. | Apr 2013 | B2 |
8413273 | Hornbach et al. | Apr 2013 | B2 |
8432287 | O'Keefe et al. | Apr 2013 | B2 |
8442738 | Patmore | May 2013 | B2 |
8442843 | Kneuer et al. | May 2013 | B2 |
8464380 | Bobey et al. | Jun 2013 | B2 |
8525682 | Dixon et al. | Sep 2013 | B2 |
8544126 | Elliott et al. | Oct 2013 | B2 |
8552880 | Kopp et al. | Oct 2013 | B2 |
8566118 | Kneuer et al. | Oct 2013 | B2 |
8572778 | Newkirk et al. | Nov 2013 | B2 |
8604917 | Collins et al. | Dec 2013 | B2 |
8606600 | Kneuer et al. | Dec 2013 | B2 |
8641301 | Yang et al. | Feb 2014 | B2 |
8650682 | Herman | Feb 2014 | B2 |
8674839 | Zerhusen et al. | Mar 2014 | B2 |
8689376 | Becker et al. | Apr 2014 | B2 |
8706527 | Kneuer et al. | Apr 2014 | B2 |
8716941 | Kim | May 2014 | B2 |
8752220 | Soderberg et al. | Jun 2014 | B2 |
8756078 | Collins, Jr. et al. | Jun 2014 | B2 |
8768520 | Oexman et al. | Jul 2014 | B2 |
8789102 | Pickelsimer et al. | Jul 2014 | B2 |
8789222 | Blanchard et al. | Jul 2014 | B2 |
8844078 | Hornbach et al. | Sep 2014 | B2 |
8847756 | Tallent et al. | Sep 2014 | B2 |
8866598 | Collins, Jr. et al. | Oct 2014 | B2 |
8868542 | Kimball et al. | Oct 2014 | B2 |
8870812 | Alberti et al. | Oct 2014 | B2 |
8896524 | Birnbaum et al. | Nov 2014 | B2 |
8923994 | Laikari et al. | Dec 2014 | B2 |
8924218 | Corpier et al. | Dec 2014 | B2 |
8926535 | Rawls-Meehan | Jan 2015 | B2 |
8984685 | Robertson et al. | Mar 2015 | B2 |
9001038 | Kasahara | Apr 2015 | B2 |
9032510 | Sampathkumaran et al. | May 2015 | B2 |
9038217 | Elliot et al. | May 2015 | B2 |
9050031 | Collins, Jr. et al. | Jun 2015 | B2 |
9088282 | Holenarsipur et al. | Jul 2015 | B2 |
9104789 | Gross et al. | Aug 2015 | B2 |
9126571 | Lemire et al. | Sep 2015 | B2 |
9138173 | Penninger et al. | Sep 2015 | B2 |
9173792 | Goffer | Nov 2015 | B2 |
9186113 | Harper et al. | Nov 2015 | B2 |
9204823 | Derenne et al. | Dec 2015 | B2 |
9220650 | Bobey et al. | Dec 2015 | B2 |
9228885 | Zerhusen et al. | Jan 2016 | B2 |
9230421 | Reeder et al. | Jan 2016 | B2 |
9233033 | Valentino et al. | Jan 2016 | B2 |
9233038 | Gibson et al. | Jan 2016 | B2 |
9259369 | Derenne et al. | Feb 2016 | B2 |
9262876 | Wood et al. | Feb 2016 | B2 |
9320664 | Newkirk et al. | Apr 2016 | B2 |
9336672 | Collins, Jr. et al. | May 2016 | B2 |
9342677 | Ali et al. | May 2016 | B2 |
9358168 | Williamson et al. | Jun 2016 | B2 |
9381125 | Herbst et al. | Jul 2016 | B2 |
9424699 | Kusens et al. | Aug 2016 | B2 |
9456938 | Blickensderfer et al. | Oct 2016 | B2 |
9463126 | Zerhusen et al. | Oct 2016 | B2 |
9466163 | Kusens et al. | Oct 2016 | B2 |
9480435 | Olsen | Nov 2016 | B2 |
9486084 | Connell et al. | Nov 2016 | B2 |
9492341 | Huster et al. | Nov 2016 | B2 |
9569591 | Vanderpohl, III | Feb 2017 | B2 |
9593833 | McMannon et al. | Mar 2017 | B2 |
9613445 | Gilger et al. | Apr 2017 | B2 |
9618383 | Zerhusen et al. | Apr 2017 | B2 |
9655798 | Zerhusen et al. | May 2017 | B2 |
9691206 | Kusens et al. | Jun 2017 | B2 |
9774991 | Kusens | Sep 2017 | B2 |
9782005 | Paul et al. | Oct 2017 | B2 |
9814410 | Kostic et al. | Nov 2017 | B2 |
9838849 | Kusens | Dec 2017 | B2 |
9844275 | Nunn et al. | Dec 2017 | B2 |
9849051 | Newkirk et al. | Dec 2017 | B2 |
9858741 | Kusens et al. | Jan 2018 | B2 |
9861321 | Collins, Jr. et al. | Jan 2018 | B2 |
9892310 | Kusens et al. | Feb 2018 | B2 |
9892311 | Kusens et al. | Feb 2018 | B2 |
9916649 | Kusens | Mar 2018 | B1 |
9934427 | Derenne et al. | Apr 2018 | B2 |
9940810 | Derenne et al. | Apr 2018 | B2 |
9984521 | Kusens et al. | May 2018 | B1 |
9998857 | Kusens | Jun 2018 | B2 |
9999555 | Magill et al. | Jun 2018 | B2 |
10004654 | Zerhusen et al. | Jun 2018 | B2 |
10034979 | Bechtel et al. | Jul 2018 | B2 |
10052249 | Elliott et al. | Aug 2018 | B2 |
10098796 | Valentino et al. | Oct 2018 | B2 |
10136841 | Alghazi | Nov 2018 | B2 |
10172752 | Goffer | Jan 2019 | B2 |
10176608 | Gilger et al. | Jan 2019 | B2 |
10188569 | Elku et al. | Jan 2019 | B2 |
10363181 | Williamson et al. | Jul 2019 | B2 |
10391008 | Zerhusen et al. | Aug 2019 | B2 |
10420688 | Furman et al. | Sep 2019 | B2 |
10561552 | Newkirk et al. | Feb 2020 | B2 |
11096850 | Bhimavarapu | Aug 2021 | B2 |
11160514 | Nahavandi et al. | Nov 2021 | B2 |
11202729 | Desaulniers | Dec 2021 | B2 |
11246776 | Moreno et al. | Feb 2022 | B2 |
11484451 | Nahavandi | Nov 2022 | B1 |
20020014951 | Kramer et al. | Feb 2002 | A1 |
20030183427 | Tojo et al. | Oct 2003 | A1 |
20040083394 | Brebner et al. | Apr 2004 | A1 |
20060077186 | Park et al. | Apr 2006 | A1 |
20060101581 | Blanchard et al. | May 2006 | A1 |
20060102392 | Johnson et al. | May 2006 | A1 |
20070163045 | Becker et al. | Jul 2007 | A1 |
20070219950 | Crawford | Sep 2007 | A1 |
20080141459 | Hamberg et al. | Jun 2008 | A1 |
20080172789 | Elliot et al. | Jul 2008 | A1 |
20080235872 | Newkirk et al. | Oct 2008 | A1 |
20090153370 | Cooper et al. | Jun 2009 | A1 |
20100039414 | Bell | Feb 2010 | A1 |
20100212087 | Leib et al. | Aug 2010 | A1 |
20110080421 | Capener | Apr 2011 | A1 |
20110144548 | Elliott et al. | Jun 2011 | A1 |
20110162067 | Shuart et al. | Jun 2011 | A1 |
20110169653 | Wang et al. | Jul 2011 | A1 |
20110214234 | Herman | Sep 2011 | A1 |
20120023670 | Zerhusen et al. | Feb 2012 | A1 |
20120089419 | Huster et al. | Apr 2012 | A1 |
20120124744 | Hornbach et al. | May 2012 | A1 |
20120137436 | Andrienko | Jun 2012 | A1 |
20120215360 | Zerhusen et al. | Aug 2012 | A1 |
20120239173 | Laikari et al. | Sep 2012 | A1 |
20130096701 | Suorajaervi et al. | Apr 2013 | A1 |
20130138452 | Cork et al. | May 2013 | A1 |
20130142367 | Berry et al. | Jun 2013 | A1 |
20130219622 | Hornbach et al. | Aug 2013 | A1 |
20130227787 | Herbst et al. | Sep 2013 | A1 |
20130238991 | Jung et al. | Sep 2013 | A1 |
20130300867 | Yoder | Nov 2013 | A1 |
20130318716 | Vanderpohl, III | Dec 2013 | A1 |
20140076644 | Derenne et al. | Mar 2014 | A1 |
20140137025 | Newkirk et al. | May 2014 | A1 |
20140259410 | Zerhusen et al. | Sep 2014 | A1 |
20140265181 | Lambarth et al. | Sep 2014 | A1 |
20140297327 | Heil et al. | Oct 2014 | A1 |
20140313700 | Connell et al. | Oct 2014 | A1 |
20140342330 | Freeman et al. | Nov 2014 | A1 |
20150002393 | Cohen et al. | Jan 2015 | A1 |
20150060162 | Goffer | Mar 2015 | A1 |
20150077534 | Derenne et al. | Mar 2015 | A1 |
20150109442 | Derenne et al. | Apr 2015 | A1 |
20150154002 | Weinstein et al. | Jun 2015 | A1 |
20150250669 | Elliott et al. | Sep 2015 | A1 |
20150317068 | Marka et al. | Nov 2015 | A1 |
20160012218 | Perna et al. | Jan 2016 | A1 |
20160022039 | Paul et al. | Jan 2016 | A1 |
20160038361 | Bhimavarapu et al. | Feb 2016 | A1 |
20160045382 | Goffer | Feb 2016 | A1 |
20160049028 | Kusens et al. | Feb 2016 | A1 |
20160050217 | Mare et al. | Feb 2016 | A1 |
20160065909 | Derenne et al. | Mar 2016 | A1 |
20160095774 | Bobey et al. | Apr 2016 | A1 |
20160140307 | Brosnan et al. | May 2016 | A1 |
20160180668 | Kusens et al. | Jun 2016 | A1 |
20160183864 | Kusens et al. | Jun 2016 | A1 |
20160193095 | Roussy et al. | Jul 2016 | A1 |
20160199240 | Newkirk et al. | Jul 2016 | A1 |
20160247342 | Kusens et al. | Aug 2016 | A1 |
20160296396 | Kolar et al. | Oct 2016 | A1 |
20160324705 | Bach Castillo | Nov 2016 | A1 |
20160338891 | Agdeppa et al. | Nov 2016 | A1 |
20160366327 | Kusens | Dec 2016 | A1 |
20160367420 | Zerhusen et al. | Dec 2016 | A1 |
20160371786 | Kusens et al. | Dec 2016 | A1 |
20170027787 | Huster et al. | Feb 2017 | A1 |
20170027789 | St.John et al. | Feb 2017 | A1 |
20170049642 | Valentino et al. | Feb 2017 | A9 |
20170055113 | Kusens | Feb 2017 | A1 |
20170076526 | Kusens et al. | Mar 2017 | A1 |
20170094477 | Kusens et al. | Mar 2017 | A1 |
20170097800 | Vanderpohl, III | Apr 2017 | A1 |
20170098048 | Brosnan et al. | Apr 2017 | A1 |
20170109770 | Kusens et al. | Apr 2017 | A1 |
20170111770 | Kusens | Apr 2017 | A1 |
20170116790 | Kusens et al. | Apr 2017 | A1 |
20170124844 | Huster et al. | May 2017 | A1 |
20170128296 | Kostic et al. | May 2017 | A1 |
20170143565 | Childs et al. | May 2017 | A1 |
20170193177 | Kusens | Jul 2017 | A1 |
20170193180 | Kusens et al. | Jul 2017 | A1 |
20170193279 | Kusens et al. | Jul 2017 | A1 |
20170193772 | Kusens et al. | Jul 2017 | A1 |
20170195637 | Kusens et al. | Jul 2017 | A1 |
20170213445 | Kusens | Jul 2017 | A1 |
20170224562 | Zerhusen et al. | Aug 2017 | A1 |
20170229009 | Foster et al. | Aug 2017 | A1 |
20170259811 | Coulter et al. | Sep 2017 | A1 |
20170281440 | Puvogel et al. | Oct 2017 | A1 |
20170352212 | Kusens et al. | Dec 2017 | A1 |
20180008052 | Krickeberg et al. | Jan 2018 | A1 |
20180017945 | Sidhu et al. | Jan 2018 | A1 |
20180039743 | Dixon et al. | Feb 2018 | A1 |
20180040091 | Kusens | Feb 2018 | A1 |
20180041864 | Kusens | Feb 2018 | A1 |
20180055418 | Kostic et al. | Mar 2018 | A1 |
20180056985 | Coulter et al. | Mar 2018 | A1 |
20180084390 | Kusens | Mar 2018 | A1 |
20180096550 | Kusens et al. | Apr 2018 | A1 |
20180104122 | Elliott et al. | Apr 2018 | A9 |
20180104123 | Newkirk et al. | Apr 2018 | A1 |
20180110445 | Bhimavarapu et al. | Apr 2018 | A1 |
20180114053 | Kusens et al. | Apr 2018 | A1 |
20180137340 | Kusens et al. | May 2018 | A1 |
20180151010 | Kusens et al. | May 2018 | A1 |
20180161225 | Zerhusen et al. | Jun 2018 | A1 |
20180167816 | Kusens et al. | Jun 2018 | A1 |
20180184984 | Zerhusen et al. | Jul 2018 | A1 |
20180189946 | Kusens et al. | Jul 2018 | A1 |
20180211464 | Kusens et al. | Jul 2018 | A1 |
20180218489 | Kusens | Aug 2018 | A1 |
20180250177 | Magill et al. | Sep 2018 | A1 |
20180271286 | Jacobs et al. | Sep 2018 | A1 |
20180271287 | Jacobs et al. | Sep 2018 | A1 |
20180303687 | Moreno et al. | Oct 2018 | A1 |
20180369035 | Bhimavarapu et al. | Dec 2018 | A1 |
20180369037 | Desaulniers et al. | Dec 2018 | A1 |
20180369038 | Bhimavarapu et al. | Dec 2018 | A1 |
20180369039 | Bhimavarapu et al. | Dec 2018 | A1 |
20180374573 | Bhimavarapu et al. | Dec 2018 | A1 |
20180374577 | Bhimavarapu | Dec 2018 | A1 |
20190008708 | Moreno et al. | Jan 2019 | A1 |
20190024882 | Jonsson et al. | Jan 2019 | A1 |
20190046373 | Coulter et al. | Feb 2019 | A1 |
20190262200 | Murai | Aug 2019 | A1 |
20190336367 | Zerhusen et al. | Nov 2019 | A1 |
20200121528 | Kaikenger et al. | Apr 2020 | A1 |
20200188204 | Newkirk et al. | Jun 2020 | A1 |
20210251579 | Nahavandi et al. | Aug 2021 | A1 |
20210346220 | Bhimavarapu | Nov 2021 | A1 |
20220071823 | Desaulniers et al. | Mar 2022 | A1 |
Number | Date | Country |
---|---|---|
101789230 | Jul 2010 | CN |
19505162 | Mar 1996 | DE |
0727298 | Aug 1996 | EP |
0727298 | Aug 1999 | EP |
2489341 | Aug 2012 | EP |
2531159 | Dec 2012 | EP |
2619724 | Jul 2013 | EP |
2918255 | Sep 2015 | EP |
2003140631 | May 2003 | JP |
20130076922 | Jul 2013 | KR |
0101913 | Jan 2001 | WO |
2006089399 | Aug 2006 | WO |
2011097569 | Aug 2011 | WO |
2012040554 | Mar 2012 | WO |
2014021873 | Feb 2014 | WO |
2015148578 | Oct 2015 | WO |
2015157402 | Oct 2015 | WO |
2015171365 | Nov 2015 | WO |
2016123595 | Aug 2016 | WO |
2016196403 | Dec 2016 | WO |
2016200556 | Dec 2016 | WO |
2017027427 | Feb 2017 | WO |
2017031111 | Feb 2017 | WO |
2017061471 | Apr 2017 | WO |
2017124056 | Jul 2017 | WO |
2017201513 | Nov 2017 | WO |
2018026979 | Feb 2018 | WO |
2018154819 | Aug 2018 | WO |
2018203476 | Nov 2018 | WO |
2018216387 | Nov 2018 | WO |
Entry |
---|
Apple, “Adjust the Brightness on you iPhone, IPad, or IPod Touch”, https://support.apple.com/en-us/HT202613, 2018, 2 pages. |
Astral Healthcare, “Opthalmology Day Surgery Chair Webpage”, Apr. 2018, http://astralhealthcare.com/?product=opthalmology-day-surgery-chair, 6 pages. |
Campbell, Mikey, “Apple Expected to Replace Touch ID With Two-Step Facial, Fingerprint Bio-Recognition Tech”, Apple Insider, Jan. 21, 2017, http://iphone.appleinsider.com/articles/17/01/21/apple-expected-to-replace-touch-id-with-two-step-facial-fingerprint-bio-recognition-tech, 4 pages. |
Doge Medical, “DOC Classic—DOC Surgery Chairs Webpage”, 2014, 2 pages, https://web.archive.org/web/20140214203605/http://www.dogemedical.com/pages/en/products/surgery-chairs/doc-classic.php?lang=EN. |
English language abstract and machine-assisted English translation for CN 101789230 extracted from espacenet.com database on Aug. 30, 2018, 31 pages. |
English language abstract and machine-assisted English translation for JP 2003-140631 extracted from espacenet.com database on Aug. 30, 2018, 19 pages. |
English language abstract and machine-assisted English translation for KR 2013-0076922 A extracted from espacenet.com database on Aug. 16, 2018, 8 pages. |
English language abstract and machine-assisted English translation for WO 2017/061471 extracted from espacenet.com database on Mar. 25, 2019, 26 pages. |
English language abstract and machine-assisted English translation for WO 2018/154819 extracted from espacenet.com database on Mar. 25, 2019, 35 pages. |
English language abstract and machine-assisted English translation for WO 2018/203476 extracted from espacenet.com database on Mar. 25, 2019, 37 pages. |
English language abstract and machine-assisted English translation for WO 2018/216387 extracted from espacenet.com database on Mar. 25, 2019, 43 pages. |
English language abstract for DE 195 05 162 C1 extracted from espacenet.com database on Aug. 16, 2018, 1 page. |
English language abstract for EP 0 727 298 A1 extracted from espacenet.com database on Aug. 16, 2018, 1 page. |
English language abstract for EP 0 727 298 B1 extracted from espacenet.com database on Aug. 16, 2018, 1 page. |
Hall, Stephen, “Nest's 3rd Generation Thermostat Gets Some New Views for Its Farsight Feature”, 9 to 5 Google, Jun. 14, 2016, https://9to5google.com/2016/06/14/nest-3rd-gen-thermostat-views-farsight/, 4 pages. |
Hill-Rom, “Centrella Smart+Bed Brochure” 2017, 11 pages. |
Imore, “How to Use Night Shift on your iPhone or iPad”, video also found at https://www.imore.com/night-shift, Nov. 1, 2017, 12 pages. |
recliners.la “Stellar 550 Large Lift Chair Recliner Webpage”, Apr. 2018, https://www.recliners.la/products/ultra-comfort-stellar-550-large-lift-chair, 4 pages. |
Stryker Medical, “EPIC II Critical Care Bed, Model 2030 Operations Manual”, 2030-309-001 Rev A, Nov. 2007, 53 pages. |
Stryker Medical, “InTouch Critical Care Bed Operations Manual”, Aug. 2014, 125 pages. |
Stryker, “InTouch Critical Care Bed Model FL27 (2130/2140) Operations Manual—Optional Pendant Control”, 2130-009-001 Rev C, Apr. 2008, p. 25. |
Supportec-Trade, “Portfolilio Webpage”, 2017, https://supportec-trade.nl/en, 2 pages. |
Umano Medical Inc., “The Ook Snow Bed Webpages and Video”, https://www.umanomedical.com/en-us/beds/ook-snow?showall=1, 2019, 6 pages. |
U.S. Appl. No. 16/019,973, filed Jun. 27, 2018, 90 pages. |
U.S. Appl. No. 16/019,986, filed Jun. 27, 2018, 57 pages. |
U.S. Appl. No. 16/020,003, filed Jun. 27, 2018, 37 pages. |
U.S. Appl. No. 16/020,052, filed Jun. 27, 2018, 48 pages. |
U.S. Appl. No. 16/020,068, filed Jun. 27, 2018, 125 pages. |
U.S. Appl. No. 16/020,085, filed Jun. 27, 2018, 67 pages. |
U.S. Appl. No. 62/525,359, filed Jun. 27, 2017. |
U.S. Appl. No. 62/525,363, filed Jun. 27, 2017. |
U.S. Appl. No. 62/525,368, filed Jun. 27, 2017. |
U.S. Appl. No. 62/525,373, filed Jun. 27, 2017. |
U.S. Appl. No. 62/525,377, filed Jun. 27, 2017. |
U.S. Appl. No. 62/783,442, filed Dec. 21, 2018. |
U.S. Appl. No. 62/783,445, filed Dec. 21, 2018. |
Youtube, “Memory Seat Escape Video”, Nov. 4, 2013, https://www.youtube.com/watch?v=xlghNmAK-7A, 1 page. |
Youtube, “Microsoft HoloLens: Partner Spotlight with Stryker Communications Video”, Feb. 21, 2017, https://www.youtube.com/watch?v=FTPxUGRGpnA, 3 pages. |
Youtube, “Scion xB 2010 Premium Head Unit with Video”, https://www.youtube.com/watch?v=HHQ9pKUEWvw, Jun. 19, 2010, 2 pages. |
Youtube, “Umano Medical Ook Snow New Hospital Bed (US Version) Video”, Apr. 6, 2015, https://www.youtube.com/watch?v=MEnRNsVFqz0, 1 page. |
U.S. Appl. No. 16/229,108, filed Dec. 21, 2018. |
Number | Date | Country | |
---|---|---|---|
20220071823 A1 | Mar 2022 | US |
Number | Date | Country | |
---|---|---|---|
62525373 | Jun 2017 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 16020003 | Jun 2018 | US |
Child | 17455038 | US |