SYSTEM AND A METHOD OF OPERATION OF THE SYSTEM INCORPORATING A GRAPHICAL USER INTERFACE ON A MOBILE COMPUTING DEVICE FOR A PASSENGER IN A VEHICLE CABIN

Information

  • Patent Application
  • 20160059954
  • Publication Number
    20160059954
  • Date Filed
    January 27, 2014
    10 years ago
  • Date Published
    March 03, 2016
    8 years ago
Abstract
A method of operation for a system incorporating a graphical user interface in a mobile computing device for a passenger within a cabin of an aircraft. The method includes displaying a menu for at least one controllable parameter, receiving a selection of the controllable parameter, displaying at least one control for the selected controllable parameter, receiving a control input for the selected controllable parameter, and adjusting the selected controllable parameter consistent with the control input. The controllable parameter include a plurality of controllable parameters selected from a group encompassing light intensity, light color, temperature, and the degree of openness of at least one window shade. A system and executable computer program product also are provided.
Description
FIELD OF THE INVENTION

The present patent application is directed to system and a method of operation of the system using a graphical user interface on a mobile computing device that is assignable to a passenger in a vehicle cabin (also referred to herein as a “passenger GUI,” “passenger input/output node,” or “passenger IO node”). The passenger IO node provides control over one or more functions within the cabin of the vehicle. The vehicle is contemplated to be an aircraft.


DESCRIPTION OF THE RELATED ART

As should be apparent to those skilled in the art, there are a number of functions that may be controlled within the cabin of an aircraft. The functions may be divided into at least two categories: (1) functions related to environment, and (2) functions related to passenger comfort and entertainment.


Environmental functions include, but are not limited to, things such as cabin temperature, the intensity of the cabin lighting, and the degree to which the window shades are open, among other variables.


Functions related to passenger comfort include those related to actuation of a personal reading light, control over the air flow through an overhead vent, positioning of the passenger seat (i.e., upright or reclined), and a remote call for a flight attendant (i.e., a flight attendant call button).


Other functions that are associated with passenger comfort include, but are not limited to control over media type (i.e., audio and/or video), content, and volume. With respect to content, selectivity may be provided so that a passenger may select a genre of music (i.e., jazz music or pop music) or a genre of movies (i.e., comedy or drama), among other variations. As should be apparent to any passenger, individuals may control the volume of the media that has been selected.


At present, selected environmental functions typically are adjusted by the flight crew for the comfort of all passengers within the aircraft. For example, temperature typically is controlled at a central location within the aircraft cabin, via a thermostat or similar temperature control device. Similarly, the main cabin lighting in the aircraft typically is controlled via a central panel available to the flight crew. As a result, the flight crew can turn on, turn off, or dim the main lights within the aircraft cabin for all of the passengers.


As should be apparent to the airplane traveler, functions associated with passenger comfort and entertainment typically are accessible directly from the passenger's seat.


This basic operational approach to aircraft cabin functions has been employed for many years. As presently configured, the control systems for the environment and for passenger comfort and entertainment within an aircraft operate independently from one another.


Recently, a desire has developed to improve the manner in which aircraft cabin functions are controlled. Specifically, a desire has arisen to develop controls for one or more functions within the cabin of an aircraft from one or more consolidated IO nodes.


SUMMARY OF THE INVENTION

The present invention provides a GUI and a method of operation of a GUI that is available to a passenger via a mobile computing device.


In one contemplated embodiment, the mobile passenger IO node provides control to passengers over one or more functions within an aircraft cabin regardless of the location of the passenger within the cabin.


The present invention provides for a method of operation for a system incorporating a graphical user interface in a mobile computing device for a passenger within a cabin of an aircraft. The method includes displaying a menu for at least one controllable parameter, receiving a selection of the controllable parameter, displaying at least one control for the selected controllable parameter, receiving a control input for the selected controllable parameter, and adjusting the selected controllable parameter consistent with the control input. The controllable parameter include a plurality of controllable parameters selected from a group encompassing light intensity, light color, temperature, media content, media volume, and the degree of openness of at least one window shade.


It is contemplated that the method also may include placing the graphical user interface into a sleep mode if selection of a controllable parameter is not received.


With respect to the method, it is contemplated that the plurality of controllable parameters may be associated with at least one of the entire cabin of the aircraft, at least one zone within the cabin of the aircraft, or at least one seat within the cabin of the aircraft.


Moreover, the plurality of controllable parameters may be controllable via an interface presenting an isometric view of at least a portion of the cabin of the aircraft.


In an alternative embodiment, it is contemplated that the method may include prioritizing the control input received from the mobile computing device in relation to control inputs received from any other input device, thereby avoiding conflicts between the control inputs.


For the method, it is contemplated that light intensity, temperature, the degree of openness of the at least one window shade, and media volume may be adjustable between a predetermined minimum and a predetermined maximum. In addition, light color may be adjustable between a predetermined warm color and a predetermined cool color.


It is contemplated that the media content may include a video library, an audio library, and a map view. A map view is contemplated to encompass a global map view and a local map view.


Concerning the method, it is contemplated that the displaying of the menu for the controllable parameter includes displaying a light icon, a media icon, a thermostat icon, and a window shade icon.


The method also is contemplated to encompass embodiments were the plurality of controllable parameters also include at least one of media type, media, content, and media volume.


For the method of the present invention, the passenger device may be at least one of a personal computer, tablet, and smartphone.


The present invention also contemplates a system incorporating a graphical user interface in a mobile computing device for a passenger within a cabin of an aircraft. The system includes a first display for displaying at least one controllable parameter, an input for receiving a selection of the controllable parameter, a second display for displaying at least one control for the selected controllable parameter, wherein the input receives a control input for the selected controllable parameter, and a controller for adjusting the selected controllable parameter consistent with the control input. The controllable parameter includes a plurality of controllable parameters selected from a group encompassing light intensity, light color, temperature, and the degree of openness of at least one window shade.


With respect to the system, it is contemplated that the plurality of controllable parameters also include at least one of media type, media, content, and media volume and that the plurality of controllable parameters may be associated with at least one of the entire cabin of the aircraft, at least one zone within the cabin of the aircraft, or at least one seat within the cabin of the aircraft.


In addition, with respect to the system, the plurality of controllable parameters may be controllable via an interface presenting an isometric view of at least a portion of the cabin of the aircraft.


Also concerning the system, it is contemplated that the control input received from the mobile computing device may be prioritized in relation to control inputs received from any other input device, thereby avoiding conflicts between the control inputs.


The present invention also provides for an executable computer program product providing instructions for operation of a system incorporating a graphical user interface for a mobile computing device for a passenger device within a cabin of an aircraft. The instructions include displaying a menu for at least one controllable parameter, receiving a selection of the controllable parameter, displaying at least one control for the selected controllable parameter, receiving a control input for the selected controllable parameter, and adjusting the selected controllable parameter consistent with the control input. The controllable parameter includes a plurality of controllable parameters selected from a group that may include light intensity, light color, temperature, and the degree of openness of at least one window shade.


It is contemplated that the executable computer program product will operate such that the plurality of controllable parameters also include at least one of media type, media, content, and media volume and that the plurality of controllable parameters are associated with at least one of the entire cabin of the aircraft, at least one zone within the cabin of the aircraft, or at least one seat within the cabin of the aircraft.


In addition, the program product is contemplated to function such that the plurality of controllable parameters is controllable via an interface presenting an isometric view of at least a portion of the cabin of the aircraft.


Concerning the executable computer program product, the instructions also may include prioritizing the control input received from the mobile computing device in relation to control inputs received from any other input device, thereby avoiding conflicts between the control inputs.


Still further aspects of the present invention will be made apparent from the drawings and description that follows.





BRIEF DESCRIPTION OF THE DRAWINGS

The present invention will now be described in connection with the figures appended hereto, in which:



FIG. 1 is a graphical overview of one embodiment of a distributed architecture with which the passenger IO node of the present invention is contemplated to cooperate;



FIG. 2 is a graphical overview of a second embodiment of a distributed architecture with which the passenger IO node of the present invention is contemplated to cooperate;



FIG. 3 is a graphical, top view of a portion of an aircraft, depicting one possible configuration for an aircraft cabin that employs the passenger IO node of the present invention;



FIG. 4 is a perspective illustration of a portion of a cabin of an aircraft, showing one contemplated position for the passenger IO node of the present invention;



FIG. 5 depicts one contemplated embodiment of a main menu displayable on the passenger IO node of the present invention;



FIG. 6 illustrates features of a video submenu displayable on the passenger IO node of the present invention;



FIG. 7 is a search GUI accessible from the video submenu that is displayable on the passenger IO node of the present invention;



FIG. 8 is a viewing options GUI that presents control options for the viewing of video programming, the viewing options GUI being displayable on the passenger IO node of the present invention;



FIG. 9 is one contemplated embodiment of an audio submenu that is displayable on the passenger IO node of the present invention;



FIG. 10 depicts one possible television submenu that is displayable on the passenger IO node of the present invention;



FIG. 11 provides one contemplated map view GUI that is displayable on the passenger IO node of the present invention;



FIG. 12 illustrates a local map GUI contemplated to be displayable on the passenger IO node of the present invention;



FIG. 13 depicts an embodiment of a cabin light GUI that may be displayed on the passenger IO node of the present invention;



FIG. 14 depicts one contemplated embodiment of a window shades GUI that may be displayed on the passenger IO node of the present invention;



FIG. 15 provides a thermostat GUI contemplated for use with the passenger IO node of the present invention;



FIG. 16 illustrates a presets GUI that is contemplated for use with the passenger IO node of the present invention;



FIG. 17 depicts a seat selector GUI that is contemplated to be displayed on the passenger IO node of the present invention;



FIG. 18 illustrates a change seat GUI that is contemplated for use with the passenger IO node of the present invention;



FIG. 19 is a flow chart that illustrates a first method contemplated for use with the passenger IO node of the present invention; and



FIGS. 20-34 are flow charts that collectively illustrate a second method contemplated for use with the passenger IO node of the present invention.





DETAILED DESCRIPTION OF EMBODIMENT(S) OF THE PRESENT INVENTION

The present invention will now be described in connection with one or more embodiments. The discussion of any one embodiment is not intended to be restrictive or limiting of the present invention. To the contrary, the embodiments described are intended to be illustrative of the broad scope of the present invention.


Among other aspects, the present invention addresses controls for parameters on board an aircraft including environmental functions and functions related to passenger comfort. As noted above, environmental functions include, but are not limited to, things such as cabin temperature, the intensity of the cabin lighting, and the degree to which the window shades are open, among other variables. Functions related to passenger comfort include those related to actuation of a personal reading light, control over the air flow through an overhead vent, positioning of the passenger seat (i.e., upright or reclined), and a remote call for a flight attendant (i.e., a flight attendant call button). Other functions that are associated with passenger comfort include, but are not limited to control over media type (i.e., audio and/or video), content, and volume. With respect to content, selectivity may be provided so that a passenger may select a genre of music (i.e., jazz music or pop music) or a genre of movies (i.e., comedy or drama), among other variations. Individuals may control the volume of the media that has been selected.


As should be apparent, and as will be made more apparent in the discussion that follows, the labels “environment” and “passenger comfort” when applied to specific functions that are controllable in an aircraft are merely provided to assist with an understanding of the present invention. Use of either of the labels is not intended to be limiting, as the labels are not considered to be mutually exclusive of one another or of other functions that are not highlighted herein. For example, control over the degree to which the window shades are opened qualifies as control over an environmental function and also over aspects of passenger comfort. The lights in the aircraft belong to the same, crossover category.


With respect to the present invention, the terms “front” (or “fore”), “rear” (or “aft”), left (or “port”), and right (or “starboard”) are used in the conventional fashion when referring to an aircraft. These conventions refer to the front, rear, left, and right sides of an aircraft as determined by its normal, forward direction of travel.


In addition, reference is made to members of the flight crew on board the aircraft. The term “flight crew” is intended to be generic to any member of the flight crew, including the pilot, co-pilot, and/or flight attendants. In other words, the term “flight crew” is intended to refer to persons other than passengers on board the aircraft.


The term “bulkhead” is used in the discussion of the present invention. A bulkhead is wall that is disposed within the aircraft. A bulkhead may or may not be a structural component of the aircraft.


It is contemplated that the passenger IO node (or passenger GUI) of the present invention may be provided on a corporate or private aircraft. In other words, it is contemplated that the present invention may be employed in an aircraft that typically has limited seating by comparison with a commercial, passenger aircraft. While corporate, business, or personal aircraft encompass the primary focus of the passenger IO node of the present invention, the present invention is not limited just to such aircraft. To the contrary, the present invention may be employed in any aircraft, including commercial passenger aircraft, without departing from the scope of the present invention.


In addition, while the passenger IO node of the present invention is contemplated to be employed on an aircraft, it is noted that the present invention may be employed in any other suitable environment. For example, the present invention may be practiced on a passenger car of a train, on board a ship, or any other suitable environment that should be apparent to those skilled in the art.


It is contemplated that the passenger IO node of the present invention will be used in conjunction with a distributed architecture 10, one embodiment of which is illustrated in FIG. 1. The distributed architecture includes a central processing unit 12 (“CPU”) that includes a processor 14 and a controller 16. The CPU 12 may be a computer, as should be apparent to those skilled in the art. However, the term CPU 12 is not intended to be limited only to a computer or any part thereof. To the contrary, the term CPU 12 is intended to encompass any type of computing device that may operate to provide the functionality described herein.


The term “processor” is intended to broadly encompass any device capable of executing machine-readable instructions. In other words, the term “processor 14” is intended to refer to any device or component that processes instructions and data. As an example, semiconductor chips within a computer are considered to fall within the definition of the term “processor 14.”


While it is contemplated that the processor 14 will be a single component of the distributed architecture 10, the distributed architecture 10 is not intended to be limited solely to such a construction. The processor 14 may include multiple devices that are separate from one another, but cooperate together to process data and execute instructions. For example, the processor 14 may include a semiconductor processing chip and/or any other peripheral devices that support the operation of the semiconductor processing chip. Alternatively, the processor 14 may encompass processing chips that are located in separate systems, but which are operatively connected to provide the desired functionality.


As also illustrated in FIG. 1, the CPU 12 includes a controller 16. In one embodiment, it is contemplated that the controller 16 may be a hardware component that is separate from the processor 14. In a second contemplated embodiment, the controller 16 may be embodied in software (i.e., operating software) that runs on the central processing unit 12. In other words, in this second embodiment, the processor 14 may be the device on which the controller 16 is executed. In a third contemplated embodiment, the controller 16 may be a combination of hardware and software. Regardless of whether the controller 16 is hardware, software, or a combination of the two, it is contemplated that the controller 16 will facilitate communication between the processor 14 and any input/output (“IO”) and/or peripheral devices connected thereto. The peripheral devices include the side ledge IO node of the present invention.


While the distributed architecture 10 is described in terms of a CPU 12, a processor 14, and a controller 16 (among other components), it is noted that this configuration is not intended to be illustrative of the breadth of the present invention. The configuration is not intended to exclude any possible server/client configurations. For example, the CPU 12 may be a server on which a client is resident. The controller 16 may be the client. In another configuration, the CPU 12 may be a server that provides access to an independent client. In still another configuration, the CPU 12 may be a router.


As should be apparent, there are many appellations that may be applied to the components comprising the distributed architecture 10. Those variations and equivalents are intended to be encompassed by the scope of the present invention.


As illustrated in FIG. 1, the processor 14 may connect to one or more databases 18. The database 18 may be a memory storage device, an IO device such as an MP3 player, a compact disc (“CD”) player, a digital video disk (“DVD”) player, or any other suitable storage and playback device. To emphasize the breadth of what is meant by the term, the database 18 may include, but is not limited to, any suitable memory on which the CPU 12 relies for its operation. The term database 18 should not be understood to be limited solely to memory devices.


It is noted that the distributed architecture 10 contemplated for use with the passenger IO node of the present invention also may be connected to other systems and processors on board the aircraft. For example, the distributed architecture 10 may receive input from a flight computer on board the aircraft. These other input devices are not illustrated for simplicity. It is noted, however, that other inputs may be provided to the distributed architecture 10, as should be apparent to those skilled in the art.


The distributed architecture 10 is intended to be specific to the passengers and flight crew on an aircraft. As a result, the CPU 12 is contemplated to connect to at least two IO nodes: (1) a passenger IO node 20 and (2) a crew IO node 22. The passenger IO node 20 receives input from and provides output to the passenger. The crew IO node 22 receives input from and provides output to members of the flight crew. Both the passenger IO node 20 and the crew IO node 22 connect to the controller 16, through which selected inputs and outputs are directed.


The passenger IO node 20 is contemplated to encompass any suitable input/output device that may be available to a passenger. Similarly, the crew IO node 22 is intended to encompass any suitable input/output device that may be available to a member of the flight crew. In other words, while the present invention will be described in connection with specific devices, the present invention is not intended to be limited thereby. Other devices may be provide or substituted for the devices described herein without departing from the scope of the present invention.


In addition, as will be made more apparent in the discussion that follows, the passenger IO node 20 and the crew IO node 22 are contemplated to provide overlapping functionality. Therefore, the discussion of a particular functionality with respect to one IO node 20, 22 does not preclude the same functionality from being provided via the other of the IO nodes 20, 22.


As illustrated in FIG. 1, the various components of the distributed architecture 10 connect to one another via communication lines 24. The communication lines 24 may be wired or wireless communication lines, as should be apparent to those skilled in the art. Wired communication lines encompass, but are not limited to, wired connections and docking stations (for one or more of the IO nodes). Wireless communication lines may be provided via any suitable data format including, but not limited to, a Bluetooth™ connection (where appropriate).


Additionally, the communication lines are illustrated as two-way communication channels. While depicted as two-way communication channels, it is noted that one-way communication channels may be employed without departing from the scope of the present invention. In addition, it is also contemplated that the communication channels 24 may encompass one or more busses that channel multiple channels of communication along a single communication line 24.



FIG. 2 illustrates a second embodiment of a distributed architecture 26 contemplated for use with the bulkhead IO node of the present invention. As will be made apparent from the discussion that follows, the second embodiment of the distributed architecture 26 may be considered as a variation of the first embodiment.


The distributed architecture 26 is directed to a location-oriented approach rather than a person-oriented approach, as detailed in connection with the distributed architecture 10. The person-oriented approach that is employed for the distributed architecture 10 encompasses an architecture where an IO node is associated with an individual, such as a passenger or a member of the flight crew. The location-oriented approach for the distributed architecture 26 encompasses an architecture that relies, at least in part, on IO nodes that are placed at specific locations with the aircraft.


As will be made apparent in discussion that follows, there is an overlap between the first distributed architecture 10 and the second distributed architecture 26.


As illustrated in FIG. 2, the second distributed architecture 26 is similar to the first distributed architecture in that the distributed architecture 26 includes the CPU 12, the processor 14, the controller 16, and the database 18. The second distributed architecture 26 differs from the first distributed architecture 10 in that additional IO nodes are provided at specific locations within the aircraft cabin, as noted above.


As illustrated in FIG. 2, the second distributed architecture is contemplated to include the passenger IO node 20 and the crew IO node 22. In addition, the second distributed architecture 26 includes a bulkhead IO node 28, a side ledge IO node 30, a table IO node 32, and a window IO node 34. Details of the bulkhead IO node 28, the side ledge IO node 30, the table IO node 32, and the window IO node 34 are provided below.


As suggested by the nomenclature employed, the IO nodes 28, 30, 32, 34 are provided at specific locations in the aircraft. The person-specific IO nodes 20, 22 are contemplated to be portable devices that are associated with individuals and, as such, are not associated with any fixed structure within the aircraft.


As illustrated in FIGS. 1 and 2, the IO nodes 20, 22, 28, 30, 32, 34 connect to the controller 16. The controller is contemplated to incorporate a hierarchical command structure that prioritizes input(s) from the different IO nodes 20, 22, 28, 30, 32, 34. For example, the controller 16 may include a hierarchical command structure where input(s) provided by a crew member override (or nullify) input(s) provided by a passenger. In another contemplated scenario, input(s) provided at one of the IO nodes 20, 22, 28, 30, 32, 34 may be given priority over any other input(s). For example, a crew member may have closed the window shades in the aircraft so that the passengers may enjoy in-flight entertainment. A passenger may wish to open his or her window shade via the window IO node 34. So that the passenger may do this, input(s) from the window IO node 34 may be placed at the top of the hierarchical command tree. Still further, the owner or operator of the aircraft may set the hierarchical command structure for the individual aircraft or a fleet of aircraft, as required or as desired.


It is noted that the window IO node 34 and the table IO node 32 are but two examples of nodes where limited space is available for control inputs and/or outputs. The present invention should not be understood to be limited to the nodes 32, 34 that are shown and described herein.


To facilitate the discussion of the distributed architectures 10, 26, a top view of an aircraft 36 is illustrated in FIG. 3. The aircraft 36 that is depicted is merely exemplary of the infinite possible configurations that are possible and should not be understood to be limiting of the configurations with which the side ledge IO node of the present invention is contemplated to operate.


As illustrated in FIG. 3, the aircraft 36 has a front end 38, a rear end 40, a left side 42, and a right side 44. The fuselage 46 of the aircraft 36 defines a cabin 48 therein. The layout of the cabin 48 illustrated in FIG. 3 may be provided for a corporate, business, or personal aircraft, such as a private jet.


The cabin 48 includes a cockpit 50, a galley 52, and a passenger area 54. The cabin 48 also includes a forward lavatory 56, a first passenger seating area 58, a second passenger seating area 60, a third passenger seating area 62, a first bedroom 64, a second bedroom 66, and an aft lavatory 68.


The first passenger seating area 58 is positioned adjacent to the galley 52 and the forward lavatory 56. The first passenger seating area 58 is immediately aft of the door 70 that provides ingress into and egress out of the aircraft 36. A first bulkhead 72 separates the area adjacent to the door 70 from the first passenger seating area 58.


The first passenger seating area 58 is defined by one passenger seat 74 and a stowable table 76. The passenger seat 74 is contemplated to be a reclining seat. However, the passenger seat 74 need not recline. The stowable table 76 is contemplated to be stowable in a side compartment adjacent to the passenger seat 74. As required by applicable aviation laws, the table 76 must be stowed for taxi, take-off, and landing.


It is noted that the first passenger seating area 58 may be reserved for one or more crew members and, therefore, be understood to be a crew seating area 58. Since the type of individual that uses the seating area 58 is not critical to operation of the present invention, the seating area 58 will be referred to herein as the first passenger seating area 58. It is also noted that, while other seating areas are indicated as being for passengers, crew members may use these areas together with the passengers.


A second bulkhead 78 separates the first passenger seating area 58 and forward lavatory 56 from the second passenger seating area 60.


The second passenger seating area 60 includes four passenger seats 74 that are positioned on opposite sides of a central aisle. Two seats 74 face one another across a table 76 on the right side 44 of the aircraft 36. Similarly, two seats 74 face one another across a stowable table 76 on the left side 42 of the aircraft.


The third passenger seating area 62 is defined by six passenger seats 74, a stowable table 76, and a stowable conference table 80. Two seats 74 face one another across the stowable table 76 on the right ride 44 of the aircraft 36. Four seats 74 face one another (in two pairs) across a stowable conference table 78. As illustrated, when the tables 76, 80 are deployed, they are contemplated to form a single conference table that extends across the width of the cabin 48.


As is apparent from FIG. 3, the second seating area 60 and the third seating area 62 are not separated from one another by any bulkhead or other barrier. Instead, these passenger areas 58, 60 are contemplated to form a continuous passenger area within the cabin 48.


The first bedroom 64 is separated from the third passenger seating area 62 by a third bulkhead 82. The first bedroom 64 includes a divan 84 on the left side 42 of the aircraft 36 and a cabinet 86, such as a media cabinet, on the right side 44 of the cabin 48. It is contemplated that the divan 84 will function both as a couch (or a sofa) and a bed, depending upon its use or configuration.


The second bedroom 66 is separated from the first bedroom 64 by a fourth bulkhead 88. The second bedroom 66 includes a divan 84 on the right side 44 of the aircraft 36. A seat 74 and stowable table 76 are provided on the left side 42 of aircraft 36. Also on the left side 42 is a cabinet 90, which may be provided with a media center, including a monitor or a television.


A fifth bulkhead 92 separates the second bedroom 66 from the rear lavatory 68.


It is noted that the fuselage 46 includes a plurality of windows 94.


In addition, at least four monitors 96 (i.e., video output screens) are provided in the aircraft 36 at various locations. The monitors 96 are contemplated to be positioned to provide video information and entertainment to the passengers in the aircraft 36. It is contemplated that entertainment also may be provided to the passengers via entertainment devices that are associated with the passenger seats 74.


As illustrated, the cabin 48 also includes several side ledges 98 that extend along the length of selected ones of the passenger seating areas 58, 60, 62. Where they are provided, the side ledges 98 are disposed between the passenger seat 74 and the wall of the fuselage 46. As is apparent from FIG. 3, the side ledges 98 are provided in the first passenger seating area 58 and the second passenger seating area 60. While side ledges 98 are not illustrated for the third passenger seating area 62, side ledges 98 may be provided in this seating area without departing from the scope of the present invention.


It is noted that the term “side ledge” is intended to encompass other furniture within the cabin 48 of the aircraft 36 in addition to the typical side ledge 98 that is identified in FIG. 3. Specifically, a cabinet or side ledge 98 may be provided adjacent to the divan 84 in the aircraft 36. While such a side ledge 98 would extend transversely to the travel direction of the aircraft 36, the side ledge 98 may be provided with control functionality. In addition, if the aircraft 36 were to include a bed with night stands, the night stands would be considered as side ledges 98 for purposes of the present invention.


As should be apparent to those skilled in the art, the configuration for the cabin 48 of the aircraft 36 that is provided in FIG. 3 is merely exemplary of the many possible configurations that may be employed in the cabin 48 of the aircraft 36. In other words, the present invention should not be understood to be limited to use on aircraft 36 with the configuration depicted in FIG. 3.


With renewed reference to the distributed architectures 10, 26, either architecture 10, 26 (or any variant thereof) may be employed onboard the aircraft 36. For purposes of the discussion herein, the aircraft 36 includes the second distributed architecture 26.


In this architecture, the passenger IO node 20 is contemplated to be a mobile electronic device, as discussed above. Mobile electronic devices include, but are not limited to, portable computers, tablets, and smartphones. As will be made apparent from the discussion that follows, it is contemplated that the passenger IO node 20 will be capable of receiving and storing a software program, such as an “app.” The app may be specific to a particular aircraft or airline, as required or desired. The app is contemplated to provide the software needed for proper interface with the controller 16 for operation of the distributed architecture 26. In other words, the software resident on the passenger IO node 20 is contemplated to be configured to provide input to the CPU 12 and to receive output from the CPU 12.


The crew IO node 22 also is contemplated to be a mobile device, such as a portable computer, tablet, or smartphone. As with the passenger IO node 20, the crew IO node 22 is contemplated to be provided with a suitable app (or resident software) for interface with the CPU 12.


Where the mobile IO nodes 20, 22 are tablets (or other suitable electronic devices), it is contemplated that the tablets 20, 22 will be provided with the delivery to the customer of the aircraft 36. In this embodiment, when a passenger boards the aircraft 36, the passenger will be assigned one of the mobile devices for use during the flight.


Alternatively, it is contemplated that a passenger may bring his or her own mobile device on board the aircraft 36. If so, the passenger (and/or crew member) may be prompted to download suitable software (i.e., the app) for interface with the controller 16 prior to boarding the aircraft. In a further contemplated embodiment, the passenger (and/or crew member) may be prompted to download suitable software after boarding the aircraft, for example. It is noted that the apps (i.e., the software) downloaded by the passenger and the crew may be the same or may be separate apps, as required or as desired.


As also discussed above, the aircraft 36 may include additional IO nodes.


As noted above, the passenger IO node 20 is the focus of the present invention. While the passenger IO node 20 is contemplated to be embodied is an electronic tablet device with a touch-sensitive surface, the passenger IO node 20 may be any other suitable alternative device without departing from the scope of the present invention. Moreover, while the present invention is described as a mobile device, meaning that it is not structurally secured to the aircraft 36, the passenger IO node 20 may be affixed in the aircraft 36 without departing from the scope of the present invention.



FIG. 4 provides a perspective illustration of a portion of an interior of the cabin 48 of an aircraft 36 that incorporates the passenger IO node 20. The passenger IO node is illustrated as a mobile computing device, such as a touch-sensitive tablet 130. Also provided in FIG. 4 is a second contemplated embodiment of the passenger IO node 20, which is a retractable knob 132 that is disposed in the side ledge 98. The retractable knob 132 is contemplated to provide at least some of (if not all of) the functionality of the tablet 130. The details of the retractable knob 132 are not the focus of the present invention and, therefore, specific details concerning the retractable knob 132 are not provided herein.


In this illustrated embodiment, the passenger IO node 20 is disposed on a retractable stand 134 that extends from the side ledge 98 adjacent to the passenger seat 74. The passenger IO node 20 is removably disposed in the stand 134. In other words, the passenger IO node 20, as embodied in the tablet 132, is not integrally connected to the stand 134. A table 76 also is illustrated in this view, to provide context for the present invention.


As should be apparent, the stand 134 need not extend from the side ledge 98. It is contemplated that the stand 134 may extend from one of the arm rests on the passenger seat 74. Alternatively, the stand 134 may extend from the table 76. As should be apparent to those skilled in the art, the exact location where the stand 134 is positioned is not critical to the present invention.


As noted above, functions associated with passenger comfort fall into two general categories: (1) media functions and (2) cabin-related environmental functions. As such, the passenger IO node 20 is contemplated to provide an interface to the user that includes these two groups of functions. As noted above, these appellations should not be considered to be limiting of the present invention.


It is noted that the term “user” is employed to refer to passengers and flight crew members, since both categories of persons are contemplated to be users of the present invention. As such, where the term “passenger” or “flight crew member” are used, the term is not intended to exclude use by any other user, as required or as desired.



FIG. 5 depicts one contemplated embodiment of a main menu 136 that is contemplated to be displayed, as a root menu, on the passenger IO node 20. The main menu 136 includes a media submenu 138 and a cabin submenu 140. Submenu icons and words (both of which are referred to as “icons” herein whether they are words or pictograms) are selectable via the touch interface on the tablet 130.


The media submenu 138 includes four options: (1) a video icon 142, (2) a music icon 144, (3) a television icon 146, and (4) a map view icon 148. Each of these separate options is accessible by touching the surface 150 of the passenger IO node 20. As should be apparent, the icons 142-148 that are available via the media submenu 138 are merely representative of the types of media that may be accessible by that menu.


The cabin submenu 136 includes nine options: (1) a cabin light icon 152, (2) a window shade icon 154, (3) an audio icon 156, (4) a thermostat icon 158, (5) a video icon 160, (6) a presets icon 162, (7) a table light icon 164, (8) a reading light icon 166, and (9) a seat icon 168. Each of these separate options is available by touching the surface 150 of the passenger IO node 20. As with the media submenu 138, the icons 152-168 that are included in the cabin submenu 140 are intended to be exemplary of the types of icons that may be available through the cabin submenu 140.


As should be apparent, the media submenu 138 and the cabin submenu 140 do not present mutually exclusive functionalities. Some functions with the cabin 48 of the aircraft 36 may be accessed from either submenu 138, 140. In other words, the menu trees for both submenus 138, 140 are contemplated to be interrelated and redundant.


The main menu 136 also includes a flight status bar 170, which extends along a top edge of the main menu 136. The flight status bar 170 provides a visual indication of the total duration of the flight, time elapsed since take off, and time remaining until landing. As should be apparent, the flight status bar 170 may provide additional information that may be of interest to the passenger.


The four icons in the media submenu 138 provide access to the four types of entertainment that are available to the passenger on board the aircraft 36.


The video icon 142 provides access to a listing of the video entertainment available to the passenger on board the aircraft 36 as well as other functionality, as discussed below.


The audio icon 144 provides access to a listing of the audio (i.e., music) entertainment available to the passenger on board the aircraft 36. Other functionality also may be made available via the audio icon 144, as discussed herein.


The television icon 146 provides access to a listing of the television programming that may be available to the passengers. Television programming is contemplated to encompass pre-recorded content. However, it is contemplated that television programming also may include real-time television programming for aircraft 36 that are equipped to receive television programming during flight.


In one contemplated embodiment, the map view icon 148 is contemplated to provide a view of the geographic position of the aircraft 36. As such, the passenger may identify where the aircraft 36 is in its flight plan. The map view icon 148 also is contemplated to permit access to local geographic maps so that the passenger may locate geographic points of interest, for example, at the destination location.


The cabin light icon 152 is intended to provide access to control over the main lighting in the cabin 48 of the aircraft 36. The main lighting in the cabin 48 is the overhead lighting and is the lighting in the general passenger area of the aircraft 36. The main cabin lighting in the aircraft 36 is distinguishable from other lighting that may be provided, such as a personal reading light, positioned over the passenger's seat 74 or a table reading light positioned over a table 76, 80 within the aircraft 36.


The window shade icon 154 provides control over one or more of the window shades that cover the windows 94 in the aircraft 36. The window shade icon 154 provides control over the degree to which the window shades in the aircraft 36 are opened or closed.


With respect to the window shades, it is noted that the window shades may be of any particular type without departing from the scope of the present invention. For example, the window shades may be made from a sheet of material that moves (via a motor, for example) in front of the window to block the transmission of light therethrough. Alternatively, the window shades may be made from an electrochromic material. Electrochromic materials respond to signals by altering their color and/or opacity.


The audio icon 156 is similar to the audio icon 144, by providing access to the audio menu, as discussed further herein.


The thermostat icon 158 provides access to a menu that permits the passenger to control the temperature within the cabin 48 of the aircraft 36.


The video icon 160 is similar to the video icon 142. This icon also provides access to the functionality of the video menu, as discussed further herein.


The presets icon 162 provides access to predetermined settings related to the cabin 48 of the aircraft 36. By accessing the presets icon 162, the passenger may select from several preset environments within the aircraft to facilitate activities such as sleep, meetings, or entertainment viewing, as discussed below.


The table light icon 164 provides control over a light that may be positioned above a stowable table 76 or a conference table 80, as may be provided in the cabin 48 of the aircraft 36.


The reading light icon 166 provides access to control over one or more reading lights above the passenger seats 74 in the cabin 48.


The seat icon 168 provides control over the comfort position of one or more of the seats 74 in the aircraft 36. Via the seat icon 168, the passenger may adjust the seat 74 between fully upright and fully reclined positions.



FIG. 6 illustrates one contemplated embodiment of a video submenu 172 according to the present invention. If the passenger accesses the video icon 142 on the main menu 136, the passenger will be directed to the video submenu 172. In this illustration, the video submenu 172 encompasses movies that are available to the passenger. However, the video submenu 172 should not be understood to be limited solely to movie content.


The video submenu 172 includes at least four separate regions, each of which provides access to different, related functionality.


As shown, the video submenu 172 includes a media bar 174 that provides access to the different types of media that are available to the passenger. Since the passenger originally selected the video icon 142, the video submenu 172 defaults to the video programming available to the user. The media bar 174 permits the passenger to change to a different media selection without having to return to the main menu 136.


The video submenu also includes an available devices section 176, a search bar section 178, and a library section 180.


The available devices section 176 provides a listing of the various video devices (i.e., the monitors 96) that are accessible on the aircraft. By selecting one or more of the icons associated with the available video devices 96, the passenger may select which of the monitors 96 will display the selected video content. For example, the passenger may elect to have a selected movie played on a nearby monitor 96 as well as a remote monitor in one of the bedrooms 64, 66. In this manner, the passenger may watch a movie from the passenger's seat 74 while his or her children watch the same movie in their bedroom 64, for example.


The search bar section 178 is provided so that the passenger may input search words to locate specific video media within the library on board the aircraft 36.


The library section 180 provides a listing of all of the video content that is available to the passenger.



FIG. 7 illustrates a search GUI 182 that may appear if the user wishes to access the search bar section 178. The search GUI 182 displays a touch-sensitive keyboard 184 so that the passenger may input key words for initiation of a search of the video library, a portion of which may remain visible in the library section 180.



FIG. 8 is a viewing options GUI 186 that may be presented to the passenger after specific video content has been selected for viewing. The viewing options GUI 186 includes a viewing area submenu 188 and a sound options submenu 190. The viewing area submenu 188 allows the passenger to select one or more devices (i.e., one or more tablets 130 and/or one or more monitors 96) where the selected video is to be shown. As suggested by the viewing area submenu 188, the cabin 48 of the aircraft 36 may be separated into various zones, consistent with the seating areas 58, 60, 62 and the bedrooms 64, 66. As a result, the passenger may control the video being displayed in one or more zones within the aircraft 36. The sound options submenu 190 permits the audio portion of the video content to be played via headphone on the armrest of the seat 74 or via speakers on the tablet 132 or speakers within the cabin 48 of the aircraft 36. As indicated, the passenger may control the sound that is played in one or more zones within the cabin 48 of the aircraft 36.



FIG. 9 illustrates one contemplated embodiment of an audio submenu 192. The audio submenu is patterned similarly to the video submenu 172. The same options are accessible via the audio submenu 192.


If the passenger accesses the audio icon 144 on the main menu 136, the passenger will be directed to the audio submenu 192. In this illustration, the audio submenu 192 encompasses audio programs that are available to the passenger. However, the audio submenu 192 should not be understood to be limited solely to movie content.


The audio submenu 192 includes at least four separate regions, each of which provides access to different, related functionality.


As shown, the audio submenu 192 includes the media bar 174 that provides access to the different types of media that are available to the passenger. Since the passenger originally selected the audio icon 144, the audio submenu 192 defaults to the audio programming available to the user. The media bar 174 permits the passenger to change to a different media selection without having to return to the main menu 136.


The audio submenu 192 also includes an available devices section 176, a search bar section 178, and a library section 180.


Submenus of the audio submenu 192 are contemplated to operate in the same manner as the viewing options GUI 186, discussed above. Specifically, audio programming may be played on one or more devices or within one or more zones in the aircraft 36. Accordingly, further discussion of this functionality is not repeated here.



FIG. 10 illustrates one contemplated embodiment of a television submenu 194. The television submenu 194 is contemplated to provide a slightly different appearance than the video submenu 172 and the audio submenu 192. In the television submenu 194, a channel listing 196 is provided. The channel listing provides a list of the different television channels that are accessible to the passenger. The television submenu 194, therefore, provides access to currently available (or real time) television channels.


If real time television stations are not available, the television submenu 194 is contemplated to default to a pre-recorded television shows library. In such a case, the television submenu 194 is contemplated to operate in the same manner as the video submenu 172 or the audio submenu 192.


Submenus of the television submenu 194 are contemplated to operate in the same manner as the viewing options GUI 186, discussed above. Specifically, television programming may be played on one or more devices or within one or more zones in the aircraft 36. Accordingly, further discussion of this functionality is not repeated here.



FIG. 11 depicts one embodiment of a map view GUI 198 according to the present invention. A map of the world and the location of the aircraft 36 are provided to the passenger.



FIG. 12 depicts a local map GUI 200. The local map GUI 200 is contemplated to provide interactive access to any selected geographic location, such as the destination of the aircraft 36. It is contemplated that the local map GUI 200 will include a search bar that permits the passenger to look for desired landmarks, restaurants, shops, etc.



FIG. 13 illustrates one embodiment of a cabin lights GUI 202 contemplated for use as a part of the present invention. The cabin lights GUI 202 includes zone designators 204, 206, 208. By selecting and highlighting one or more of the zone designators 204, 206, 208, the passenger is able to control the cabin lighting in the selected zones within the aircraft 36.


Two controls over the cabin lighting are provided via the cabin lights GUI 202. The passenger is provided with control over the intensity (or brightness) of the cabin lights via the intensity control menu 210. Cabin light intensity is contemplated to be controllable from a minimum of 0 lumens to a predetermined maximum. The passenger also may be provided with control over the color of the cabin lights via a color control menu 212. Color refers to the “warmness” of the light, as should be apparent to those skilled in the art. Warmer light includes more yellow light elements. Cool light includes a bluer appearance. It is contemplated that the passenger may be provided control over the coolness or warmness of the light, as indicated by the color control menu 212. Both the intensity control menu 210 and the color control menu 212 are contemplated to be presented as slider bars, with slider elements 213, 215, that assist the passenger to appreciate where the controls are in relation to the extremes.


The cabin lights GUI 202 also includes a window shades up icon 214 and a window shades down icon 216. These icons provide control over the degree of openness of one or more of the window shades in the cabin 48. The table light icon 164 also is provided to the passenger. As should be apparent, other controls for other lighting also may be provided on the cabin lights GUI 202. Control over any lights in the cabin 48 is contemplated to include control over the intensity of the light and the warmness or coolness of the light. With respect to the warmness (i.e., the yellow or amber content) or coolness (i.e., the blue content) of the light, it is contemplated that the user will adjust the color of the light between two standard colors for the light. As should be apparent, the colors may be set according to standards for lighting or they may be selected by the aircraft owner or user, as appropriate.


In an alternate embodiment, it is contemplated that the passenger may be provided with even greater control over the color of the lights in the aircraft 36. It is contemplated, for example, that the passenger may be able to control the red, green, and blue (“RGB”) values for the lights in the cabin 48. If so, RGB controllers are anticipated to be displayed on the tablet 130. As should be apparent, for control over the color of the lights, it is contemplated that the lights will be light emitting diodes (“LEDs”), where control over the saturation of the RGB values for the LEDs is permissible.



FIG. 14 illustrates one embodiment of a window shades GUI 218 contemplated for use as a part of the present invention. The window shades GUI 218 includes window designators 220. By selecting and highlighting one or more of the window designators 220, the passenger is able to control the window shade in the selected window 94 within the aircraft 36. In a further contemplated embodiment, the passenger may be provided with control over the window shades in selected zones in the aircraft 36.


Control over the degree of openness of the window shades is contemplated to be provided via a control bar 222 with a slider 224. The slider 224 is contemplated to provide control over the window shades from a fully closed to a full opened condition.



FIG. 15 illustrates one contemplated embodiment of a thermostat GUI 226. The thermostat GUI 226 includes zone indicators 228 so that the passenger may select one or more zones for which the temperature in the aircraft 36 is to be adjusted. The temperature is contemplated to be changed using a temperature control bar 230 with a slider 232. The temperature is contemplated to be controllable within 5-10° C. of the standard ambient of 25° C. Of course, a greater or lesser control may be provided as required or as desired.



FIG. 16 depicts one contemplated embodiment of a presets GUI 234. The presets GUI 234 includes zone indicators 236, as in previous embodiments. Each zone may be controlled according to a predetermined list of environmental conditions (i.e., presets). In the illustrated example, there are three presets: (1) a dining preset 238, (2) a movie preset 240, and (3) a sleep preset 242. An off switch 244 also is provided to disable one or more of the selected presets. Each preset is contemplated to have a lighting intensity, color, etc. associated therewith, where the presets are conducive to the activity listed, such as “dining.”



FIG. 17 provides a seat selector GUI 246, which permits the passenger to identify his or her seat 74 via the seat indicator 248. Any environmental selections that are made by the user are then applied to the selected seat 74. Alternatively, it is contemplated that the passenger may be provided control over the environmental and comfort conditions of other seats 74. For example, a parent may wish to adjust comfort parameters for a child.



FIG. 18 illustrates a change seat GUI 250. The change seat GUI 250 permits the passenger to move from an initial seat 74 to a new seat 74. Any selected comfort variables that the passenger selected may then be applied to the passenger's new seat.



FIG. 19 illustrates one method 252 contemplated by the present invention. The method 252 is considered to be generic to the operation of the passenger IO node 20 of the present invention. In the discussion that follows, reference is made to the passenger tablet 130, because the passenger tablet 130 is considered to embody one preferred embodiment of the present invention. As noted above, the tablet 130 is but one embodiment of the passenger IO node 20 of the present invention. The passenger IO node 20 may be embodied in other electronic devices, such as smart phones. Reference to the tablet 130, therefore, should not be understood to limit the present invention solely to a mobile tablet 130.


The method 252 begins at step 254. From the start 254, the method 252 proceeds to step 256 where the method 252 optionally receives input activating the user interface associated with the passenger IO node 20. As noted above, this includes, but is not limited to, activation of the tablet 130.


It is contemplated that the tablet 130 might not provide any display until activated. As noted above, a passenger may activate the tablet 130 by touching the touch-sensitive surface 150 thereof. Alternatively, a switch (not shown) may be provided to turn on or turn off the tablet 130.


Separately, it is contemplated that the tablet 130 may operate such that the tablet 130 remains in a constant on mode of operation. In this contemplated mode of operation, the tablet 130 may provide a display at all times during flight.


From optional step 256, the method 252 proceeds to step 258, where a menu for controllable parameters is displayed. The menu includes, but is not limited to, a display of the cabin light icon 152, the window shade icon 154, the audio icon 156, the thermostat icon 158, the video icon 160, the presets icon 162, the table light icon 164, the reading light icon 166, and the seat icon 168. As discussed above, each of these icons is associated with a controllable parameter on board the aircraft 36.


The method 252 then proceeds to step 260, where a selection of one of the controllable parameters is received by the method 252. As noted above, the input may be received when a person taps on a particular icon 152-168. In an alternative contemplated operation, the user may use a swiping motion to access the menus associated with the icons 152-168. Specifically, the user may use a swiping motion, by dragging his or her finger across the surface 150 of the tablet 130, to navigate through the different menus associated with each of the icons 152-168.


If no input is received at step 260, the method 252 proceeds to an optional step 262 where the tablet 130 is placed into a sleep mode. In the sleep mode, the tablet 130 may go dark. Alternatively, it may continue to display the screen last selected by a user. In still another embodiment, the tablet 130 may default to the main menu 136.


If the user selects one of the controllable parameters by selecting one of the icons 152-168, the method 252 proceeds to step 264. At step 264, the method 252 displays the controls appropriate for the selected controllable parameter. For example, if the table light icon 164 is selected, the light intensity menu 210 may be displayed. A color light menu 212 also may be displayed as another lighting option for the table light.


Once the control(s) are displayed, the method 252 proceeds to step 266. At step 266, the method 252 receives control input(s) from the user to adjust one or more of the controllable parameters in the cabin 48 of the aircraft 36.


After receiving the input at step 266, the method 252 proceeds to step 268, where the selected, controllable parameters are adjusted according to the input provided by the user.


After step 268, the method 252 is contemplated to return to step 258 and display the main menu 136.


As noted above, it is contemplated that the tablet 130 will operate after being awakened by a person's touch. In keeping with this mode of operation, it is contemplated that the tablet 130 will enter into a sleep mode (or go dark) after the expiry of a predetermined time period. For example, if the tablet 130 has not received tactile input for a period of 2 minutes, the tablet 130 will be instructed to enter into the sleep mode where it will await the next command.



FIGS. 20-34 illustrate a second method 270 of operation of the tablet 130 of the present invention.


The method 270 starts at step 272. The method 270 then proceeds to optional step 274, where the tablet 130 receives an input activating the tablet 130. As noted above, the activation input may be a touch on the surface 150 of the tablet 130. Other inputs may be employed to wake the tablet 130 from a sleep mode without departing from the scope of the present invention.


After being awakened at step 274, the method 270 proceeds to step 276, where the tablet 130 displays a menu of parameters that are controllable within the cabin 48 of the aircraft 36. As noted above, the controllable parameters may be divided into two separate categories including, but not limited to, a media submenu 138 and a cabin submenu 140. As should be apparent, the media submenu 138 and the cabin submenu 140 are one contemplated embodiment of the present invention. The selection of a media submenu 138 and a cabin submenu 140 should not be considered to be limiting of the present invention.


If the user selects an option under the media submenu 138, the method 270 proceeds to the media subroutine 280, which is illustrated in FIG. 21. The connector 282 connects step 278 with the media subroutine 280.


If the user does not select one of the options available in the media submenu 138, the method 270 proceeds to step 284. If the user selects one of the options associated with cabin parameters, the method 270 proceeds to the cabin subroutine 286 via the connector 288. The cabin subroutine 286 is illustrated in FIG. 22.


It is noted that steps 278 and 284 are illustrated in series. However, these steps 278, 284 need not occur in the order presented. Moreover, the steps 278, 284 need not occur in series. It is contemplated that the steps 278, 284 may operate in parallel or in any other suitable order without departing from the scope of the present invention.


If the user does not select one of the cabin parameters in step 284, the method 270 proceeds to step 290, where the method 270 places the tablet 130 into a sleep mode. As noted, this step 290 is optional. It is contemplated that the tablet 130 may not enter a sleep mode. Instead, it is contemplated that the tablet 130 may remain in a constant on condition during operation of the aircraft 36.



FIG. 21 illustrates the media subroutine 280, which connects to the portion of the method 270 illustrated in FIG. 20 via the connector 282.


The media subroutine 280 starts at step 292, where the method 270 awaits selection of video control(s). If video control(s) are selected, the method 270 proceeds to the video subroutine 294 via the connector 296. The video subroutine 194 is illustrated in FIG. 23.


If the user does not select the video control(s), the method 270 proceeds to step 298, where the method 270 awaits selection of the audio control(s). If the user selects the audio controls(s), the method 270 proceeds to the audio subroutine 300 via the connector 302. The audio subroutine 300 is illustrated in FIG. 24.


If the user does not select the audio control(s) in step 298, the method 270 proceeds to step 304, where the method 270 awaits selection of the television control(s). If the user selects the television control(s), the method 270 proceeds to the television subroutine 306 via the connector 308. The television subroutine 306 is illustrated in FIG. 25.


If the user does not select the television control(s), the method 270 proceeds to step 310, where the method 270 awaits selection of the map view control(s). If the user selects the map view control(s), the method 270 proceeds to the map subroutine 312 via the connector 314.


As should be apparent, while the steps 292, 298, 304, 310 are illustrated in a particular order, the present invention does not require that the steps 292, 298, 304, 310 be executed in this order. The steps 292, 298, 304, 310 may be executed in any order without departing from the scope of the present invention. In an alternative contemplated embodiment, the steps 292, 298, 304, 310 may proceed in parallel.


If the user does not select the map view control(s), the method 270 proceeds to step 316, where the method 270 optionally places the tablet 130 into sleep mode. From step 316, the method 270 returns to step 274 via the connector 318.



FIG. 22 illustrates the cabin subroutine 286. As discussed in the paragraphs that follow, the cabin subroutine 286 illustrates one contemplated subroutine for processing input and output related to the parameters associated with functions that are controllable within the cabin 48 of the aircraft 36 from the tablet 130.


The cabin subroutine 286 connects to the portion of the method 270 illustrated in FIG. 20 via the connector 288.


The subroutine 286 then proceeds to step 320, where the subroutine 286 authenticates if the user of the tablet 130 is a flight crew member. If the person operating the tablet 130 is a flight crew member, the method 270 proceeds to step 322. At step 322, the method 270 makes flight crew control(s) available to the flight crew member. It is noted that, if the user is authenticated as a flight crew member, the tablet 130 transitions to a crew IO node 22 and additional functionality becomes available to the flight crew member. Since this aspect of the method 270 is not the focus of the present invention, further details are not provided herein.


With the understanding that the flight crew member will have additional features available to him or her, the method 270 proceeds to step 324. If the user is not a flight crew member but is a passenger, the method 270 proceeds to step 324 without additional functionality being provided to the passenger.


At step 324, the method 270 awaits receipt of the selection of cabin light control(s). The cabin light control(s) are made available if the user accesses the cabin light icon 152. If the method 270 receives the cabin light control(s), the method 270 proceeds to the cabin light subroutine 326 via the connector 328. The cabin light subroutine 326 is illustrated in FIG. 27.


If the method 270 does not receive any selection of cabin light control(s), the method 270 proceeds to step 330. At step 330, the method awaits input of window shade control(s). The window shade control(s) are available through activation of the window shade icon 154, for example. If the method 270 receives input for the window shade control(s), the method proceeds to the window shade subroutine 332 via the connector 334. The window shade subroutine 332 is illustrated in FIG. 28.


If the method 270 does not receive inputs for the window shade control(s), the method 270 proceeds to step 336, where the method 270 awaits input for audio control(s). If the user accesses the audio control(s), the method 270 proceeds to the audio subroutine 300 via the connector 302. The audio subroutine 300 is illustrated in FIG. 24.


If the method 270 does not receive any selection of audio control(s) in step 336, the method proceeds to step 338, where the method 270 awaits selection of the thermostat controls. If the method 270 receives a selection of the thermostat control(s), such as by receiving a selection of the thermostat icon 158, the method 270 proceeds to the thermostat subroutine 340, which is illustrated in FIG. 29. The thermostat subroutine 340 connects to the portion of the method 270 depicted in FIG. 22 via the connector 342.


If the method 270 does not receive a selection of the thermostat control(s), the method proceeds, via the connector 344, to step 346, which is illustrated in FIG. 30. At step 346, the method 270 awaits input selecting the video control(s) that are made available by the selection of the video icon 160, for example.


If the method receives a selection of video control(s) at step 346, the method 270 proceeds to the video subroutine 294, which is illustrated in FIG. 23. The connector 350 indicates the connection to the video subroutine 294. The video subroutine 294 may be accessed via the video icon 160, for example.


If the method 270 does not receive the selection of video control(s) at step 346, the method proceeds to step 352. At step 352, the method awaits selection of the presets control(s) via the tablet 130. The user may access the presets control(s) by selecting the presets icon 162, for example. If the user accesses the presets control(s), the method 270 transitions to the presets subroutine 354 via the connector 356. The presets subroutine 354 is illustrated in FIG. 31.


If the method 270 does not receive any input indicating the selection of the presets control(s), the method proceeds to step 358. At step 358, the method 270 determines if the user provides input selecting the table light control(s). If so, the method 270 proceeds to the table light subroutine 360, which is illustrated in FIG. 32. The table light control(s) are accessible via the table light icon 164, for example. The connector 362 connects step 358 with the table light subroutine 360.


If the method does not receive any input from the user that the user has selected the table light icon 164, the method 270 proceeds to step 364 where the method 270 awaits input of the selection of the reading light control(s). If the user selects the reading light control(s) by accessing the reading light icon 166, for example, the method proceeds to the reading light subroutine 366 via the connector 368. The reading light subroutine 366 is illustrated in FIG. 33.


If the method 270 does not receive any input from the user selecting the reading light control(s), the method 270 proceeds to step 370. At step 370, the method 270 awaits input of the selection of the seat control(s). The seat controls may be accessed by selecting the seat icon 168. If the method 270 receives the selection of the seat control(s), the method 270 proceeds to step 372 via the connector 374. The seat subroutine is illustrated in FIG. 34.


If the method 270 does not receive input regarding the seat, the method 270 proceeds to step 376, where the method 270 optionally places the tablet 130 into a sleep mode. From step 376, the method 270 returns to step 274 via the connector 318.


It is noted that the steps 324, 330, 336, 338, 346, 352, 358, 364, 370 need not be executed in the order described in connection with FIGS. 22 and 30. To the contrary, the steps may be performed in a different order without departing from the scope of the present invention. Alternatively, one or more of the steps 324, 330, 336, 338, 346, 352, 358, 364, 370 may be performed in parallel without departing from the scope of the present invention.



FIG. 23 illustrates the video subroutine 294, as discussed above.


The video subroutine 294 starts at step 378, which follows from the connector 296 that is illustrated in FIG. 23.


At step 378, the method displays the video library 180, which is contemplated to encompass all of the video files that are accessible by the user. The video files may be stored in the database 18, for example. While the video files may be displayed in any particular order and according to any particular sorting parameter(s), it is contemplated that the video files will be presented in alphabetical order.


From step 378, the method 270 proceeds to step 380 where the method 270 determines if there has been a selection of specific video content.


If specific video content has been selected, the method 270 proceeds to step 382. At step 382, the selected video content is played to the user until the video content is exhausted. In other words, at step 382, the video content is anticipated to be played from the beginning to the end of the video file. As should be apparent, control options may be provided to the user to start, stop, advance, and retard the play back of the video file at any point during the playback of the video content. As indicated above, the video content may be provided in the form of an electronic file, a file read from a storage medium (i.e., a digital video disk), etc.


After the video file is played, the method 270 returns to step 274 via the connector 318. Since the user has control over the playback of the video content, the method 270 may return to the step 274 at any time after the user elects to stop the playback, as appropriate.


If the user does not select a particular video from the video library 180, the method proceeds to step 384, where the method 270 awaits the user's selection of search controls. The search controls and search terms may be entered, for example, in the search GUI 182.


If the user enters search parameters, the method 270 proceeds to step 386 where the method 270 displays search control(s). In this step 386, the search controls and search terms may be entered, for example, in the search GUI 182. Searching is contemplated to be performed based on words, phrases, or other suitable search parameters.


At step 388, the method 270 receives the search parameter(s) from the user.


From step 388, the method 270 proceeds to step 390, where the method 270 displays the result(s) of the search to the user.


After the search results are displayed, the method returns to step 380, where the user is permitted to select one of the results from the results that are displayed at step 390.


If the method 270 does not receive the selection of search control(s) at step 384, the method proceeds to step 392, where the method 270 receives a selection of volume controls. If the user does not select the volume controls, the method returns to step 274 via the connector 318. If the user does select the volume controls, the method 270 proceeds to step 394, where the volume controls are displayed to the user.


At step 396, the method 270 receives input for the volume controls.


The method 270 then proceeds to step 398, where the method 270 adjusts the volume according to the input provided by the user.


After step 398, the method 270 returns to step 274 via the connector 318.



FIG. 24 illustrates the audio subroutine 300, as discussed above.


The audio subroutine 300 starts at step 400, which follows from the connector 302 that is illustrated in FIG. 24.


At step 400, the method 270 displays the audio library 180 in the audio submenu 192, which is contemplated to encompass all of the audio files that are accessible by the user. The audio files may be stored in the database 18, for example. While the audio files may be displayed in any particular order and according to any particular sorting parameter(s), it is contemplated that the audio files will be presented in alphabetical order.


From step 400, the method 270 proceeds to step 402 where the method 270 determines if there has been a selection of specific audio content.


If specific audio content has been selected, the method 270 proceeds to step 404. At step 404, the selected audio content is played to the user until the audio content is exhausted. In other words, at step 404, the audio content is anticipated to be played from the beginning to the end of the audio file. As should be apparent, control options may be provided to the user to start, stop, advance, and retard the play back of the audio file at any point during the playback of the audio content. As indicated above, the audio content may be provided in the form of an electronic file, a file read from a storage medium (i.e., a compact disk, digital audio disk, or mp3 file, etc.).


After the audio file is played, the method 270 returns to step 274 via the connector 318. Since the user has control over the playback of the audio content, the method 270 may return to the step 274 at any time after the user elects to stop the playback, as appropriate.


If the user does not select a particular audio from the audio library 180, the method proceeds to step 406, where the method 270 awaits the user's selection of search controls. The search controls and search terms may be entered, for example, in the search GUI 192.


If the user enters search parameters, the method 270 proceeds to step 408 where the method 270 displays search control(s). In this step 408, the search controls and search terms may be entered, for example, in the search GUI 192. Searching is contemplated to be performed based on words, phrases, or other suitable search parameters.


At step 410, the method 270 receives the search parameter(s) from the user.


From step 410, the method 270 proceeds to step 412, where the method 270 displays the result(s) of the search to the user.


After the search results are displayed, the method returns to step 402, where the user is permitted to select one of the results from the results that are displayed at step 412.


If the method 270 does not receive the selection of search control(s) at step 406, the method proceeds to step 414, where the method 270 receives a selection of volume controls. If the user does not select the volume controls, the method returns to step 274 via the connector 318. If the user does select the volume controls, the method 270 proceeds to step 416, where the volume controls are displayed to the user.


At step 418, the method 270 receives input for the volume controls.


The method 270 then proceeds to step 420, where the method 270 adjusts the volume according to the input provided by the user.


After step 420, the method 270 returns to step 274 via the connector 318.



FIG. 25 illustrates the steps comprising the television subroutine 306 of the method 270. The television subroutine 306 begins from the connector 308, as illustrated.


The television subroutine 306 of the method 270 of the present invention starts with a display of the television submenu 194 at step 422. One contemplated embodiment of the television submenu 194 is shown in FIG. 10.


After the display of the television library in step 422, the method 270 proceeds to step 424, where the method 270 awaits receipt of the selection of television content. Television content may include the selection of a particular television channel or pre-recorded television content. If the method 270 receives selected television content from the user, the method 270 proceeds to step 426, where the selected television content is played. After the selected television content is played, the method 270 returns to step 274 via the connector 318.


If the method 270 does not receive a selection of television content, the method proceeds to step 428 where the method receives a selection of search controls. If the method 270 does not receive a selection of search controls, the method 270 returns to step 274 via the connector 318.


If the method 270 receives a selection of search controls, the method 270 proceeds to step 430 where the method displays search controls.


At step 432, the method 270 receives input of search parameters. The user may search for specific content, for a genre of television programs, etc.


After receiving the search parameters, the method 270 proceeds to step 434, where the method displays the search results. The user may then select content from the displayed results. As such, the method returns to step 424 from step 434.



FIG. 26 illustrates the global map subroutine 312 according to one contemplated embodiment of the present invention. The global map subroutine 312 starts at step 436, which follows from the connector 314. At step 436, the method displays the global map GUI 198. One contemplated embodiment of the global map GUI 198 is shown in FIG. 11.


The method 270 proceeds to step 438 where the method 270 awaits receipt of a selection of a local map view. If the method 270 does not receive any selection of a local map view, the method 270 proceeds to step 440.


At step 440, the method 270 awaits selection of local map search parameters via the local map GUI 200. The local map GUI 200 may be configured to receive search parameters associated with the destination of the flight, for example. The user may wish to search for restaurants, museums, and other points of interest at the destination location for the flight, for example.


If the method 270 does not receive a selection of local map search parameters at step 440, the method 270 returns to step 274 via the connector 318.


If the method 270 receives a selection of local map search parameters at step 440, the method 270 proceeds to step 442. At step 442, the method 270 displays the results for the local map search.



FIG. 27 illustrates the cabin lights subroutine 326. The cabin lights subroutine 326 is contemplated to provide control over the cabin lights in the aircraft 36.


The cabin lights subroutine 326 begins at step 444, which is connected to step 324, for example, via the connector 328. At step 444, the method 270 displays the controls for cabin light intensity and/or color. As noted above, the intensity of the cabin lights may be altered to provide a desirable brightness for the lights in the cabin 48. In addition, it is contemplated that the color of the cabin lights may be adjusted between “warm” and “cool” tones.


After step 444, the method 270 proceeds to step 446 where the method receives controls from the user over the cabin lights. The control inputs may be over light intensity and/or color. Controls may be possible via a suitable touch-sensitive control bar, as discussed above.


At step 448, the method 270 adjusts the cabin light intensity and/or color based on the inputs received from the user.



FIG. 28 illustrates the window shade subroutine 332. The window shade subroutine 332 provides control over the degree of openness of the window shades in the cabin 48 of the aircraft 36.


The window shade subroutine 332 begins at step 450, which follows the connector 334. At step 450, the method displays the controls for input of the degree to which one or more of the window shades is to be opened. The control may be by a control slider as discussed above.


At step 452, the method receives control input from a user regarding the degree to which the window shades are to be opened on the aircraft 36. As noted above, the control may be provided over a single window shade or a group of window shades.


At step 454, the method 270 adjusts the degree to which the window shades are opened based on the input provided by the user.



FIG. 29 illustrates the thermostat subroutine 340, which connects to the remainder of the method 270 via the connector 342.


At step 456, the method 270 displays the control inputs for controlling the temperature on board the aircraft 36. The thermostat controls are contemplated to include a control bar a slider, but the controls are not limited to this arrangement.


At step 458, the method 270 receives input for the thermostat controls. Specifically, the method 270 receives temperature inputs for one or more of the regions within the cabin 48 of the aircraft 36.


At step 460, the method 270 adjusts the temperature within the cabin 48 of the aircraft 36 according to the control inputs provided by the user. The method then returns to step 274 via the connector 318.



FIG. 30 illustrates the remainder of the method 270 that is illustrated in FIG. 22. This portion of the method 270 continues after step 338, to which a connection is made via the connector 344.


This portion of the method 270 has already been described.



FIG. 31 illustrates the presets subroutine 354. The presets subroutine connects to the method 270 via the connector 356.


The presets subroutine 354 begins at step 462 where the method 270 displays the control inputs for the presets. One contemplated embodiment for this display is the presets GUI 234 that is illustrated in FIG. 16, for example.


At step 464, the method 270 receives input for the control presets. As discussed above, one of the presets may include a lighting level and environmental controls that are suitable for viewing a meeting. Another preset may include environmental controls for assisting with sleep.


At step 466, the method 270 adjusts that cabin parameters according to the inputs provided by the user.



FIG. 32 illustrates the table light subroutine 360. The table light subroutine 360 provides access to and control over one or more lights that may be positioned above a retractable table 76 or a conference table 80.


The table light subroutine 360 begins at step 468, where the method 270 displays the controls for the table light. The controls may include a control bar and slider as previously described. The controls may include one or both of intensity of the table light and the color, as discussed above.


From step 468, the method 270 proceeds to step 470 where the method 270 receives input regarding the intensity and/or color of the table light. The input may be provided by the user.


At step 472, the method 270 adjusts the table light according to the input received at step 470.



FIG. 33 illustrates a reading light subroutine 366. The reading light subroutine provides control over a reading light that is contemplated to be local to the passenger seat 74. In particular, the reading light is contemplated to be over the seat 74 of the passenger. The reading light subroutine 366 is contemplated to provide control over at least one of the light's intensity and/or color.


At step 474, the controls for the reading light are displayed by the method 270 of the present invention. The controls are contemplated to encompass a control bar with a slider as discussed herein. Of course, other control schemes may be employed without departing from the scope of the present invention.


At step 476, the method 270 receives input concerning the light intensity and/or color.


At step 478, the method 270 adjusts the light intensity and/or color in accordance with the inputs received at step 476.



FIG. 34 illustrates a seat subroutine 372, which connects with step 370 in the method 270, as illustrated in FIG. 30.


The seat subroutine 372 starts at step 482, where the method 270 awaits a request from a user to change his or her seat assignment. If the method 270 receives a request for a passenger to change his or her seat assignment, the method 270 proceeds to step 484. At step 484, the method changes the seat assignment for the passenger according to the input received. A change in seat assignment includes a change in any preferences and settings previously provided for the seat of origin to the changes seat. Accordingly, it is contemplated that, if a passenger changes his or her seat, the comfort parameters previously entered will be transferred to the passenger's new seat 74.


If the method 270 does not receive a request for a passenger to change seats, the method 270 proceeds to step 486. At step 486, the method receives input from the user to adjust the seat. If the method 270 does not receive input to adjust the seat 74, the method proceeds to step 274 via the connector 318. If the method 270 receives input to adjust the seat, the method 270 proceeds to step 488. At step 488, the method 270 adjusts the seat 74 according to the input provided. After the seat 74 is adjusted, the method 270 returns to step 274 via the connector 318.


As should be apparent from FIGS. 13-17, the present invention is contemplated to provide general, localized, and individualized control via the passenger IO node 20, such as the tablet 130. Control may be provided for the cabin 48 as a whole. Selective control may alternatively be provided for zones within the aircraft 36. Finally, the user is provided with control over functionality associated with a passenger's seat 74.


As discussed above, inputs provided by any of the IO nodes 20, 22 and 28-34 are first provided to the controller 16. The reason for this is simple: the controller 16 provides overall control for the functions that are available to passengers in the cabin 48. Without a centralized control, it is possible that passengers might issue instructions that are contrary to one another. The controller 16 may be programmed to address these conflicts or issue an alarm when conflicts arise.


As noted above, it is contemplated that the controller 16 will incorporate a command hierarchy that will resolve any conflicts between the various inputs received from the various nodes 20, 22, 28, 30, 32, 34. The command hierarchy may be based on the status of the person (i.e., crew versus passenger) or based on the location of the IO node (i.e., window IO node 34 versus bulkhead IO node 28). It is also noted that the command and control functions need not be incorporated solely in the controller 16 but may be incorporated into other features without departing from the scope of the present invention.


As also noted above, the present invention contemplates reliance on an isometric view of the cabin 48 of the aircraft 36. The isometric view permits a user to select specific controllable features and zones within the aircraft 36. For example, the user may select one of the passenger seating areas 58, 60, 62 over which control is to be asserted. Alternatively, the user may select an individual seat 74 over which controls are to be asserted. Still further, by selecting a suitable icon from an isometric view of the cabin 48 of the aircraft 36, the user may assert control over one or more of the monitors 96 within the aircraft 36. The isometric view of the cabin 48 of the aircraft 36 provides an easily understood interface for a user to direct inputted commands and assert control over one or more controllable parameters within the cabin 48 of the aircraft 36.


As noted above, the present invention is not intended to be limited solely to the embodiment(s) described herein. To the contrary, those skilled in the art should appreciate that the present invention may be embodied in one or more variations and equivalents to the embodiment(s) described herein. The present invention is intended to encompass those variations and equivalents.

Claims
  • 1. A method of operation for a system within an aircraft comprising a processor, a controller operatively connected to the processor, and a graphical user interface operatively connected to the controller, the method comprising: displaying, on the graphical user interface, a menu of regions in a cabin of the aircraft;receiving, by the controller from the graphical user interface, a selected region from the menu of regions;upon receipt of the selected region from the menu of regions, displaying, on the graphical user interface, a menu of controllable parameters associated with the selected region;receiving, by the controller from the graphical user interface, a selected controllable parameter from the menu of the controllable parameters;displaying, on the graphical user interface, at least one control for the selected controllable parameter;receiving, by the controller from the graphical user interface, a control input for the selected controllable parameter; andadjusting, by the controller, the selected controllable parameter consistent with the control input,wherein the graphical user interface is embodied in a mobile computing device for a passenger, andwherein the selected controllable parameter comprises one of light intensity, light color, temperature, and a degree of openness of at least one window shade.
  • 2. The method of claim 1, further comprising: placing the graphical user interface into a sleep mode if selection of the at least one controllable parameter is not received.
  • 3. The method of claim 1, wherein the menu of regions comprises at least two of: the entire cabin of the aircraft, at least one zone within the cabin of the aircraft, and at least one seat within the cabin of the aircraft.
  • 4. The method of claim 1, wherein the menu of regions in the cabin of the aircraft comprises an isometric representation of at least a portion of the cabin of the aircraft.
  • 5. The method of claim 1, further comprising: prioritizing, by the controller, the control input received from the mobile computing device in relation to control inputs received from any other input device, thereby avoiding conflicts between the control inputs.
  • 6. The method of claim 1, wherein light intensity, temperature, and a degree of openness of the at least one window shade is adjustable between a predetermined minimum value and a predetermined maximum value.
  • 7. The method of claim 1, wherein the light color is adjustable between a predetermined warm white color value and a predetermined cool white color value.
  • 8. The method of claim 11, wherein the media content includes a video library, an audio library, and a map view.
  • 9. The method of claim 8, wherein the map view comprises a global map view and a local map view.
  • 10. The method of claim 1, wherein the displaying of the menu of controllable parameters includes displaying a light icon, a media icon, a thermostat icon, and a window shade icon.
  • 11. The method of claim 1, wherein the selected controllable parameter further comprises one of media type, media content, and media volume.
  • 12. The method of claim 1, wherein the mobile computing device is at least one of a personal computer, tablet, and smartphone.
  • 13. A system within an aircraft, comprising: a processor;a controller operatively connected to the processor;a graphical user interface operatively connected to the controller;a first display for displaying, on the graphical user interface, a menu of regions in a cabin of the aircraft;a first input for receiving, via the graphical user interface, a selected region from the menu of regions;a second display for displaying, on the graphical user interface upon receipt of the selected region from the menu of regions, a menu of controllable parameters associated with the selected region;a second input for receiving, via the graphical user interface, a selected controllable parameter from the menu of the controllable parameters;a third display for displaying, on the graphical user interface, at least one control for the selected controllable parameter, wherein the third display also receives a control input for the selected controllable parameter; anda controller for adjusting the selected controllable parameter consistent with the control input,wherein the graphical user interface is embodied in a mobile computing device for a passenger, andwherein the selected controllable parameter comprises one of light intensity, light color, temperature, and a degree of openness of at least one window shade.
  • 14. The system of claim 13, wherein the selected controllable parameter further comprises one of media type, media, content, and media volume, and wherein the menu of regions comprises at least two of the entire cabin of the aircraft, at least one zone within the cabin of the aircraft, or at least one seat within the cabin of the aircraft.
  • 15. The system of claim 13, wherein the menu of regions in the cabin of the aircraft comprises an isometric representation of at least a portion of the cabin of the aircraft.
  • 16. The system of claim 13, wherein the control input received from the mobile computing device is prioritized, by the controller, in relation to control inputs received from any other input device, thereby avoiding conflicts between the control inputs.
  • 17. A non-transitory computer readable medium comprising instructions for operation of a system within an aircraft comprising a processor, a controller operatively connected to the processor, and a graphical user interface operatively connected to the controller, wherein the instructions comprise: displaying, on the graphical user interface, a menu of regions in a cabin of the aircraft;receiving, by the controller from the graphical user interface, a selected region from the menu of regions;upon receipt of the selected region from the menu of regions, displaying, on the graphical user interface, a menu of controllable parameters associated with the selected region;receiving, by the controller from the graphical user interface, a selected controllable parameter from the menu of the controllable parameters;displaying, on the graphical user interface, at least one control for the selected controllable parameter;receiving, by the controller from the graphical user interface, a control input for the selected controllable parameter; andadjusting, by the controller, the selected controllable parameter consistent with the control input,wherein the graphical user interface is embodied in a mobile computing device for a passenger, andwherein the selected controllable parameter comprises one of light intensity, light color, temperature, media content, media volume, and a degree of openness of at least one window shade.
  • 18. The non-transitory computer readable medium of claim 17, wherein the selected controllable parameter also includes at least one of media type, media, content, and media volume, and wherein the menu of regions comprises at least two of the entire cabin of the aircraft, at least one zone within the cabin of the aircraft, or at least one seat within the cabin of the aircraft.
  • 19. The non-transitory computer readable medium of claim 17, wherein the menu of regions in the cabin of the aircraft comprises an isometric representation of at least a portion of the cabin of the aircraft.
  • 20. The non-transitory computer readable medium of claim 17, wherein the instructions further comprise: prioritizing, by the controller, the control input received from the mobile computing device in relation to control inputs received from any other input device, thereby avoiding conflicts between the control inputs.
CROSS-REFERENCE TO RELATED APPLICATION(S)

This United States Non-Provisional patent application relies for priority on U.S. Provisional Patent Application Ser. No. 61/759,156, filed on Jan. 31, 2013, the entire contents of which are incorporated herein by reference.

Provisional Applications (1)
Number Date Country
61759156 Jan 2013 US