This invention relates to the field of training and more particularly to a system that provides hints for improving skills of the trainee.
Driving training simulators are well known. Such simulators often included controls that simulate the target vehicle (e.g. car, truck, bus, etc). It is known that such simulators improve skills and safety by familiarizing the trainee with operation of the vehicle by presenting simulated situations in which, making the wrong decision does not result in a potential accident or bodily harm. In this way, the trainee learns basic driving skills before they eventually need to perform using the actual target vehicle and before they have to perform using that vehicle while operating in traffic.
There are many types of simulators known. The simplest simulator is a typical driving video game having a display screen and a hand controller. In some systems, a simulated steering wheel is provided. A mock-vehicle is displayed on the display screen and the driver uses the hand controller to keep the mock-vehicle on a simulated, moving roadway on the display screen. This type of simulator helps build driver hand and eye coordination, but does not provide the true control operation of the real steering wheel, brake, clutch, shifter, windshield views and mirror views. Such simulators are more of a game than an actual driver training system.
Another type of simulator includes a video display screen to simulate a windshield view, a steering wheel, a gas pedal, a brake pedal, a shifter and, optionally, a clutch pedal. A road situation is displayed on the display screen and the driver uses the controls to drive the simulated vehicle, moving down a roadway that is displayed on the display screen. This type of simulator helps build driver skills, but does not include interaction with speedometers, tachometers, etc. Such simulators don't provide feedback from the shifter such as gear grinding when the clutch isn't operated correctly. Furthermore, such simulators have a fixed configuration relating to a single type/layout of vehicle. In some such simulators, certain gauges are provided to simulate the operation and information provided to a driver of this singular vehicle. All current simulators provide fixed scenarios to the trainee and evaluate the trainee responses in a fixed program, progressing from scenario to scenario in a linear progress.
None of the current driver training simulators provide training simulations that provide real-time driving hints to the trainee.
Today, simulators provide a realistic environment in which a trainee is able to manipulate a virtual vehicle in a virtual reality (e.g. road system) without any danger of causing an accident and damaging other vehicles or hurting other drivers. Such simulators often run a pre-programmed simulation, monitor actions and responses of the trainee, and report results when the simulation is complete. The simulators of the past do not provide real-time coaching during the simulation to improve the learning experience of the trainee.
What is needed is a driver training system that provides real-time coaching during training scenarios that adapt to objects or events presented in the simulation.
In one embodiment, a method of driver/operator training is disclosed including in a training system, while presenting a simulation segment: collecting data from at least one input device and analyzing the data to predict a potential future issue. If this analysis indicates that there exists the potential future issue, an informational message is presented to warn of the trainee of the potential future issue.
In another embodiment, method of driver operated training of a trainee on a training system is disclosed. The training system includes a computer that includes a storage device and a plurality of simulation segments; the simulation segments are stored on the storage device and accessible by the computer; the computer is operatively interfaced to one or more graphics displays, at least one of the graphics displays simulates a windshield of a target vehicle and the computer generates images on the graphics displays. Additionally, the computer is operatively interfaced to one of more input devices and one or more output devices, the input devices and output devices corresponding to control/indicator devices of the target vehicle that are in communication with the trainee. The method includes (a) the computer selecting a first simulation segment of the simulation segments as a current simulation segment and (b) presenting the current simulation segment on one or more of the graphics displays and/or one or more output devices, thereby the computer simulating an operational scenario of the target vehicle. (c) Collecting data from the one or more input devices by the computer. (d) The computer analyzes the data to predict a potential future issue and (e) if there exists the potential future issue; the computer presents an informational message to warn of the trainee of the potential future issue. (f) The computer determines a next simulation segments from the simulation segments and (g) repeats steps b-f until the first simulation segment is complete.
In another embodiment, a training system is disclosed. The training system includes a computer that includes data storage. A plurality of simulation segments are stored in the data storage and are accessible by the computer. There are one or more graphics displays, at least one which simulates a windshield of a target vehicle. The computer has hardware/software for generating images on the graphics displays. One or more input devices that correspond to control devices of the target vehicle are operatively coupled to the computer. Software running on the computer causes the computer to present a simulation segment of the plurality of simulation segments. In such, the software running on the computer causes the computer to display a content of the simulation segment on one or more of the graphics displays, thereby the software running on the computer causes the computer to simulate an operation of the target vehicle. Concurrently, the software running on the computer causes the computer to collect data from the one or more input devices; to analyze the data and determines if a potential future issue exists; and if the potential future issue exists, the software running on the computer causes the computer to emit an informational message that provides a hint regarding the potential future issue.
The invention can be best understood by those having ordinary skill in the art by reference to the following detailed description when considered in conjunction with the accompanying drawings in which:
Reference will now be made in detail to the presently preferred embodiments of the invention, examples of which are illustrated in the accompanying drawings. Throughout the following detailed description, the same reference numerals refer to the same elements in all figures. In general, the training system 10 is often known as a driving/flying/boating/engineering/etc simulator, depending upon the target vehicle (car/truck, airplane, boat, train, etc). The training system 10 is any system for training a trainee 5 (e.g. truck driver trainee) that simulates some or all of the operator controls (e.g. steering, brake, shifter) and visuals (e.g. mirrors, windows, dash boards, etc) without requiring the trainee to operate the actual vehicle (e.g., drive the actual truck). Although not limited to any particular target vehicle, the remainder of this description will use a truck as an example of such target vehicle for brevity reasons. Note that some of the controls described (e.g. shifter, clutch, steering wheel) are related to certain types of target vehicles and not necessarily to others. For example, many automobiles have automatic transmissions and, therefore, do not have a clutch. In another example, an airplane does not have rear-view mirrors, shifters, clutches, etc. Likewise, a truck driving simulator has rear-view mirrors, shifters, clutches, but does not have ailerons, thrust, altitude gauges, etc.
Referring to
The exemplary training system 10 has a windshield display 12 on which a simulated driving situation is presented as the trainee 5 would see through the windshield of the target vehicle. The windshield display 12 shows, for example, the road being driven upon, the grade of the road, obstacles such as other vehicles, tress, parked cars, pot holes, etc. In some training scenarios, the windshield is fogged or distorted by simulated weather condition s such as rain, snow, sleet, etc.
The trainee 5 typically sits on a seat 20 that, preferably, though not required, mimics a seat of the target vehicle. The trainee has controls similar to those in the target vehicle such as a steering wheel 30, horn 31, gas pedal 32, brake pedal 34, clutch 36 (see
In a preferred embodiment, though not required, the steering wheel 30 and shifter handle 98 have touch sensors that detect if and when the trainee 5 has is grasping the steering wheel 30 and/or shifter handle 98. The touch sensors are any known touch sensor such as a mechanical switch or switches, capacitive or resistive detectors, etc. In some embodiments, the position of the trainee's hands is determined by the camera(s) 13 in conjunction with or instead of the touch sensors.
In a position similar to that of a dashboard of the target vehicle is a dashboard (e.g. display) 14 (details of an exemplary dashboard 14 are shown in
In some embodiments, one or more side-view mirror displays 42/44 are mounted on or in the cabinet 8. When provided, the side-view mirror displays 42/44 show a simulated view of what is visible to the trainee 5 such as vehicles being passed and/or approaching vehicles. In some embodiments, the side-view mirror displays 42/44 display objects as they would appear in a real side-view mirror simulating concave or convex mirrors as appropriate. Additionally, in some embodiments, the image displayed includes simulated dirt, etc, as often occurs in real life.
In some embodiments, a rear-view mirror display 43 that is center-mounted is also provided (not shown). When provided, the rear-view mirror display 43 shows a simulated view of what is visible to the trainee 5 such as approaching vehicles and/or oncoming vehicles.
In some embodiments, an information console 16 is provided. The information console 16 does not simulate something from the target vehicle. Instead, the information console 16 presents menus, status information, audio messages, and auxiliary information to the trainee 5 and accepts inputs such as scenario selection, study chapter selection, login data, etc.
In some embodiments, an audio system 18 is provided to enhance realism and provide simulations of sounds that are normally heard when operating the target vehicle such as engine noise, tire noise, other vehicles, rain or sleet hitting the target vehicle, emergency vehicles, sounds of a collision, etc.
In some embodiments, one or more cameras 13 (e.g. trainee sensors) are provided to detect various aspects of the trainee 5 such as position upon the seat 20, head angle, attention, drowsiness, where the trainee is looking, and in some embodiments, verify that the trainee 5 is who he/she says he/she is. This information is used to make sure the trainee is properly performing the task at hand. The cameras 13 (e.g. trainee sensors) are, for example, cameras, light detectors, ultrasonic transducers, or any other detector as known in the industry. The cameras 13 (e.g. trainee sensors) are coupled to the computer 100 (see
Referring to
In
In
In
In
Referring to
In
In some embodiments, the windshield display 12 is also touch sensitive. This provides even more capabilities for testing the trainee's 5 ability to identify environmental (e.g. roadway) objects such as signs, barriers, etc. For example, the trainee is asked to touch the stop sign or touch the lane in which is most appropriate for his/her vehicle, etc.
Again, in some embodiments, one or more cameras 13 (e.g. trainee sensors) are integrated into the training system 10. The trainee sensors (e.g. camera 13 or cameras 13) are coupled to the computer 100. The computer 100 analyzes data from the cameras 13 (e.g. trainee sensors) to determine what the trainee 5 is doing and/or where the trainee 5 is looking to provide feedback to the trainee 5 and evaluate the trainee's abilities (e.g. the cameras 13 are used to determine if the trainee 5 looked in the right mirror display 42 before changing lanes). The cameras 13 (e.g. trainee sensors) are positioned as needed to determine the position, stance and view of the trainee 5.
Referring to
It is also preferred that each sub-component of the dashboard 14 is touch-sensitive. In the example of
In some embodiments, status or identification information 50 is provided on the dashboard 14 such as the vehicle details and, perhaps, the name of the trainee 5, etc.
Referring to
Central to the training system 10 is a computer 100. Many different types of computers 100 are anticipated such as personal computers, dedicated computers and server computers. It is anticipated that computers 100 of one training system 10 are connected by local or wide area networks to other training systems 10 and/or to central data collection and control systems (not shown). In some embodiments, the computer has a motherboard with multiple PCI-Ex16 slots that provide multiple simulator display channels with 2D and/or 3D capability. A video processor card is optionally installed in each of these slots. The video cards run the simulation in multi channel mode with low transient delay times. It is anticipated, though not required, that a single image generator (single motherboard computer) can drive multiple displays. Although any number of display channels is anticipated, the training system typically is configured with from 3 to 8 real time interactive screens.
The computer 100 includes, in some embodiments, a display device or terminal device 140. This terminal device 140 has a display screen, a keyboard and/or a touch screen and is primarily used by an administrator to operate the computer 100, for example, performing backups and other system administration function. In some embodiments, these functions are performed using one or more of the other components/displays 12/14/16.
The computer 100 also includes one or more persistent storage devices 110/120 such as hard drives, flash memory, etc. for storage of, for example, courseware 110 and user data 120. In a preferred embodiment, the one or more persistent storage devices 110/120 is one or more hard drives or solid-state drives. In some embodiments, the one or more storage devices 110/120 include a raid system to provide more reliable data storage.
Interfaced to the computer 100 are several components of the training system 10. The windshield display 12, dashboard (e.g. dashboard graphics display and touch screen) 14 and information console 16 are all interfaced to the computer 100 as known in the industry. The mirror displays 42/43/44 (when present) are also interfaced to the computer 100 as known in the industry. All specialized hardware devices such as the shifter touch detector 118 (also the X-position, Y-position, switch status not shown for brevity reasons), clutch 36 (position and force), gas pedal 32 (position and force), brake pedal 34 (position and force) and steering wheel 30 (rotation and touch) are also interfaced to the computer 100 as known in the industry. It is preferred that some or all of such interfaces are bi-directional to provide control of the device (e.g. vary the counter-force of the brake pedal 34 or gates of the transmission-shifter subsystem 80) and to receive feedback from the device (e.g. sufficient pressure was applied to the brake pedal 34, hands are on the steering wheel 30 or the trainee 5 successfully shifted from first gear into second gear).
In embodiments that have trainee sensors such as cameras 13, etc., the trainee sensors (e.g. cameras 13) are interfaced to the computer 100 as known in the industry.
In embodiments that have hand proximity sensors 118 (on shifter handle 98 or steering wheel 30), the hand proximity sensors 123 are interfaced to the computer 100 as known in the industry.
In embodiments that have shifter force sensors 123 (on shifter shaft 99, the shifter force sensors 123 are interfaced to the computer 100 as known in the industry.
In some embodiments, one or more biometric sensors 15 are interfaced to the computer 100. The biometric sensors 15 sense, for example, fingerprints, retina, face characteristics, etc, of a user of the training system 10 to make sure the training and results correspond to the correct trainee 5, thereby preventing one trainee 5 from intentionally or unintentionally scoring/learning for another trainee 5.
In embodiments having a audio system 18, the audio system 18 is interfaced to the computer 100 as known in the industry such as audio outputs connected to amplifiers and speakers, TOSLINK, USB, etc.
In embodiments having a transmission transducer 106, the transmission transducer 106 is interfaced to the computer 100 as known in the industry such as through audio outputs connected to amplifiers and speakers, TOSLINK, USB, etc or over a local area network (see
Referring to
Also connected to the processor 610 is a system bus 630 for connecting to peripheral subsystems such as a network interface (not shown), a persistent storage 640 (e.g. a hard disk, semiconductor storage such as flash, a raid system, etc), a disk drive 650 (e.g. DVD), one or more graphics adapter(s) 660, a keyboard/mouse 670 and/or one or more touch screen interfaces 675. The graphics adapter(s) 660 receives commands and display information from the system bus 630 and generates a display image that is displayed on one or more of the graphic display devices 12/14/16/42/43/44.
In general, the hard disk 640 may be used to store programs, executable code and data (e.g. courseware 110 and user data 120) persistently. For data security and reliability, in some embodiments, the hard disk 640 is multiple disks or a raid system, etc. The removable disk drive 650 is often used to load CD/DVD/Blue ray disks having programs, executable code and data onto the hard disk 640. These peripherals are examples of input/output devices, persistent storage and removable media storage. Other examples of persistent storage include core memory, FRAM, flash memory, etc. Other examples of removable disk drives 650 include CDRW, DVD, DVD writeable, Blue ray, compact flash, other removable flash media, floppy disk, etc. In some embodiments, other devices are connected to the system through the system bus 630 or with other input-output connections. Examples of these devices include printers; graphics tablets; joysticks; audio components; and communications adapters such as modems and Ethernet adapters.
Although there are many ways anticipated for connecting training system components 13/30/32/34/36/106/118/123 to the processor, one preferred interface is a bi-directional local area network such as Car Area Network (CAN) 685 connected to the bus 630 by a Car Area Network (CAN) interface 680 as known in the industry. Any connection scheme to the system components 13/30/32/34/36/106/118/123 is anticipated including direct wiring, any local area network (e.g. Ethernet, CAN or VAN) and wireless (e.g. Bluetooth).
Referring to
Now a prediction engine or stored data in the scenario (courseware 110) determines 465 if a potential future issue is possible. For example, a prediction is made that it will be dark in 15 minutes or that a person might step into the street ahead in the scenario. If the prediction engine or stored date determines 465 that a potential future issue is possible 466 such as the trainee 5 has not turned on the headlights as dark is approaching or has not slowed down or changed lanes in view of the pedestrian, the trainee 5 is provided an informational message (e.g., either a text message on screen or an audio message). In this, the trainee 5 has not made any mistakes; it is just a proper time for the trainee 5 to start thinking about an action such as turning on the headlights, slowing down, or changing lanes. The informational message is a hint to the trainee 5 to consider what they should do before an incident occurs such as not being able to see because it is dark or hitting a pedestrian, etc.
Next, the data is analyzed 468 to determine the performance of the trainee 5 in the given scenario/segment meets passing requirements and information is displayed 470 to the trainee 5 on one or more of the display devices 12/14/16. If the performance indicates that the trainee 5 didn't perform the task sufficiently 472, a new scenario/segment is selected 474 (adaptive training). The new scenario/segment is selected 474 based upon elements of the prior scenario/segment that were not adequately performed. Since the method is adaptive, the method uses any existing or modified scenario/segment to fortify the element that was not adequately performed. For example, if while the training system 10 is in the simulation mode, the trainee 5 avoided the collision but the trainee 5 didn't step on the clutch while applying the brakes, therefore stalling the engine, one or more training scenarios/segments or chapters related to proper use of the clutch while braking are selected 474 to be presented to the trainee 5 either during the current session (e.g. the training system 10 enters a training mode and the training scenario/segment is presented while the trainee 5 remains in the training system 10) or during a future session.
If the trainee's 5 performance meets passing requirements 472, the data (e.g. results) are stored 476 for later reporting/analysis/grading and it is determined if there are more scenarios/segments 478 for the trainee 5 (e.g. scenarios/segments are often grouped in chapters and the trainee 5 is finished when he/she complete a chapter, etc.). If there are more scenarios/segments 478, the next scenario/segment is retrieved 480 and the above steps 462-478 are repeated until there are more scenarios/segments planned for the trainee 5.
The methods of the prior art do not predict potential issues and present informational messages to warn the trainee 5 of such potential issues. As discussed above, in some scenarios, the potential issues are predicted 465 by the prediction engine using data from the scenarios/segments (courseware 110) such as a darkness level or location of a pedestrian. The prediction engine then reviews a status of the simulation (e.g. headlights=off or speed=X or lane=right) and calculates whether there will be a potential issue using the data from the scenario in view of the status of the current simulation. For example, if the darkness level of the current scenario/segment is “dusk” and the status of the simulation indicates headlights=on, no informational messages in needed to warn the trainee 5 of such potential issues. Now, say the darkness level of the current scenario/segment is “dusk” and the status of the simulation indicates headlights=off, an informational messages in needed, warning the trainee 5 of such potential issues.
Equivalent elements can be substituted for the ones set forth above such that they perform in substantially the same manner in substantially the same way for achieving substantially the same result.
It is believed that the system and method as described and many of its attendant advantages will be understood by the foregoing description. It is also believed that it will be apparent that various changes may be made in the form, construction and arrangement of the components thereof without departing from the scope and spirit of the invention or without sacrificing all of its material advantages. The form herein before described being merely exemplary and explanatory embodiment thereof. It is the intention of the following claims to encompass and include such changes.
Number | Name | Date | Kind |
---|---|---|---|
3251142 | Jazbutis | May 1966 | A |
3309794 | Greenshields | Mar 1967 | A |
3479750 | Swanson | Nov 1969 | A |
3583079 | Koci | Jun 1971 | A |
3611589 | Wiltse | Oct 1971 | A |
3896564 | Dewey et al. | Jul 1975 | A |
3936955 | Gruen | Feb 1976 | A |
4034484 | Radice | Jul 1977 | A |
4464117 | Foerst | Aug 1984 | A |
4683891 | Cornellier et al. | Aug 1987 | A |
4846686 | Adams | Jul 1989 | A |
5015189 | Wenzinger, Jr. | May 1991 | A |
5116051 | Moncrief et al. | May 1992 | A |
5184956 | Langlais et al. | Feb 1993 | A |
5197003 | Moncrief et al. | Mar 1993 | A |
5277584 | DeGroat et al. | Jan 1994 | A |
5287489 | Nimmo et al. | Feb 1994 | A |
5366376 | Copperman | Nov 1994 | A |
5399091 | Mitsumoto | Mar 1995 | A |
5415550 | Aoki | May 1995 | A |
5626362 | Mottola | May 1997 | A |
5707237 | Takemoto | Jan 1998 | A |
5823876 | Unbehand | Oct 1998 | A |
5888074 | Staplin et al. | Mar 1999 | A |
5921780 | Myers | Jul 1999 | A |
5951018 | Mamitsu | Sep 1999 | A |
6053737 | Babbitt et al. | Apr 2000 | A |
6105737 | Weigert | Aug 2000 | A |
6186794 | Brown et al. | Feb 2001 | B1 |
6200139 | Clapper | Mar 2001 | B1 |
6206700 | Brown et al. | Mar 2001 | B1 |
6227862 | Harkness | May 2001 | B1 |
6270349 | Ng | Aug 2001 | B1 |
6270350 | Christopher | Aug 2001 | B1 |
6361321 | Huston et al. | Mar 2002 | B1 |
6427063 | Cook et al. | Jul 2002 | B1 |
6431872 | Shiraishi | Aug 2002 | B1 |
6915104 | Yonezu et al. | Jul 2005 | B2 |
6916074 | Jung | Jul 2005 | B2 |
6948398 | Dybro | Sep 2005 | B2 |
6994361 | Nishimura | Feb 2006 | B2 |
7052277 | Kellman | May 2006 | B2 |
D531221 | Shiraishi | Oct 2006 | S |
7467121 | Hagelin | Dec 2008 | B2 |
7512577 | Slemmer et al. | Mar 2009 | B2 |
7625287 | Champagne | Dec 2009 | B2 |
7895136 | Slemmer et al. | Feb 2011 | B2 |
7949295 | Slemmer et al. | May 2011 | B2 |
8113844 | Huang et al. | Feb 2012 | B2 |
8170976 | Dargue et al. | May 2012 | B2 |
8770980 | Best et al. | Jul 2014 | B2 |
20020146667 | Dowdell et al. | Oct 2002 | A1 |
20030039948 | Donahue | Feb 2003 | A1 |
20030233155 | Slemmer et al. | Dec 2003 | A1 |
20040158476 | Blessinger | Aug 2004 | A1 |
20040161731 | Arington | Aug 2004 | A1 |
20040259059 | Aoki | Dec 2004 | A1 |
20050277099 | Van Schaack et al. | Dec 2005 | A1 |
20060040239 | Cummins | Feb 2006 | A1 |
20060078856 | Kellman | Apr 2006 | A1 |
20060195412 | Slemmer et al. | Aug 2006 | A1 |
20070048690 | Stricek et al. | Mar 2007 | A1 |
20070072154 | Akatsuka et al. | Mar 2007 | A1 |
20070122778 | Beitel | May 2007 | A1 |
20080064014 | Wojtczak | Mar 2008 | A1 |
20080082372 | Burch | Apr 2008 | A1 |
20080145830 | Huang et al. | Jun 2008 | A1 |
20080180523 | Stratton et al. | Jul 2008 | A1 |
20090035733 | Meitar et al. | Feb 2009 | A1 |
20090098519 | Byerly | Apr 2009 | A1 |
20090163283 | Childress | Jun 2009 | A1 |
20090181349 | Harkness | Jul 2009 | A1 |
20090187519 | Slemmer et al. | Jul 2009 | A1 |
20090208910 | Brueckner | Aug 2009 | A1 |
20090306880 | Gomi | Dec 2009 | A1 |
20100003659 | Edmonds | Jan 2010 | A1 |
20100100520 | Dargue et al. | Apr 2010 | A1 |
20100182432 | Augst | Jul 2010 | A1 |
20100190142 | Gal et al. | Jul 2010 | A1 |
20100190143 | Gal et al. | Jul 2010 | A1 |
20100190145 | Singer et al. | Jul 2010 | A1 |
20100217573 | Hatcherson | Aug 2010 | A1 |
20110121938 | Slemmer et al. | May 2011 | A1 |
20110283866 | Hogan | Nov 2011 | A1 |
20120135382 | Winston et al. | May 2012 | A1 |
20140186810 | Falash | Jul 2014 | A1 |
20140188574 | Luca | Jul 2014 | A1 |
20150187224 | Moncrief | Jul 2015 | A1 |
20160163217 | Harkness | Jun 2016 | A1 |
Number | Date | Country |
---|---|---|
0483991 | May 1992 | EP |
2439718 | Apr 2012 | EP |
10-2002-0030305 | Apr 2002 | KR |
10-2002-0090751 | Dec 2002 | KR |
48661 | Oct 2005 | RU |
2010044982 | Apr 2010 | WO |
2010-083563 | Jul 2010 | WO |
2013033723 | Mar 2013 | WO |
Entry |
---|
Dr. Painter et al., “General Aviation Pilot Advisory and Training System Semiannual Technical Progress Report”, pp. 1-36, Sep. 16, 1997. |
April Broomer, “CAE Simfinity™Suite of training devices”, retrieved Mar. 14, 2011. |
Ok-Choon Park, Jung Lee, “Adaptive Instructional Systems”, Handbook of Research on Educational 1 Communications and Technology, Chapter 25, Association for Educational Communications and Technology, http://Dlearngen.org/-aust/EdTecheBooks/AECT HAN DBOOK 2ND/25 .pdf. |
Richard C. Atkinson, “Adaptive Instructional Systems”, Psychology and Education Series, Technical Report No. 240, Nov. 20, 1974, Stanford University, Stanford, California, http://suppes-corpus.stanford.edu/techreports/ D IMSSS_240.pdf. |
Small Business Innovation Research, “Adaptive Instructional Systems”, U.S. Army Research Institute Rotary-Wing Aviation Research Unit, Fort Rucker, AL. |
Third CLIPS Conference Proceedings, NASA Conference Publication 10162 vol. I, Washington, D.C., Nasa Center for AeroSpace Information, 1994. |
Zhang et al., “Program-level Adaptive Memory Management”, ISMM'06, Jun. 10-11, 2006, Ottawa, Ontario, Canada, http://anon.cs.rochester.edu/u/zhangchl/publications/ismm06.pdf. |
Ludwig et al., “Developing an Adaptive Intelligent Flight Trainer”, 2005, www.dtic.mil/cgi-bin/GetTRDoc?AD=ADA439799. |
“Software to Detect Malfunctions in Anesthesia Equipment”, Nasa Tech Briefs, Lyndon B. Johnson Space Center, Houston, Texas, Aug. 1, 2000. |
8 Espacenet English language abstract of KR 10-2002-0030305, published Apr. 25, 2002. |
9 Espacenet English language abstract of KR 10-2002-0090751, published Dec. 5, 2002. |
Number | Date | Country | |
---|---|---|---|
20200020244 A1 | Jan 2020 | US |
Number | Date | Country | |
---|---|---|---|
61277768 | Sep 2009 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 12889448 | Sep 2010 | US |
Child | 14285017 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 14285017 | May 2014 | US |
Child | 16576822 | US |