The present disclosure relates generally to a projector system, and more particularly, to a projector system for a bumper of a vehicle.
Vehicles are often required to relay information to individuals outside of the vehicle. The messages can be as simple as honking a horn to get attention or actuating a turn signal to indicate a lane change of the vehicle. These means of providing information are inherently limited in the complexity of the information that can be provided. For example, the indications are typically limited to a monotone audible output or a single blinking light.
The disclosed projector system is directed to mitigating or overcoming one or more of the problems set forth above and/or other problems in the prior art.
One aspect of the present disclosure is directed to a projector system for a vehicle. The projector system includes an external member including at least a portion formed of a transparent material, the portion including a translucent surface, and a projector configured to project an image onto the translucent surface.
Another aspect of the present disclosure is directed to a method of displaying an image onto a vehicle. The method includes receiving, with a projector, a signal indicative of a message, and projecting, with the projector, the image based on the signal onto an external surface of the vehicle.
Yet another aspect of the present disclosure is directed to a vehicle. The vehicle includes a projector system having an external member including a translucent surface, and a projector configured to project an image onto the translucent surface, the external member including at least a portion formed of a transparent material and having the translucent surface.
The disclosure is generally directed to a projector system for projecting images onto at least a portion of a surface of an external member of a vehicle. In some embodiments, the external surface includes an area on a front of the vehicle and/or an area on a rear of the vehicle. More particularly, in some embodiments, the system includes a short throw high definition projector positioned in line with a grill area or bumper on the front of the vehicle and/or a trunk portion or bumper on the rear of the vehicle. In an illustrated embodiment, the system includes a short throw high definition projector positioned in line with at least one of the front and rear bumpers, as the external member(s) to project an image on at least a portion of a surface of the bumper(s). The bumper may include an automotive grade flexible and transparent polymer. That is, at least a portion or the entirety of the bumper may be formed of the transparent material. The portion or the entirety of the bumper formed of the transparent polymer may be back painted with a pearlescent coating to provide a desirable translucent surface or lens. The projector system may project images onto the translucent surface related to safety, aesthetics, vehicle status, entertainment, and/or car sharing.
As illustrated in
User interface 28 may be any device configured to receive input from the user and transmit data. For example, user interface 28 may have a display including an LCD, an LED, a plasma display, or any other type of display. User interface 28 may also provide a graphical user interface (GUI) presented on the display for user input and data display. User interface 28 may further include input devices, such as a touchscreen, a keyboard, a mouse, and/or a tracker ball.
Camera 32 may include any device configured to capture videos and/or images of cabin 26 and generate signals and/or data to be processed to visually detect the surrounding environment. For example, camera 32 may be used in conjunction with image processing software, e.g., image recognition software, such that the software may distinguish a person from inanimate objects, and distinguish identities, facial features, and/or other biometric aspects of users. The software may also be configured to detect gestures by the people based on relative positions and movements of facial features, heads, arms, fingers, legs, and/or feet. In some embodiments, the software may be configured to process the image to determine conditions of the people, such as sleep deprivation, intoxication, attentiveness, distraction, and/or other physical or mental conditions. Camera 32 may be configured to be adjusted by a motor (not shown) to improve an image of the user. For example, the motor may be configured to tilt camera 32 in a horizontal and/or vertical plane to substantially center on occupants of vehicle 10. The motor may also be configured to adjust the focal point of the camera 32 to enhance detection of the occupant(s).
Vehicle 10 can optionally include one or more cameras and/or sensors mounted to capture video and/or images and/or movement in areas exterior to vehicle 10. For example, additional cameras 33 and 35 can be mounted at front end 12 and rear end 13, respectively, of vehicle 10. Video and/or images captured by cameras 33, 35 can be displayed on user interface 28.
Manual input 34 is configured to receive a user input and indicate a vehicle operation. For example, if the driver moves manual control 34 in a first direction (e.g., by depressing manual control 34), vehicle 10 provides an indication of an intention to move vehicle 10 in the first direction, and if the driver moves manual control 34 in the second direction (e.g., by raising manual control 34), vehicle 10 provides an indication of an intention to move vehicle 10 in a second direction. It also is contemplated that varying degrees of actuation of manual input 34 provide different indications. For example, if the driver depresses/raises the manual control 34 to a certain extent, vehicle 10 may indicate that the driver intends to change lanes. However, if the driver depresses/raises manual control 34 more drastically (e.g., past a detent or to a stop), vehicle 10 may indicate that the driver intends to make a turn, for example, onto a cross-street. In some embodiments, the indications may be provided by an image (e.g., images 60, 62 depicted in
Projector(s) 42, 44 may be any number of devices configured to project an image. In some embodiments, projector(s) 42, 44 includes a short throw projector, such that projector(s) 42, 44 are placed within a few feet from bumper(s) 14, 20 and project on a significant portion (e.g., entire outer surface) of bumper(s) 14, 20. Projector(s) 42, 44 may be configured to provide a high definition image onto a curved surface of at least a portion of bumpers 14, 20.
Bumpers 14, 20 may be constructed of several different materials configured to provide the desired flexibility and transparency, and may be formed in a variety of shapes to provide safety, aerodynamics, and/or aesthetics. For example, bumpers 14, 20 may each include at least a portion, or the entirety formed of, a transparent material, such as poly(methyl methacrylate) (PMMA), which is known under its trade name Plexiglas. In some embodiments, bumpers 14, 20 include a coating on an inner surface of the transparent material to form a translucent surface enhancing the image, color, and resolution of images projected by projector(s) 42, 44. In some embodiments, the coating includes a pearlescent coating configured to refract the wavelengths emitted by projector(s) 42, 44. An exemplary method of coating bumpers 14, 20 is discussed with regard to
Projector 42 is controlled to display image 60 onto an outer surface of vehicle 10 (e.g., outer surface 58 of bumper 14), and image 60 may provide a variety of types of information. For example, image 60 may be projected through the inner surface of bumper 14 to be displayed on at least a portion of outer surface 58 of bumper 10, that is viewable by a person outside of vehicle. Image 60 may include text (e.g., “Exiting Left”) that may be stationary or moving relative to external surface 58. Image 60 may additionally or alternatively include video, animation, color, shading, gauges, meters, symbols, and/or schematic indicators. For example, image 60 may display information from a variety of sources as discussed below regarding
Projector 44 is controlled to display image 62 onto an outer surface of vehicle 10 (e.g., outer surface 59 of bumper 20), and image 62 may provide a variety of types of information. For example, image 62 may be projected through the inner surface of bumper 14 to be displayed on at least a portion of outer surface 59 of bumper 20, that is viewable by a person outside of vehicle. Image 62 may include text (e.g., “Exiting Left”) that may be stationary or moving relative to external surface 59. Image 62 may additionally or alternatively include video, animation, color, shading, gauges, meters, symbols, and/or schematic indicators. For example, image 62 may display information from a variety of sources as discussed below regarding
Method 1000 may include applying one or more layers of pearlescent paint, in a number of different orders. The layers may include a base paint, such as a Xyrellic Pearl Basecoat (XPB-Base) such as, for example, manufactured by PPG. The layers may also include an Ice White F1 Paint, produced, for example, by thoroughly mixing 40 grams of the XPB-Base with 3 grams of Ecket OEM fine silver. The layers may further include an Ice White F1 Pearl Paint, produced, for example, by diluting the Ice White F1 Paint at a ratio of 1 part to 1.5 parts of XPS90030 Durethane Solvent Reducer. One or more of the coatings may be applied using a spray gun at a pressure between about 18 and 22 psi.
Although discussed regarding bumpers 14, 20, method 1000 may be applied to any number of other surfaces of vehicle 10. For example, it may be applied to a rear or side surface of a bus to display advertisements. Furthermore, modifications to the one or more layers may vary depending on the degree of desired color matching to the rest of vehicle 10.
In step 1010, the inner surface of bumpers 14, 20 is lightly scuffed in order to aid in paint adhesion.
In step 1020, at least one pearlescent coat is applied to the lightly scuffed inner surface of bumpers 14, 20. For example, in some embodiments, three coats of Ice White F1 Pearl Paint are applied to the inner surface of bumpers 14, 20. The pearlescent coat is configured to refract specific wavelengths of light to create a favorable image, color, and resolution.
In step 1030, at least one base coat is applied to the inner surface of bumpers 14, 20 on the pearlescent coat. For example, in some embodiments, two coats of XPB-Base are applied to the inner surface of bumpers 14, 20.
In step 1040, at least one coat of ice white paint is applied on the base coat. For example, in some embodiments, two coats of Ice White F1 Pearl Paint are applied to the inner surface of bumpers 14, 20 on the XPB-Base coat.
In some embodiments, steps 1030 and 1040 are omitted.
While method 1000 includes use of Ice White F1 paint in both steps 1020 and 1040, the disclosure is not so limited. Method 1000 may be practiced with equal effectiveness using a paint color other than Ice White.
I/O interface 102 is configured for two-way communication between controller 100 and various components of projector system 11, such as battery pack 25, GUI 28, microphone 30, cameras 32, 33, 35, manual input 34, projector(s) 42, 44, and/or a server 80 via communication cables, wireless networks, or other communication mediums. I/O interface may also send and receive operating signals to and from battery pack 25, GUI 28, microphone 30, cameras 32, 33, 35, manual input 34, projector(s) 42, 44, and/or server 80.
Processor 104 is configured to receive and process signals to determine a plurality of conditions of the operation of vehicle 10. Processor 104 may also be configured to generate and transmit command signals, via I/O interface 102, in order to actuate the devices in communication. For example, processor 104 may be configured to display images on GUI 28, actuate cameras 32, 33, 35, and/or actuate projector(s) 42, 44 to display images 60, 62.
In some embodiments, processor 104 may be configured to receive signals indicating a condition that the vehicle has been unlocked and/or started, and control projectors 40, 42 to display images or animation indicating the condition.
In some embodiments, processor 104 is configured to receive an input from occupants and control projectors 40, 42 to project images 60, 62 according to the input. For example, processor 104 may be configured to receive a signal indicative of an intended movement of the vehicle and actuate projector(s) 42, 44 to project the image based on the signal. For instance, if the occupant raises/depresses manual input 34, projector(s) 42, 44 projects image(s) 60, 62, such as “CHANGING LANES” or “TURNING LEFT/RIGHT.” Processor 104 may also receive an input from occupants through GUI 28, microphone 30, and/or camera 32. For example, the occupant may request turn-by-turn directions to a destination through GUI 28, which is relayed to projector(s) 42, 44. For example, when vehicle 10 is approaching an exit ramp en route to the destination, images 60, 62 may display “EXITING LEFT.” Processor 104 may receive verbal commands from occupant(s) through microphone 30, and display images 60, 62 via projector(s) 42, 44. For example, if the occupant says “I want to exit left” or “Tell the vehicle behind me that their brights are on,” processor 104 may generate a signal to projector(s) 42, 44 to project an indicative images 60, 62. Processor 104 may also be configured to generate a signal to projector(s) 42, 44 based on detected gestures, according to image recognition software. For example, processor 104 may be configured to detect hand gestures and/or facial expressions that are predefined to indicate an images 60, 62 to be projected by projector(s) 42, 44.
In some embodiments, processor 104 is configured to control projectors 42, 44 to display images 60, 62 in order to provide information about the occupants. For example, processor 104 may be configured to aggregate data from microphone 30 and camera 32 to determine physical and/or mental conditions of occupant(s), such as sleep deprivation, intoxication, attentiveness, distraction, and/or other activities. For example, processor 104 may be configured to detect eye movement of the driver, and/or based on images from camera 33 or a movement sensor mounted on front end 12, determine that a pedestrian is within the driver's sightline, and indicate to the pedestrian that it is safe for them to cross the pathway of vehicle 10. For example, processor 104 may be configured to generate a signal to project images 60, 62, which can be animations that represent the movement of pedestrians as they cross the pathway of the vehicle. Similarly, processor 104 may be configured to detect movement behind vehicle 10 based on images from camera 35 or a movement sensor mounted on rear end 13, to determine the presence of a pedestrian, and generate a signal to project images to indicate it is safe for the pedestrian to cross behind vehicle 10. Processor 104 may also process data from microphone 30 and/or camera 32 to detect slurred speech, dilated pupils, closing eyelids, and/or misdirected eye movement, and generate a signal to project warning images 60, 62, such as “DISTRACTED DRIVER” or “DRUNK DRIVER.” In some embodiments, images 60, 62 may improve aesthetics, such as creating virtual shadows representing people that are crossing in front of vehicle 10.
In some embodiments, processor 104 is configured to receive data from server 80 and project images 60, 62. For example, processor 104 may be configured to receive data from server 80 related to social media, advertisements, traffic information, road hazards, weather, and/or emergency information (e.g., amber alerts). Processor 104 may then control projectors 42, 44 to project images 60, 62 based on the received data to be viewed by people outside of vehicle 10. For example, processor 104 may be configured to integrate social media feeds of passengers, and project images 60, 62.
In some embodiments, processor 104 is configured to receive data from sensors of vehicle 10, and project the data in images 60, 62. For example, processor 104 may be configured to project information, such as a current battery life of battery pack 25, charging status, scheduled maintenance, fluid levels, temperatures, tire status, and/or other similar types of data. In this manner, processor 104 may provide information to the person without have to enter cabin 26. Processor 104 may also be configured to project information.
For example, with regard to charging status, processor 104 can be configured to control projectors 42, 44 to project images 60, 62 such as an animation indicating that the vehicle is charging, including text that indicates an estimated time until vehicle 10 is charged.
In some embodiments, processor 104 may be configured to detect the presence of a known mobile device of a person associated with vehicle 10 (e.g., an owner) and control projectors 42, 44 to project information based on the data received from the sensors of vehicle 10, e.g., battery life, charging status, fluid levels, etc.
In some embodiments, processor 104 is configured to provide information for ride sharing functions, such as taxi cabs, Uber™, and/or Zipcar™. For example, when a driver is approaching a person that he intends to pick up, the images 60, 62 may display the person's name, the driver's name, a destination of the vehicle, the number of passenger seats available, a rating of the driver, and/or advertise a fare. Also, for example, processor 104 may be configured to detect that a person the driver is approaching is a ride share user, and control projectors 42, 44 to project images 60, 62 to suitably notify the rideshare user.
In some embodiments, processor 104 may be applied to the race car industry. For example, images 60, 62 may display an assigned number, livery colors (e.g., team colors), overall position of the vehicle or team, driver response time, breaking force, speed, g-forces, and/or any other similar types of data.
Storage unit 106 is configured to store one or more computer programs for execution by controller 100 to perform functions of projector system 11. For example, storage unit 106 may be configured to store software for image processing software, e.g., image recognition software, such that the software may distinguish a person from inanimate objects, and distinguish identities, facial features, and/or other biometric aspects of users. Storage unit 106 may also be configured to store look-up charts including data, such as, a maintenance schedule, indicative levels of battery life, and/or personal information of occupants.
Another aspect of the disclosure is directed to a non-transitory computer-readable medium storing instructions which, when executed, cause one or more processors to perform the methods, as discussed above. The computer-readable medium may include volatile or non-volatile, magnetic, semiconductor, tape, optical, removable, non-removable, or other types of computer-readable medium or computer-readable storage devices. For example, the computer-readable medium may be the storage unit or the memory module having the computer instructions stored thereon, as disclosed. In some embodiments, the computer-readable medium may be a disc or a flash drive having the computer instructions stored thereon.
It will be apparent to those skilled in the art that various modifications and variations can be made to the disclosed system and related methods. Other embodiments will be apparent to those skilled in the art from consideration of the specification and practice of the disclosed remote control system and related methods. It is intended that the specification and examples be considered as exemplary only, with a true scope being indicated by the following claims and their equivalents.
The present application claims the benefit of U.S. Provisional Application No. 62/322,914, filed Apr. 15, 2016, the entirety of which is hereby incorporated by reference.
Number | Date | Country | |
---|---|---|---|
62322914 | Apr 2016 | US |