This disclosure relates to a system and method for detecting a vehicle event and generating review criteria that correspond to the vehicle event.
Systems configured to record, store, and transmit video, audio, and sensor data associated with a vehicle responsive to an incident involving the vehicle are known. The video from the incident is usually analyzed by a user at a later time after the incident. Vehicle Engine Control Module (ECM) systems are known. Such systems interface with external computers (e.g., at an automobile mechanic) where the data stored by the ECM system is analyzed.
One aspect of the disclosure relates to system and method for detecting a vehicle event and generating review criteria based on the vehicle event. The system is configured to detect a vehicle event, record information representing the vehicle event, and generate review criteria relevant to the information recorded. Some or all of the system may be installed in the vehicle and/or be otherwise coupled with the vehicle. In some implementations, the system may include multiple video capture devices configured to acquire visual output information representing a vehicle environment, one or more sensors configured to generate output signals conveying vehicle operation information related to the vehicle, and/or other components. The multiple video capture devices may have one or more fields of view. The fields of view may include areas of the vehicle environment from which the multiple video capture devices are configured to acquire visual output information. The fields of view may be associated with a direction (e.g., forward looking, rear looking), a location (on top of a vehicle, on the side of a road), and/or other portions of the vehicle environment. An individual video capture device may have an individual field of view that corresponds to one or more other fields of view from other video capture devices. For example, a first field of view from a front looking video capture device and a second field of view from a video capture device positioned on the side of a road may both capture visual information related to the same vehicle event. In some implementations, the system may detect a vehicle event and record information based on the detection of a vehicle event. In some implementations, the system may determine a vehicle event type and generate review criteria based on the vehicle event type and/or fields of view corresponding to the video capture devices. In some implementations, the system may include one or more of a video capture device, a sensor, a processor, an operator interface, external resources, electronic storage, a remote computing device, a reviewer interface, and/or other components.
The video capture devices may be configured to acquire visual output information representing the vehicle environment. In some implementations, visual output information representing the vehicle environment may be acquired from system-integrated and/or non-system integrated vehicle capture devices installed in the vehicle and/or the vehicle environment. The vehicle environment may include spaces in and around the interior and exterior of the vehicle. In some implementations, video capture devices may include individual cameras, a given one of the individual cameras corresponding to an individual field of view.
The sensors may be configured to generate output signals conveying vehicle operation information related to the operation and/or the context of the vehicle. Vehicle operation information related to the operation of the vehicle may include feedback information from one or more of the mechanical systems of the vehicle, one or more safety systems of the vehicle, one or more non-standard aftermarket sensors installed in the vehicle, and/or other information. In some implementations, at least one of the sensors may be a vehicle system sensor included in an ECM system of the vehicle. Information related to the context of the vehicle may include information related to the vehicle environment in and/or around the vehicle. In some implementations, the output signals conveying the vehicle operation information related to the context of the vehicle may be generated via non-standard aftermarket sensors installed in the vehicle, and/or other sensors. For example, one or more individual sensors may be and/or include a proximity sensor. Based on an analysis of the output signals from this sensor, the system may automatically determine a person, vehicle, object and/or other objects are in proximity to the vehicle. This example sensor may be one of a plurality of sensors in the system.
One or more processors may be configured to execute one or more computer program components. The computer program components may comprise one or more of a parameter component, a detection component, a classification component, a control component, and/or other components.
The parameter component may be configured to determine one or more vehicle event parameters. The parameter component may determine the vehicle event parameters based on the information conveyed by the output signals from the sensors, visual output information acquired from the video capture devices, information acquired from external resources, and/or other information.
The detection component may be configured to detect the occurrence of the vehicle event. The detection component may be configured to detect the vehicle event in real-time or near real-time. The detection component may be configured to detect the vehicle event based on the information conveyed by the output signals generated by the sensors, the visual output information acquired by the video capture devices, information from external resources, information determined by the parameter component, and/or based on other information. For example, the vehicle event may be detected based on a comparison of the determined vehicle event parameters to predetermined vehicle event criteria.
The classification component may be configured to determine a vehicle event type for individual vehicle events. A vehicle event may have one or more vehicle event types. The classification component may determine the vehicle event type in response to the detection of the vehicle event by the vehicle detection component. In some implementations, the vehicle event type may be determined based on pre-set criteria which may be customizable for the system.
The control component may be configured to generate review criteria that correspond to the vehicle event. The review criteria may be generated based on the determined vehicle event type and the fields of view corresponding to the video capture devices. In some implementations, the control component will generate review criteria that correspond to specific fields of view and specific vehicle event types. The review criteria may be generated for a manual reviewer, for example.
The control component may be configured to facilitate wireless communication of information conveyed by the output signals, the visual output information, information related to a detected vehicle event, and/or other information to the remote computing device and/or other devices. The control component may be configured to facilitate communication responsive to the detection of the vehicle event. The control component may be configured to facilitate communication in real-time or near real-time. The control component may be configured to facilitate wireless communication of the information conveyed by the output signals, the visual output information, information related to the detected vehicle event, and/or other information from a period of time that includes at least the vehicle event.
In some implementations, the control component may be configured to facilitate recording of the vehicle event data based on detection of the vehicle event. Vehicle event data may include video, audio, ECM, metadata, GPS, and/or other sensor data for a period of time that lasts from before a given vehicle event starts until after it ends. The period of time for which data is recorded may be determined based on the timing of the vehicle event, the point in time at which the vehicle event is detected, the type of vehicle event, and/or other information.
In some implementations, the control component may be configured to selectively determine one or more types of data that are recorded based on vehicle event types and fields of view corresponding to the video capture devices. For example, the control component may selectively determine whether to record visual output information from devices having one or more specific fields of view, types of data, and/or other vehicle and/or visual output information. Continuing with the example, if the detected vehicle event involved the opening of a passenger door of the vehicle, the control component may cause a video stream from an interior camera with an interior field of view to be included in the event record for monitoring the loading and unloading of passengers. Whereas the interior field of view may not be needed when recording vehicle situations for other (non-interior) vehicle events.
In some implementations, the control component may be configured to determine how much of the event data (video, audio, sensor, etc.) to record before the vehicle event (e.g., pre-vehicle event duration) and after the vehicle event (e.g. post-vehicle event duration) based on the vehicle event itself and/or other information. This may allow the visual output information and/or vehicle operation information to reflect a relevant and/or desired time frame relative to the vehicle event.
In some implementations, the external resources may include an adaptor for adapting analog visual output information to digital visual output information. This may facilitate adapting visual output information acquired from non-system integrated video capture devices so that it has the same format as visual output information acquired from system cameras, and/or other devices. In some implementations, the adaptor may facilitate adapting one or more of a frame rate, a resolution, a brightness, a saturation, and/or an image correction of visual output information acquired from non-system integrated video capture devices.
These and other objects, features, and characteristics of the system and/or method disclosed herein, as well as the methods of operation and functions of the related elements of structure and the combination of parts and economies of manufacture, will become more apparent upon consideration of the following description and the appended claims with reference to the accompanying drawings, all of which form a part of this specification, wherein like reference numerals designate corresponding parts in the various figures. It is to be expressly understood, however, that the drawings are for the purpose of illustration and description only and are not intended as a definition of the limits of the invention. As used in the specification and in the claims, the singular form of “a”, “an”, and “the” include plural referents unless the context clearly dictates otherwise.
Vehicle environment 124 may include spaces in and/or around an interior and/or an exterior of vehicle 122. System 100 may include one or more sensors 112 configured to generate output signals conveying information related to the vehicle 122. In some implementations system 100 may detect the vehicle event based on the information conveyed by the output signals from sensors 112 and/or other information. System 100 may determine a vehicle event type based on information conveyed by the output signals from sensors 112, visual output information from multiple video capture devices 114, and/or other information. In some implementations, system 100 may generate review criteria that correspond to the vehicle event. The review criteria may be generated based on the vehicle event type and the fields of view corresponding to the video capture devices 114.
Review criteria may prompt a reviewer to review a vehicle event based on the vehicle event type, the fields of view corresponding to the video capture devices, and/or any other information relevant to the vehicle event. Review criteria may include questions, inquiries, prompts, and/or other review criteria related to the vehicle event. The review criteria may guide a reviewer to review relevant aspects of the vehicle event.
System 100 may be configured to monitor vehicles, equipment, drivers, operators, and/or other information to facilitate reduction of risky behaviors, facilitate coaching and/or training, mitigate damage to a vehicle, and/or for other purposes. System 100 may detect and record vehicle events and facilitate review of the vehicle events by a manual reviewer based on the review criteria. In some implementations, system 100 may determine one or more video capture devices 114 with corresponding fields of view relevant to the vehicle event detected. System 100 may generate review criteria for review by a manual reviewer based on these fields of view and/or other information.
In some implementations, system 100 may include one or more of a sensor 112, a video capture device 114, an external resource 116, an operator interface 118, electronic storage 120, a processor 102, a remote computing device 128, and/or other components.
In some implementations, one or more of the components of system 100 may form at least a portion of a vehicle event recorder system such as the vehicle event recorder system described in U.S. patent application Ser. No. 13/957,810 filed Aug. 2, 2013 entitled “Vehicle Event Recorder Systems and Networks Having Integrated Cellular Wireless Communications Systems”, which is incorporated herein by reference.
One or more video capture devices 114 may be configured to acquire visual output information representing a vehicle environment (e.g., vehicle environment 124). Video capture devices 114 may include systems for compiling and/or collecting video information, for example, video cameras, digital cameras, analog cameras, still cameras, infrared sensors, and/or other video capture devices 114 that acquire visual output information. In some implementations, video capture devices 114 may include system-integrated video capture devices, non-system integrated video capture devices (e.g., video capture devices included in external resources 116), and/or other video capture devices. Visual output information may include videos, images, clips, broadcasts, pictures, visual data and/or other visual output information representing vehicle environment 124. In some implementations, multiple video capture devices 114 may be positioned in and/or around vehicle 122 and synchronized together to provide a 360 degree view of the inside of vehicle 122 and/or a 360 degree view of the outside of vehicle 122.
In some implementations, video capture devices may be enabled and/or disabled based on user configurability, randomly determined time intervals, the detection of a vehicle event, information from one or more sensors, and or other information. For example, one or more cameras may be enabled and/or disabled based on one or more geo-fences. An individual geo-fence may be a virtual boundary line that defines, represents, and/or is otherwise related to a physical area (e.g., a storage yard where fleet vehicles are stored), a point of interest, a segment of a road/highway/etc., and/or other physical locations. Geo-fences may define areas where a particular action and/or maneuver is allowed. Geo-fences may define areas where particular driving behaviors are disallowed and/or particularly dangerous (e.g., a narrow country road). Specific geo-fences may be associated with specific vehicle events. For example, the payment of roadway tolls may be expected in a geo-fenced area near a tollbooth. In some implementations, geo-fences may be associated with a time of day. For example, a geo-fence may surround a large parking lot at a mall. The geo-fence may apply only during hours that the mall is open, and not, for example, during the early morning or late night hours. In some implementations, one or more cameras may be enabled and/or disabled based on an operator login identification information, dispatch information, and/or other information.
Although video capture devices 114 are depicted in
Referring to
Individual ones of the one or more video capture devices 114 may have separate corresponding fields of view of the vehicle environment 124. In some implementations, the separate corresponding fields of view may include, for example, one or more of a forward view, a rear view, a passenger's side view, a driver's side view, a driver view, a cargo view, a left view, a right view, and/or other suitable corresponding fields of view.
One or more sensors 112 may be configured to generate output signals conveying information related to the operation and/or the context of vehicle 122. Information related to the operation of vehicle 122 may include feedback information from one or more of the mechanical systems of vehicle 122, one or more non-standard aftermarket sensors 112 coupled with vehicle 122, and/or other feedback information from other sensors. The mechanical systems of vehicle 122 may include, for example, the engine, the drive train, the lighting systems (e.g., headlights, brake lights), the braking system, the transmission, fuel delivery systems, door/compartment locking systems and/or other mechanical systems. The mechanical systems of vehicle 122 may include one or more mechanical sensors, electronic sensors, and/or other sensors that generate the output signals (e.g., seat belt sensors, tire pressure sensors, etc.).
In some implementations, at least one of sensors 112 may be a vehicle system sensor included in an ECM system of vehicle 122. For example, in some implementations, system 100 may be configured such that a first sensor 112 may be a fuel level sensor which may be associated with an ECM system of vehicle 122. In some implementations, non-standard aftermarket sensors 112 may be coupled with vehicle 122 such that they generate output signals that convey information related to the mechanical systems of vehicle 122. For example, a non-standard aftermarket sensor 112 may be configured to generate output signals that convey information related to braking force. Information related to the context of vehicle 122 may include information related to vehicle 122 relative to the vehicle environment 124. The information related to the context of vehicle 122 may include information related to movement of vehicle 122 (e.g., acceleration, velocity, etc.), an orientation of vehicle 122, a geographic position of vehicle 122, a spatial position of vehicle 122 relative to other objects, a tilt angle of vehicle 122, an inclination/declination angle of vehicle 122, other objects or persons in proximity to vehicle 122, weather conditions surrounding vehicle 122, and/or other information. In some implementations, the output signals conveying the information related to the context of vehicle 122 may be generated via non-standard aftermarket sensors installed in vehicle 122, and/or other sensors. The non-standard aftermarket sensor may include, for example, a video camera (e.g., a video capture device 114), a microphone, an accelerometer, a gyroscope, a geolocation sensor (e.g., a GPS device), a magnetometer, radar (e.g. for measuring distance of leading vehicle), a temperature sensor, a light sensor, and/or other sensors.
Although sensors 112 are depicted in
Although video capture devices 114 and sensors 112 are depicted in
As shown in
In some implementations, control component 110 may be configured to associate the visual output information from multiple video capture devices 114, the vehicle operation information conveyed by the output signals of the one or more sensors 112, the vehicle event type, the vehicle event data, the review criteria, and/or the information from any external resources 116, with information related to operation and/or context of the vehicle that corresponds to a given vehicle event into a vehicle event record for that vehicle event.
Parameter component 104 may be configured to determine one or more vehicle event parameters of vehicle 122. Parameter component 104 may determine the vehicle event parameters based on the information conveyed by the output signals from sensors 112, the visual output information acquired by the video capture devices 114, information provided by external systems and/or databases, and/or other information. The one or more vehicle event parameters may be related to the operation of vehicle 122, the context of vehicle 122, vehicle environment 124, and/or other information. For example, the one or more vehicle event parameters may be related to one or more of an acceleration, a direction of travel, a vehicle speed, an engine speed (e.g. RPM), a duration of time, a closing distance, a lane departure from an intended travelling lane of the vehicle, a following distance, the proximity of another car or person to vehicle 122, the opening of a door or compartment of vehicle 122, the arrival of vehicle 122 at a specific destination, and/or other parameters. Vehicle event parameters may have associated predetermined criteria.
Detection component 106 may be configured to detect a vehicle event. Detection of a vehicle event may be based on the information conveyed by the output signals from the one or more sensors 112, the visual output information from the multiple video capture devices 114, parameters determined by parameter component 104, and/or other sources of vehicle operation information. Detection component 106 may detect a vehicle event in real-time, near-real time, and/or other time. In some implementations, vehicle events may have associated predetermined detection criteria. The detection of a vehicle event may be based on one or more determined vehicle event parameters satisfying predetermined criteria. In some implementations, a vehicle event may be detected based on, for example, information, output signals, determined parameters, and/or other information related to one or more of acceleration or deceleration, a direction of travel, a vehicle speed, an engine speed (e.g. RPM), a duration of time, a closing distance, a lane departure, a following distance, another vehicle in close proximity, a person in close proximity, an open door or compartment, a geolocation, a collision, an assault on the operator of vehicle 122, the loading and unloading of cargo from and onto vehicle 122, a passenger incident, property theft, damage to vehicle 122, operator engagement in unsafe or non-compliant activities, fare payments, a video capture device trigger, and/or other vehicle events related to vehicle 122. For example, the detection of a vehicle event such as backing may be based on a video capture device that detects backing presenting and/or capturing visual output information. In some implementations, a vehicle event may include external events not related to vehicle 122. For example, an external event may include an explosion at a job site located within one or more fields of view corresponding to one or more video capture devices of vehicle 122.
In some implementations, detection component 106 may be configured to detect a vehicle event based on output signals from one or more sensors 112 remotely triggered by a user and/or operator. A user and/or operator may remotely trigger one or more of sensors 112 to output information detected by detection component 106 indicating a vehicle event. For example, if a user and/or operator notice something amiss, the user and/or operator may use a remote push bottom to wirelessly trigger one or more sensors to indicate a vehicle event and record visual output information from one or more video capture devices.
By way of a first non-limiting example, detection component 106 may be configured to detect a vehicle event such as the arrival of vehicle 122 at a designated geolocation based on the output signals from a geolocation sensor 112. Detection component 106 may be configured to detect the arrival of vehicle 122 at a designated geolocation in order to monitor a situation such as the payment of fares. Detection component 106 may detect the arrival of vehicle 122 at a designated geolocation based on geolocation parameters determined by parameter component 104 satisfying predetermined geolocation criteria.
By way of a second non-limiting example, detection component 106 may be configured to detect a vehicle event such as a departure from the intended vehicle lane to monitor a potential situation such as a collision.
By way of a third non-limiting example, detection component 106 may be configured to detect a vehicle event such as the opening of a door and/or compartment of vehicle 122 to monitor the loading or unloading of cargo and/or passengers.
By way of a forth non-limiting example, detection component 106 may be configured to detect a vehicle event such as the opening of a cargo compartment and/or motion within a cargo compartment while vehicle 122 is not located at a designated delivery location to monitor a possible situation such as theft.
Classification component 108 may be configured to determine a vehicle event type based on the information conveyed by the output signals from one or more sensors 112, the visual output information from multiple video capture devices 114, information determined by parameter component 104, information determined by detection component 106, information from external resources 116, and/or other information. In some implementations, determining the vehicle event type may include classification of the vehicle event into one or more vehicle event type categories. In some implementations, for example, the vehicle event type may be one or more of a rear vehicle event type, a front vehicle event type, a side vehicle event type, an interior vehicle event type, an exterior vehicle event type, a damage vehicle event type, a fare vehicle event type, a loading/unloading vehicle event type, an operator vehicle event type, and/or other vehicle event types. In some implementations, vehicle event types may be customizable and/or determined by a user. The user may include the operator, reviewer, and/or other users. The user may customize the vehicle event types, for example, a user may determine the vehicle event types include front vehicle events, rear vehicle events, side vehicle events, and/or other vehicle events determined by the user. In some implementations, a vehicle event type may be based on visual output information being available from one or more non-system video capture devices and/or other external resources 116.
Control component 110 may be configured to trigger vehicle event recording (e.g., recording of visual output information and/or other sensor information associated with a vehicle event) based on vehicle event detection. Visual output information and/or sensor information may include video, audio, ECM, metadata, and/or other sensor data for a period of time that lasts from before a given vehicle event starts until after it ends. The period of time for which data is recorded may be determined based on the timing of the vehicle event, the point in time at which the vehicle event is detected, and/or other information. Control component 110 may be configured to trigger recording of substantially all vehicle event data (e.g., video, audio, sensor output signals, etc.), and not just the signals, visual output information, and/or parameters used for vehicle event detection. In some implementations, control component 110 may be configured to trigger recording of visual output information from individual ones of the one or more video capture devices that may have acquired visual output information relevant to the detected vehicle event based on its corresponding field of view and/or the vehicle operation information conveyed by the output signals from sensors 112.
Control component 110 may be configured such that the visual output information and/or vehicle operation information is saved to a non-volatile memory (e.g., included in electronic storage 120) and later offloaded wirelessly via a vehicle event recorder system such as the vehicle event recorder system described in U.S. patent application Ser. No. 11/377,167 filed Mar. 16, 2006 and entitled, “Vehicle Event Recorder Systems And Networks Having Integrated Cellular Wireless Communications Systems” (incorporated by reference above), and/or the vehicle event recorder system described in U.S. patent application Ser. No. 11/377,157 filed Mar. 16, 2006 and entitled, “Vehicle Event Recorder Systems and Networks Having Parallel Communications Links,” which is incorporated herein by reference.
Control component 110 may be configured to generate review criteria corresponding to the vehicle event. The generation of review criteria may be responsive to the detection of the vehicle event by detection component 106. Review criteria may be generated based on the vehicle event type, fields of view corresponding to the video capture devices 114, and/or other information. Control component 110 may determine which video capture devices 114 may have acquired visual output information relevant to the detected vehicle event based on their corresponding fields of view. Control component 110 may generate review criteria for the visual output information acquired by these individual video capture devices 114. In some implementations, the review criteria for a first individual field of view corresponding to a first individual video capture device 114 may be different than the review criteria for a second individual field of view corresponding to a second individual video capture device 114.
By way of a first non-limiting example, review criteria generated for a rear field of view may be different than review criteria generated based on a driver field of view.
By way of a second non-limiting example, review criteria generated based on an exterior vehicle event type may be different than review criteria based on an interior vehicle event type.
By way of a third non-limiting example, review criteria generated based on an exterior vehicle event type and the field of view corresponding to a rear video capture device may be different from review criteria generated based on an interior vehicle event type and the field of view corresponding to a rear video capture device, which may be further different from review criteria generated based on an exterior vehicle event type and the field of view corresponding to a front video capture device.
Control component 110 may be configured to facilitate wireless communication of vehicle operation information, the visual output information, the vehicle event detection information, the vehicle event type, the vehicle event data, the review criteria, and/or other information to remote computing device 128 and/or other devices. In some implementations, control component 110 may be configured to facilitate wireless communication of visual output information as a live feed to a wireless handheld device.
Control component 110 may be configured to facilitate communication responsive to the detection of a vehicle event. Control component 110 may be configured to facilitate communication, in real-time, near-real time, and/or other time, of individual visual output information from individual video output devices. The individual visual output information communicated, the timing of the communication, and/or other characteristics of the wireless communication may be based on one or more of the vehicle event type, an intensity, the one or more fields of view available, and/or other information. The one or more fields of view available may be based on the determination of which video capture devices 114 are enabled and/or may have acquired visual output information relevant to the detected vehicle event based on their corresponding fields of view. By way of non-limiting example, visual output information from a video capture device having a rear field of view may be communicated if the vehicle event type is a rear collision.
Control component 110 may be configured to facilitate communication in real-time, near real-time, and/or other time. For example, control component 110 may facilitate one or more individual communications during the operation of vehicle 122. Each individual communication may be responsive to a detected vehicle event and/or may occur just after detection of a vehicle event. In some implementations, control component 110 may be configured to facilitate communication after use of vehicle 110 has ceased such that the vehicle operation information, visual output information, information identifying the vehicle event detected by detection component 106, the vehicle event type, the vehicle event data, the review criteria, and/or other information is communicated in a single communication and/or multiple communications.
Control component 110 may be configured to facilitate storage and/or wireless communication of the vehicle operation information conveyed by the output signals, the visual output information, the information identifying the vehicle event detected by detection component 106, the vehicle event type, the vehicle event data, the review criteria, information from any external resources 116, and/or other information from a period of time that includes at least the vehicle event. For example, control component 110 may facilitate communication and/or storage of the information conveyed by the output signals, the visual output information, and/or other information from a period of time that begins before the beginning of the detected vehicle event, lasts throughout the vehicle event, and ends after the end of the detected vehicle event. Control component 110 may be configured such that the period of time that includes at least the vehicle event is centered around the specific vehicle event and includes a pre-vehicle event time, the vehicle event, and a post-vehicle event time, wherein the pre-vehicle event time and the post vehicle event time are approximately equal. In some implementations, control component 110 may be configured such that the period of time that includes at least the vehicle event begins with the vehicle event and continues for a post-vehicle event period of time.
In some implementations, control component 110 may be configured to determine how much of the vehicle event data (video, audio, sensor, etc.) to record before the vehicle event (e.g., pre-vehicle event duration) and after the vehicle event (e.g. post-vehicle event) based on the vehicle event itself. This may allow the event data to show a relevant and/or desired time frame.
For example, responsive to detection of the opening of a passenger door by detection component 106, control component 110 may be configured to facilitate wireless communication and/or electronic storage of the vehicle operation information conveyed by the output signals, the visual output information, the vehicle event type, the vehicle event data, the review criteria, and/or other information. The communicated and/or stored information may include, for example, video images showing the operator performing the vehicle event, video images before the performance of the vehicle event, video images after the performance of the vehicle event, the vehicle event type, the corresponding fields of view of video capture devices 114, and/or other information.
In some implementations, control component 110 may be configured to facilitate temporary electronic storage of the vehicle information conveyed by the output signals, the visual output information, the vehicle event type, the corresponding fields of view or the video capture devices and/or other information in electronic storage (as described above). Control component 110 may be configured to facilitate temporary electronic storage until the information is communicated to remote computing device 128. Control component 110 may be configured to erase the data temporarily stored in electronic storage 120 responsive to the communication to remote computing device 128.
In some implementations, control component 110 may be configured to cause one or more sensors 112 and/or one or more video capture devices 114 to operate continuously. For example, control component 110 may cause a video camera (e.g., video capture device 114) to continuously generate video images (e.g., visual output information) of vehicle 122, the vehicle environment 124, and/or other vehicle information. Control component 110 may be configured to cause system 100 to continuously buffer visual output information and vehicle operation information which may include, for example, video, audio, sensor, and/or other information. Control component 110 may be configured such that once the vehicle event occurs, the buffered data related to the vehicle event is transferred from a volatile to a non-volatile memory. Control component 110 is configured such that a determination is made in real-time as to whether or not the data related to the vehicle event should be offloaded immediately or kept on the device as backup.
In some implementations, control component 110 may facilitate display of a customizable reviewer interface 126. For example, customizable reviewer interface 126 may facilitate review of the vehicle events by a manual reviewer based on the review criteria. Customizable reviewer interface 126 may include multiple views and may display a portion of the vehicle operation information a portion of the visual output information, and/or other information. The display of the portion of the visual output information may include display of individual ones of the fields of view corresponding to the video capture devices 114. In some implementations, control component 110 may be configured to automatically determine a quantity, the corresponding fields of view, and/or location of visual output information that appears in the customizable reviewer interface. For example, control component 110 may dynamically configure video streams (e.g., visual output information) in groups of two or more. Labeling and/or mapping of visual output information from one or more fields of view may be user configurable.
In some implementations, vehicle operation information and/or visual output information may be routed to one or more given manual reviewers based on a given vehicle event, vehicle event type, information source, information type, and/or other characteristics of the vehicle operation information and/or visual output information. For example, vehicle operation information and/or visual output information corresponding to a rear vehicle event type may be routed to one or more of a set of manual reviewers trained to analyze and/or review vehicle events of the rear vehicle event type. In some implementations, review of vehicle operation information and/or visual output information may be prioritized based on a portion of the vehicle operation information, a portion of the visual output information, a vehicle event type, and/or other characteristics of the vehicle event and/or information.
In some implementations, control component 110 may be configured to synchronize visual output information from one or more video output devices (e.g., synchronize multiple sets of video streams together). For example, visual output information may be synchronized via a field programmable gate arrays located in one or more of video capture devices 114, external resources 116, and/or other locations.
Referring to
Referring to
Referring to
Returning to
In some implementations, users (e.g., customers) may indicate preferences such that system 100 prioritizes individual vehicle event types. Control component 110 may be configured to include a feedback mechanism for users to rate events and receive more or less events of a given vehicle event type via reviewer interface 126, operator interface 118, and/or another interface.
In some implementations, control component 110 may be configured to alert a user if one or more video capture devices are plugged-in, enabled, and/or functioning based on the user configuration.
In some implementations, control component 110 may generate a performance score for a vehicle operator of vehicle 122 at the time of the vehicle event. The performance score may be generated based on information received and/or selected during a manual review of the specific vehicle event facilitated by the review criteria. In some implementations, the performance score may be normalized based on the one or more fields of view corresponding to the video capture devices 114 that acquire visual output information relevant to the specific vehicle event. In some implementations, the performance scores from multiple vehicle events may be averaged or combined to generate an overall performance score. In some implementations, the vehicle operation information, visual output data, and/or information from any external resources, may be associated with the generated performance score in a given event record. The associated vehicle operation information, visual output data, and/or information from any external resources 116, may be compiled and/or filtered based on the generated performance scores for use in a coaching session.
Returning to
In some implementations, the adapter of system 100 may control and/or adapt one or more of a frame rate, a resolution, a brightness, a contrast, a saturation, an image correction, and/or other characteristics of a given visual output information which may be acquired from one or more non-system integrated video capture devices (e.g., external resources 116). By way of non-limiting example, the adaptor of system 100 may be configured to control and/or adapt characteristics of visual output information based on a vehicle event type.
In some implementations, external devices 116 may include a connection device and/or media hub (e.g., a registered jack). The connection device may include strain relief features and/or environment protection features (e.g., IP67).
In some implementations, method 600 may be implemented in one or more processing devices (e.g., a digital processor, an analog processor, a digital circuit designed to process information, an analog circuit designed to process information, a state machine, and/or other mechanisms for electronically processing information). The one or more processing devices may include one or more devices executing some or all of the operations of method 600 in response to instructions stored electronically on one or more electronic storage mediums. The one or more processing devices may include one or more devices configured through hardware, firmware, and/or software to be specifically designed for execution of one or more of the operations of method 600.
Referring to
At an operation 604, output signals conveying vehicle operation information related to the operation and/or context of the vehicle may be generated by one or more sensors. The vehicle operation information conveyed by the output signals may be related to one or more of mechanical systems of the vehicle, movement of the vehicle, an orientation of the vehicle, a geographic position of the vehicle, a spatial position of the vehicle relative to other objects, and/or other operational/contextual characteristics of the vehicle. In some implementations, the sensors may include a camera and the output signals may include visual output information. In some implementations, the output signals may convey information related to mechanical systems of the vehicle. The output signals conveying information related to mechanical systems of the vehicle may be generated by one or more vehicle subsystem sensors included in a vehicle on-board data system. In some implementations, operation 604 may be performed by one or more sensors the same as or similar to sensors 112 (shown in
At an operation 606, a vehicle event may be detected based on information conveyed by the output signals, and/or other information. In some implementations, the vehicle event may be detected based on, for example, information or signals related to one or more of acceleration and/or deceleration, a direction of travel, a vehicle speed, an engine speed (e.g. RPM), a duration of time, a closing distance, a lane departure, a following distance, another vehicle in close proximity, a person in close proximity, an open door or compartment, a geolocation, and/or other vehicle events of vehicle. The detection may be in real-time or near real-time. The detection may be based on the information conveyed by the output signals from the one or more sensors, the visual output information from multiple video capture devices, and/or other information. In some implementations, a set of parameters may describe the vehicle event and the vehicle event may be detected based on those parameters. In some implementations, operation 606 may be performed by a processor component the same as or similar to detection component 106 (shown in
In some implementations, operation 606 may include wireless communication of the vehicle operation information and/or the visual output information. The facilitation of the wireless communication may be in real-time, near real-time, and/or other time. The facilitation of the wireless communication may be responsive to the detection of the vehicle event. The vehicle operation information and the visual output information that is wirelessly communicated may include information for a period of time that may include the vehicle event and/or other information. The information may include, for example, video and/or other data associated with a vehicle event and/or the vehicle environment. In some implementations, operation 606 may be performed by a processor component the same as or similar to control component 110 (shown in
In some implementations, operation 606 may include electronic storage of the vehicle operation information and/or the visual output information. Facilitation of the electronic storage may be responsive to the detection of the vehicle event. The electronically stored information may include the information for the period of time that may include the vehicle event and/or other information. In some implementations, operation 606 may be performed by a processor component and/or electronic storage the same as or similar to control component 110 and/or electronic storage 118 (shown in
At an operation 608, a vehicle event type may be determined based on the vehicle operation information conveyed by the output signals, the visual output information from multiple video capture devices, information from external resources, and/or other information. The determination of a vehicle event type may be responsive to the detection of the vehicle event. In some implementations, the vehicle event type may be determined based on visual output information that may be acquired from multiple video capture devices. The video capture devices may have individual corresponding fields of view. In some implementations, the vehicle event type may be determined based on information conveyed by the individual corresponding fields of view of the video capture devices. The vehicle event type may include more than one vehicle event type and/or one or more vehicle event subtypes. Operation 608 may be performed by a processor component the same as or similar to classification component 108 (shown in
At an operation 610, review criteria may be generated that correspond to the vehicle event. The review criteria may be based on the vehicle event type, the fields of view corresponding to the video capture devices, and/or other information. In some implementations, at operation 610, it may be determined which video capture devices may have acquired visual output information relevant to the detected vehicle event based on their corresponding fields of view. Review criteria may be generated for the visual output information acquired by these individual video capture devices 114. For example, review criteria generated for a rear view may be different that review criteria generated for a side view. In some implementations, the review criteria may be generated based on corresponding fields of view of non-system integrated video capture devices. Non-system integrated video capture devices may be located, for example, in the vehicle environment, on the vehicle, on other vehicles, and/or other locations. Operation 610 may be performed by a processor component similar to/or the same as control component 110 (shown in
In some implementations, method 700 may be implemented in one or more processing devices (e.g., a digital processor, an analog processor, a digital circuit designed to process information, an analog circuit designed to process information, a state machine, and/or other mechanisms for electronically processing information). The one or more processing devices may include one or more devices executing some or all of the operations of method 700 in response to instructions stored electronically on one or more electronic storage mediums. The one or more processing devices may include one or more devices configured through hardware, firmware, and/or software to be specifically designed for execution of one or more of the operations of method 700.
Referring to
At an operation 704, output signals conveying vehicle operation information related to the operation and/or context of the vehicle may be generated by one or more sensors. The vehicle operation information conveyed by the output signals may be related to one or more of mechanical systems of the vehicle, movement of the vehicle, an orientation of the vehicle, a geographic position of the vehicle, a spatial position of the vehicle relative to other objects, and/or other operational/contextual characteristics of the vehicle. In some implementations, the sensors may include a camera and the output signals may include visual output information. In some implementations, the output signals may convey information related to mechanical systems of the vehicle. The output signals conveying information related to mechanical systems of the vehicle may be generated by one or more vehicle subsystem sensors included in a vehicle on-board data system. In some implementations, operation 704 may be performed by one or more sensors the same as or similar to sensors 112 (shown in
At operation 706, the given visual output information having an analog format may be adapted into a given visual output having a digital format. An adapted given visual output and a non-adapted given visual output may then have the same format. In some implementations, one or more given visual outputs may be converged onto a timeline. The timeline may present the visual output information according to when it was acquired. In some implementations, operation 706 may be performed by one or more external resources the same as or similar to external resources 116 (shown in
At an operation 708, a vehicle event may be detected based on information conveyed by the output signals. In some implementations, the vehicle event may include, for example, information or signals related to one or more of acceleration or deceleration, a direction of travel, a vehicle speed, an engine speed (e.g. RPM), a duration of time, a closing distance, a lane departure, a following distance, another vehicle in close proximity, a person in close proximity, an open door or compartment, a geolocation, and/or other vehicle events of vehicle. The detection may be in real-time or near real-time. The detection may be based on the information conveyed by the output signals from the one or more sensors, the visual output information from multiple video capture devices, and/or other information. In some implementations, the vehicle event may have associated predetermined criteria. The detection of the vehicle event may be based on one or more determined vehicle event parameters satisfying predetermined criteria. In some implementations, operation 708 may be performed by a processor component the same as or similar to detection component 106 (shown in
At an operation 708, wireless communication of the vehicle operation information and the visual output information may be facilitated. The facilitation of the wireless communication may be in real-time or near real-time. The facilitation of the wireless communication may be responsive to the detection of the vehicle event. The vehicle operation information and/or the visual output information that is wirelessly communicated may include information for a period of time that may include the vehicle event and/or other information. The information may include, for example, video and/or other data associated with a vehicle event and/or the vehicle surroundings. In some implementations, operation 708 may be performed by a processor component the same as or similar to control component 110 (shown in
At an operation 708, electronic storage of the vehicle operation information and/or the visual output information may be facilitated. Facilitation of the electronic storage may be responsive to the detection of the vehicle event. The electronically stored information may include the information for the period of time that may include the vehicle event and/or other information. In some implementations, operation 708 may be performed by a processor component and/or electronic storage the same as or similar to control component 110 and/or electronic storage 118 (shown in
At an operation 710, a vehicle event type may be determined based on information conveyed by the output signals. The determination of a vehicle event type may be responsive to the detection of a vehicle event. In some implementations, the vehicle event type may be determined based on visual output information that may be acquired from multiple video capture devices. The video capture devices may have individual corresponding fields of view. In some implementations, the vehicle event type may be determined based on the individual corresponding fields of view of the video capture devices. The vehicle event type may include more than one vehicle event types and/or one or more vehicle event subtypes. A vehicle event may have more than one vehicle event type. In some implementations, responsive to a review, by a manual reviewer, of the vehicle event and/or previous vehicle events of the same vehicle event type determined at operation 710, method 700 may include generating coaching session information that corresponds to the vehicle event for use in a coaching session.
At an operation 712, review criteria may be generated that correspond to the vehicle event. The review criteria may be based on the vehicle event type and the fields of view corresponding to the video capture devices. For example, review criteria generated for a rear view may be different that review criteria generated for a side view. In some implementations, the review criteria may be generated based on corresponding fields of view of non-system integrated video capture devices. Non-system integrated video capture devices may be located, for example, in the vehicle environment, on the vehicle, on other vehicles, and/or in other locations. In some implementations, operation 712 may be performed by a processor component the same as or similar to control component 110 (shown in
Returning to
In some embodiments, operator interface 118 and/or reviewer interface 126 may be included in remote computing device 128. Examples of interface devices suitable for facilitating an operator and/or reviewer interface 126 comprise a keypad, buttons, switches, a keyboard, knobs, levers, a display screen, a touch screen, speakers, a microphone, an indicator light, an audible alarm, a printer, a tactile feedback device, and/or other interface devices. In one implementation, remote computing device 128 comprises a plurality of separate interfaces. In one implementation, reviewer interface 126 comprises at least one interface that is provided integrally with processor 102 and/or electronic storage 120.
It is to be understood that other communication techniques, either hard-wired or wireless, are also contemplated by the present disclosure as operator interface 118 and/or reviewer interface 126 included in remote computing device 128. For example, the present disclosure contemplates that operator interface 118 and/or the reviewer interface 126 included in remote computing device 128 may be integrated with a cellular and/or wireless (Wi-Fi) connection. In some implementations, operator interface 118 and/or the reviewer interface 126 included in remote computing device 128, may be included in a removable storage interface provided by electronic storage 120 (although this implementation may not be as scalable as integration with a cellular and/or wireless connection). In these examples, information may be loaded into system 100 wirelessly from a remote location, from removable storage (e.g., a smart card, a flash drive, a removable disk, etc.), and/or other sources that enable the user(s) to customize the implementation of system 100. Other exemplary input devices and techniques adapted for use with system 100 as reviewer interface 126 and/or operator interface 118 comprise, but are not limited to, an RS-232 port, RF link, an IR link, modem (telephone, cable, and/or other modems), a cellular network, a Wi-Fi network, a local area network, and/or other devices and/or systems. In short, any technique for communicating information with system 100 is contemplated by the present disclosure as user interface 118 and/or the user interface included in remote computing device 128.
Electronic storage 120 may comprise electronic storage media that electronically stores information. The electronic storage media of electronic storage 120 may comprise one or both of system storage that is provided integrally (e.g., substantially non-removable) with system 100 and/or removable storage that is removably connectable to system 100 via, for example, a port (e.g., a USB port, a firewire port, etc.) or a drive (e.g., a disk drive, etc.). Electronic storage 120 may comprise one or more of optically readable storage media (e.g., optical disks, etc.), magnetically readable storage media (e.g., magnetic tape, magnetic hard drive, floppy drive, etc.), electrical charge-based storage media (e.g., EEPROM, RAM, etc.), solid-state storage media (e.g., flash drive, etc.), and/or other electronically readable storage media. Electronic storage 120 may store software algorithms, recorded video event data, information determined by processor 102, information received via operator interface 118, information received via reviewer interface 126, and/or other information that enables system 100 to function properly. Electronic storage 120 may be (in whole or in part) a separate component within system 100, or electronic storage 120 may be provided (in whole or in part) integrally with one or more other components of system 100.
Remote computing device 128 may include one or more processors, reviewer interface 126, operator interface 118, electronic storage, and/or other components. Remote computing device 128 may be configured to enable an operator and/or reviewer to interface with system 100 (e.g., as described above), and/or provide other functionality attributed herein to remote computing device 128. Remote computing device 128 may be configured to communicate with processor 102 via a network such as the internet, cellular network, Wi-Fi network, Ethernet, and other interconnected computer networks. Remote computing device 128 may facilitate viewing and/or analysis of the information conveyed by the output signals of sensors 112, the visual output information from video capture devices 114, the information determined by processor 102, the information stored by electronic storage 120, information from any external resources 116, and/or other information. By way of non-limiting example, remote computing device 128 may include one or more of a server, a server cluster, desktop computer, a laptop computer, a handheld computer, a tablet computing platform, a NetBook, a Smartphone, a gaming console, and/or other computing platforms.
In some implementations, remote computing device 128 may be and/or include a server. The server may include communication lines, or ports to enable the exchange of information with a network, processor 102 of system 100, and/or other computing platforms. The server may include a plurality of processors, electronic storage, hardware, software, and/or firmware components operating together to provide the functionality attributed herein to remote computing device 128. For example, the server may be implemented by a cloud of computing platforms operating together as a system server.
As described above, processor 102 may be configured to provide information processing capabilities in system 100. As such, processor 102 may comprise one or more of a digital processor, an analog processor, a digital circuit designed to process information, an analog circuit designed to process information, a state machine, and/or other mechanisms for electronically processing information. Although processor 102 is shown in
Processor 102 may be configured to execute components 104, 106, 108, and/or 110 by software; hardware; firmware; some combination of software, hardware, and/or firmware; and/or other mechanisms for configuring processing capabilities on processor 102. It should be appreciated that although components 104, 106, 108, and 110 are illustrated in
Although the system(s) and/or method(s) of this disclosure have been described in detail for the purpose of illustration based on what is currently considered to be the most practical and preferred implementations, it is to be understood that such detail is solely for that purpose and that the disclosure is not limited to the disclosed implementations, but, on the contrary, is intended to cover modifications and equivalent arrangements that are within the spirit and scope of the appended claims. For example, it is to be understood that the present disclosure contemplates that, to the extent possible, one or more features of any implementation can be combined with one or more features of any other implementation.
Number | Name | Date | Kind |
---|---|---|---|
673203 | Freund | Apr 1901 | A |
673795 | Hammer | May 1901 | A |
673907 | Johnson | May 1901 | A |
676075 | McDougall | Jun 1901 | A |
679511 | Richards | Jul 1901 | A |
681036 | Burg | Aug 1901 | A |
681283 | Waynick | Aug 1901 | A |
681998 | Swift | Sep 1901 | A |
683155 | Thompson | Sep 1901 | A |
683214 | Mansfield | Sep 1901 | A |
684276 | Lonergan | Oct 1901 | A |
685082 | Wood | Oct 1901 | A |
685969 | Campbell | Nov 1901 | A |
686545 | Selph | Nov 1901 | A |
689849 | Brown | Dec 1901 | A |
691982 | Sturgis | Jan 1902 | A |
692834 | Davis | Feb 1902 | A |
694781 | Prinz | Mar 1902 | A |
2943141 | Knight | Jun 1960 | A |
3634866 | Meyer | Jan 1972 | A |
3781824 | Caiati | Dec 1973 | A |
3812287 | Lemelson | May 1974 | A |
3885090 | Rosenbaum | May 1975 | A |
3992656 | Joy | Nov 1976 | A |
4054752 | Dennis | Oct 1977 | A |
4072850 | McGlynn | Feb 1978 | A |
4258421 | Juhasz | Mar 1981 | A |
4271358 | Schwarz | Jun 1981 | A |
4276609 | Patel | Jun 1981 | A |
4280151 | Tsunekawa | Jul 1981 | A |
4281354 | Conte | Jul 1981 | A |
4401976 | Stadelmayr | Aug 1983 | A |
4409670 | Herndon | Oct 1983 | A |
4420773 | Toyoda | Dec 1983 | A |
4425097 | Owens | Jan 1984 | A |
4456931 | Toyoda | Jun 1984 | A |
4489351 | d'Alayer de Costemore | Dec 1984 | A |
4496995 | Colles | Jan 1985 | A |
4500868 | Tokitsu | Feb 1985 | A |
4528547 | Rodney | Jul 1985 | A |
4533962 | Decker | Aug 1985 | A |
4558379 | Hutter | Dec 1985 | A |
4588267 | Pastore | May 1986 | A |
4593313 | Nagasaki | Jun 1986 | A |
4621335 | Bluish | Nov 1986 | A |
4625210 | Sagl | Nov 1986 | A |
4630110 | Cotton | Dec 1986 | A |
4632348 | Keesling | Dec 1986 | A |
4638289 | Zottnik | Jan 1987 | A |
4646241 | Ratchford | Feb 1987 | A |
4651143 | Yamanaka | Mar 1987 | A |
4671111 | Lemelson | Jun 1987 | A |
4718685 | Kawabe | Jan 1988 | A |
4754255 | Sanders | Jun 1988 | A |
4758888 | Lapidot | Jul 1988 | A |
4763745 | Eto | Aug 1988 | A |
4785474 | Bernstein | Nov 1988 | A |
4789904 | Peterson | Dec 1988 | A |
4794566 | Richards | Dec 1988 | A |
4804937 | Barbiaux | Feb 1989 | A |
4806931 | Nelson | Feb 1989 | A |
4807096 | Skogler | Feb 1989 | A |
4814896 | Heitzman | Mar 1989 | A |
4837628 | Sasaki | Jun 1989 | A |
4839631 | Tsuji | Jun 1989 | A |
4843463 | Michetti | Jun 1989 | A |
4843578 | Wade | Jun 1989 | A |
4853856 | Hanway | Aug 1989 | A |
4853859 | Morita | Aug 1989 | A |
4866616 | Takeuchi | Sep 1989 | A |
4876597 | Roy | Oct 1989 | A |
4883349 | Mittelhaeuser | Nov 1989 | A |
4896855 | Furnish | Jan 1990 | A |
4926331 | Windle | May 1990 | A |
4930742 | Schofield | Jun 1990 | A |
4936533 | Adams | Jun 1990 | A |
4939652 | Steiner | Jul 1990 | A |
4942464 | Milatz | Jul 1990 | A |
4945244 | Castleman | Jul 1990 | A |
4949186 | Peterson | Aug 1990 | A |
4980913 | Skret | Dec 1990 | A |
4987541 | Levente | Jan 1991 | A |
4992943 | McCracken | Feb 1991 | A |
4993068 | Piosenka | Feb 1991 | A |
4995086 | Lilley | Feb 1991 | A |
5012335 | Cohodar | Apr 1991 | A |
5027104 | Reid | Jun 1991 | A |
5046007 | McCrery | Sep 1991 | A |
5050166 | Cantoni | Sep 1991 | A |
5056056 | Gustin | Oct 1991 | A |
5057820 | Markson | Oct 1991 | A |
5096287 | Kakinami | Mar 1992 | A |
5100095 | Haan | Mar 1992 | A |
5111289 | Lucas | May 1992 | A |
5140434 | Van | Aug 1992 | A |
5140436 | Blessinger | Aug 1992 | A |
5140438 | Blessinger | Aug 1992 | A |
5144661 | Shamosh | Sep 1992 | A |
5178448 | Adams | Jan 1993 | A |
5185700 | Bezos | Feb 1993 | A |
5196938 | Blessinger | Mar 1993 | A |
5223844 | Mansell | Jun 1993 | A |
5224211 | Roe | Jun 1993 | A |
5262813 | Scharton | Nov 1993 | A |
5283433 | Tsien | Feb 1994 | A |
5294978 | Katayama | Mar 1994 | A |
5305214 | Komatsu | Apr 1994 | A |
5305216 | Okura | Apr 1994 | A |
5308247 | Dyrdek | May 1994 | A |
5309485 | Chao | May 1994 | A |
5311197 | Sorden | May 1994 | A |
5321753 | Gritton | Jun 1994 | A |
5327288 | Wellington | Jul 1994 | A |
5330149 | Haan | Jul 1994 | A |
5333759 | Deering | Aug 1994 | A |
5343527 | Moore | Aug 1994 | A |
5353023 | Mitsugi | Oct 1994 | A |
5361326 | Aparicio | Nov 1994 | A |
5387926 | Bellan | Feb 1995 | A |
5388045 | Kamiya | Feb 1995 | A |
5388208 | Weingartner | Feb 1995 | A |
5404330 | Lee | Apr 1995 | A |
5408330 | Squicciarini | Apr 1995 | A |
5422543 | Weinberg | Jun 1995 | A |
5430431 | Nelson | Jul 1995 | A |
5430432 | Camhi | Jul 1995 | A |
5435184 | Pineroli | Jul 1995 | A |
5445024 | Riley | Aug 1995 | A |
5445027 | Zoerner | Aug 1995 | A |
5446659 | Yamawaki | Aug 1995 | A |
5455625 | Englander | Oct 1995 | A |
5455716 | Suman | Oct 1995 | A |
5465079 | Bouchard | Nov 1995 | A |
5473729 | Bryant | Dec 1995 | A |
5477141 | Naether | Dec 1995 | A |
5495242 | Kick | Feb 1996 | A |
5495243 | McKenna | Feb 1996 | A |
5497419 | Hill | Mar 1996 | A |
5499182 | Ousborne | Mar 1996 | A |
5504482 | Schreder | Apr 1996 | A |
5505076 | Parkman | Apr 1996 | A |
5513011 | Matsumoto | Apr 1996 | A |
5515285 | Garrett | May 1996 | A |
5519260 | Washington | May 1996 | A |
5521633 | Nakajima | May 1996 | A |
5523811 | Wada | Jun 1996 | A |
5526269 | Ishibashi | Jun 1996 | A |
5530420 | Tsuchiya | Jun 1996 | A |
5532678 | Kin | Jul 1996 | A |
5537156 | Katayama | Jul 1996 | A |
5539454 | Williams | Jul 1996 | A |
5541590 | Nishio | Jul 1996 | A |
5544060 | Fujii | Aug 1996 | A |
5546191 | Hibi | Aug 1996 | A |
5546305 | Kondo | Aug 1996 | A |
5548273 | Nicol | Aug 1996 | A |
5552990 | Ihara | Sep 1996 | A |
5559496 | Dubats | Sep 1996 | A |
5568211 | Bamford | Oct 1996 | A |
5570087 | Lemelson | Oct 1996 | A |
5570127 | Schmidt | Oct 1996 | A |
5574424 | Nguyen | Nov 1996 | A |
5574443 | Hsieh | Nov 1996 | A |
D376571 | Kokat | Dec 1996 | S |
5581464 | Woll | Dec 1996 | A |
5586130 | Doyle | Dec 1996 | A |
5590948 | Moreno | Jan 1997 | A |
5596382 | Bamford | Jan 1997 | A |
5596647 | Wakai | Jan 1997 | A |
5600775 | King | Feb 1997 | A |
5608272 | Tanguay | Mar 1997 | A |
5610580 | Lai | Mar 1997 | A |
5612686 | Takano | Mar 1997 | A |
5631638 | Kaspar | May 1997 | A |
5638273 | Coiner | Jun 1997 | A |
5642106 | Hancock | Jun 1997 | A |
5646856 | Kaesser | Jul 1997 | A |
5652706 | Morimoto | Jul 1997 | A |
RE35590 | Bezos | Aug 1997 | E |
5654892 | Fujii | Aug 1997 | A |
5659355 | Barron | Aug 1997 | A |
5666120 | Kline | Sep 1997 | A |
5667176 | Zamarripa | Sep 1997 | A |
5669698 | Veldman | Sep 1997 | A |
5671451 | Takahashi | Sep 1997 | A |
5677979 | Squicciarini | Oct 1997 | A |
5680117 | Arai | Oct 1997 | A |
5680123 | Lee | Oct 1997 | A |
5686765 | Washington | Nov 1997 | A |
5686889 | Hillis | Nov 1997 | A |
5689442 | Swanson | Nov 1997 | A |
5696705 | Zykan | Dec 1997 | A |
5706362 | Yabe | Jan 1998 | A |
5706909 | Bevins | Jan 1998 | A |
5712679 | Coles | Jan 1998 | A |
5717456 | Rudt | Feb 1998 | A |
5719554 | Gagnon | Feb 1998 | A |
5758299 | Sandborg | May 1998 | A |
5781101 | Stephen | Jul 1998 | A |
5781145 | Williams | Jul 1998 | A |
5784007 | Pepper | Jul 1998 | A |
5784021 | Oliva | Jul 1998 | A |
5784521 | Nakatani | Jul 1998 | A |
5790403 | Nakayama | Aug 1998 | A |
5790973 | Blaker | Aug 1998 | A |
5793308 | Rosinski | Aug 1998 | A |
5793420 | Schmidt | Aug 1998 | A |
5793739 | Tanaka | Aug 1998 | A |
5793985 | Natarajan | Aug 1998 | A |
5794165 | Minowa | Aug 1998 | A |
5797134 | McMillan | Aug 1998 | A |
5798458 | Monroe | Aug 1998 | A |
5800040 | Santo | Sep 1998 | A |
5802545 | Coverdill | Sep 1998 | A |
5802727 | Blank | Sep 1998 | A |
5805079 | Lemelson | Sep 1998 | A |
5813745 | Fant | Sep 1998 | A |
5815071 | Doyle | Sep 1998 | A |
5815093 | Kikinis | Sep 1998 | A |
5819198 | Peretz | Oct 1998 | A |
5825284 | Dunwoody | Oct 1998 | A |
5825412 | Hobson | Oct 1998 | A |
5844505 | Van | Dec 1998 | A |
5845733 | Wolfsen | Dec 1998 | A |
5867802 | Borza | Feb 1999 | A |
5877897 | Schofield | Mar 1999 | A |
5896167 | Omae | Apr 1999 | A |
5897602 | Mizuta | Apr 1999 | A |
5897606 | Miura | Apr 1999 | A |
5899956 | Chan | May 1999 | A |
5901806 | Takahashi | May 1999 | A |
5914748 | Parulski | Jun 1999 | A |
5919239 | Fraker | Jul 1999 | A |
5926210 | Hackett | Jul 1999 | A |
5928291 | Jenkins | Jul 1999 | A |
5938321 | Bos | Aug 1999 | A |
5946404 | Bakshi | Aug 1999 | A |
5948038 | Daly | Sep 1999 | A |
5959367 | Ofarrell | Sep 1999 | A |
5978017 | Tino | Nov 1999 | A |
6002326 | Turner | Dec 1999 | A |
6006148 | Strong | Dec 1999 | A |
6008723 | Yassan | Dec 1999 | A |
6008841 | Charlson | Dec 1999 | A |
6009370 | Minowa | Dec 1999 | A |
6011492 | Garesche | Jan 2000 | A |
6028528 | Lorenzetti | Feb 2000 | A |
6037860 | Zander | Mar 2000 | A |
6037977 | Peterson | Mar 2000 | A |
6041410 | Hsu | Mar 2000 | A |
6049079 | Noordam | Apr 2000 | A |
6057754 | Kinoshita | May 2000 | A |
6060989 | Gehlot | May 2000 | A |
6064792 | Fox | May 2000 | A |
6067488 | Tano | May 2000 | A |
6076026 | Jambhekar | Jun 2000 | A |
6084870 | Wooten | Jul 2000 | A |
6088635 | Cox | Jul 2000 | A |
6092008 | Bateman | Jul 2000 | A |
6092021 | Ehlbeck | Jul 2000 | A |
6092193 | Loomis | Jul 2000 | A |
6100811 | Hsu | Aug 2000 | A |
6111254 | Eden | Aug 2000 | A |
6118768 | Bhatia | Sep 2000 | A |
6122738 | Millard | Sep 2000 | A |
6141611 | Mackey | Oct 2000 | A |
6144296 | Ishida | Nov 2000 | A |
6147598 | Murphy | Nov 2000 | A |
6151065 | Steed | Nov 2000 | A |
6163338 | Johnson | Dec 2000 | A |
6163749 | McDonough | Dec 2000 | A |
6167186 | Kawasaki | Dec 2000 | A |
6170742 | Yacoob | Jan 2001 | B1 |
6181373 | Coles | Jan 2001 | B1 |
6182010 | Berstis | Jan 2001 | B1 |
6185490 | Ferguson | Feb 2001 | B1 |
6195605 | Tabler | Feb 2001 | B1 |
6200139 | Clapper | Mar 2001 | B1 |
6208919 | Barkesseh | Mar 2001 | B1 |
6211907 | Scaman | Apr 2001 | B1 |
6218960 | Ishikawa | Apr 2001 | B1 |
6246933 | Baque | Jun 2001 | B1 |
6246934 | Otake | Jun 2001 | B1 |
6252544 | Hoffberg | Jun 2001 | B1 |
6253129 | Jenkins | Jun 2001 | B1 |
6259475 | Ramachandran | Jul 2001 | B1 |
6263265 | Fera | Jul 2001 | B1 |
6266588 | McClellan | Jul 2001 | B1 |
6298290 | Abe | Oct 2001 | B1 |
6300875 | Schafer | Oct 2001 | B1 |
6317682 | Ogura | Nov 2001 | B1 |
6324450 | Iwama | Nov 2001 | B1 |
6333759 | Mazzilli | Dec 2001 | B1 |
6337622 | Sugano | Jan 2002 | B1 |
6349250 | Hart | Feb 2002 | B1 |
6353734 | Wright | Mar 2002 | B1 |
6356823 | Iannotti | Mar 2002 | B1 |
6360147 | Lee | Mar 2002 | B1 |
6366207 | Murphy | Apr 2002 | B1 |
6389339 | Just | May 2002 | B1 |
6389340 | Rayner | May 2002 | B1 |
6400835 | Lemelson | Jun 2002 | B1 |
6405112 | Rayner | Jun 2002 | B1 |
6405132 | Breed | Jun 2002 | B1 |
6408232 | Cannon | Jun 2002 | B1 |
6411874 | Morgan | Jun 2002 | B2 |
6421080 | Lambert | Jul 2002 | B1 |
6429773 | Schuyler | Aug 2002 | B1 |
6434510 | Callaghan | Aug 2002 | B1 |
6449540 | Rayner | Sep 2002 | B1 |
6456321 | Ito | Sep 2002 | B1 |
6459988 | Fan | Oct 2002 | B1 |
6470241 | Yoshikawa | Oct 2002 | B2 |
6472771 | Frese | Oct 2002 | B1 |
6490513 | Fish | Dec 2002 | B1 |
6493650 | Rodgers | Dec 2002 | B1 |
6505106 | Lawrence | Jan 2003 | B1 |
6507838 | Syeda-Mahmood | Jan 2003 | B1 |
6508400 | Ishifuji | Jan 2003 | B1 |
6516256 | Hartmann | Feb 2003 | B1 |
6518881 | Monroe | Feb 2003 | B2 |
6525672 | Chainer | Feb 2003 | B2 |
6526352 | Breed | Feb 2003 | B1 |
6529159 | Fan | Mar 2003 | B1 |
6535804 | Chun | Mar 2003 | B1 |
6552682 | Fan | Apr 2003 | B1 |
6553308 | Uhlmann | Apr 2003 | B1 |
6556905 | Mittelsteadt | Apr 2003 | B1 |
6559769 | Anthony | May 2003 | B2 |
6574538 | Sasaki | Jun 2003 | B2 |
6575902 | Burton | Jun 2003 | B1 |
6580373 | Ohashi | Jun 2003 | B1 |
6580983 | Laguer-Diaz | Jun 2003 | B2 |
6593848 | Atkins | Jul 2003 | B1 |
6594576 | Fan | Jul 2003 | B2 |
6611740 | Lowrey | Aug 2003 | B2 |
6611755 | Coffee | Aug 2003 | B1 |
6624611 | Kirmuss | Sep 2003 | B2 |
6629029 | Giles | Sep 2003 | B1 |
6629030 | Klausner | Sep 2003 | B2 |
6636791 | Okada | Oct 2003 | B2 |
6664922 | Fan | Dec 2003 | B1 |
6665613 | Duvall | Dec 2003 | B2 |
6679702 | Rau | Jan 2004 | B1 |
6684137 | Takagi | Jan 2004 | B2 |
6694483 | Nagata | Feb 2004 | B1 |
6701234 | Vogelsang | Mar 2004 | B1 |
6714894 | Tobey | Mar 2004 | B1 |
6718239 | Rayner | Apr 2004 | B2 |
6721640 | Glenn | Apr 2004 | B2 |
6721652 | Sanqunetti | Apr 2004 | B1 |
6728612 | Carver | Apr 2004 | B1 |
6732031 | Lightner | May 2004 | B1 |
6732032 | Banet | May 2004 | B1 |
6735503 | Ames | May 2004 | B2 |
6737954 | Chainer | May 2004 | B2 |
6738697 | Breed | May 2004 | B2 |
6739078 | Morley | May 2004 | B2 |
6741168 | Webb | May 2004 | B2 |
6745153 | White | Jun 2004 | B2 |
6747692 | Patel | Jun 2004 | B2 |
6748305 | Klausner | Jun 2004 | B1 |
6760757 | Lundberg | Jul 2004 | B1 |
6762513 | Landgraf | Jul 2004 | B2 |
6779716 | Grow | Aug 2004 | B1 |
6795017 | Puranik | Sep 2004 | B1 |
6795111 | Mazzilli | Sep 2004 | B1 |
6795759 | Doyle | Sep 2004 | B2 |
6798743 | Ma | Sep 2004 | B1 |
6804590 | Sato | Oct 2004 | B2 |
6810362 | Adachi | Oct 2004 | B2 |
6812831 | Ikeda | Nov 2004 | B2 |
6819989 | Maeda | Nov 2004 | B2 |
6831556 | Boykin | Dec 2004 | B1 |
6832140 | Fan | Dec 2004 | B2 |
6832141 | Skeen | Dec 2004 | B2 |
6836712 | Nishina | Dec 2004 | B2 |
6842762 | Raithel | Jan 2005 | B2 |
6847873 | Li | Jan 2005 | B1 |
6850823 | Eun | Feb 2005 | B2 |
6859695 | Klausner | Feb 2005 | B2 |
6859705 | Rao | Feb 2005 | B2 |
6862524 | Nagda | Mar 2005 | B1 |
6865457 | Mittelsteadt | Mar 2005 | B1 |
6867733 | Sandhu | Mar 2005 | B2 |
6873261 | Anthony | Mar 2005 | B2 |
6882313 | Fan | Apr 2005 | B1 |
6882912 | DiLodovico | Apr 2005 | B2 |
6894606 | Forbes | May 2005 | B2 |
6895248 | Akyol | May 2005 | B1 |
6898492 | De Leon | May 2005 | B2 |
6898493 | Ehrman | May 2005 | B2 |
6919823 | Lock | Jul 2005 | B1 |
6922566 | Puranik | Jul 2005 | B2 |
6928348 | Lightner | Aug 2005 | B1 |
6931309 | Phelan | Aug 2005 | B2 |
6947817 | Diem | Sep 2005 | B2 |
6950122 | Mirabile | Sep 2005 | B1 |
6954223 | Miyazawa | Oct 2005 | B2 |
6954689 | Hanson | Oct 2005 | B2 |
6988034 | Marlatt | Jan 2006 | B1 |
7003289 | Kolls | Feb 2006 | B1 |
7012632 | Freeman | Mar 2006 | B2 |
7020548 | Saito | Mar 2006 | B2 |
7023333 | Blanco | Apr 2006 | B2 |
7027621 | Prokoski | Apr 2006 | B1 |
7039510 | Gumpinger | May 2006 | B2 |
7076348 | Bucher | Jul 2006 | B2 |
7079927 | Tano | Jul 2006 | B1 |
7082359 | Breed | Jul 2006 | B2 |
7082382 | Rose et al. | Jul 2006 | B1 |
7088387 | Freeman | Aug 2006 | B1 |
7095782 | Cohen | Aug 2006 | B1 |
7098812 | Hirota | Aug 2006 | B2 |
7100190 | Johnson | Aug 2006 | B2 |
7113853 | Hecklinger | Sep 2006 | B2 |
7117075 | Larschan | Oct 2006 | B1 |
7119832 | Blanco | Oct 2006 | B2 |
7138904 | Dutu | Nov 2006 | B1 |
7155321 | Bromley | Dec 2006 | B2 |
7177738 | Diaz | Feb 2007 | B2 |
7209833 | Isaji | Apr 2007 | B2 |
7239252 | Kato | Jul 2007 | B2 |
7254482 | Kawasaki | Aug 2007 | B2 |
7265663 | Steele | Sep 2007 | B2 |
7266507 | Simon | Sep 2007 | B2 |
7272179 | Siemens | Sep 2007 | B2 |
7308341 | Schofield | Dec 2007 | B2 |
7317974 | Luskin | Jan 2008 | B2 |
7343306 | Bates | Mar 2008 | B1 |
7348895 | Lagassey | Mar 2008 | B2 |
7349027 | Endo | Mar 2008 | B2 |
7370261 | Winarski | May 2008 | B2 |
7382933 | Dorai | Jun 2008 | B2 |
7386376 | Basir | Jun 2008 | B2 |
7389178 | Raz | Jun 2008 | B2 |
7457693 | Olsen | Nov 2008 | B2 |
7471189 | Vastad | Dec 2008 | B2 |
7471192 | Hara | Dec 2008 | B2 |
7536457 | Miller | May 2009 | B2 |
7548586 | Mimar | Jun 2009 | B1 |
7561054 | Raz | Jul 2009 | B2 |
7584033 | Mittelsteadt | Sep 2009 | B2 |
7596439 | Oesterling | Sep 2009 | B2 |
7623754 | McKain | Nov 2009 | B1 |
7659827 | Gunderson | Feb 2010 | B2 |
7659835 | Jung | Feb 2010 | B2 |
7667731 | Kreiner | Feb 2010 | B2 |
7689001 | Kim | Mar 2010 | B2 |
7702442 | Takenaka | Apr 2010 | B2 |
7725216 | Kim | May 2010 | B2 |
7768548 | Silvernail | Aug 2010 | B2 |
7769499 | McQuade | Aug 2010 | B2 |
7783956 | Ko | Aug 2010 | B2 |
7804426 | Etcheson | Sep 2010 | B2 |
7821421 | Tamir | Oct 2010 | B2 |
7845560 | Emanuel | Dec 2010 | B2 |
7853376 | Peng | Dec 2010 | B2 |
7868912 | Venetianer | Jan 2011 | B2 |
7893958 | DAgostino | Feb 2011 | B1 |
7904219 | Lowrey | Mar 2011 | B1 |
7940250 | Forstall | May 2011 | B2 |
7941258 | Mittelsteadt | May 2011 | B1 |
7974748 | Goerick | Jul 2011 | B2 |
7984146 | Rozak | Jul 2011 | B2 |
8054168 | McCormick | Nov 2011 | B2 |
8068979 | Breed | Nov 2011 | B2 |
8090598 | Bauer | Jan 2012 | B2 |
8113844 | Huang | Feb 2012 | B2 |
8139820 | Plante | Mar 2012 | B2 |
8140265 | Grush | Mar 2012 | B2 |
8140358 | Ling | Mar 2012 | B1 |
8152198 | Breed | Apr 2012 | B2 |
8239092 | Plante | Aug 2012 | B2 |
8269617 | Cook | Sep 2012 | B2 |
8311858 | Everett | Nov 2012 | B2 |
8314708 | Gunderson | Nov 2012 | B2 |
8321066 | Becker | Nov 2012 | B2 |
8373567 | Denson | Feb 2013 | B2 |
8417562 | Siemens | Apr 2013 | B1 |
8423009 | Srinivasan | Apr 2013 | B2 |
8442690 | Goldstein | May 2013 | B2 |
8471701 | Yariv | Jun 2013 | B2 |
8508353 | Cook | Aug 2013 | B2 |
8538696 | Cassanova | Sep 2013 | B1 |
8538785 | Coleman | Sep 2013 | B2 |
8564426 | Cook | Oct 2013 | B2 |
8564446 | Gunderson | Oct 2013 | B2 |
8571755 | Plante | Oct 2013 | B2 |
8577703 | McClellan | Nov 2013 | B2 |
8606492 | Botnen | Dec 2013 | B1 |
8634958 | Chiappetta | Jan 2014 | B1 |
8635557 | Geise | Jan 2014 | B2 |
8676428 | Richardson | Mar 2014 | B2 |
8744642 | Nemat-Nasser | Jun 2014 | B2 |
8775067 | Cho | Jul 2014 | B2 |
8781292 | Ross | Jul 2014 | B1 |
8803695 | Denson | Aug 2014 | B2 |
8805110 | Rhoads | Aug 2014 | B2 |
8849501 | Cook | Sep 2014 | B2 |
8855847 | Uehara | Oct 2014 | B2 |
8862395 | Richardson | Oct 2014 | B2 |
8868288 | Plante | Oct 2014 | B2 |
8880279 | Plante | Nov 2014 | B2 |
8892310 | Palmer | Nov 2014 | B1 |
8989959 | Plante | Mar 2015 | B2 |
8996234 | Tamari | Mar 2015 | B1 |
8996240 | Plante | Mar 2015 | B2 |
9047721 | Botnen | Jun 2015 | B1 |
9085362 | Kilian | Jul 2015 | B1 |
9183679 | Plante | Nov 2015 | B2 |
9201842 | Plante | Dec 2015 | B2 |
9208129 | Plante | Dec 2015 | B2 |
9226004 | Plante | Dec 2015 | B1 |
9240079 | Lambert | Jan 2016 | B2 |
9373203 | Fields | Jun 2016 | B1 |
9607526 | Hsu-Hoffman | Mar 2017 | B1 |
9610955 | Palmer | Apr 2017 | B2 |
9715711 | Konrardy | Jul 2017 | B1 |
9754325 | Konrardy | Sep 2017 | B1 |
9767516 | Konrardy | Sep 2017 | B1 |
9786154 | Potter | Oct 2017 | B1 |
9792656 | Konrardy | Oct 2017 | B1 |
9805423 | Konrardy | Oct 2017 | B1 |
9805601 | Fields | Oct 2017 | B1 |
9858621 | Konrardy | Jan 2018 | B1 |
9868394 | Fields | Jan 2018 | B1 |
9870649 | Fields | Jan 2018 | B1 |
10222228 | Chan | Mar 2019 | B1 |
10360739 | Palmer | Jul 2019 | B2 |
10404951 | Plante | Sep 2019 | B2 |
20010005217 | Hamilton | Jun 2001 | A1 |
20010005804 | Rayner | Jun 2001 | A1 |
20010018628 | Jenkins | Aug 2001 | A1 |
20010020204 | Runyon | Sep 2001 | A1 |
20010052730 | Baur | Dec 2001 | A1 |
20020019689 | Harrison | Feb 2002 | A1 |
20020027502 | Mayor | Mar 2002 | A1 |
20020029109 | Wong | Mar 2002 | A1 |
20020035422 | Sasaki | Mar 2002 | A1 |
20020044225 | Rakib | Apr 2002 | A1 |
20020059453 | Eriksson | May 2002 | A1 |
20020061758 | Zarlengo | May 2002 | A1 |
20020067076 | Talbot | Jun 2002 | A1 |
20020087240 | Raithel | Jul 2002 | A1 |
20020091473 | Gardner | Jul 2002 | A1 |
20020105438 | Forbes | Aug 2002 | A1 |
20020107619 | Klausner | Aug 2002 | A1 |
20020111725 | Burge | Aug 2002 | A1 |
20020111756 | Modgil | Aug 2002 | A1 |
20020118206 | Knittel | Aug 2002 | A1 |
20020120374 | Douros | Aug 2002 | A1 |
20020135679 | Seaman | Sep 2002 | A1 |
20020138587 | Koehler | Sep 2002 | A1 |
20020156558 | Hanson | Oct 2002 | A1 |
20020163532 | Thomas | Nov 2002 | A1 |
20020169529 | Kim | Nov 2002 | A1 |
20020169530 | Laguer-Diaz | Nov 2002 | A1 |
20020170064 | Monroe | Nov 2002 | A1 |
20020183905 | Maeda | Dec 2002 | A1 |
20030016753 | Kim | Jan 2003 | A1 |
20030028298 | Macky | Feb 2003 | A1 |
20030053433 | Chun | Mar 2003 | A1 |
20030055557 | Dutta | Mar 2003 | A1 |
20030055666 | Roddy | Mar 2003 | A1 |
20030065805 | Barnes | Apr 2003 | A1 |
20030067541 | Joao | Apr 2003 | A1 |
20030079041 | Parrella | Apr 2003 | A1 |
20030080713 | Kirmuss | May 2003 | A1 |
20030080878 | Kirmuss | May 2003 | A1 |
20030081121 | Kirmuss | May 2003 | A1 |
20030081122 | Kirmuss | May 2003 | A1 |
20030081123 | Rupe | May 2003 | A1 |
20030081127 | Kirmuss | May 2003 | A1 |
20030081128 | Kirmuss | May 2003 | A1 |
20030081934 | Kirmuss | May 2003 | A1 |
20030081935 | Kirmuss | May 2003 | A1 |
20030095688 | Kirmuss | May 2003 | A1 |
20030112133 | Webb | Jun 2003 | A1 |
20030125854 | Kawasaki | Jul 2003 | A1 |
20030137194 | White | Jul 2003 | A1 |
20030144775 | Klausner | Jul 2003 | A1 |
20030152145 | Kawakita | Aug 2003 | A1 |
20030154009 | Basir | Aug 2003 | A1 |
20030158638 | Yakes | Aug 2003 | A1 |
20030177187 | Levine | Sep 2003 | A1 |
20030187704 | Hashiguchi | Oct 2003 | A1 |
20030191568 | Breed | Oct 2003 | A1 |
20030195678 | Betters et al. | Oct 2003 | A1 |
20030201875 | Kuo | Oct 2003 | A1 |
20030214585 | Bakewell | Nov 2003 | A1 |
20030220835 | Barnes | Nov 2003 | A1 |
20030222880 | Waterman | Dec 2003 | A1 |
20040008255 | Lewellen | Jan 2004 | A1 |
20040032493 | Franke | Feb 2004 | A1 |
20040033058 | Reich | Feb 2004 | A1 |
20040039503 | Doyle | Feb 2004 | A1 |
20040039504 | Coffee | Feb 2004 | A1 |
20040044452 | Bauer | Mar 2004 | A1 |
20040044592 | Ubik | Mar 2004 | A1 |
20040054444 | Abeska | Mar 2004 | A1 |
20040054513 | Laird | Mar 2004 | A1 |
20040054689 | Salmonsen | Mar 2004 | A1 |
20040064245 | Knockeart | Apr 2004 | A1 |
20040070926 | Boykin | Apr 2004 | A1 |
20040083041 | Skeen | Apr 2004 | A1 |
20040088090 | Wee | May 2004 | A1 |
20040103008 | Wahlbin | May 2004 | A1 |
20040103010 | Wahlbin | May 2004 | A1 |
20040104842 | Drury | Jun 2004 | A1 |
20040111189 | Miyazawa | Jun 2004 | A1 |
20040117638 | Monroe | Jun 2004 | A1 |
20040135979 | Hazelton | Jul 2004 | A1 |
20040138794 | Saito | Jul 2004 | A1 |
20040145457 | Schofield et al. | Jul 2004 | A1 |
20040153244 | Kellum | Aug 2004 | A1 |
20040153362 | Bauer | Aug 2004 | A1 |
20040167689 | Bromley | Aug 2004 | A1 |
20040179600 | Wells | Sep 2004 | A1 |
20040181326 | Adams | Sep 2004 | A1 |
20040184548 | Kerbiriou | Sep 2004 | A1 |
20040203903 | Wilson | Oct 2004 | A1 |
20040209594 | Naboulsi | Oct 2004 | A1 |
20040210353 | Rice | Oct 2004 | A1 |
20040230345 | Tzamaloukas | Nov 2004 | A1 |
20040230370 | Tzamaloukas | Nov 2004 | A1 |
20040230373 | Tzamaloukas | Nov 2004 | A1 |
20040230374 | Tzamaloukas | Nov 2004 | A1 |
20040233284 | Lesesky | Nov 2004 | A1 |
20040236474 | Chowdhary | Nov 2004 | A1 |
20040243285 | Gounder | Dec 2004 | A1 |
20040243308 | Irish | Dec 2004 | A1 |
20040243668 | Harjanto | Dec 2004 | A1 |
20040254689 | Blazic | Dec 2004 | A1 |
20040254698 | Hubbard | Dec 2004 | A1 |
20040257208 | Huang | Dec 2004 | A1 |
20040267419 | Jeng | Dec 2004 | A1 |
20050021199 | Zimmerman | Jan 2005 | A1 |
20050038581 | Kapolka | Feb 2005 | A1 |
20050043869 | Funkhouser | Feb 2005 | A1 |
20050060070 | Kapolka | Mar 2005 | A1 |
20050060071 | Winner | Mar 2005 | A1 |
20050065682 | Kapadia | Mar 2005 | A1 |
20050065716 | Timko | Mar 2005 | A1 |
20050068417 | Kreiner | Mar 2005 | A1 |
20050073585 | Ettinger | Apr 2005 | A1 |
20050078423 | Kim | Apr 2005 | A1 |
20050083404 | Pierce | Apr 2005 | A1 |
20050088291 | Blanco | Apr 2005 | A1 |
20050099498 | Lao | May 2005 | A1 |
20050100329 | Lao | May 2005 | A1 |
20050102074 | Kolls | May 2005 | A1 |
20050125117 | Breed | Jun 2005 | A1 |
20050131585 | Luskin | Jun 2005 | A1 |
20050131595 | Luskin | Jun 2005 | A1 |
20050131597 | Raz | Jun 2005 | A1 |
20050136949 | Barnes | Jun 2005 | A1 |
20050137757 | Phelan | Jun 2005 | A1 |
20050137796 | Gumpinger | Jun 2005 | A1 |
20050146458 | Carmichael | Jul 2005 | A1 |
20050149238 | Stefani | Jul 2005 | A1 |
20050149259 | Cherveny | Jul 2005 | A1 |
20050152353 | Couturier | Jul 2005 | A1 |
20050159964 | Sonnenrein | Jul 2005 | A1 |
20050166258 | Vasilevsky | Jul 2005 | A1 |
20050168258 | Vasilevsky | Jul 2005 | A1 |
20050171663 | Mittelsteadt | Aug 2005 | A1 |
20050171692 | Hamblen | Aug 2005 | A1 |
20050174217 | Basir | Aug 2005 | A1 |
20050182538 | Phelan | Aug 2005 | A1 |
20050182824 | Cotte | Aug 2005 | A1 |
20050185052 | Raisinghani | Aug 2005 | A1 |
20050185936 | Lao | Aug 2005 | A9 |
20050192749 | Flann | Sep 2005 | A1 |
20050197748 | Hoist | Sep 2005 | A1 |
20050200714 | Marchese | Sep 2005 | A1 |
20050203683 | Olsen | Sep 2005 | A1 |
20050205719 | Hendrickson | Sep 2005 | A1 |
20050206741 | Raber | Sep 2005 | A1 |
20050209776 | Ogino | Sep 2005 | A1 |
20050212920 | Evans | Sep 2005 | A1 |
20050216144 | Baldassa | Sep 2005 | A1 |
20050228560 | Doherty | Oct 2005 | A1 |
20050233805 | Okajima | Oct 2005 | A1 |
20050243171 | Ross, Sr. | Nov 2005 | A1 |
20050251304 | Cancellara | Nov 2005 | A1 |
20050256681 | Brinton | Nov 2005 | A1 |
20050258942 | Manasseh | Nov 2005 | A1 |
20050264691 | Endo | Dec 2005 | A1 |
20050283284 | Grenier | Dec 2005 | A1 |
20060001671 | Kamijo | Jan 2006 | A1 |
20060007151 | Ram | Jan 2006 | A1 |
20060011399 | Brockway | Jan 2006 | A1 |
20060015233 | Olsen | Jan 2006 | A1 |
20060022842 | Zoladek | Feb 2006 | A1 |
20060025897 | Shostak | Feb 2006 | A1 |
20060025907 | Kapolka | Feb 2006 | A9 |
20060030986 | Peng | Feb 2006 | A1 |
20060040239 | Cummins | Feb 2006 | A1 |
20060047380 | Welch | Mar 2006 | A1 |
20060053038 | Warren | Mar 2006 | A1 |
20060055521 | Blanco | Mar 2006 | A1 |
20060057543 | Roald | Mar 2006 | A1 |
20060058950 | Kato | Mar 2006 | A1 |
20060072792 | Toda | Apr 2006 | A1 |
20060078853 | Lanktree | Apr 2006 | A1 |
20060082438 | Bazakos | Apr 2006 | A1 |
20060092043 | Lagassey | May 2006 | A1 |
20060095175 | Dewaal | May 2006 | A1 |
20060095199 | Lagassey | May 2006 | A1 |
20060095349 | Morgan | May 2006 | A1 |
20060103127 | Lie | May 2006 | A1 |
20060106514 | Liebl | May 2006 | A1 |
20060111817 | Phelan | May 2006 | A1 |
20060122749 | Phelan | Jun 2006 | A1 |
20060129578 | Kim | Jun 2006 | A1 |
20060142913 | Coffee | Jun 2006 | A1 |
20060143435 | Kwon | Jun 2006 | A1 |
20060147187 | Takemoto | Jul 2006 | A1 |
20060158349 | Oesterling | Jul 2006 | A1 |
20060161960 | Benoit | Jul 2006 | A1 |
20060168271 | Pabari | Jul 2006 | A1 |
20060178793 | Hecklinger | Aug 2006 | A1 |
20060180647 | Hansen | Aug 2006 | A1 |
20060184295 | Hawkins | Aug 2006 | A1 |
20060192658 | Yamamura | Aug 2006 | A1 |
20060200008 | Moore-Ede | Sep 2006 | A1 |
20060200305 | Sheha | Sep 2006 | A1 |
20060204059 | Ido | Sep 2006 | A1 |
20060209090 | Kelly et al. | Sep 2006 | A1 |
20060209840 | Paatela | Sep 2006 | A1 |
20060212195 | Veith | Sep 2006 | A1 |
20060215884 | Ota | Sep 2006 | A1 |
20060226344 | Werth | Oct 2006 | A1 |
20060229780 | Underdahl | Oct 2006 | A1 |
20060242680 | Johnson | Oct 2006 | A1 |
20060247833 | Malhotra | Nov 2006 | A1 |
20060253307 | Warren | Nov 2006 | A1 |
20060259218 | Wu | Nov 2006 | A1 |
20060259933 | Fishel | Nov 2006 | A1 |
20060261931 | Cheng | Nov 2006 | A1 |
20070001831 | Raz | Jan 2007 | A1 |
20070005404 | Raz | Jan 2007 | A1 |
20070027583 | Tamir | Feb 2007 | A1 |
20070027726 | Warren | Feb 2007 | A1 |
20070035632 | Silvernail | Feb 2007 | A1 |
20070043487 | Krzystofczyk | Feb 2007 | A1 |
20070088488 | Reeves | Apr 2007 | A1 |
20070100509 | Piekarz | May 2007 | A1 |
20070120948 | Fujioka | May 2007 | A1 |
20070124332 | Ballesty | May 2007 | A1 |
20070127833 | Singh | Jun 2007 | A1 |
20070132773 | Plante | Jun 2007 | A1 |
20070135979 | Plante | Jun 2007 | A1 |
20070135980 | Plante | Jun 2007 | A1 |
20070136078 | Plante | Jun 2007 | A1 |
20070142986 | Alaous | Jun 2007 | A1 |
20070143499 | Chang | Jun 2007 | A1 |
20070150138 | Plante | Jun 2007 | A1 |
20070150140 | Seymour | Jun 2007 | A1 |
20070159309 | Ito | Jul 2007 | A1 |
20070173994 | Kubo | Jul 2007 | A1 |
20070179691 | Grenn | Aug 2007 | A1 |
20070183635 | Weidhaas | Aug 2007 | A1 |
20070208494 | Chapman | Sep 2007 | A1 |
20070213920 | Igarashi | Sep 2007 | A1 |
20070216521 | Guensler | Sep 2007 | A1 |
20070219685 | Plante | Sep 2007 | A1 |
20070219686 | Plante | Sep 2007 | A1 |
20070236474 | Ramstein | Oct 2007 | A1 |
20070241874 | Okpysh | Oct 2007 | A1 |
20070244614 | Nathanson | Oct 2007 | A1 |
20070253307 | Mashimo | Nov 2007 | A1 |
20070256481 | Nishiyama | Nov 2007 | A1 |
20070257781 | Denson | Nov 2007 | A1 |
20070257782 | Etcheson | Nov 2007 | A1 |
20070257804 | Gunderson | Nov 2007 | A1 |
20070257815 | Gunderson | Nov 2007 | A1 |
20070260677 | Demarco | Nov 2007 | A1 |
20070262855 | Zuta | Nov 2007 | A1 |
20070263984 | Sterner | Nov 2007 | A1 |
20070268158 | Gunderson | Nov 2007 | A1 |
20070271105 | Gunderson | Nov 2007 | A1 |
20070273480 | Burkman | Nov 2007 | A1 |
20070280677 | Demarco | Nov 2007 | A1 |
20070279214 | Buehler | Dec 2007 | A1 |
20070299612 | Kimura | Dec 2007 | A1 |
20080035108 | Ancimer | Feb 2008 | A1 |
20080059019 | Delia | Mar 2008 | A1 |
20080071827 | Hengel | Mar 2008 | A1 |
20080111666 | Plante | May 2008 | A1 |
20080122603 | Plante | May 2008 | A1 |
20080137912 | Kim | Jun 2008 | A1 |
20080143834 | Comeau | Jun 2008 | A1 |
20080147267 | Plante et al. | Jun 2008 | A1 |
20080157510 | Breed | Jul 2008 | A1 |
20080167775 | Kuttenberger | Jul 2008 | A1 |
20080169914 | Albertson | Jul 2008 | A1 |
20080177436 | Fortson | Jul 2008 | A1 |
20080195261 | Breed | Aug 2008 | A1 |
20080204556 | de Miranda | Aug 2008 | A1 |
20080211779 | Pryor | Sep 2008 | A1 |
20080234920 | Nurminen | Sep 2008 | A1 |
20080243389 | Inoue | Oct 2008 | A1 |
20080252412 | Larsson | Oct 2008 | A1 |
20080252485 | Lagassey | Oct 2008 | A1 |
20080252487 | McClellan | Oct 2008 | A1 |
20080269978 | Shirole | Oct 2008 | A1 |
20080281485 | Plante | Nov 2008 | A1 |
20080309762 | Howard et al. | Dec 2008 | A1 |
20080319604 | Follmer | Dec 2008 | A1 |
20090009321 | McClellan | Jan 2009 | A1 |
20090043500 | Satoh | Feb 2009 | A1 |
20090043971 | Kim | Feb 2009 | A1 |
20090051510 | Follmer | Feb 2009 | A1 |
20090138191 | Engelhard | May 2009 | A1 |
20090157255 | Plante | Jun 2009 | A1 |
20090216775 | Ratliff et al. | Aug 2009 | A1 |
20090224869 | Baker | Sep 2009 | A1 |
20090290848 | Brown | Nov 2009 | A1 |
20090299622 | Denaro | Dec 2009 | A1 |
20090312998 | Berckmans | Dec 2009 | A1 |
20090326796 | Prokhorov | Dec 2009 | A1 |
20090327856 | Mouilleseaux | Dec 2009 | A1 |
20100030423 | Nathanson | Feb 2010 | A1 |
20100045451 | Periwal | Feb 2010 | A1 |
20100047756 | Schneider | Feb 2010 | A1 |
20100049516 | Talwar | Feb 2010 | A1 |
20100054709 | Misawa | Mar 2010 | A1 |
20100057342 | Muramatsu | Mar 2010 | A1 |
20100063672 | Anderson | Mar 2010 | A1 |
20100063680 | Tolstedt | Mar 2010 | A1 |
20100063850 | Daniel | Mar 2010 | A1 |
20100070175 | Soulchin | Mar 2010 | A1 |
20100076621 | Kubotani | Mar 2010 | A1 |
20100085193 | Boss | Apr 2010 | A1 |
20100085430 | Kreiner | Apr 2010 | A1 |
20100087984 | Joseph | Apr 2010 | A1 |
20100100315 | Davidson | Apr 2010 | A1 |
20100103165 | Lee | Apr 2010 | A1 |
20100104199 | Zhang | Apr 2010 | A1 |
20100149418 | Freed | Jun 2010 | A1 |
20100153146 | Angell | Jun 2010 | A1 |
20100157061 | Katsman | Jun 2010 | A1 |
20100191411 | Cook | Jul 2010 | A1 |
20100201875 | Rood | Aug 2010 | A1 |
20100220892 | Kawakubo | Sep 2010 | A1 |
20100250020 | Lee | Sep 2010 | A1 |
20100250021 | Cook | Sep 2010 | A1 |
20100250022 | Hines | Sep 2010 | A1 |
20100250060 | Maeda | Sep 2010 | A1 |
20100250116 | Yamaguchi | Sep 2010 | A1 |
20100253918 | Seder | Oct 2010 | A1 |
20100268415 | Ishikawa | Oct 2010 | A1 |
20100283633 | Becker | Nov 2010 | A1 |
20100312464 | Fitzgerald | Dec 2010 | A1 |
20110035139 | Konlditslotis | Feb 2011 | A1 |
20110043624 | Haug | Feb 2011 | A1 |
20110060496 | Nielsen | Mar 2011 | A1 |
20110077028 | Wilkes | Mar 2011 | A1 |
20110091079 | Yu-Song | Apr 2011 | A1 |
20110093159 | Boling | Apr 2011 | A1 |
20110112995 | Chang | May 2011 | A1 |
20110121960 | Tsai | May 2011 | A1 |
20110125365 | Larschan | May 2011 | A1 |
20110130916 | Mayer | Jun 2011 | A1 |
20110140884 | Santiago | Jun 2011 | A1 |
20110153367 | Amigo | Jun 2011 | A1 |
20110161116 | Peak | Jun 2011 | A1 |
20110166773 | Raz | Jul 2011 | A1 |
20110169625 | James | Jul 2011 | A1 |
20110172864 | Syed | Jul 2011 | A1 |
20110173015 | Chapman | Jul 2011 | A1 |
20110208428 | Matsubara | Aug 2011 | A1 |
20110212717 | Rhoads | Sep 2011 | A1 |
20110213628 | Peak | Sep 2011 | A1 |
20110224891 | Iwuchukwu | Sep 2011 | A1 |
20110251752 | DeLarocheliere | Oct 2011 | A1 |
20110251782 | Perkins | Oct 2011 | A1 |
20110254676 | Marumoto | Oct 2011 | A1 |
20110257882 | McBurney | Oct 2011 | A1 |
20110273568 | Lagassey | Nov 2011 | A1 |
20110282542 | Nielsen | Nov 2011 | A9 |
20110283223 | Vattinen et al. | Nov 2011 | A1 |
20110304446 | Basson | Dec 2011 | A1 |
20120021386 | Anderson | Jan 2012 | A1 |
20120035788 | Trepagnier | Feb 2012 | A1 |
20120041675 | Juliver | Feb 2012 | A1 |
20120046803 | Inou | Feb 2012 | A1 |
20120065834 | Senart | Mar 2012 | A1 |
20120071140 | Oesterling | Mar 2012 | A1 |
20120078063 | Moore-Ede | Mar 2012 | A1 |
20120081567 | Cote | Apr 2012 | A1 |
20120100509 | Gunderson | Apr 2012 | A1 |
20120109447 | Yousefi | May 2012 | A1 |
20120123806 | Schumann | May 2012 | A1 |
20120134547 | Jung | May 2012 | A1 |
20120150436 | Rossano | Jun 2012 | A1 |
20120176234 | Taneyhill | Jul 2012 | A1 |
20120190001 | Knight | Jul 2012 | A1 |
20120198317 | Eppolito | Aug 2012 | A1 |
20120210252 | Fedoseyeva | Aug 2012 | A1 |
20120277950 | Plante | Nov 2012 | A1 |
20120280835 | Raz | Nov 2012 | A1 |
20120283895 | Noda | Nov 2012 | A1 |
20120330528 | Schwindt | Dec 2012 | A1 |
20130004138 | Kilar | Jan 2013 | A1 |
20130006469 | Green | Jan 2013 | A1 |
20130021148 | Cook | Jan 2013 | A1 |
20130028320 | Gardner | Jan 2013 | A1 |
20130030660 | Fujimoto | Jan 2013 | A1 |
20130046449 | Yuecel | Feb 2013 | A1 |
20130052614 | Mollicone | Feb 2013 | A1 |
20130073112 | Phelan | Mar 2013 | A1 |
20130073114 | Nemat-Nasser | Mar 2013 | A1 |
20130096731 | Tamari | Apr 2013 | A1 |
20130127980 | Haddick | May 2013 | A1 |
20130145269 | Latulipe | Jun 2013 | A1 |
20130151980 | Lee | Jun 2013 | A1 |
20130170762 | Marti | Jul 2013 | A1 |
20130189649 | Mannino | Jul 2013 | A1 |
20130197774 | Denson | Aug 2013 | A1 |
20130209968 | Miller | Aug 2013 | A1 |
20130274950 | Richardson | Oct 2013 | A1 |
20130278631 | Border | Oct 2013 | A1 |
20130317711 | Plante | Nov 2013 | A1 |
20130332004 | Gompert et al. | Dec 2013 | A1 |
20130345927 | Cook | Dec 2013 | A1 |
20130345929 | Bowden | Dec 2013 | A1 |
20140025225 | Armitage | Jan 2014 | A1 |
20140025254 | Plante | Jan 2014 | A1 |
20140032062 | Baer | Jan 2014 | A1 |
20140046550 | Palmer | Feb 2014 | A1 |
20140047371 | Palmer | Feb 2014 | A1 |
20140058583 | Kesavan | Feb 2014 | A1 |
20140089504 | Scholz | Mar 2014 | A1 |
20140094992 | Lambert | Apr 2014 | A1 |
20140098228 | Plante | Apr 2014 | A1 |
20140152828 | Plante | Jun 2014 | A1 |
20140226010 | Molin | Aug 2014 | A1 |
20140232863 | Paliga | Aug 2014 | A1 |
20140270684 | Jayaram | Sep 2014 | A1 |
20140279707 | Joshua | Sep 2014 | A1 |
20140280204 | Avery | Sep 2014 | A1 |
20140300739 | Mimar | Oct 2014 | A1 |
20140309849 | Ricci | Oct 2014 | A1 |
20140335902 | Guba | Nov 2014 | A1 |
20140336916 | Yun | Nov 2014 | A1 |
20150035665 | Plante | Feb 2015 | A1 |
20150057836 | Plante | Feb 2015 | A1 |
20150105934 | Palmer | Apr 2015 | A1 |
20150112542 | Fuglewicz | Apr 2015 | A1 |
20150112545 | Binion | Apr 2015 | A1 |
20150134226 | Palmer | May 2015 | A1 |
20150135240 | Shibuya | May 2015 | A1 |
20150156174 | Fahey | Jun 2015 | A1 |
20150170428 | Harter | Jun 2015 | A1 |
20150189042 | Sun | Jul 2015 | A1 |
20150022449 | Salinger | Aug 2015 | A1 |
20150222449 | Salinger | Aug 2015 | A1 |
20150317846 | Plante | Nov 2015 | A1 |
20160054733 | Hollida | Feb 2016 | A1 |
20190176837 | Williams | Jun 2019 | A1 |
20190176847 | Palmer | Jun 2019 | A1 |
20190180524 | Palmer | Jun 2019 | A1 |
Number | Date | Country |
---|---|---|
2469728 | Dec 2005 | CA |
2469728 | Dec 2005 | CA |
2692415 | Aug 2011 | CA |
2692415 | Aug 2011 | CA |
20311262 | Sep 2003 | DE |
102004004669 | Dec 2005 | DE |
0708427 | Apr 1996 | EP |
0840270 | May 1998 | EP |
0848270 | May 1998 | EP |
1170697 | Jan 2002 | EP |
1324274 | Jul 2003 | EP |
1355278 | Oct 2003 | EP |
1427165 | Jun 2004 | EP |
1818873 | Aug 2007 | EP |
2320387 | May 2011 | EP |
2407943 | Jan 2012 | EP |
2268608 | Jan 1994 | GB |
2402530 | Dec 2004 | GB |
2402530 | Dec 2004 | GB |
2451485 | Feb 2009 | GB |
2447184 | Jun 2011 | GB |
2446994 | Aug 2011 | GB |
58085110 | May 1983 | JP |
62091092 | Apr 1987 | JP |
02056197 | Feb 1990 | JP |
5294188 | Nov 1993 | JP |
10076880 | Mar 1998 | JP |
1000588169 | Dec 2000 | KR |
20000074416 | Dec 2000 | KR |
8809023 | Nov 1988 | WO |
9005076 | May 1990 | WO |
9427844 | Dec 1994 | WO |
9600957 | Jan 1996 | WO |
9701246 | Jan 1997 | WO |
9726750 | Jul 1997 | WO |
9937503 | Jul 1999 | WO |
9940545 | Aug 1999 | WO |
9962741 | Dec 1999 | WO |
0007150 | Feb 2000 | WO |
0028410 | May 2000 | WO |
0048033 | Aug 2000 | WO |
0077620 | Dec 2000 | WO |
0123214 | Apr 2001 | WO |
0125054 | Apr 2001 | WO |
0146710 | Jun 2001 | WO |
03045514 | Jun 2003 | WO |
2004066275 | Aug 2004 | WO |
2006022824 | Mar 2006 | WO |
2006022824 | Mar 2006 | WO |
2007067767 | Jun 2007 | WO |
2007109091 | Sep 2007 | WO |
2009081234 | Jul 2009 | WO |
2011055743 | May 2011 | WO |
2013072939 | May 2013 | WO |
2013159853 | Oct 2013 | WO |
Entry |
---|
PCT International Search Report and Written Opinion for PCT/IB16/51863, dated Sep. 16, 2016. |
USPTO Non-Final Office Action dated Jan. 4, 2016 in U.S. Appl. No. 14/529,134, filed Oct. 30, 2014 (65 pgs). |
PCT International Search Report and Written Opinion for PCT/US15/60721 dated Feb. 26, 2016, 11 pages. |
Notice of Allowance for U.S. Appl. No. 13/957,810, dated Jun. 8, 2015, 10 pages. |
Trivinci Systems, LLC, “Race-Keeper System User Guide”, V1 .1.02, Jan. 2011, p. 21. |
USPTO Non-Final Office Action dated Nov. 27, 2013 in U.S. Appl. No. 13/957,810, filed Aug. 2, 2013. |
Adaptec published and sold its VideoOh! DVD software USB 2.0 Edition in at least Jan. 24, 2003. |
Ambulance Companies Use Video Technology to Improve Driving Behavior, Ambulance Industry Journal, Spring 2003. |
Amended Complaint for Patent Infringement, Trade Secret Misappropriation, Unfair Competition and Conversion in DriveCam, Inc. v. SmartDrive Systems, Inc., Case No. 3:11-CV-00997-H-RBB, for the Southern District of California, Document 34, filed Oct. 20, 2011, pp. 1-15. |
Amendment filed Dec. 23, 2009 during prosecution of U.S. Appl. No. 11/566,424. |
Answer to Amended Complaint; Counterclaims; and Demand for Jury Trial in DriveCam, Inc. v. SmartDrive Systems, Inc., Case No. 3:11-CV-00997 H (RBB), for the Southern District of California, Document 47, filed Dec. 13, 2011, pp. 1-15. |
U.S. Appl. No. 11/296,906, filed Dec. 8, 2005, File History. |
U.S. Appl. No. 11/297,669, filed Dec. 8, 2005, File History. |
U.S. Appl. No. 11/297,889, filed Dec. 8, 2005, File History. |
U.S. Appl. No. 11/298,069, filed Dec. 9, 2005, File History. |
U.S. Appl. No. 11/299,028, filed Dec. 9, 2005, File History. |
U.S. Appl. No. 11/593,659, filed Nov. 7, 2006, File History. |
U.S. Appl. No. 11/593,682, filed Nov. 7, 2006, File History. |
U.S. Appl. No. 11/593,882, filed Nov. 7, 2006, File History. |
U.S. Appl. No. 11/595,015, filed Nov. 9, 2006, File History. |
U.S. Appl. No. 11/637,754, filed Dec. 13, 2006, File History. |
U.S. Appl. No. 11/637,755, filed Dec. 13, 2006, File History. |
Bill, ‘DriveCam—FAQ’, Dec. 12, 2003. |
Bill Siuru, ‘DriveCam Could Save You Big Bucks’, Land Line Magazine, May-Jun. 2000. |
Chris Woodyard, ‘Shuttles save with DriveCam’, Dec. 9, 2003. |
Dan Carr, Flash Video Template: Video Presentation with Navigation, Jan. 16, 2006, http://www.adobe.com/devnet/fiash/articles/vidtemplate_mediapreso_flash8.html. |
David Cullen, ‘Getting a real eyeful’, Fleet Owner Magazine, Feb. 2002. |
David Maher, ‘DriveCam Brochure Folder’, Jun. 6, 2005. |
David Maher, “DriveCam Brochure Folder”, Jun. 8, 2005. |
David Vogeleer et al., Macromedia Flash Professional 8UNLEASHED (Sams Oct. 12, 2005). |
Del Lisk, ‘DriveCam Training Handout Ver4’, Feb. 3, 2005. |
Drivecam, Inc., User's Manual for Drivecam Video Systems' Hindsight 20/20 Software Version 4.0 (2003). |
DriveCam, Inc.'s Infringement Contentions Exhibit A, U.S. Pat. No. 6,389,340, Document 34.1, Oct. 20, 2011. |
DriveCam, Inc.'s Infringement Contentions Exhibit B, U.S. Pat. No. 7,659,827. Aug. 19, 2011. |
DriveCam, Inc.'s Infringement Contentions Exhibit B, U.S. Pat. No. 7,804,426, Document 34.2, Oct. 20, 2011. |
DriveCam, Inc.'s Infringement Contentions Exhibit C, U.S. Pat. No. 7,659,827, Document 34.3, Oct. 20, 2011. |
DriveCam, Inc.'s Infringement Contentions Exhibit C, U.S. Pat. No. 7,804,426. Aug. 19, 2011. |
DriveCam, Inc.'s Infringement Contentions Exhibit D, Document 34.4, Oct. 20, 2011. |
DriveCam—Illuminator Data Sheet, Oct. 2, 2004. |
Drivecam.com as retrieved by the Internet Wayback Machine as of Mar. 5, 2005. |
DriveCam's Disclosure of Asserted Claims and Preliminary Infringement Contentions in DriveCam, Inc. v. SmartDrive Systems, Inc., Case No. 3:11-CV-00997-H-RBB, for the Southern District of California. Aug. 19, 2011. |
DriveCam Driving Feedback System, Mar. 15, 2004. |
DriveCam Extrinsic Evidence with Patent LR 4.1 .a Disclosures, Nov. 3, 2011. |
DriveCam Extrinsic Evidence with Patent LR 4.1 .a Disclosures, Nov. 8, 2011. |
Driver Feedback System, Jun. 12, 2001. |
First Amended Answer to Amended Complaint and First Amended Counterclaims; and Demand for Jury Trial in DriveCam, Inc. v. SmartDrive Systems, Inc., Case No. 3:11-CV-00997 H (RBB), for the Southern District of California, Document 53, filed Dec. 20, 2011, pp. 1-48. |
First Amended Answer to Amended Complaint and First Amended Counterclaims; and Demand for Jury Trial in DriveCam, Inc. v. SmartDrive Systems, Inc., Case No. 3:11-CV-00997 H (RBB), for the Southern District of California, Document 55, filed Jan. 1, 2012, pp. 86-103. |
First Amended Answer to Amended Complaint and First Amended Counterclaims; and Demand for Jury Trial in DriveCam, Inc. v. SmartDrive Systems, Inc., Case No. 3:11-CV-00997 H (RBB), for the Southern District of California, Document 55, filed Jan. 3, 2012, pp. 86-103. |
First Amended Answer to Amended Complaint and First Amended Counterclaims; and Demand for Jury Trial in DriveCam, Inc. v. SmartDrive Systems, Inc., Case No. 3:11-CV-00997 H (RBB), for the Southern District of California, Exhibit A, Document 55, filed Jan. 3, 2012, pp. 49-103. |
Franke, U., et al., Autonomous Driving Goes Downtown, IEEE Intelligent Systems, 13(6):40-48 (1988); Digital Object Identifier 10.1109/5254.736001. |
Gallagher, B., et al., Wireless Communications for Vehicle Safety: Radio Link Performance and Wireless Connectivity Methods, Vehicular Technology Magazine, IEEE, 1(4):4-24 (2006); Digital Object Identifier 10.1109/MVT.2006.343641. |
Gandhi, T., et al., Pedestrian Protection Systems: Issues, Survey, and Challenges, IEEE Transactions on Intelligent Transportation Systems, 8(3):413-430 (2007); Digital Object Identifier 10.1109/TITS.2007.903444. |
Gary and Sophia Rayner, Final Report for Innovations Deserving Exploratory Analysis (IDEA) Intelligent Transportation Systems (ITS) Programs' Project 84, I-Witness Black Box Recorder, San Diego, CA. Nov. 2001. |
GE published its VCR User's Guide for Model VG4255 in 1995. |
Glenn Oster, ‘Hindsight 20/20 v4.0 Software Installation’, 1 of 2, Jun. 20, 2003. |
Glenn Oster, ‘HindSight 20/20 v4.0 Software Installation’, 2 of 2, Jun. 20, 2003. |
Glenn Oster, ‘Illuminator Installation’, Oct. 3, 2004. |
Hans Fantel, Video; Search Methods Make a Difference in Picking VCR's, NY Times, Aug. 13, 1989. |
I/O Port Racing Supplies' website discloses using Traqmate's Data Acquisition with Video Overlay system in conjunction with professional driver coaching sessions (available at http://www.ioportracing.com/Merchant2/merchant.mvc?Screen=CTGY&Categorys- ub.--Code=coaching)., printed from site on Jan. 11, 2012. |
Interior Camera Data Sheet, Oct. 26, 2001. |
International Search Report and Written Opinion issued in PCT/US07/68325 dated Feb. 27, 2008. |
International Search Report and Written Opinion issued in PCT/US07/68328 dated Oct. 15, 2007. |
International Search Report and Written Opinion issued in PCT/US07/68329 dated Mar. 3, 2008. |
International Search Report and Written Opinion issued in PCT/US07/68332 dated Mar. 3, 2008. |
International Search Report and Written Opinion issued in PCT/US07/68334 dated Mar. 5, 2008. |
International Search Report for PCT/US2006/47055, dated Mar. 20, 2008 (2 pages). |
International Search Report issued in PCT/US2006/47042 dated Feb. 25, 2008. |
J. Gallagher, ‘Lancer Recommends Tech Tool’, Insurance and Technology Magazine, Feb. 2002. |
Jean (DriveCam vendor), ‘DC Data Sheet’, Nov. 6, 2002. |
Jean (DriveCam vendor), ‘DriveCam brochure’, Nov. 6, 2002. |
Jean (DriveCam vendor), ‘Feedback Data Sheet’, Nov. 6, 2002. |
Jean (DriveCam vendor), ‘Hindsight 20-20 Data Sheet’, Nov. 4, 2002. |
Jessyca Wallace, ‘Analyzing and Processing DriveCam Recorded Events’, Oct. 6, 2003. |
Jessyca Wallace, ‘Overview of the DriveCam Program’, Dec. 15, 2005. |
Jessyca Wallace, ‘The DriveCam Driver Feedback System’, Apr. 6, 2004. |
Joint Claim Construction Chart, U.S. Pat. No. 6,389,340, ‘Vehicle Data Recorder’ for Case No. 3:11-CV-00997-H-RBB, Document 43-1, filed Dec. 1, 2011, pp. 1-33. |
Joint Claim Construction Chart in DriveCam, Inc. v. SmartDrive Systems, Inc., Case No. 11-CV-0997-H (RBB), for the Southern District of California, Document 43, filed Dec. 1, 2011, pp. 1-2. |
Joint Claim Construction Worksheet, U.S. Pat. No. 6,389,340, ‘Vehicle Data Reporter’ for Case No. 3:11-CV-00997-H-RBB, Document 44-1, filed Dec. 1, 2011, pp. 1-10. |
Joint Claim Construction Worksheet in DriveCam, Inc. v. SmartDrive Systems, Inc., Case No. 3:11-CV-00997 H (RBB), for the Southern District of California, Document 44, filed Dec. 1, 2011, pp. 1-2. |
Joint Motion for Leave to Supplement Disclosure of Asserted Claims and Preliminary Infringement Contentions in DriveCam, Inc. v. SmartDrive Systems, Inc., Case No. 3:11-cv-00997-H-RBB, Document 29, filed Oct. 12, 2011, pp. 1-7. |
Julie Stevens, ‘DriveCam Services’, Nov. 15, 2004. |
Julie Stevens, ‘Program Support Roll-Out & Monitoring’, Jul. 13, 2004. |
Jung, Sang-Hack, et al., Egomotion Estimation in Monocular Infra-red Image Sequence for Night Vision Applications, IEEE Workshop on Applications of Computer Vision (WACV '07), Feb. 2007, pp. 8-8; Digital Object Identifier 10.1109/WACV.2007.20. |
JVC Company of America, JVC Video Cassette Recorder HR-IP820U Instructions (1996). |
Kamijo, S., et al., A Real-Time Traffic Monitoring System by Stochastic Model Combination, IEEE International Conference on Systems, Man and Cybernetics, 4:3275-3281 (2003). |
Kamijo, S., et al., An Incident Detection System Based on Semantic Hierarchy, Proceedings of the 7th International IEEE Intelligent Transportation Systems Conference, Oct. 3-6, 2004, pp. 853-858; Digital Object Identifier 10.1109/ITSC.2004.1399015. |
Karen, ‘Downloading Options to HindSight 20120’, Aug. 6, 2002. |
Karen, ‘Managers Guide to the DriveCam Driving Feedback System’, Jul. 30, 2002. |
Kathy Latus (Latus Design), ‘Case Study—Cloud 9 Shuttle’, Sep. 23, 2005. |
Kathy Latus (Latus Design), ‘Case Study—Lloyd Pest Control’, Jul. 19, 2005. |
Kathy Latus (Latus Design), ‘Case Study—Time Warner Cable’, Sep. 23, 2005. |
Ki, Yong-Kul, et al., A Traffic Accident Detection Model using Metadata Registry, Proceedings of the Fourth International Conference on Software Engineering Research, Management and Applications; Aug. 9-11, 2006 pp. 255-259 Digital Object Identifier 10.1109/SFRA.2006.8. |
Kitchin, Charles. “Understanding accelerometer scale factor and offset adjustments.” Analog Devices (1995). |
Lin, Chin-Teng et al., EEG-based drowsiness estimation for safety driving using independent component analysis; IEEE Transactions on Circuits and Systems-I: Regular Papers, 52(12):2726-2738 (2005); Digital Object Identifier 10.1109/TCSI.2005.857555. |
Lisa Mckenna, ‘A Fly on the Windshield?’, Pest Control Technology Magazine, Apr. 2003. |
Miller, D.P., Evaluation of Vision Systems for Teleoperated Land Vehicles. Control Systems Magazine, IEEE, 8(3):37-41 (1988); Digital Identifier 10.1109/37.475. |
Munder, S., et al., Pedestrian Detection and Tracking Using a Mixture of View-Based Shape-Texture Models, IEEE Transactions on Intelligent Transportation Systems, 9(2):333-343 (2008); Digital Identifier 10.1109/TITS.2008.922943. |
Panasonic Corporation, Video Cassette Recorder (VCR) Operating Instructions for Models No. PV-V4020/PV-V4520. |
Passenger Transportation Mode Brochure, May 2, 2005. |
Patent Abstracts of Japan vol. 007, No. 180 (P-215), Aug. 9, 1983 (Aug. 9, 1983) & JP 58 085110 A (Mitsuhisa Ichikawa), May 21, 1983 (May 21, 1983). |
Patent Abstracts of Japan vol. 011, No. 292 (E-543), Sep. 19, 1987 (Sep. 19, 1987) & JP 62 091092 A (OK Eng:KK), Apr. 25, 1987 (Apr. 25, 1987). |
Patent Abstracts of Japan vol. 012, No. 001 (M-656), Jan. 6, 1988 (Jan. 6, 1988) & JP 62 166135 A (Fuji Electric Co Ltd), Jul. 22, 1987 (Jul. 22, 1987). |
Patent Abstracts of Japan vol. 014, No. 222 (E-0926), May 10, 1990 (May 10, 1990) & JP 02 056197 A (Sanyo Electric Co Ltd), Feb. 26, 1990 (Feb. 26, 1990). |
Patent Abstracts of Japan vol. 017, No. 039 (E-1311), Jan. 25, 1993 (Jan. 25, 1993) & JP 04 257189 A (Sony Corp), Sep. 11, 1992 (Sep. 11, 1992). |
Patent Abstracts of Japan vol. 017, No. 521 (E-1435), Sep. 20, 1993 (Sep. 20, 1993) & JP 05 137144 A (Kyocera Corp), Jun. 1, 1993 (Jun. 1, 1993). |
Patent Abstracts of Japan vol. 1996, No. 09, Sep. 30, 1996 (Sep. 30, 1996) & JP 08 124069 A (Toyota Motor Corp), May 17, 1996 (May 17, 1996). |
Patent Abstracts of Japan vol. 1997, No. 10, Oct. 31, 1997 (Oct. 31, 1997) & JP 09 163357 A (Nippon Soken Inc), Jun. 20, 1997 (Jun. 20, 1997). |
Patent Abstracts of Japan vol. 1998, No. 02, Jan. 30, 1998 (Jan. 30, 1998) & JP 09 272399 A (Nippon Soken Inc), Oct. 21, 1997 (Oct. 21, 1997). |
Patent Abstracts of Japan vol. 1998, No. 8, Jun. 30, 1998 (Jun. 30, 1998) & JP 10 076880 A (Muakami Corp), Mar. 24, 1998 (Mar. 24, 1998). |
PCT/US2010/022012, Invitation to Pay Additional Fees with Communication of Partial International Search, dated Jul. 21, 2010. |
Peter G. Thurlow, Letter (including exhibits) Regarding Patent Owner's Response to Initial Office Action in Ex Parte Reexamination, dated Mar. 27, 2012. |
Preliminary Claim Construction and Identification of Extrinsic Evidence of Defendant/Counterclaimant SmartDriveSystems, Inc.' in DriveCam, Inc. v. SmartDrive Systems, Inc., Case No. 3:11-CV-00997-H (RBB), for the Southern District of California. Nov. 8, 2011. |
Quinn Maughan, ‘DriveCam Enterprise Services’, Jan. 5, 2006. |
Quinn Maughan, ‘DriveCam Managed Services’, Jan. 5, 2006. |
Quinn Maughan, ‘DriveCam Standard Edition’, Jan. 5, 2006. |
Quinn Maughan, ‘DriveCam Unit Installation’, Jul. 21, 2005. |
Quinn Maughan, ‘Enterprise Services’, Apr. 17, 2006. |
Quinn Maughan, ‘Enterprise Services’, Apr. 7, 2006. |
Quinn Maughan, ‘Hindsight Installation Guide’, Sep. 29, 2005. |
Quinn Maughan, ‘Hindsight Users Guide’, Jun. 7, 2005. |
Ronnie Rittenberry, ‘Eyes on the Road’, Jul. 2004. |
SmartDrives Systems, Inc's Production, SO14246-S014255, Nov. 16, 2011. |
Supplement to DriveCam's Disclosure of Asserted Claims and Preliminary Infringement Contentions' in DriveCam, Inc. v. SmartDrive Systems, Inc., Case No. 3:11-CV-00997-H-RBB, for the Southern District of California. Oct. 14, 2011. |
The DriveCam, Nov. 6, 2002. |
The DriveCam, Nov. 8, 2002. |
Traqmate GPS Data Acquisition's Traqmate Data Acquisition with Video Overlay system was used to create a video of a driving event on Oct. 2, 2005 (available at http://www.trackvision.net/phpBB2/viewtopic.php?t=51&sid=1184fbbcbe3be5c87ffa0f2ee6e2da76), printed from site on Jan. 11, 2012. |
U.S. Appl. No. 12/691,639, entitled ‘Driver Risk Assessment System and Method Employing Selectively Automatic Event Scoring’, filed Jan. 21, 2010. |
U.S. Appl. No. 11/377,167, Final Office Action dated Nov. 8, 2013. |
U.S. Appl. No. 11/377,157, filed Mar. 16, 2006 entitled, “Vehicle Event Recorder Systems and Networks Having Parallel Communications Links”. |
U.S. Appl. No. 11/377,167, filed Mar. 16, 2006 entitled, “Vehicle Event Recorder Systems and Networks Having Integrated Cellular Wireless Communications Systems”. |
USPTO Final Office Action for U.S. Appl. No. 11/297,669, dated Nov. 7, 2011, 15 pages. |
USPTO Final Office Action for U.S. Appl. No. 13/957,810, dated Jun. 27, 2014, 24 pages. |
USPTO Non-Final Office Action for U.S. Appl. No. 11/296,906, dated Apr. 2, 2009, 7 pages. |
USPTO Non-Final Office Action for U.S. Appl. No. 11/296,906, dated Nov. 6, 2009, 9 pages. |
USPTO Non-Final Office Action for U.S. Appl. No. 11/297,669, dated Apr. 28, 2011, 11 pages. |
USPTO Non-Final Office Action for U.S. Appl. No. 11/299,028, dated Apr. 24, 2008, 9 pages. |
USPTO Non-Final Office Action for U.S. Appl. No. 11/377,164, dated Nov. 19, 2007, 7 pages. |
USPTO Non-Final Office Action for U.S. Appl. No. 11/377,164, dated Nov. 25, 2011, 9 pages. |
USPTO Non-Final Office Action for U.S. Appl. No. 11/377,164, dated Sep. 11, 2008, 8 pages. |
USPTO Non-Final Office Action for U.S. Appl. No. 11/377,167, dated Jun. 5, 2008, 11 pages. |
USPTO Non-Final Office Action for U.S. Appl. No. 11/800,876, dated Dec. 1, 2010, 12 pages. |
USPTO Non-Final Office Action for U.S. Appl. No. 11/800,876, dated Dec. 20, 2011, 8 pages. |
USPTO Non-Final Office Action for U.S. Appl. No. 12/096,591, dated May 20, 2014, 19 pages. |
USPTO Non-Final Office Action for U.S. Appl. No. 11/296,907, dated Mar. 22, 2007 ( 17 pages). |
USPTO Non-final Office Action dated Aug. 27, 2009 during prosecution of U.S. Appl. No. 11/566,424. |
Veeraraghavan, H., et al., Computer Vision Algorithms for Intersection Monitoring, IEEE Transactions on Intelligent Transportation Systems, 4(2):78-89 (2003); Digital Object Identifier 10.1109/TITS.2003.821212. |
Wijesoma, W.S., et al., Road Curb Tracking in an Urban Environment, Proceedings of the Sixth International Conference of Information Fusion, 1:261-268 (2003). |
World News Tonight, CBC Television New Program discussing teen drivers using the DriveCam Program and DriveCam Technology, Oct. 10, 2005, on PC formatted CD-R, World News Tonight.wmv, 7.02 MB, Created Jan. 12, 2011. |
Written Opinion issued in PCT/US07/68328 dated Oct. 15, 2007. |
Written Opinion of the International Searching Authority for PCT/US2006/47042, dated Feb. 25, 2008 (5 pages). |
Written Opinion of the International Searching Authority for PCT/US2006/47055, dated Mar. 20, 2008 (5 pages). |
Inovate Motorsports, OT-1 16 Channel OBD-II Interface User Manual, Version 1.0, Nov. 28, 2007, pp. 3, 4, 21 & 27. |
Trivinci Systems, LLC, Race-Keeper Systems User Guide, Jan. 2011, v1, 1.02, pp. 34 and 39. |
USPTO Non-Final Office Action for U.S. Appl. No. 14/036,299, dated Aug. 12, 2014. |
DriveCam, Inc.'s Infringement Contentions Exhibit B, U.S. Pat. No. 7,659,827. Aug. 19, 2011. (29 pgs.). |
DriveCam, Inc.'s Infringement Contentions Exhibit C, U.S. Pat. No. 7,804,426. Aug. 19, 2011. (47 pgs.). |
DriveCam's Disclosure of Asserted Claims and Preliminary Infringement Contentions in DriveCam, Inc. v. SmartDrive Systems, Inc., Case No. 3:11-CV-00997-H-RBB, for the Southern District of California. Aug. 19, 2011. (6 pgs.). |
Preliminary Claim Construction and Identification of Extrinsic Evidence of Defendant/Counterclaimant SmartDriveSystems, Inc.' in DriveCam, Inc. v. SmartDrive Systems, Inc., Case No. 3:11-CV-00997-H (RBB), for the Southern District of California. Nov. 8, 2011. (13 pgs.). |
Supplement to DriveCam's Disclosure of Asserted Claims and Preliminary Infringement Contentions' in DriveCam, Inc. v. SmartDrive Systems, Inc., Case No. 3:11-CV-00997-H-RBB, for the Southern District of California. Oct. 14, 2011. (7 pgs.). |
Edwin Olson, A Passive Solution to the Sensor Synchronization Problem, the 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems, Oct. 18-22, 2010, 6 pages. |
“DriveCam, Inc's Disclosure of Proposed Constructions and Extrinsic Evidence Pursuant to Patent L.R. 4.1.a & 4.1.b” Disclosure and Extrinsic Evidence in DriveCam, Inc. v. SmartDrive Systems, Inc., Case No. 3:11-CV-00997-H-RBB, for the Southern District of California. Nov. 8, 2011. |
“DriveCam Driving Feedback System”, DriveCam brochure, Jun. 12, 2001, Document #6600128. |
“DriveCam Driving Feedback System” DriveCam brochure, Mar. 15, 2004. |
“DriveCam Passenger Transportation Module”, DriveCam brochure, Oct. 26, 2001. |
“DriveCam Video Event Data Recorder”, DriveCam brochure, Nov. 6, 2002, Document #6600127. |
“Responsive Claim Construction and Identification of Extrinsic Evidence of Defendani/Counterclaimant SmartDrive Systems, Inc.” Claim Construction and and Extrinsic Evidence in DriveCam, Inc. v. SmartDrive Systems, Inc., Case No. 3:11-CV-00997-H (RBB), for the Southern District of California. Nov. 15, 2011. |
“Sonic MyDVD 4.0: Tutorial: Trimming video segments”. Tutorial for software bundled with Adaptec VideoOh! DVD USB 2.0 Edition, 2003. |
“User's Manual for DriveCam Video Systems' HindSight 20/20 Software Version 4.0” DriveCam Manual, San Diego, 2003, Document #6600141-1. |
Canadian Office Action issued in Application No. 2,632,685 dated Jan. 30, 2015; 5 pages. |
Dan Maher, “DriveCam Taking Risk Out of Driving”, DriveCam brochure folder, Jun. 6, 2005. |
Del Lisk, “DriveCam Training Seminar” Handout, 2004. |
European Examination Report issued in EP 07772812.9 dated Jan. 22, 2015; 5 pages. |
Jean (DriveCam vendor) “DriveCam Driving Feedback System”, DriveCam brochure, Nov. 6, 2002, Document #6600128-1. |
Notice of Allowance Allowance for U.S. Appl. No. 14/036,299, dated Mar. 20, 2015, xx pages. |
Notice of Allowance Application for U.S. Appl. No. 11/566,424, dated Feb. 26, 2010, 6 pages. |
Notice of Allowance for U.S. Appl. No. 11/377,164, dated Dec. 3, 2014, 5 pages. |
Notice of Allowance for U.S. Appl. No. 11/377,164, dated Feb. 13, 2015, 2 pages. |
Notice of Allowance for U.S. Appl. No. 11/377,164, dated Feb. 25, 2014, 2 pages. |
Notice of Allowance for U.S. Appl. No. 11/377,164, dated Nov. 18, 2013, 7 pages. |
Notice of Allowance for U.S. Appl. No. 11/377,167, dated Apr. 1, 2015, 7 pages. |
Notice of Allowance for U.S. Appl. No. 11/800,876, dated Apr. 19, 2012, 8 pages. |
USPTO Final Office Action for U.S. Appl. No. 11/296,906, dated Aug. 8, 2012, 15 pages. |
USPTO Final Office Action for U.S. Appl. No. 12/096,591, dated Dec. 5, 2014, 23 pages. |
USPTO Final Office Action for U.S. Appl. No. 12/096,591, dated Jul. 18, 2012, 15 pages. |
USPTO Final Office Action for U.S. Appl. No. 12/096,591, dated Nov. 7, 2013, 14 pages. |
USPTO Final Office Action for U.S. Appl. No. 13/957,810, dated Jun. 27, 2014, 22 pages. |
USPTO Final Office Action for U.S. Appl. No. 14/036,299, dated Feb. 24, 2015, 9 pages. |
USPTO Non-Final Office Action for U.S. Appl. No. 11/296,906, dated Apr. 8, 2014, 19 pages. |
USPTO Non-Final Office Action for U.S. Appl. No. 11/296,906, dated Jun. 12, 2012, 13 pages. |
USPTO Non-Final Office Action for U.S. Appl. No. 11/377,164, dated Apr. 7, 2014, 7 pages. |
USPTO Non-Final Office Action for U.S. Appl. No. 11/377,164, dated Aug. 18, 2014, 5 pages. |
USPTO Non-Final Office Action for U.S. Appl. No. 11/377,164, dated Sep. 10, 2012, 10 pages. |
USPTO Non-Final Office Action for U.S. Appl. No. 11/377,167, dated Jun. 27, 2013, 11 pages. |
USPTO Non-Final Office Action for U.S. Appl. No. 12/096,591, dated Jun. 14, 2011, 8 pages. |
USPTO Non-Final Office Action for U.S. Appl. No. 12/096,591, dated Mar. 27, 2013, 16 pages. |
USPTO Non-Final Office Action for U.S. Appl. No. 13/957,810, dated Apr. 17, 2015, 6 pages. |
USPTO Non-final Office Action for U.S. Appl. No. 13/957,810, dated Nov. 27, 2013, 18 pages. |
History of the web browser, Wikipedia (Year: 2019) (8 pages). |
Olson, E, A passive solution to the sensor synchronization problem, Intelligent Robots and Systems (IROS), Technical Fields 2010 IEEE/RSJ International Conference on, IEEE, Piscataway, NJ, USA, Searched (IPC) Oct. 18, 2010 (Oct. 18, 2010), pp. 1059-1064, XP031920438, DOI: 10.1109/IROS.2010.5650579 ISBN: 978-1-4244-6674-0. |
Gary A. Rayner, U.S. Appl. No. 09/020,700, filed Feb. 9, 1998. |
Gary A. Rayner, U.S. Appl. No. 09/405,857, filed Sep. 24, 1999. |
Gary A. Rayner, U.S. Appl. No. 09/611,891, filed Jul. 7, 2000. |
Gary A. Rayner, U.S. Appl. No. 09/669,449, filed Sep. 25, 2000. |
Gary A. Rayner, U.S. Appl. No. 09/732,813, filed Dec. 11, 2000. |
Charlie Gunderson, U.S. Appl. No. 11/382,222, filed May 8, 2006. |
Charlie Gunderson, U.S. Appl. No. 11/382,239, filed May 9, 2006. |
Charlie Gunderson, U.S. Appl. No. 11/382,325, filed May 9, 2006. |
Charlie Gunderson, U.S. Appl. No. 11/382,328, filed May 9, 2006. |
David Stanley, U.S. Appl. No. 11/465,765, filed Aug. 18, 2006. |
Larry Richardson, U.S. Appl. No. 11/467,486, filed Aug. 25, 2006. |
Craig Denson, U.S. Appl. No. 11/467,694, filed Aug. 28, 2006. |
Carl Miller, U.S. Appl. No. 11/566,526, filed Dec. 4, 2006. |
Jamie Etcheson, U.S. Appl. No. 11/566,424, filed Dec. 4, 2006. |
Jamie Etcheson, U.S. Appl. No. 11/566,539, filed Dec. 4, 2006. |
Bryan Cook, U.S. Appl. No. 12/359,787, filed Jan. 26, 2009. |
Bryan Cook, U.S. Appl. No. 12/691,639, filed Jan. 21, 2010. |
Bryan Cook, U.S. Appl. No. 12/793,362, filed Jun. 3, 2010. |
Bryan Cook, U.S. Appl. No. 12/814,117, filed Jun. 11, 2010. |
Charlie Gunderson, U.S. Appl. No. 13/234,103, filed Sep. 15, 2011. |
Syrus C. Nemat-Nasser, U.S. Appl. No. 13/235,263, filed Sep. 16, 2011. |
Roni Tamari, U.S. Appl. No. 13/271,417, filed Oct. 12, 2011. |
Bryan Cook, U.S. Appl. No. 13/586,750, filed Aug. 15, 2012. |
Craig Denson, U.S. Appl. No. 13/736,709, filed Jan. 8, 2013. |
Bryan Cook, U.S. Appl. No. 13/923,130, filed Jun. 20, 2013. |
DriveCam, Inc., U.S. Appl. No. 13/914,339, filed Jun. 10, 2013. |
DriveCam, Inc., U.S. Appl. No. 14/027,038, filed Sep. 13, 2013. |
Larry Richardson, U.S. Appl. No. 13/448,725, filed Apr. 17, 2012. |
DriveCam, Inc., U.S. Appl. No. 14/034,296, filed Sep. 23, 2013. |
Joshua Donald Botnen, U.S. Appl. No. 13/222,301, filed Aug. 31, 2011. |
DriveCam, Inc., U.S. Appl. No. 14/070,206, filed Nov. 1, 2013. |
DriveCam, Inc., U.S. Appl. No. 90/011,951, filed Oct. 11, 2011. |
DriveCam, Inc., U.S. Appl. No. 95/001,779, filed Oct. 11, 2011. |
James Plante, U.S. Appl. No. 11/296,906, filed Dec. 8, 2005. |
SmartDrive Systems, Inc., WO2006US47042, Jun. 14, 2007. |
James Plante, U.S. Appl. No. 12/096,591, filed Oct. 3, 2008. |
SmartDrive Systems, Inc., CA2632689, Dec. 8, 2006. |
SmartDrive Systems, Inc., EP068475375, Aug. 20, 2008. |
James Plante, U.S. Appl. No. 11/296,907, filed Dec. 8, 2005. |
James Plante, U.S. Appl. No. 11/595,015, filed Nov. 9, 2006. |
SmartDrive Systems, Inc., WO2006US47055, Jun. 14, 2007. |
James Plante, U.S. Appl. No. 12/096,592, filed Oct. 3, 2008. |
James Plante, U.S. Appl. No. 13/734,800, filed Jan. 4, 2013. |
SmartDrive Systems, Inc., CA2632738, Dec. 7, 2006. |
SmartDrive Systems, Inc., GB2632738, Dec. 7, 2006. |
James Plante, U.S. Appl. No. 11/297,669, filed Dec. 8, 2005. |
SmartDrive Systems, Inc., WO2006US47029, Jun. 14, 2007. |
SmartDrive Systems, Inc., CA2632685, Dec. 7, 2006. |
SmartDrive Systems, Inc., GB08113383, Dec. 7, 2006. |
James Plante, U.S. Appl. No. 11/377,157, filed Mar. 16, 2006. |
SmartDrive Systems, Inc., WO2007US06536, Sep. 27, 2007. |
SmartDrive Systems, Inc., AU2007227555, Mar. 14, 2007. |
SmartDrive Systems, Inc., CA2645896, Mar. 14, 2007. |
SmartDrive Systems, Inc., EP077531838, Dec. 17, 2008. |
SmartDrive Systems, Inc., NZ571542, Mar. 14, 2007. |
James Plante, U.S. Appl. No. 11/377,164, filed Mar. 16, 2006. |
James Plante, WO2007US06404, Oct. 18, 2007. |
James Plante, EP077728129, Dec. 17, 2008. |
James Plante, U.S. Appl. No. 11/377,167, filed Mar. 16, 2006. |
James Plante, WO2007US06265, Sep. 27, 2007. |
James Plante, EP077529295, Dec. 17, 2008. |
James Plante, U.S. Appl. No. 11/593,659, filed Nov. 7, 2006. |
James Plante, WO2007US83997, May 15, 2008. |
James Plante, U.S. Appl. No. 13/568,151, filed Aug. 7, 2012. |
James Plante, U.S. Appl. No. 13/570,283, filed Aug. 9, 2012. |
James Plante, U.S. Appl. No. 11/593,682, filed Nov. 7, 2006. |
James Plante, WO2007US83998, May 15, 2008. |
James Plante, U.S. Appl. No. 11/637,754, filed Dec. 13, 2006. |
James Plante, U.S. Appl. No. 11/800,876, filed May 8, 2007. |
James Plante, U.S. Appl. No. 13/539,312, filed Jun. 30, 2012. |
James Plante, U.S. Appl. No. 11/298,069, filed Dec. 9, 2005. |
James Plante, U.S. Appl. No. 13/957,810, filed Aug. 2, 2013. |
James Plante, U.S. Appl. No. 11/299,028, filed Dec. 9, 2005. |
Jason Palmer, U.S. Appl. No. 14/076,511, filed Nov. 11, 2013. |
Jason Palmer, U.S. Appl. No. 14/055,833, filed Oct. 16, 2013. |
Jason Palmer, U.S. Appl. No. 14/186,416, filed Feb. 21, 2014. |
James Plante, U.S. Appl. No. 14/036,299, filed Sep. 25, 2013. |
James Plante, U.S. Appl. No. 14/177,047, filed Feb. 10, 2014. |
SmartDrive Systems, Inc., PCT/US07/084366, Nov. 9, 2007. |
DriveCam, Inc., U.S. Appl. No. 95/001,781, filed Oct. 11, 2011. |
DriveCam, Inc., U.S. Appl. No. 95/001,802, filed Nov. 3, 2011. |
DriveCam, Inc., U.S. Appl. No. 14/162,678. |
DriveCam, Inc., PCT/US07/68324, May 7, 2007. |
DriveCam, Inc., PCT/US07/68325, Nov. 22, 2007. |
DriveCam, Inc., PCT/US07/68328, Nov. 22, 2007. |
DriveCam, Inc., PCT/US07/68329, Nov. 22, 2007. |
DriveCam, Inc., PCT/US07/68331, Nov. 22, 2007. |
DriveCam, Inc., PCT/US07/68332, Nov. 22, 2007. |
DriveCam, Inc., PCT/US07/68333, Nov. 22, 2007. |
DriveCam, Inc., PCT/US07/68334, Nov. 22, 2007. |
DriveCam, Inc., PCT/US07/75397, Feb. 28, 2008. |
DriveCam, Inc., PCT/US10/22012, Jul. 29, 2010. |
DriveCam, Inc., PCT/US11/22087, Jul. 28, 2011. |
DriveCam, Inc., PCT/US12/55060, Apr. 18, 2013. |
DriveCam, Inc., PCT/US12/55063, Mar. 21, 2013. |
DriveCam, Inc., PCT/US13/36244, Apr. 11, 2013. |
DriveCam, Inc., PCT/US99/01810, Aug. 12, 1999. |
DriveCam, Inc., PCT/US99/29382, Apr. 5, 2001. |
Jason Palmer, U.S. Appl. No. 13/571,416, filed Aug. 10, 2012. |
Jason Palmer, U.S. Appl. No. 13/571,445, filed Aug. 10, 2012. |
Number | Date | Country | |
---|---|---|---|
20160140872 A1 | May 2016 | US |