This disclosure relates to the field of automated railroad assessment systems. More particularly, this disclosure relates to a system and method for generating point clouds of a rail corridor along a survey path of a test vehicle.
In the last decade, light detection and ranging or “LiDAR” technology has been used in the railroad assessment industry to acquire 3D laser scans of the surroundings of a rail corridor. For example, PCT Publication Number WO2018/208153 entitled “System and Method for Mapping a Railway Track” to Fugro Technologies B.V. describes the use of a 3D laser scanner mounted on the front of a locomotive to gather and generate geo-referenced 3D point cloud data which includes point data corresponding to the two rails on which the vehicle is moving as well as the surroundings of the railroad track. In order to gain a broad view for the LiDAR equipment, such equipment is placed on the very front of a train on the locomotive.
A similar system for gathering 3D data using a device on the front of a locomotive is described in U.S. Patent Application Publication Number 2018/0370552 entitled “Real Time Machine Vision System for Vehicle Control and Protection” to Solfice Research, Inc. The system described can gather point cloud data including data gathered using LiDAR.
U.S. Patent Application Publication Number 2019/0135315 entitled “Railway Asset Tracking and Mapping System” to Herzog Technologies, Inc. describes the use of a LiDAR system for gathering and storing the positions and imagery of physical railroad assets along a rail corridor. Unlike the previously described systems, the Herzog system is mounted on a hi-rail vehicle which can be disruptive to normal train traffic along a railroad.
U.S. Pat. No. 9,175,998 entitled “Ballast Delivery and Computation System and Method” to Georgetown Rail Equipment Company describes a system mounted on a hi-rail vehicle using LiDAR to determine ballast profiles and whether ballast needs to be replaced.
All of the examples discussed above are either mounted on the front of a locomotive or on a hi-rail vehicle. In the former examples, the system is necessarily attached to a locomotive. In the latter examples, hi-rail vehicles are used which can be disruptive to railroads causing downtime while the hi-rail vehicle is operating.
What is needed, therefore, is an alternative to the different ways LiDAR has been used in the past to gather data related to railroads and their surroundings.
The above and other needs are met by a system for generating and interpreting point clouds of a rail corridor along a survey path while moving on a railroad corridor assessment platform. In one embodiment, the system includes a railroad corridor assessment platform; a first LiDAR sensor configured to scan along a first scan plane, the first LiDAR sensor attached to the railroad corridor assessment platform in a rear-facing direction; a second LiDAR sensor configured to scan along a second scan plane, the second LiDAR sensor attached to the railroad corridor assessment platform in a rear-facing direction, wherein the first scan plane crosses the second scan plane but is not coplanar at all points with the second scan plane and wherein neither the first scan plane nor the second scan plane intersect a main body of any rail car adjoined to a rear end of the railroad corridor assessment platform; a data storage device; an Inertial Measurement Unit (IMU); a geo-location device; and a high performance computing system in electrical communication with the first LiDAR sensor, the second LiDAR sensor, the data storage device, the IMU, and the geo-location device, the computing system comprising a high-performance processor wherein the processor controls operations of the first LiDAR sensor and the second LiDAR sensor, and wherein the processor performs a method for generating and interpreting point clouds of a rail corridor, the method comprising operations of (i) obtaining a first set of point cloud data using the first LiDAR sensor; (ii) obtaining a second set of point cloud data using the second LiDAR sensor; (iii) obtaining railroad corridor assessment platform attitude information using the IMU; (iv) obtaining geo-location information using the geo-location device; (v) combining the first set of point cloud data together, the second set of point cloud data, the railroad corridor platform attitude information, and the geo-location information to generate a combined point cloud; (vi) identifying rail corridor features of interest found in the combined point cloud; (vii) creating an inventory of the identified rail corridor features of interest; and (viii) storing the combined point cloud on the data storage device.
The system for generating and interpreting point clouds of a rail corridor described above preferably further includes (A) a first sensor enclosure further including (i) a first sensor enclosure outer shell comprising a first sensor enclosure outer shell first aperture and a first sensor enclosure outer shell second aperture; (ii) a first sensor enclosure first high resolution camera in electrical communication with the computing system, the first sensor enclosure first high resolution camera oriented to view from the inside of the first sensor enclosure through the first sensor enclosure outer shell first aperture to gather digital image data of a rail corridor; and (ii) a first sensor enclosure second high resolution camera in electrical communication with the computing system, the first sensor enclosure second high resolution camera oriented to view from the inside of the first sensor enclosure through the first sensor enclosure outer shell second aperture to gather digital image data of a rail corridor; and (B) a second sensor enclosure including (i) a second sensor enclosure outer shell comprising a second sensor enclosure outer shell first aperture and a second sensor outer shell second aperture; (ii) a second sensor enclosure first high resolution camera in electrical communication with the computing system, the second sensor enclosure first high resolution camera oriented to view from the inside of the second sensor enclosure through the second sensor enclosure outer shell first aperture to gather digital image data of a rail corridor; and (iii) a second sensor enclosure second high resolution camera in electrical communication with the computing system, the second sensor enclosure second high resolution camera oriented to view from the inside of the second sensor enclosure through the second sensor enclosure outer shell second aperture to gather digital image data of a rail corridor. The system for generating and interpreting point clouds of a rail corridor may further include (A) a wheel mounted shaft encoder for sending trigger signals to the first sensor enclosure first high resolution camera, the first sensor enclosure second high resolution camera, the second sensor enclosure first high resolution camera, and the second sensor enclosure second high resolution camera as the railroad corridor assessment platform moves along a survey path; (B) the high-performance processor wherein the processor controls operations of the first sensor enclosure first high resolution camera, the first sensor enclosure second high resolution camera, the second sensor enclosure first high resolution camera, and the second sensor enclosure second high resolution camera, and wherein the processor performs a method for generating and interpreting digital image data, the method comprising operations of (i) receiving pulses from the shaft encoder and triggering the first sensor enclosure first high resolution camera, the first sensor enclosure second high resolution camera, the second sensor enclosure first high resolution camera, and the second sensor enclosure second high resolution camera to obtain digital image data at the same time instances; (ii) obtaining a first set of digital image data using the first sensor enclosure first high resolution camera; (iii) obtaining a second set of digital image data using the first sensor enclosure second high resolution camera; (iv) obtaining a third set of digital image data using the second sensor enclosure first high resolution camera; (v) obtaining a fourth set of digital image data using the second sensor enclosure second high resolution camera; (vi) combining the first set of digital image data, the second set of digital image data, the third set of digital image data, and the fourth set of digital image data to form a combined set of digital image data comprising a plurality of digital images and generating a combined panoramic digital image of the rail corridor; and (vii) storing the combined set of digital image data on the data storage device. Alternatively or additionally, the system for generating and interpreting point clouds of a rail corridor may further include (A) the first sensor enclosure further including (i) a first sensor enclosure inner shell comprising a first sensor enclosure inner shell first aperture and a first sensor enclosure inner shell second aperture, wherein the first sensor enclosure inner shell is configured to move relative to the first sensor enclosure outer shell from a first sensor enclosure inner shell open position wherein the first sensor enclosure outer shell first aperture is in line with the first sensor enclosure inner shell first aperture and the first sensor enclosure outer shell second aperture is in line with the first sensor enclosure inner shell second aperture to a first sensor enclosure inner shell closed position wherein the first sensor enclosure outer shell first aperture is not in line with the first sensor enclosure inner shell first aperture and the first sensor enclosure outer shell second aperture is not in line with the first sensor enclosure inner shell second aperture resulting in (1) the first sensor enclosure outer shell first aperture being blocked by the first sensor enclosure inner shell to protect the first sensor enclosure first high resolution camera and (2) the first sensor enclosure outer shell second aperture being blocked by the first sensor enclosure inner shell to protect the first sensor enclosure second high resolution camera; (ii) a first inner shell motorized linear actuator in electrical communication with the computing system and connected to the first sensor enclosure inner shell for moving the first sensor enclosure inner shell from the first sensor enclosure inner shell open position to the first sensor enclosure inner shell closed position or from the first sensor enclosure inner shell closed position to the first sensor enclosure inner shell open position depending upon control instructions from the computing system; (iii) a first sensor enclosure LiDAR cap configured to move from a first sensor enclosure cap open position in which the first LiDAR sensor is exposed to a first sensor enclosure cap closed position in which the first LiDAR sensor is not exposed, and from the first sensor enclosure cap closed position to the first sensor enclosure cap open position; and (iv) a first LiDAR cap motorized actuator in electrical communication with the computing system and connected to the first sensor enclosure LiDAR cap for moving the first sensor enclosure LiDAR cap from the first sensor enclosure cap closed position to the first sensor enclosure cap open position or from the first sensor enclosure cap open position to the first sensor enclosure cap closed position depending on control instructions from the computing system; and (B) the second sensor enclosure further including (i) a second sensor enclosure inner shell comprising a second sensor enclosure inner shell first aperture and a second sensor enclosure inner shell second aperture, wherein the second sensor enclosure inner shell is configured to move relative to the second sensor enclosure outer shell from a second sensor enclosure inner shell open position wherein the second sensor enclosure outer shell first aperture is in line with the second sensor enclosure inner shell first aperture and the second sensor enclosure outer shell second aperture is in line with the second sensor enclosure inner shell second aperture to a second sensor enclosure inner shell closed position wherein the second sensor enclosure outer shell first aperture is not in line with the second sensor enclosure inner shell first aperture and the second sensor enclosure outer shell second aperture is not in line with the second sensor enclosure inner shell second aperture resulting in (1) the second sensor enclosure outer shell first aperture being blocked by the second sensor enclosure inner shell to protect the second sensor enclosure first high resolution camera and (2) the second sensor enclosure outer shell second aperture being blocked by the second sensor enclosure inner shell to protect the second sensor enclosure second high resolution camera; (ii) a second inner shell motorized linear actuator in electrical communication with the computing system and connected to the second sensor enclosure inner shell for moving the second sensor enclosure inner shell from the second sensor enclosure inner shell open position to the second sensor enclosure inner shell closed position or from the second sensor enclosure inner shell closed position to the second sensor enclosure inner shell open position depending upon control instructions from the computing system; (iii) a second sensor enclosure LiDAR cap configured to move from a second sensor enclosure cap open position in which the second LiDAR sensor is exposed to a second sensor enclosure cap closed position in which the second LiDAR sensor is not exposed, and from the second sensor enclosure cap closed position to the second sensor enclosure cap open position; and (iV) a second LiDAR cap motorized actuator in electrical communication with the computing system and connected to the second sensor enclosure LiDAR cap for moving the second sensor enclosure LiDAR cap from the second sensor enclosure cap closed position to the second sensor enclosure cap open position or from the second sensor enclosure cap open position to the second sensor enclosure cap closed position depending on control instructions from the computing system.
Some versions of the system described above may further include (A) a climatic sensor on the railroad corridor assessment platform, the climatic sensor in electrical communication with the computing system; and (B) the high-performance processor wherein the processor controls operations of the first inner shell motorized linear actuator and the second inner shell motorized linear actuator, and wherein the processor performs a method for protecting the first sensor enclosure first high resolution camera, the first sensor enclosure second high resolution camera, the second sensor enclosure first high resolution camera, and the second sensor enclosure second high resolution camera, the method comprising operations of (i) receiving climatic conditions data from the climatic sensor; (ii) activating the first inner shell motorized linear actuator to move the first sensor enclosure inner shell from the first sensor enclosure inner shell open position to the first sensor enclosure inner shell closed position based on the received climatic conditions data; (iii) activating the second inner shell motorized linear actuator to move the second sensor enclosure inner shell from the second sensor enclosure inner shell open position to the second sensor enclosure inner shell closed position based on the received climatic conditions data; (iv) activating the first LiDAR cap motorized actuator to move the first sensor enclosure LiDAR cap from the first sensor enclosure cap open position to the first sensor enclosure cap closed position; and (v) activating the second LiDAR cap motorized actuator to move the second sensor enclosure LiDAR cap from the second sensor enclosure cap open position to the second sensor enclosure cap closed position.
Some versions of the system described above may further include (A) a motion sensor for sensing motion of the railroad corridor assessment platform, the motion sensor in electrical communication with the computing system; and (B) the high-performance processor wherein the processor controls operations of the first inner shell motorized linear actuator and the second inner shell motorized linear actuator, and wherein the processor performs a method for protecting the first sensor enclosure first high resolution camera, the first sensor enclosure second high resolution camera, the second sensor enclosure first high resolution camera, and the second sensor enclosure second high resolution camera, the method comprising operations of (i) receiving a motion sensor signal from the motion sensor indicating that the railroad corridor assessment platform is moving relative to a railroad track below a minimum speed threshold programmed into the computing system; (ii) activating the first inner shell motorized linear actuator to move the first sensor enclosure inner shell from the first sensor enclosure inner shell open position to the first sensor enclosure inner shell closed position based on the received motion sensor signal; (iii) activating the second inner shell motorized linear actuator to move the second sensor enclosure inner shell from the second sensor enclosure inner shell open position to the second sensor enclosure inner shell closed position based on the received motion sensor signal; (iv) activating the first LiDAR cap motorized actuator to move the first sensor enclosure LiDAR cap from the first sensor enclosure cap open position to the first sensor enclosure cap closed position; and (v) activating the second LiDAR cap motorized actuator to move the second sensor enclosure LiDAR cap from the second sensor enclosure cap open position to the second sensor enclosure cap closed position.
Some versions of the system described above may further include (A) a temperature sensor on the railroad corridor assessment platform in electrical communication with the computing system and proximate to the first sensor enclosure and the second sensor enclosure; (B) a heating and cooling system in electrical communication with the computing system, the heating and cooling system further including (i) an air blower; (ii) a heater for heating air blown from the air blower; (iii) an air chiller for cooling air blown from the air blower; and (iv) an air duct for channeling air from the air blower to the first sensor enclosure and the second sensor enclosure; and (C) the high-performance processor wherein the processor controls operations of the heating and cooling system, and wherein the processor performs a method for regulating air temperature in the first sensor enclosure and the second sensor enclosure, the method comprising operations of (i) receiving temperature data from the temperature sensor; (ii) activating the air blower; and (iii) activating the heater or the air chiller based on the received temperature data.
Some versions of the system described above may further include (A) the first sensor enclosure further comprising at least one first sensor air aperture through which air can directed across the first LiDAR sensor; and (B) the second sensor enclosure further comprising at least one second sensor air aperture through which air can directed across the second LiDAR sensor.
Some versions of the system described above may further include (A) a first sensor enclosure in which the first LiDAR sensor is housed; (B) a second sensor enclosure in which the second LiDAR sensor is housed; (C) a temperature sensor in electrical communication with the computing system and located proximate to the first LiDAR sensor and the second LiDAR sensor; (D) a heating and cooling system in electrical communication with the computing system, the heating and cooling system further including (i) an air blower; (ii) a heater for heating air blown from the air blower; (iii) an air chiller for cooling air blown from the air blower; and (iv) a duct for channeling air from the air blower to the first sensor enclosure and the second sensor enclosure depending on temperature data sent by the temperature sensor to the computing system.
In another aspect, a method for generating and interpreting point clouds of a rail corridor is disclosed. In one embodiment, the method includes (A) obtaining a first set of point cloud data using a processor and a first LiDAR sensor oriented to scan along a first scan plane and attached to a railroad corridor assessment platform in a rear-facing orientation wherein the first LiDAR sensor is in electrical communication with the processor; (B) obtaining a second set of point cloud data using the processor and a second LiDAR sensor oriented to scan along a second scan plane and attached to the railroad corridor assessment platform in a rear-facing orientation wherein the second LiDAR sensor is in electrical communication with the processor, wherein the first scan plane crosses the second scan plane but is not coplanar at all points with the second scan plane, and wherein neither the first scan plane nor the second scan plane intersect a main body of any rail car adjoined to a rear end of the railroad corridor assessment platform; (C) obtaining railroad corridor assessment platform attitude data using an Inertial Measurement Unit (IMU) in electrical communication with the processor; (D) obtaining geo-location data of the railroad corridor assessment platform using a geo-location device in electrical communication with the processor; (E) combining the first set of point cloud data, the second set of point cloud data, the railroad corridor assessment platform attitude data, and the geo-location data to generate a combined point cloud using the processor; (F) identifying rail corridor features of interest found in the combined point cloud using the processor; (G) creating an inventory of the identified rail corridor features of interest using the processor; and (H) storing the combined point cloud on a data storage device in electrical communication with the processor. In some embodiments, the method may further include (A) receiving pulses from a wheel mounted shaft encoder in electrical communication with the processor and triggering a first sensor enclosure first high resolution camera, a first sensor enclosure second high resolution camera, a second sensor enclosure first high resolution camera, and a second sensor enclosure second high resolution camera to obtain digital image data at the same time instances; (B) obtaining a first set of digital image data using the first sensor enclosure first high resolution camera; (C) obtaining a second set of digital image data using the first sensor enclosure second high resolution camera; (D) obtaining a third set of digital image data using the second sensor enclosure first high resolution camera; (E) obtaining a fourth set of digital image data using the second sensor enclosure second high resolution camera; (F) combining the first set of digital image data, the second set of digital image data, the third set of digital image data, and the fourth set of digital image data to form a combined set of digital image data comprising a plurality of digital images and generating a combined panoramic digital image of a rail corridor; and (G) storing the combined set of digital image data on a data storage device in electrical communication with the processor. Such methods may further include colorizing the combined point cloud using the combined set of digital image data and the processor.
Some versions of the method described above may further include (A) triggering a first sensor enclosure third high resolution camera and a second sensor enclosure third high resolution camera to obtain digital image data at the same time instances as the first sensor enclosure first high resolution camera, the first sensor enclosure second high resolution camera, the second sensor enclosure first high resolution camera, and the second sensor enclosure second high resolution camera; (B) obtaining a fifth set of digital image data using a first sensor enclosure third high resolution camera; (C) obtaining a sixth set of digital image data using a second sensor enclosure third high resolution camera; and (D) combining the first set of digital image data, the second set of digital image data, the third set of digital image data, the fourth set of digital image data, the fifth set of digital image data and the sixth set of digital image data to form a combined set of digital image data comprising a plurality of digital images and generating a combined panoramic digital image of a rail corridor. Such methods may further include geo-referencing the colorized combined point cloud using a geo-referencing device.
Some versions of the methods described above may further include (A) housing the first LiDAR sensor, the first sensor enclosure first high resolution camera, the first sensor enclosure second high resolution camera, and the first sensor enclosure third high resolution camera in a first sensor enclosure including a first sensor enclosure LiDAR cap for protecting the first LiDAR sensor; (B) housing the second LiDAR sensor, the second sensor enclosure first high resolution camera, the second sensor enclosure second high resolution camera, and the second sensor enclosure third high resolution camera in a second sensor enclosure including a second sensor enclosure LiDAR cap for protecting the second LiDAR sensor; and (C) blowing temperature-controlled air to the first sensor enclosure and the second sensor enclosure using a heating and cooling system including an air blower wherein temperature-controlled air is blown through an air duct to the first sensor enclosure and the second sensor enclosure and wherein the heating and cooling system is controlled by the computing system based on temperature data received by a temperature sensor proximate to the first LiDAR sensor and the second LiDAR sensor.
Some versions of the methods described above may further include (A) blowing temperature-controlled air through an aperture in the first sensor enclosure adjacent to the first LiDAR sensor for blowing away flying debris and precipitation from the first LiDAR sensor and to maintain the first sensor enclosure LiDAR cap at a temperature above freezing to eliminate the accumulation of frozen precipitation; and (B) blowing temperature-controlled air through an aperture in the second sensor enclosure adjacent to the second LiDAR sensor for blowing away flying debris and precipitation from the second LiDAR sensor and to maintain the second sensor enclosure LiDAR cap at a temperature above freezing to eliminate the accumulation of frozen precipitation
Some versions of the methods described above may further include (A) housing the first LiDAR sensor in a first sensor enclosure including (i) a first sensor enclosure LiDAR cap configured to move from a first sensor enclosure cap open position in which the first LiDAR sensor is exposed to a first sensor enclosure cap closed position in which the first LiDAR sensor is not exposed, and from the first sensor enclosure cap closed position to the first sensor enclosure cap open position; and (ii) a first LiDAR cap motorized linear actuator in electrical communication with the computing system and connected to the first sensor enclosure LiDAR cap for moving the first sensor enclosure LiDAR cap from the first sensor enclosure cap closed position to the first sensor enclosure cap open position or from the first sensor enclosure cap open position to the first sensor enclosure cap closed position depending on control instructions from the computing system; (B) housing the first sensor enclosure first high-resolution camera in the first sensor enclosure including (i) a first sensor enclosure outer shell; (ii) a first sensor enclosure outer shell first aperture through which the first sensor enclosure first high-resolution camera obtains digital image data; (ii) a first sensor enclosure inner shell configured to move relative to the first sensor outer shell from a first sensor enclosure inner shell open position wherein the first sensor enclosure outer shell first aperture is open and the first sensor enclosure first high-resolution camera is exposed to weather outside the first sensor enclosure to a first sensor enclosure inner shell closed position wherein the first sensor enclosure outer shell first aperture is blocked by the first sensor inner shell and the first sensor enclosure first high-resolution camera is not exposed to weather outside the first sensor enclosure; (C) housing the first sensor enclosure second high-resolution camera in the first sensor enclosure including (i) a first sensor enclosure outer shell second aperture through which the first sensor enclosure second high-resolution camera obtains digital image data; (ii) the first sensor enclosure inner shell configured to move relative to the first sensor enclosure outer shell from the first sensor enclosure inner shell open position wherein the first sensor enclosure outer shell second aperture is open and the first sensor enclosure second high-resolution camera is exposed to weather outside the first sensor enclosure to the first sensor enclosure inner shell closed position wherein the first sensor enclosure outer shell second aperture is blocked by the first sensor inner shell and the first sensor enclosure second high-resolution camera is not exposed to weather outside the first sensor enclosure; (iii) a first inner shell motorized linear actuator connected to the first sensor enclosure inner shell and in electrical communication with the processor for moving the first sensor enclosure inner shell from the first sensor enclosure inner shell open position to the first sensor enclosure inner shell closed position and from the first sensor enclosure inner shell closed position to the first sensor enclosure inner shell open position depending on instructions from the processor; (D) housing the second LiDAR sensor in a second sensor enclosure including (i) a second sensor enclosure LiDAR cap configured to move from a second sensor enclosure cap open position in which the second LiDAR sensor is exposed to a second sensor enclosure cap closed position in which the second LiDAR sensor is not exposed, and from the second sensor enclosure cap closed position to the second sensor enclosure cap open position; and (ii) a second LiDAR cap motorized linear actuator in electrical communication with the computing system and connected to the second sensor enclosure LiDAR cap for moving the second sensor enclosure LiDAR cap from the second sensor enclosure cap closed position to the second sensor enclosure cap open position or from the second sensor enclosure cap open position to the second sensor enclosure cap closed position depending on control instructions from the computing system; (E) housing the second sensor enclosure first high-resolution camera in the second sensor enclosure including (i) a second sensor enclosure outer shell; (ii) a second sensor enclosure outer shell first aperture through which the second sensor enclosure first high-resolution camera obtains digital image data; (iii) a second sensor enclosure inner shell configured to move relative to the second sensor outer shell from a second sensor enclosure inner shell open position wherein the second sensor enclosure outer shell first aperture is open and the second sensor enclosure first high-resolution camera is exposed to weather outside the second sensor enclosure to a second sensor inner shell closed position wherein the second sensor enclosure outer shell first aperture is blocked by the second sensor inner shell and the second sensor enclosure first high-resolution camera is not exposed to weather outside the second sensor enclosure; and (F) housing the second sensor enclosure second high-resolution camera in the second sensor enclosure including (i) a second sensor enclosure outer shell second aperture through which the second sensor enclosure second high-resolution camera obtains digital image data; (ii) the second sensor enclosure inner shell configured to move relative to the second sensor enclosure outer shell from the second sensor enclosure inner shell open position wherein the second sensor enclosure outer shell second aperture is open and the second sensor enclosure second high-resolution camera is exposed to weather outside the second sensor enclosure to the second sensor enclosure inner shell closed position wherein the second sensor enclosure outer shell second aperture is blocked by the second sensor inner shell and the second sensor enclosure second high-resolution camera is not exposed to weather outside the second sensor enclosure; and (iii) a second inner shell motorized linear actuator connected to the second sensor enclosure inner shell and in electrical communication with the processor for moving the second sensor enclosure inner shell from the second sensor enclosure inner shell open position to the second sensor enclosure inner shell closed position and from the second sensor enclosure inner shell closed position to the second sensor enclosure inner shell open position depending on instructions from the processor;
Some versions of the methods described above may further include (A) detecting weather conditions outside the first sensor enclosure and the second sensor enclosure using a climatic sensor in electrical communication with the processor; (B) activating the first inner shell motorized linear actuator to move the first sensor enclosure inner shell from the first sensor enclosure inner shell open position to the first sensor enclosure inner shell closed position based on information received by the processor from the climatic sensor; (C) activating the second inner shell motorized linear actuator to move the second sensor enclosure inner shell from the second sensor enclosure inner shell open position to the second sensor enclosure inner shell closed position based on information received by the processor from the climatic sensor; (D) activating the first LiDAR cap motorized linear actuator to move the first sensor enclosure LiDAR cap from the first sensor enclosure cap open position to the first sensor enclosure cap closed position; and (E) activating the second LiDAR cap motorized linear actuator to move the second sensor enclosure LiDAR cap from the second sensor enclosure cap open position to the second sensor enclosure cap closed position.
Some versions of the methods described above may further include (A) detecting movement of the railroad corridor assessment platform using a motion sensor; (B) activating the first inner shell motorized linear actuator to move the first sensor enclosure inner shell from the first sensor enclosure inner shell open position to the first sensor enclosure inner shell closed position based on information received by the processor from the motion sensor; (C) activating the second inner shell motorized linear actuator to move the second sensor enclosure inner shell from the second sensor enclosure inner shell open position to the second sensor enclosure inner shell closed position based on information received by the processor from the motion sensor; (D) activating the first LiDAR cap motorized actuator to move the first sensor enclosure LiDAR cap from the first sensor enclosure cap open position to the first sensor enclosure cap closed position; and (E) activating the second LiDAR cap motorized actuator to move the second sensor enclosure LiDAR cap from the second sensor enclosure cap open position to the second sensor enclosure cap closed position.
In another aspect, a system for generating and interpreting point clouds of a rail corridor along a survey path while moving on a railroad corridor assessment platform is disclosed. In some embodiments, the system includes (A) a railroad corridor assessment platform; (B) a first LiDAR sensor configured to scan along a first scan plane, the first LiDAR sensor attached to the railroad corridor assessment platform in a rear-facing direction; (C) a second LiDAR sensor configured to scan along a second scan plane, the second LiDAR sensor attached to the railroad corridor assessment platform in a rear-facing direction, wherein the first scan plane crosses the second scan plane but is not coplanar at all points with the second scan plane and wherein neither the first scan plane nor the second scan plane intersect a main body of any rail car adjoined to a rear end of the railroad corridor assessment platform; (D) a data storage device; (E) an Inertial Measurement Unit (IMU); (F) a geo-location device; and (G) a high performance computing system in electrical communication with the first LiDAR sensor, the second LiDAR sensor, the data storage device, the IMU, and the geo-location device, the computing system comprising a high-performance processor wherein the processor controls operations of the first LiDAR sensor and the second LiDAR sensor for obtaining and storing point cloud data.
Some versions of the methods described above may further include (A) a first sensor enclosure in which the first LiDAR sensor is housed; (B) a second sensor enclosure in which the second LiDAR sensor is housed; (C) a temperature sensor in electrical communication with the computing system and located proximate to the first LiDAR sensor and the second LiDAR sensor; (D) a heating and cooling system in electrical communication with and controlled by the computing system, the heating and cooling system further including (i) an air blower; (ii) a heater for heating air blown from the air blower; (iii) an air chiller for cooling air blown from the air blower; and (iv) a duct for channeling air from the air blower to the first sensor enclosure and the second sensor enclosure depending on temperature data sent by the temperature sensor to the computing system.
Some versions of the methods described above may further include (A) the first sensor enclosure further including (i) a first sensor enclosure inner shell comprising a first sensor enclosure inner shell first aperture and a first sensor enclosure inner shell second aperture, wherein the first sensor enclosure inner shell is configured to move relative to the first sensor enclosure outer shell from a first sensor enclosure inner shell open position wherein the first sensor enclosure outer shell first aperture is in line with the first sensor enclosure inner shell first aperture and the first sensor enclosure outer shell second aperture is in line with the first sensor enclosure inner shell second aperture to a first sensor enclosure inner shell closed position wherein the first sensor enclosure outer shell first aperture is not in line with the first sensor enclosure inner shell first aperture and the first sensor enclosure outer shell second aperture is not in line with the first sensor enclosure inner shell second aperture resulting in (1) the first sensor enclosure outer shell first aperture being blocked by the first sensor enclosure inner shell to protect the first sensor enclosure first high resolution camera and (2) the first sensor enclosure outer shell second aperture being blocked by the first sensor enclosure inner shell to protect the first sensor enclosure second high resolution camera; and (ii) a first inner shell motorized linear actuator in electrical communication with the computing system and connected to the first sensor enclosure inner shell for moving the first sensor enclosure inner shell from the first sensor enclosure inner shell open position to the first sensor enclosure inner shell closed position and from the first sensor enclosure inner shell closed position to the first sensor enclosure inner shell open position depending upon control instructions from the computing system; (iii) a first sensor enclosure LiDAR cap configured to move from a first sensor enclosure cap open position in which the first LiDAR sensor is exposed to a first sensor enclosure cap closed position in which the first LiDAR sensor is not exposed, and from the first sensor enclosure cap closed position to the first sensor enclosure cap open position; and (iv) a first LiDAR cap motorized actuator in electrical communication with the computing system and connected to the first sensor enclosure LiDAR cap for moving the first sensor enclosure LiDAR cap from the first sensor enclosure cap closed position to the first sensor enclosure cap open position or from the first sensor enclosure cap open position to the first sensor enclosure cap closed position depending on control instructions from the computing system; and (B) the second sensor enclosure further including (i) a second sensor enclosure inner shell comprising a second sensor enclosure inner shell first aperture and a second sensor enclosure inner shell second aperture, wherein the second sensor enclosure inner shell is configured to move relative to the second sensor enclosure outer shell from a second sensor enclosure inner shell open position wherein the second sensor enclosure outer shell first aperture is in line with the second sensor enclosure inner shell first aperture and the second sensor enclosure outer shell second aperture is in line with the second sensor enclosure inner shell second aperture to a second sensor enclosure inner shell closed position wherein the second sensor enclosure outer shell first aperture is not in line with the second sensor enclosure inner shell first aperture and the second sensor enclosure outer shell second aperture is not in line with the second sensor enclosure inner shell second aperture resulting in (1) the second sensor enclosure outer shell first aperture being blocked by the second sensor enclosure inner shell to protect the second sensor enclosure first high resolution camera and (2) the second sensor enclosure outer shell second aperture being blocked by the second sensor enclosure inner shell to protect the second sensor enclosure second high resolution camera; (ii) a second inner shell motorized linear actuator in electrical communication with the computing system and connected to the second sensor enclosure inner shell for moving the second sensor enclosure inner shell from the second sensor enclosure inner shell open position to the second sensor enclosure inner shell closed position and from the second sensor enclosure inner shell closed position to the second sensor enclosure inner shell open position depending upon control instructions from the computing system; (iii) a second sensor enclosure LiDAR cap configured to move from a second sensor enclosure cap open position in which the second LiDAR sensor is exposed to a second sensor enclosure cap closed position in which the second LiDAR sensor is not exposed, and from the second sensor enclosure cap closed position to the second sensor enclosure cap open position; and (iv) a second LiDAR cap motorized actuator in electrical communication with the computing system and connected to the second sensor enclosure LiDAR cap for moving the second sensor enclosure LiDAR cap from the second sensor enclosure cap closed position to the second sensor enclosure cap open position or from the second sensor enclosure cap open position to the second sensor enclosure cap closed position depending on control instructions from the computing system.
Some versions of the methods described above may further include a climatic sensor in electrical communication with the computing device.
Some versions of the methods described above may further include a motion sensor for sensing motion of the railroad corridor assessment platform and wherein the motion sensor is in electrical communication with the computing system.
The summary provided herein is intended to provide examples of particular disclosed embodiments and is not intended to cover all potential embodiments or combinations of embodiments. Therefore, this summary is not intended to limit the scope of the invention disclosure in any way, a function which is reserved for the appended claims.
Further features, aspects, and advantages of the present disclosure will become better understood by reference to the following detailed description, appended claims, and accompanying figures, wherein elements are not to scale so as to more clearly show the details, wherein like reference numbers indicate like elements throughout the several views, and wherein:
The figures are provided to illustrate concepts of the invention disclosure and are not intended to embody all potential embodiments of the invention. Therefore, the figures are not intended to limit the scope of the invention disclosure in any way, a function which is reserved for the appended claims.
Various terms used herein are intended to have particular meanings. Some of these terms are defined below for the purpose of clarity. The definitions given below are meant to cover all forms of the words being defined (e.g., singular, plural, present tense, past tense). If the definition of any term below diverges from the commonly understood and/or dictionary definition of such term, the definitions below control.
Air or Gas: broadly defined as any gas or mixtures thereof
Data Communication: a first feature is said to be in data communication with a second feature if the first feature is configured to transmit information to the second feature and the second feature is configured to receive such data, whether such data is transmitted through one or more electrical conductors (e.g., wires), cables (including optical fiber), wirelessly, or a combination thereof.
Electrical Communication: a first feature is said to be in electrical communication with a second feature if there is a conductive path for electricity in any form to flow between the first feature and the second feature thereby electrically connecting the first feature with the second feature. Being in electrical communication does not necessarily mean that electricity is actively flowing but that such structures are configured so that electricity could flow easily from the first feature to the second feature. Features that are in electrical communication may also be in data communication with one another. Therefore, for features that normally transfer or receive data, if such features are said to be in electrical communication with one another, it can be inferred that such features are also in data communication with one another.
Fluid Communication: a first feature is said to be in fluid communication with a second feature if there is a duct or path for air to flow between the first feature and the second feature.
Proximate: a first feature is said to be proximate to a second feature if the first feature is attached to or otherwise extends all the way to the second feature or if the first feature is located close to or extends to a location close to the second feature.
An important aspect of the railroad corridor assessment system 100 is the orientation of the first LiDAR sensor 108A and the second LiDAR sensor 108B. The first LiDAR sensor 108A is oriented and configured to scan a first scan plane 120A. The second LiDAR sensor 108B is oriented and configured to scan a second scan plane 120B which, although intersecting with the first scan plane 120A is not in the same plane as the first scan plane 120A. The first scan plane 120A and the second scan plane 120B intersect as shown in
The LiDAR sensors 108 are used to gather point cloud data along a railroad corridor. Such point cloud data is used by the processor 106 to generate and interpret point clouds, revealing various features along a railroad corridor including signage, furniture, adjoining tracks, ballast profile, drainage ditch profile, embankments and tunnel walls. Real-time point clouds are generated by the processor, preferably in LAS file format. The point cloud data that is gathered and the generated point clouds are stored on a data storage device 130 in electrical communication with the computing system 104. The data storage system is preferably in the form of network-attached storage (NAS) computer data storage server. In order to produce a correctly referenced point cloud some additional devices are included in the system 100 including an Inertial Measurement Unit (IMU) 131 in electrical communication with the processor 106 as well as a geolocation device such as, for example, a GPS device 132 in electrical communication with the processor 106. These additional devices help provide real-time LiDAR sensor 108 attitude information (based on the attitude of the boxcar 110 on which the LiDAR sensors 108 are installed) and real-time GPS position information in conjunction with the gathered LiDAR sensor point cloud data. The processor 106 controls operations of the first LiDAR sensor 108A and the second LiDAR sensor 108B and performs a method for generating and interpreting point clouds of a rail corridor. The method includes operations of obtaining a first set of point cloud data using the first LiDAR sensor 108A; obtaining a second set of point cloud data using the second LiDAR sensor 108B; obtaining boxcar 110 attitude information using the IMU 131; obtaining GPS information using the geo-location device 132; combining the first set of point cloud data together, the second set of point cloud data, the IMU attitude information, and GPS information to generate a combined point cloud using the processor 106; identifying rail corridor features of interest found in the combined point cloud using the processor 106; creating an inventory of the identified rail corridor features of interest using the processor 106; and storing the combined point cloud and the inventory of identified rail corridor features on the data storage device 130.
In addition to the processor 106, the computing system preferably further includes one or more LiDAR controllers, a local data storage server and a high-performance compute graphics processing unit (GPU) server. The LiDAR sensors 108 used are preferably Riegl VUX-1HA sensors available from RIEGL Laser Measurement Systems GmbH based in Horn, Austria. Raw and post-processed trajectory and LiDAR sensor data are archived to the data storage device 130 for back-office re-processing and analysis. The railroad corridor assessment system 100 is capable of gathering data and generating point clouds with survey grade accuracy while the railroad corridor assessment platform 102 is moving at speeds up to and even greater than 70 miles per hour on a railroad track. Electrical power for the various devices described in this disclosure can be provided by a diesel generator onboard the railroad corridor assessment platform 102 and/or photovoltaic solar panels, or set of batteries on the railroad corridor assessment platform 102. Preferably, batteries are available on the railroad corridor assessment platform 102 and are charged by an onboard generator and one or more photovoltaic solar panels mounted on the roof of the railroad corridor assessment platform 102. In addition, individual devices may include individualized backup battery power from smaller batteries in electrical communication with individualized devices.
In addition to the LiDAR sensors 108, the railroad corridor assessment system 100 also preferably includes a plurality of first sensor enclosure high resolution cameras 134 and a plurality of second sensor enclosure high resolution cameras 136. The plurality of first sensor enclosure high resolution cameras 134 preferably includes at least a first sensor enclosure first high-resolution camera 134A and a first sensor enclosure second high resolution camera 134B. In the embodiments shown in
The first sensor enclosure 116A includes a first sensor enclosure outer shell 138A including a first sensor enclosure outer shell first aperture 140A, a first sensor enclosure outer shell second aperture 140B, and a first sensor enclosure outer shell third aperture 140C. The first sensor enclosure first high-resolution camera 134A is oriented to view from the inside of the first sensor enclosure 116A through the first sensor enclosure outer shell first aperture 140A to gather digital image data of a rail corridor. The first sensor enclosure second high resolution camera 134B is oriented to view from the inside of the first sensor enclosure 116A through the first sensor enclosure outer shell second aperture 140B to gather digital image data of a rail corridor. The first sensor enclosure third high resolution camera 134C is oriented to view from the inside of the first sensor enclosure 116A through the first sensor enclosure outer shell third aperture 140C to gather digital image data of a rail corridor.
The second sensor enclosure 116B includes a second sensor enclosure outer shell 138B including a second sensor enclosure outer shell first aperture 142A, a second sensor enclosure outer shell second aperture 142B, and a second sensor enclosure outer shell third aperture 142C. The second sensor enclosure first high-resolution camera 136A is oriented to view from the inside of the second sensor enclosure 116B through the second sensor enclosure outer shell first aperture 142A to gather digital image data of a rail corridor. The second sensor enclosure second high resolution camera 136B is oriented to view from the inside of the second sensor enclosure 116B through the second sensor enclosure outer shell second aperture 142B to gather digital image data of a rail corridor. The second sensor enclosure third high resolution camera 136C is oriented to view from the inside of the second sensor enclosure 116B through the second sensor enclosure outer shell third aperture 142C to gather digital image data of a rail corridor.
As shown in the Figures, in embodiments in which three high-resolution digital cameras are used, preferably one of the three cameras is facing up, one of the three cameras is facing out to the side away from the railroad boxcar 110, and one of the three cameras is facing down. Using all six high-resolution digital cameras, it is possible to generate a combined 360-degree panoramic digital image of a rail corridor using the processor 106. The digital image data from each camera (134A, 134B, 134C, 136A, 136B, and 136C) are synchronized using a boxcar wheel mounted shaft encoder 143. Preferably, the shaft encoder 143 uses a 10,000 pulse per revolution producing a pulse every 0.287 millimeter (mm). The encoder pulses are divided to produce a camera trigger every 1.5 to 2 meters while the railroad boxcar 110 is moving at 70 miles per hour. This trigger is used to acquire an image from all six cameras (134A, 134B, 134C, 136A, 136B, and 136C) at the same instance and position so the images can be combined into a single panoramic image. They cannot be accurately combined, nor geo-referenced as a panoramic image, if they are not acquired at the same instance. The processor performs a method for generating and interpreting digital image data. The method includes operations of obtaining a first set of digital image data using the first sensor enclosure first high resolution camera 134A being triggered by signals from the shaft encoder 143; obtaining a second set of digital image data using the first sensor enclosure second high resolution camera 134B being triggered by signals from the shaft encoder 143; obtaining a third set of digital image data using the first sensor enclosure third high resolution camera 134C being triggered by signals from the shaft encoder 143; obtaining a fourth set of digital image data using the second sensor enclosure first high resolution camera 136A being triggered by signals from the shaft encoder 143; obtaining a fifth set of digital image data using the second sensor enclosure second high resolution camera 136B being triggered by signals from the shaft encoder 143; obtaining a sixth set of digital image data using the second sensor enclosure third high resolution camera 136C being triggered by signals from the shaft encoder 143; combining the first set of digital image data, the second set of digital image data, the third set of digital image data, the fourth set of digital image data, the fifth set of digital image data, and the sixth set of digital image data to form a combined set of digital image data including a plurality of digital images and generating a combined panoramic digital image of the rail corridor using the processor 106; time stamping the plurality of digital images using the processor 106; and storing the combined set of digital image data on the data storage device 130. The time stamping of the digital image data allows for geo-referencing and/or coloring a generated LiDAR point cloud by superimposing the generated LiDAR point cloud with the combined panoramic digital image of the rail corridor. Acquired images are preferably able to resolve text with a font height as small as 2 inches at a distance of from about 2 meters to about 15 meters. The combined panoramic digital image of the rail corridor provides a way to visually assess site conditions at the point and time an image of a specific site is obtained and generated by the railroad corridor assessment system 100.
Because high-resolution digital cameras and their attached optic lens are sensitive equipment, it is undesirable to have such devices exposed to the elements in inclement weather, airborne debris or other detrimental conditions. The sensor enclosures 116 shown in the
In addition to protecting the digital cameras (134A, 134B, 134C, 136A, 136B, and 136C) and associated lenses, the sensor enclosures 116 are also configured to protect the LiDAR sensors 108 as well.
In some embodiments, the processor performs a method for protecting the LiDAR sensors (108A and 108B) and the high-resolution cameras (134A, 134B, 134C, 136A, 136B, and 136C). The method includes operations of receiving climatic conditions data from the climatic sensor 144; using the processor 106 to activate the first inner shell motorized linear actuator 152A to move the first sensor enclosure inner shell 146A from the first sensor enclosure inner shell open position to the first sensor enclosure inner shell closed position and activate the first LiDAR cap motorized linear actuator 156A to move the first sensor enclosure LiDAR cap 155A from the first sensor enclosure inner shell open position to the first sensor enclosure inner shell closed position based on the received climatic conditions data; and using the processor to activate the second inner shell motorized linear actuator 152B to move the second sensor enclosure inner shell 146B from the second sensor enclosure inner shell open position to the second sensor enclosure inner shell closed position and activate the second LiDAR cap motorized linear actuator 156B to move the second sensor enclosure LiDAR cap 155B from the second sensor enclosure cap open position to the second sensor enclosure cap closed position based on the received climatic conditions data. This usually would occur in inclement weather. If the climatic sensor 144 sends data to the processor indicating that it is safe to expose the high-resolution cameras (134A, 134B, 134C, 136A, 136B, and 136C) to allow for the cameras to obtain data, a method includes operations of receiving climatic conditions data from the climatic sensor 144; using the processor 106 to activate the first inner shell motorized linear actuator 152A to move the first sensor enclosure inner shell 146A from the first sensor enclosure inner shell closed position to the first sensor enclosure inner shell open position and activate the first LiDAR cap motorized linear actuator 156A to move the first sensor enclosure LiDAR cap 155A from the first sensor enclosure cap closed position to the first sensor enclosure cap open position based on the received climatic conditions data; and using the processor to activate the second inner shell motorized linear actuator 152B to move the second sensor enclosure inner shell 146B from the second sensor enclosure inner shell closed position to the second sensor enclosure inner shell open position and activate the second LiDAR cap motorized linear actuator 156B to move the second sensor enclosure LiDAR cap 155B from the second sensor enclosure cap closed position to the second sensor enclosure cap open position based on the received climatic conditions data.
The railroad corridor assessment system 100 also preferably includes a motion sensor 157 for detecting motion of the railroad corridor assessment platform 102. Depending on data received from the motion sensor 157 to the computing device 104, the processor 106 preferably (1) activates the first inner shell motorized linear actuator 152A to move the first sensor enclosure inner shell 146A to the first sensor enclosure inner shell open position (see
As indicated above, the processor 106 controls operations of the first inner shell motorized linear actuator 152A and the second inner shell motorized linear actuator 152B. In some embodiments, the processor performs a method for protecting the high-resolution cameras (134 and 136). The method includes operations of receiving a motion sensor signal from the motion sensor 157 indicating that the railroad corridor assessment platform 102 has stopped moving relative to a railroad track or is moving below a minimum speed threshold programmed into the computing system 104; using the processor 106 to activate the first inner shell motorized linear actuator 152A to move the first sensor enclosure inner shell 146A from the first sensor enclosure inner shell open position to the first sensor enclosure inner shell closed position based on the received motion sensor signal; and using the processor 106 to activate the second inner shell motorized linear actuator 152B to move the second sensor enclosure inner shell 146B from the second sensor enclosure inner shell open position to the second sensor enclosure inner shell closed position based on the received motion sensor signal. These steps are usually performed if the railroad corridor assessment platform 102 has ceased moving along a railroad track. If, on the other hand, the railroad corridor assessment platform 102 starts moving from a stalled or stopped state, in some embodiments, the processor performs a method for exposing the high-resolution cameras (134 and 136) so that they can gather data. The method includes operations of receiving a motion sensor signal from the motion sensor 157 indicating that the railroad corridor assessment platform 102 has started moving relative to a railroad track at or above a minimum speed threshold programmed into the computing system 104; using the processor 106 to activate the first inner shell motorized linear actuator 152A to move the first sensor enclosure inner shell 146A from the first sensor enclosure inner shell closed position to the first sensor enclosure inner shell open position based on the received motion sensor signal; and using the processor 106 to activate the second inner shell motorized linear actuator 152B to move the second sensor enclosure inner shell 146B from the second sensor enclosure inner shell closed position to the second sensor enclosure inner shell open position based on the received motion sensor signal.
As indicated above, the railroad corridor assessment system 100 and other systems like it are vulnerable to extreme weather conditions and flying dirt and debris in disturbed air created by a consist moving at speed along a rail corridor. The features described above regarding the first sensor enclosure 116A and the second sensor enclosure 116B address some weather concerns. In addition to these features, the railroad corridor assessment system 100 includes a temperature sensor 158 on the railroad corridor assessment platform 102 in electrical communication with the computing system 104 and proximate to the first sensor enclosure 116A and the second sensor enclosure 116B. In some embodiments there are separate temperature sensors including a first temperature sensor 158A in the first sensor enclosure 116A and a second temperature sensor 158B in the second sensor enclosure 116B. The railroad corridor assessment system 100 preferably includes a heating and cooling system 160 in electrical communication with the computing system 104. The heating and cooling system 160 preferably includes an air blower 162, a heater 164 for heating air blown from the air blower 162, an air chiller 166 for cooling air blown from the air blower 162, and an air duct 168 in fluid communication with the heating and cooling system 160 and the sensor enclosures 116 for channeling air from the air blower 162 to the first sensor enclosure 116A and the second sensor enclosure 116B. Preferably, the heater 164 and chiller 166 include a combination heater/chiller 167 such as, for example, a Peltier thermoelectric heating and cooling device capable of providing heating or cooling depending on electrical control signals received by such device from the processor 106. Depending on temperature readings from the first temperature sensor 158A and the second temperature sensor 158B sent to the computing system 104, the processor 106 can be programmed to activate the air blower 162 in addition to either the heating or cooling function of the combination heater/chiller 164 depending on whether the sensor enclosures 116 need to be heated or cooled. If temperatures in the sensor enclosures 116 are within an acceptable range, the processor 106 optionally can activate only the air blower 162 to circulate air to and through the sensor enclosures 116. The climate control features allow the system 100 to operate in extreme weather and climate conditions.
The sensor enclosures 116 each include a sensor enclosure outer cap (170A and 170B) and the LiDAR sensor caps (155A and 155B) shown in
With reference to
The previously described embodiments of the present disclosure have many advantages, including gathering LiDAR point cloud data from a standard-sized railroad boxcar that can easily be added to or removed from a consist. The specific orientation of the LiDAR sensors 108 allows for the gathering of point cloud data even if another boxcar is coupled directly behind the railroad corridor assessment platform 102. The LiDAR sensors enclosures 116 are situated below a roofline of the railroad corridor assessment platform 102. Point clouds can be assessed in real-time by the processor 106 to identify and inventory various features along a rail corridor such as signage, furniture, adjoining tracks, and Positive Train Control (PTC) assets. The processor can use the generated point cloud(s) to measure ballast profiles, measure drainage ditch profiles, and identify and measure of embankments and tunnel walls. If the same survey path is run more than once, on the later survey(s), the processor 106 can be programmed to detect changes in the features that were previously detected in the point clouds of a prior survey by comparing the new and old point clouds. The point cloud data is geo-referenced and time stamped to correlate such data with new or additional data.
An additional advantage is the use of a plurality of digital cameras to gather a 360-degree panoramic ribbon digital image of a rail corridor. The image data is preferably time-stamped and can be combined with the LiDAR point cloud data to add color to the point cloud(s). The sensitive cameras are protected in the sensor enclosures 116 which automatically open or close depending on (1) the weather and (2) whether the railroad corridor assessment platform 102 is moving at a minimum speed. The heating and cooling system 160 provides temperature-controlled air to the sensor enclosures 116 so that the system 100 can keep operating even in extreme heat or cold. Additionally, the enclosures 116 include cap apertures through which the temperature-controlled air can exit the enclosures 116 ad, while exiting, act like an air curtain to blow away any flying dust or debris from the devise in the enclosures 116.
The foregoing description of preferred embodiments of the present disclosure has been presented for purposes of illustration and description. The described preferred embodiments are not intended to be exhaustive or to limit the scope of the disclosure to the precise form(s) disclosed. For example, various features said to be in electrical communication with one another may be communicating wirelessly and powered locally by batteries or other power sources. Obvious modifications or variations are possible in light of the above teachings. The embodiments are chosen and described in an effort to provide the best illustrations of the principles of the disclosure and its practical application, and to thereby enable one of ordinary skill in the art to utilize the concepts revealed in the disclosure in various embodiments and with various modifications as are suited to the particular use contemplated. All such modifications and variations are within the scope of the disclosure as determined by the appended claims when interpreted in accordance with the breadth to which they are fairly, legally, and equitably entitled.
Any element in a claim that does not explicitly state “means for” performing a specified function, or “step for” performing a specific function, is not to be interpreted as a “means” or “step” clause as specified in 35 U.S.C. § 112, ¶6. In particular, the use of “step of” in the claims herein is not intended to invoke the provisions of 35 U.S.C. § 112, ¶6.
This application is a nonprovisional application claiming priority to (1) provisional U.S. Provisional Patent Application No. 62/848,630 invented by Darel Mesher and entitled “Autonomous Track Assessment System” which was filed on May 16, 2019; provisional U.S. Provisional Patent Application No. 62/988,630 invented by Darel Mesher and entitled “Autonomous Track Assessment System” which was filed on Mar. 12, 2020; and provisional U.S. Provisional Patent Application No. 63/016,661 invented by Darel Mesher and entitled “Autonomous Track Assessment System” which was filed on Apr. 28, 2020, the entireties of which are incorporated herein by reference.
Number | Name | Date | Kind |
---|---|---|---|
3562419 | Stewart et al. | Feb 1971 | A |
3942000 | Dieringer | Mar 1976 | A |
4040738 | Wagner | Aug 1977 | A |
4198164 | Cantor | Apr 1980 | A |
4265545 | Slaker | May 1981 | A |
4330775 | Iwamoto et al. | May 1982 | A |
4490038 | Theurer et al. | Dec 1984 | A |
4531837 | Panetti | Jul 1985 | A |
4554624 | Wickham et al. | Nov 1985 | A |
4600012 | Kohayakawa et al. | Jul 1986 | A |
4653316 | Fukuhara | Mar 1987 | A |
4676642 | French | Jun 1987 | A |
4691565 | Theurer | Sep 1987 | A |
4700223 | Shoutaro et al. | Oct 1987 | A |
4731853 | Hata | Mar 1988 | A |
4775238 | Weber | Oct 1988 | A |
4781060 | Berndt | Nov 1988 | A |
4899296 | Khattak | Feb 1990 | A |
4900153 | Weber et al. | Feb 1990 | A |
4915504 | Thurston | Apr 1990 | A |
4974168 | Marx | Nov 1990 | A |
5199176 | Theurer et al. | Apr 1993 | A |
5203089 | Trefouel et al. | Apr 1993 | A |
5221044 | Guins | Jun 1993 | A |
5245855 | Burgel et al. | Sep 1993 | A |
5247338 | Danneskiold-Samsoe et al. | Sep 1993 | A |
5353512 | Theurer et al. | Oct 1994 | A |
5433111 | Hershey et al. | Jul 1995 | A |
5487341 | Newman | Jan 1996 | A |
5493499 | Theurer et al. | Feb 1996 | A |
5612538 | Hackel et al. | Mar 1997 | A |
5623244 | Cooper | Apr 1997 | A |
5627508 | Cooper et al. | May 1997 | A |
5671679 | Straub et al. | Sep 1997 | A |
5721685 | Holland et al. | Feb 1998 | A |
5743495 | Welles | Apr 1998 | A |
5744815 | Gurevich et al. | Apr 1998 | A |
5757472 | Wangler et al. | May 1998 | A |
5786750 | Cooper | Jul 1998 | A |
5787815 | Andersson et al. | Aug 1998 | A |
5791063 | Kesler et al. | Aug 1998 | A |
5793491 | Wangler et al. | Aug 1998 | A |
5793492 | Vanaki | Aug 1998 | A |
5804731 | Jaeggi | Sep 1998 | A |
5808906 | Sanchez-Revuelta et al. | Sep 1998 | A |
5912451 | Gurevich et al. | Jun 1999 | A |
5969323 | Gurevich | Oct 1999 | A |
5970438 | Clark et al. | Oct 1999 | A |
6025920 | Dec | Feb 2000 | A |
6055322 | Salganicoff | Apr 2000 | A |
6055862 | Martens | May 2000 | A |
6062476 | Stern et al. | May 2000 | A |
6064428 | Trosino et al. | May 2000 | A |
6069967 | Rozmus et al. | May 2000 | A |
6128558 | Kernwein | Oct 2000 | A |
6243657 | Tuck et al. | Jun 2001 | B1 |
6252977 | Salganicoff | Jun 2001 | B1 |
6324912 | Woon | Dec 2001 | B1 |
6347265 | Bidaud | Feb 2002 | B1 |
6356299 | Trosino et al. | Mar 2002 | B1 |
6357297 | Makino et al. | Mar 2002 | B1 |
6405141 | Carr et al. | Jun 2002 | B1 |
6416020 | Gronskov | Jul 2002 | B1 |
6496254 | Bostrom | Dec 2002 | B2 |
6523411 | Mian et al. | Feb 2003 | B1 |
6540180 | Anderson | Apr 2003 | B2 |
6570497 | Puckette, IV | May 2003 | B2 |
6600999 | Clark et al. | Jul 2003 | B2 |
6615648 | Ferguson et al. | Sep 2003 | B1 |
6647891 | Holmes et al. | Nov 2003 | B2 |
6665066 | Nair et al. | Dec 2003 | B2 |
6681160 | Bidaud | Jan 2004 | B2 |
6698279 | Stevenson | Mar 2004 | B1 |
6715354 | Wooh | Apr 2004 | B2 |
6768551 | Mian et al. | Jul 2004 | B2 |
6768959 | Ignagni | Jul 2004 | B2 |
6804621 | Pedanckar | Oct 2004 | B1 |
6854333 | Wooh | Feb 2005 | B2 |
6862936 | Kenderian et al. | Mar 2005 | B2 |
6873998 | Dorum | Mar 2005 | B1 |
6909514 | Nayebi | Jun 2005 | B2 |
7023539 | Kowalski | Apr 2006 | B2 |
7034272 | Leonard | Apr 2006 | B1 |
7036232 | Casagrande | May 2006 | B2 |
7054762 | Pagano et al. | May 2006 | B2 |
7084989 | Johannesson et al. | Aug 2006 | B2 |
7130753 | Pedanekar | Oct 2006 | B2 |
7164476 | Shima et al. | Jan 2007 | B2 |
7208733 | Mian et al. | Apr 2007 | B2 |
7213789 | Matzan | May 2007 | B1 |
7298548 | Mian | Nov 2007 | B2 |
7328871 | Mace et al. | Feb 2008 | B2 |
7355508 | Mian et al. | Apr 2008 | B2 |
7357326 | Hattersley et al. | Apr 2008 | B2 |
7392117 | Bilodeau et al. | Jun 2008 | B1 |
7392595 | Heimann | Jul 2008 | B2 |
7403296 | Farritor et al. | Jul 2008 | B2 |
7412899 | Mian et al. | Aug 2008 | B2 |
7463348 | Chung | Dec 2008 | B2 |
7499186 | Waisanen | Mar 2009 | B2 |
7502670 | Harrison | Mar 2009 | B2 |
7516662 | Nieisen et al. | Apr 2009 | B2 |
7555954 | Pagano et al. | Jul 2009 | B2 |
7564569 | Mian et al. | Jul 2009 | B2 |
7602937 | Mian et al. | Oct 2009 | B2 |
7616329 | Villar et al. | Nov 2009 | B2 |
7659972 | Magnus et al. | Feb 2010 | B2 |
7680631 | Selig et al. | Mar 2010 | B2 |
7681468 | Verl et al. | Mar 2010 | B2 |
7698028 | Bilodeau et al. | Apr 2010 | B1 |
7755660 | Nejikovsky et al. | Jul 2010 | B2 |
7755774 | Farritor et al. | Jul 2010 | B2 |
7832281 | Mian et al. | Nov 2010 | B2 |
7869909 | Harrison | Jan 2011 | B2 |
7882742 | Martens | Feb 2011 | B1 |
7899207 | Mian et al. | Mar 2011 | B2 |
8006559 | Mian et al. | Aug 2011 | B2 |
8079274 | Mian et al. | Dec 2011 | B2 |
8081320 | Villar et al. | Dec 2011 | B2 |
8111387 | Douglas et al. | Feb 2012 | B2 |
8140250 | Mian et al. | Mar 2012 | B2 |
8150105 | Mian et al. | Apr 2012 | B2 |
8155809 | Bilodeau et al. | Apr 2012 | B1 |
8180590 | Szwilski et al. | May 2012 | B2 |
8188430 | Mian et al. | May 2012 | B2 |
8190377 | Fu | May 2012 | B2 |
8209145 | Paglinco et al. | Jun 2012 | B2 |
8263953 | Fomenkar et al. | Sep 2012 | B2 |
8289526 | Kilian et al. | Oct 2012 | B2 |
8326582 | Mian et al. | Dec 2012 | B2 |
8335606 | Mian et al. | Dec 2012 | B2 |
8345948 | Zarembski et al. | Jan 2013 | B2 |
8345099 | Bloom et al. | Feb 2013 | B2 |
8365604 | Kahn | Feb 2013 | B2 |
8405837 | Nagle, II et al. | Mar 2013 | B2 |
8412393 | Anderson | Apr 2013 | B2 |
8418563 | Wigh et al. | Apr 2013 | B2 |
8423240 | Mian et al. | Apr 2013 | B2 |
8424387 | Wigh et al. | Apr 2013 | B2 |
8478480 | Mian et al. | Jul 2013 | B2 |
8485035 | Wigh et al. | Jul 2013 | B2 |
8490887 | Jones | Jul 2013 | B2 |
8514387 | Scherf et al. | Aug 2013 | B2 |
8615110 | Landes | Dec 2013 | B2 |
8625878 | Haas et al. | Jan 2014 | B2 |
8649932 | Mian et al. | Feb 2014 | B2 |
8655540 | Mian et al. | Feb 2014 | B2 |
8682077 | Longacre, Jr. | Mar 2014 | B1 |
8700924 | Mian et al. | Apr 2014 | B2 |
8711222 | Aaron et al. | Apr 2014 | B2 |
8724904 | Fujiki | May 2014 | B2 |
8806948 | Kahn et al. | Aug 2014 | B2 |
8818585 | Bartonek | Aug 2014 | B2 |
8820166 | Wigh et al. | Sep 2014 | B2 |
8868291 | Mian et al. | Oct 2014 | B2 |
8875635 | Turner | Nov 2014 | B2 |
8903574 | Cooper et al. | Dec 2014 | B2 |
8925873 | Gamache et al. | Jan 2015 | B2 |
8934007 | Snead | Jan 2015 | B2 |
8942426 | Bar-am | Jan 2015 | B2 |
8958079 | Kainer et al. | Feb 2015 | B2 |
9036025 | Haas et al. | May 2015 | B2 |
9049433 | Prince | Jun 2015 | B1 |
9050984 | Li et al. | Jun 2015 | B2 |
9121747 | Mian et al. | Sep 2015 | B2 |
9134185 | Mian et al. | Sep 2015 | B2 |
9175998 | Turner | Nov 2015 | B2 |
9177210 | King | Nov 2015 | B2 |
9187104 | Fang et al. | Nov 2015 | B2 |
9195907 | Longacre, Jr. | Nov 2015 | B1 |
9205849 | Cooper et al. | Dec 2015 | B2 |
9205850 | Shimada | Dec 2015 | B2 |
9212902 | Enomoto et al. | Dec 2015 | B2 |
9222904 | Harrison | Dec 2015 | B2 |
9234786 | Groll et al. | Jan 2016 | B2 |
9255913 | Kumar et al. | Feb 2016 | B2 |
9297787 | Fisk | Mar 2016 | B2 |
9310340 | Mian et al. | Apr 2016 | B2 |
9336683 | Inomata et al. | May 2016 | B2 |
9340219 | Gamache et al. | May 2016 | B2 |
9346476 | Dargy et al. | May 2016 | B2 |
9389205 | Mian et al. | Jul 2016 | B2 |
9415784 | Bartonek et al. | Aug 2016 | B2 |
9423415 | Nanba et al. | Aug 2016 | B2 |
9429545 | Havira et al. | Aug 2016 | B2 |
9441956 | Kainer et al. | Sep 2016 | B2 |
9446776 | Cooper et al. | Sep 2016 | B2 |
9454816 | Mian et al. | Sep 2016 | B2 |
9469198 | Cooper et al. | Oct 2016 | B2 |
9518947 | Bartonek et al. | Dec 2016 | B2 |
9533698 | Warta | Jan 2017 | B2 |
9562878 | Graham et al. | Feb 2017 | B2 |
9571796 | Mian et al. | Feb 2017 | B2 |
9575007 | Rao et al. | Feb 2017 | B2 |
9580091 | Kraeling et al. | Feb 2017 | B2 |
9581998 | Cooper et al. | Feb 2017 | B2 |
9607446 | Cooper et al. | Mar 2017 | B2 |
9618335 | Mesher | Apr 2017 | B2 |
9619725 | King | Apr 2017 | B2 |
9628762 | Farritor | Apr 2017 | B2 |
9664567 | Sivathanu et al. | May 2017 | B2 |
9671358 | Cooper et al. | Jun 2017 | B2 |
9689760 | Lanza di Scalea et al. | Jun 2017 | B2 |
9714043 | Mian et al. | Jul 2017 | B2 |
9752993 | Thompson et al. | Sep 2017 | B1 |
9771090 | Warta | Sep 2017 | B2 |
9796400 | Puttagunta et al. | Oct 2017 | B2 |
9810533 | Fosburgh et al. | Nov 2017 | B2 |
9825662 | Mian et al. | Nov 2017 | B2 |
9849894 | Mesher | Dec 2017 | B2 |
9849895 | Mesher | Dec 2017 | B2 |
9860962 | Mesher | Jan 2018 | B2 |
9873442 | Mesher | Jan 2018 | B2 |
9921584 | Rao et al. | Mar 2018 | B2 |
9922416 | Mian et al. | Mar 2018 | B2 |
9950716 | English | Apr 2018 | B2 |
9950720 | Mesher | Apr 2018 | B2 |
9981671 | Fraser et al. | May 2018 | B2 |
9981675 | Cooper et al. | May 2018 | B2 |
9983593 | Cooper et al. | May 2018 | B2 |
9989498 | Lanza di Scalea et al. | Jun 2018 | B2 |
10035498 | Richardson et al. | Jul 2018 | B2 |
10040463 | Singh | Aug 2018 | B2 |
10043154 | King | Aug 2018 | B2 |
10077061 | Schmidt et al. | Sep 2018 | B2 |
10081376 | Singh | Sep 2018 | B2 |
10086857 | Puttagunta et al. | Oct 2018 | B2 |
10167003 | Bilodeau | Jan 2019 | B1 |
10286877 | Lopez Galera et al. | May 2019 | B2 |
10322734 | Mesher | Jun 2019 | B2 |
10349491 | Mesher | Jul 2019 | B2 |
10352831 | Kondo et al. | Jul 2019 | B2 |
10362293 | Mesher | Jul 2019 | B2 |
10384697 | Mesher | Aug 2019 | B2 |
10401500 | Yang et al. | Sep 2019 | B2 |
10408606 | Raab | Sep 2019 | B1 |
10414416 | Hampapur | Sep 2019 | B2 |
10502831 | Eichenholz | Dec 2019 | B2 |
10518791 | Singh | Dec 2019 | B2 |
10582187 | Mesher | Mar 2020 | B2 |
10611389 | Khosla | Apr 2020 | B2 |
10613550 | Khosla | Apr 2020 | B2 |
10616556 | Mesher | Apr 2020 | B2 |
10616557 | Mesher | Apr 2020 | B2 |
10616558 | Mesher | Apr 2020 | B2 |
10618537 | Khosla | Apr 2020 | B2 |
10625760 | Mesher | Apr 2020 | B2 |
10730538 | Mesher | Aug 2020 | B2 |
10796192 | Fernandez | Oct 2020 | B2 |
10816347 | Wygant et al. | Oct 2020 | B2 |
10822008 | Wade | Nov 2020 | B2 |
20010045495 | Olson et al. | Nov 2001 | A1 |
20020065610 | Clark et al. | May 2002 | A1 |
20020070283 | Young | Jun 2002 | A1 |
20020093487 | Rosenberg | Jul 2002 | A1 |
20020099507 | Clark et al. | Jul 2002 | A1 |
20020150278 | Wustefeld | Oct 2002 | A1 |
20020196456 | Komiya et al. | Dec 2002 | A1 |
20030059087 | Waslowski et al. | Mar 2003 | A1 |
20030062414 | Tsikos et al. | Apr 2003 | A1 |
20030072001 | Mian et al. | Apr 2003 | A1 |
20030075675 | Braune et al. | Apr 2003 | A1 |
20030140509 | Casagrande | Jul 2003 | A1 |
20030160193 | Sanchez Reveulta et al. | Aug 2003 | A1 |
20030164053 | Ignagni | Sep 2003 | A1 |
20040021858 | Shima et al. | Feb 2004 | A1 |
20040084069 | Woodard | May 2004 | A1 |
20040088891 | Theurer | May 2004 | A1 |
20040122569 | Bidaud | Jun 2004 | A1 |
20040189452 | Li | Sep 2004 | A1 |
20040247157 | Lages | Dec 2004 | A1 |
20040263624 | Nejikovsky | Dec 2004 | A1 |
20050121539 | Takada et al. | Jun 2005 | A1 |
20050279240 | Pedanekar et al. | Dec 2005 | A1 |
20060017911 | Villar | Jan 2006 | A1 |
20060098843 | Chew | May 2006 | A1 |
20060171704 | Bingle | Aug 2006 | A1 |
20060231685 | Mace et al. | Oct 2006 | A1 |
20070136029 | Selig et al. | Jun 2007 | A1 |
20070150130 | Welles | Jun 2007 | A1 |
20070211145 | Kilian et al. | Sep 2007 | A1 |
20070265780 | Kesler et al. | Nov 2007 | A1 |
20070289478 | Becker et al. | Dec 2007 | A1 |
20080007724 | Chung | Jan 2008 | A1 |
20080177507 | Mian et al. | Jul 2008 | A1 |
20080212106 | Hoffmann | Sep 2008 | A1 |
20080298674 | Baker | Dec 2008 | A1 |
20080304065 | Hesser | Dec 2008 | A1 |
20080304083 | Farritor et al. | Dec 2008 | A1 |
20090040503 | Kilian | Feb 2009 | A1 |
20090073428 | Magnus | Mar 2009 | A1 |
20090196486 | Distante et al. | Aug 2009 | A1 |
20090250533 | Akiyama et al. | Oct 2009 | A1 |
20090273788 | Nagle et al. | Nov 2009 | A1 |
20090319197 | Villar et al. | Dec 2009 | A1 |
20100007551 | Pagliuco | Jan 2010 | A1 |
20100026551 | Szwilski | Feb 2010 | A1 |
20100106309 | Grohman et al. | Apr 2010 | A1 |
20100289891 | Akiyama | Nov 2010 | A1 |
20110064273 | Zarembski et al. | Mar 2011 | A1 |
20110209549 | Kahn | Sep 2011 | A1 |
20120026352 | Natroshvilli et al. | Feb 2012 | A1 |
20120051643 | Ha et al. | Mar 2012 | A1 |
20120062731 | Enomoto et al. | Mar 2012 | A1 |
20120192756 | Miller et al. | Aug 2012 | A1 |
20120218868 | Kahn et al. | Aug 2012 | A1 |
20120245908 | Berggren | Sep 2012 | A1 |
20120263342 | Haas | Oct 2012 | A1 |
20120300060 | Farritor | Nov 2012 | A1 |
20130070083 | Snead | Mar 2013 | A1 |
20130092758 | Tanaka et al. | Apr 2013 | A1 |
20130096739 | Landes et al. | Apr 2013 | A1 |
20130170709 | Distante et al. | Jul 2013 | A1 |
20130191070 | Kainer | Jul 2013 | A1 |
20130230212 | Landes | Sep 2013 | A1 |
20130313372 | Gamache et al. | Nov 2013 | A1 |
20130317676 | Cooper et al. | Nov 2013 | A1 |
20140069193 | Graham et al. | Mar 2014 | A1 |
20140129154 | Cooper | May 2014 | A1 |
20140142868 | Bidaud | May 2014 | A1 |
20140151512 | Cooper | Jun 2014 | A1 |
20140177656 | Mian et al. | Jun 2014 | A1 |
20140333771 | Mian et al. | Nov 2014 | A1 |
20140339374 | Mian et al. | Nov 2014 | A1 |
20150131108 | Kainer et al. | May 2015 | A1 |
20150219487 | Maraini | Aug 2015 | A1 |
20150225002 | Branka et al. | Aug 2015 | A1 |
20150268172 | Naithani et al. | Sep 2015 | A1 |
20150269722 | Naithani et al. | Sep 2015 | A1 |
20150284912 | Delmonic et al. | Oct 2015 | A1 |
20150285688 | Naithani et al. | Oct 2015 | A1 |
20160002865 | English et al. | Jan 2016 | A1 |
20160039439 | Fahmy et al. | Feb 2016 | A1 |
20160059623 | Kilian | Mar 2016 | A1 |
20160121912 | Puttagunta | May 2016 | A1 |
20160159381 | Fahmy | Jun 2016 | A1 |
20160207551 | Mesher | Jul 2016 | A1 |
20160209003 | Mesher | Jul 2016 | A1 |
20160212826 | Mesher | Jul 2016 | A1 |
20160221592 | Puttagunta | Aug 2016 | A1 |
20160249040 | Mesher | Aug 2016 | A1 |
20160282108 | Martinod Restrepo et al. | Sep 2016 | A1 |
20160304104 | Witte et al. | Oct 2016 | A1 |
20160305915 | Witte et al. | Oct 2016 | A1 |
20160312412 | Schrunk, III | Oct 2016 | A1 |
20160318530 | Johnson | Nov 2016 | A1 |
20160321513 | Mitti et al. | Nov 2016 | A1 |
20160325767 | LeFebvre et al. | Nov 2016 | A1 |
20160368510 | Simon et al. | Dec 2016 | A1 |
20170029001 | Berggren | Feb 2017 | A1 |
20170034892 | Mesher | Feb 2017 | A1 |
20170066459 | Singh | Mar 2017 | A1 |
20170106885 | Singh | Apr 2017 | A1 |
20170106887 | Mian et al. | Apr 2017 | A1 |
20170182980 | Davies et al. | Jun 2017 | A1 |
20170203775 | Mesher | Jul 2017 | A1 |
20170205379 | Prince et al. | Jul 2017 | A1 |
20170267264 | English et al. | Sep 2017 | A1 |
20170297536 | Giraud et al. | Oct 2017 | A1 |
20170305442 | Viviani | Oct 2017 | A1 |
20170313286 | Giraud et al. | Nov 2017 | A1 |
20170313332 | Paget et al. | Nov 2017 | A1 |
20170336293 | Kondo et al. | Nov 2017 | A1 |
20180038957 | Kawazoe et al. | Feb 2018 | A1 |
20180039842 | Schuchmann et al. | Feb 2018 | A1 |
20180057030 | Puttagunta et al. | Mar 2018 | A1 |
20180079433 | Mesher | Mar 2018 | A1 |
20180079434 | Mesher | Mar 2018 | A1 |
20180106000 | Fruehwirt | Apr 2018 | A1 |
20180120440 | O'Keeffe | May 2018 | A1 |
20180127006 | Wade | May 2018 | A1 |
20180220512 | Mesher | Aug 2018 | A1 |
20180222504 | Birch et al. | Aug 2018 | A1 |
20180276494 | Fernandez | Sep 2018 | A1 |
20180281829 | Euston et al. | Oct 2018 | A1 |
20180339720 | Singh | Nov 2018 | A1 |
20180370552 | Puttagunta et al. | Dec 2018 | A1 |
20180372875 | Juelsgaard et al. | Dec 2018 | A1 |
20190039633 | Li | Feb 2019 | A1 |
20190054937 | Graetz | Feb 2019 | A1 |
20190107607 | Danziger | Apr 2019 | A1 |
20190135315 | Dargy et al. | May 2019 | A1 |
20190156569 | Jung | May 2019 | A1 |
20190179026 | Englard et al. | Jun 2019 | A1 |
20190310470 | Weindort et al. | Oct 2019 | A1 |
20190349563 | Mesher | Nov 2019 | A1 |
20190349564 | Mesher | Nov 2019 | A1 |
20190349565 | Mesher | Nov 2019 | A1 |
20190349566 | Mesher | Nov 2019 | A1 |
20190357337 | Mesher | Nov 2019 | A1 |
20190367060 | Mesher | Dec 2019 | A1 |
20190367061 | Mesher | Dec 2019 | A1 |
20200025578 | Wygant | Jan 2020 | A1 |
20200034637 | Olson et al. | Jan 2020 | A1 |
20200086903 | Mesher | Mar 2020 | A1 |
20200116865 | Yang et al. | Apr 2020 | A1 |
20200156677 | Mesher | May 2020 | A1 |
20200160733 | Dick et al. | May 2020 | A1 |
20200164904 | Dick et al. | May 2020 | A1 |
20200180667 | Kim et al. | Jun 2020 | A1 |
20200198672 | Underwood et al. | Jun 2020 | A1 |
20200221066 | Mesher | Jul 2020 | A1 |
20200231193 | Chen et al. | Jul 2020 | A1 |
20200239049 | Dick et al. | Jul 2020 | A1 |
Number | Date | Country |
---|---|---|
2061014 | Aug 1992 | CA |
2574428 | Feb 2006 | CA |
2607634 | Apr 2008 | CA |
2574428 | Oct 2009 | CA |
2782341 | Jun 2011 | CA |
2844113 | Feb 2013 | CA |
2986580 | Sep 2014 | CA |
2867560 | Apr 2015 | CA |
2607634 | Jun 2015 | CA |
2945614 | Oct 2015 | CA |
2732971 | Jan 2016 | CA |
2996128 | Mar 2016 | CA |
2860073 | May 2016 | CA |
2867560 | Jul 2017 | CA |
106291538 | Jan 2017 | CN |
106364503 | Feb 2017 | CN |
106373191 | Feb 2017 | CN |
106384190 | Feb 2017 | CN |
104535652 | Jun 2017 | CN |
107688024 | Feb 2018 | CN |
206984011 | Feb 2018 | CN |
108009484 | May 2018 | CN |
108657222 | Oct 2018 | CN |
19831176 | Jan 2000 | DE |
19831215 | Jan 2000 | DE |
10040139 | Jul 2002 | DE |
19826422 | Sep 2002 | DE |
60015268 | Mar 2005 | DE |
19943744 | Jan 2006 | DE |
19919604 | Aug 2009 | DE |
102012207427 | Jul 2013 | DE |
102009018036 | Feb 2014 | DE |
102014119056 | Jun 2016 | DE |
0274081 | Jul 1988 | EP |
1079322 | Feb 2001 | EP |
1146353 | Oct 2001 | EP |
1158460 | Nov 2001 | EP |
1168269 | Jan 2002 | EP |
1197417 | Apr 2002 | EP |
1098803 | Jan 2003 | EP |
1600351 | Jan 2007 | EP |
1892503 | Jul 2007 | EP |
1918702 | May 2008 | EP |
1964026 | Sep 2008 | EP |
1992167 | May 2016 | EP |
3024123 | May 2016 | EP |
2806065 | Sep 2016 | EP |
3138753 | Mar 2017 | EP |
3138754 | Mar 2017 | EP |
2697738 | Aug 2017 | EP |
2697738 | Aug 2017 | EP |
2998927 | Sep 2018 | EP |
3561501 | Oct 2019 | EP |
3105599 | Apr 2020 | EP |
3433154 | Jun 2020 | EP |
3658439 | Jun 2020 | EP |
3689706 | Aug 2020 | EP |
2674809 | Oct 1992 | FR |
3049255 | Sep 2017 | FR |
3077553 | Feb 2018 | FR |
3049255 | Apr 2018 | FR |
3052416 | Jul 2019 | FR |
3077553 | Aug 2019 | FR |
2265779 | Oct 1993 | GB |
2378344 | Feb 2003 | GB |
2383635 | Jun 2005 | GB |
2536746 | Sep 2016 | GB |
2536746 | Mar 2017 | GB |
60039555 | Mar 1985 | JP |
63302314 | Dec 1988 | JP |
6011316 | Jan 1994 | JP |
06322707 | Nov 1994 | JP |
H07146131 | Jun 1995 | JP |
7280532 | Oct 1995 | JP |
H07294443 | Nov 1995 | JP |
H07294444 | Nov 1995 | JP |
10332324 | Dec 1998 | JP |
11172606 | Jun 1999 | JP |
2000221146 | Aug 2000 | JP |
2000241360 | Sep 2000 | JP |
H0924828 | Jul 2002 | JP |
2002294610 | Oct 2002 | JP |
2003074004 | Mar 2003 | JP |
2003121556 | Apr 2003 | JP |
2004132881 | Apr 2004 | JP |
2007240342 | Sep 2007 | JP |
4008082 | Nov 2007 | JP |
2010229642 | Oct 2010 | JP |
5283548 | Sep 2013 | JP |
5812595 | Nov 2015 | JP |
2015209205 | Nov 2015 | JP |
2016191264 | Nov 2016 | JP |
6068012 | Jan 2017 | JP |
2017020862 | Jan 2017 | JP |
6192717 | Sep 2017 | JP |
6327413 | May 2018 | JP |
6425990 | Nov 2018 | JP |
2019065650 | Apr 2019 | JP |
6530979 | Jun 2019 | JP |
101562635 | Oct 2015 | KR |
101706271 | Feb 2017 | KR |
1020180061929 | Jun 2018 | KR |
2142892 | Dec 1999 | RU |
101851 | Jan 2011 | RU |
1418105 | Aug 1988 | SU |
200005576 | Feb 2000 | WO |
200008459 | Feb 2000 | WO |
2000-73118 | Dec 2000 | WO |
2001066401 | Sep 2001 | WO |
2001066401 | May 2003 | WO |
2005036199 | Apr 2005 | WO |
2005036199 | Apr 2005 | WO |
2005098352 | Oct 2005 | WO |
2006008292 | Jan 2006 | WO |
2006014893 | Feb 2006 | WO |
2011002534 | Jan 2011 | WO |
2012142548 | Oct 2012 | WO |
2013146502 | Mar 2013 | WO |
2013177393 | Nov 2013 | WO |
2014017015 | Jan 2014 | WO |
2015160300 | Oct 2015 | WO |
2015165560 | Nov 2015 | WO |
2016008201 | Jan 2016 | WO |
2016027072 | Feb 2016 | WO |
2016007393 | Jul 2016 | WO |
2016168576 | Oct 2016 | WO |
2016168623 | Oct 2016 | WO |
2017159701 | Sep 2017 | WO |
2018158712 | Sep 2018 | WO |
2018207469 | Nov 2018 | WO |
2018208153 | Nov 2018 | WO |
2018210441 | Nov 2018 | WO |
2019023613 | Jan 2019 | WO |
2019023658 | Jan 2019 | WO |
2019023613 | Jan 2019 | WO |
2019023658 | Jan 2019 | WO |
2019086158 | May 2019 | WO |
2019212693 | Nov 2019 | WO |
2020078703 | Apr 2020 | WO |
Entry |
---|
US 8,548,242 B1, 10/2013, Longacre, Jr. (withdrawn) |
US Patent and Trademark Office, Non-Final Office Action for U.S. Appl. No. 14/725,490 dated Feb. 23, 2018. |
Shawn Landers et al., “Development and Calibration of a Pavement Surface Performance Measure and Prediction Models for the British Columbia Pavement Management System” (2002). |
Zheng Wu, “Hybrid Multi-Objective Optimization Models for Managing Pavement Assetts” (Jan. 25, 2008). |
“Pavement Condition Index 101”, OGRA's Milestones (Dec. 2009). |
“Rail Radar Bringing the Track Into the Office” presentation given to CN Rail Engineering on Jan. 21, 2011. |
Rail Radar, Inc. Industrial Research Assistance Program Application (IRAP) (Aug. 10, 2012). |
“Rail Radar Automated Track Assessment” paper distributed at the Association of American Railways (AAR) Transportation Test Center in Oct. 2010 by Rail Radar, Inc. |
US Patent and Trademark Office, Non-Final Office Action for U.S. Appl. No. 14/725,490 dated Mar. 30, 2017. |
US Patent and Trademark Office, Final Office Action for U.S. Appl. No. 14/725,490 dated Aug. 16, 2017. |
Kantor, et al., “Automatic Railway Classification Using Surface and Subsurface Measurements” Proceedings of the 3rd International Conference on Field and Service Robitics, pp. 43-48 (2001). |
Magnes, Daniel L., “Non-Contact Technology for Track Speed Rail Measurements (ORIAN)” SPIE vol. 2458, pp. 45-51 (1995). |
Ryabichenko, et al. “CCD Photonic System for Rail Width Measurement” SPIE vol. 3901, pp. 37-44 (1999). |
Gingras, Dennis, “Optics and Photonics Used in Road Transportation” (1998). |
Liviu Bursanescu and François Blais, “Automated Pavement Distress Data Collection and Analysis: a 3-D Approach” (1997). |
US Patent and Trademark Office, Non-Final Office Action for U.S. Appl. No. 14/724,925 dated Feb. 26, 2016. |
US Patent and Trademark Office, Non-Final Office Action for U.S. Appl. No. 14/724,890 dated Jul. 29, 2016. |
US Patent and Trademark Office, Final Office Action for U.S. Appl. No. 14/724,890 dated Nov. 10, 2016. |
US Patent and Trademark Office, Non-Final Office Action for U.S. Appl. No. 14/724,890 dated Mar. 24, 2017. |
Korean Intellectual Property Office, International Search Report for Int. App. No. PCT/IB2018/058574 dated Feb. 27, 2019. |
Korean Intellectual Property Office, Written Opinion of the International Searching Authority for Int. App. No. PCT/IB2018/058574 dated Feb. 27, 2019. |
US Patent and Trademark Office, Non-Final Office Action for U.S. Appl. No. 16/255,928 dated Oct. 18, 2019. |
US Patent and Trademark Office, Final Office Action for U.S. Appl. No. 16/127,956 dated Jul. 9, 2019. |
International Report on Patentability, PCT App. Ser. No. PCT/US2020/033374 dated Aug. 14, 2020. |
US Patent and Trademark Office, Non-Final Office Action for U.S. Appl. No. 16/127,956 dated Dec. 31, 2018. |
D.D. Davis et al., “Tie Performance—A Progress Report of the Des Plaines Test Site,” Report No. R-746, Association of American Railroads Research and Test Department (Apr. 1990). |
Mattias Johanneson, “Architectures for Sheet-of-Light Range Imaging,” Report No. LiTH-ISY-I-1335, Image Processing Group, Department of Electrical Engineering, Linköping University (Feb. 27, 1992). |
U.S. Appl. No. 60/584,769, “System & Method for Inspecting Railroad Track” by John Nagle & Steven C. Orrell. |
Mattias Johannesson, “Sheet-of-light Range Imaging,” Linköping Studies in Science and Technology. Dissertations No. 399 (1995). |
M. Johannesson, SIMD Architectures for Range and Radar Imaging, PhD thesis, University of Linköping (1995). |
Erik Åstrand, “Automatic Inspection of Sawn Wood,” Linköping Studies in Science and Technology. Dissertations. No. 424 (1996). |
Mattias Johannesson, “Sheet-of-Light range imaging experiments with MAPP2200,” Report No. LiTH-ISY-I-1401, Image Processing Group, Department of Electrical Engineering, Linköping University (Sep. 28, 1992). |
M. de Bakker et al., “A Smart Range Image Sensor,” Proceedings of the 24th European Solid-State Circuits Conference (1998):208-11;xii+514. |
Dr. Mats Gokstorp et al., “Smart Vision Sensors,” International Conference on Image Processing (Oct. 4-7, 1998), Institute of Electrical and Electronics Engineers, Inc. |
Mattias Johanneson, et al., “An Image Sensor for Sheet-of-Light Range Imaging,” IAPR Workshop on Machine Vision Applications (Dec. 7-9, 1992). |
Mattias Johannesson, “Can Sorting using sheet-of-light range imaging and MAPP2200,” Institute of Electrical and Electronics Engineers; International Conference on Systems, Man and Cybernetics (Oct. 17-20, 1993). |
Michiel de Bakker, et al., “Smart PSD array for sheet-of-light range imaging,” The International Society for Optical Engineering. Sensors and Camera Systems for Scientific, Industrial, and Digital Photography Applications (Jan. 24-26, 2000). |
Umayal Chidambaram, “Edge Extraction of Color and Range Images,” (Dec. 2003). |
Franz Pernkopf et al., “Detection of surface defects on raw milled steel blocks using range imaging” The International Society for Optical Engineering. Machine Vision Applications in Industrial Inspection X (Jan. 21-22, 2002). |
Murhed, Anders, “IVP Integrated Vision Products,” Pulp and Paper International 44.12 (Dec. 1, 2002). |
Anders Astrand, “Smart Image Sensors,” Linköping Studies in Science and Technology. Dissertations. No. 319 (1993). |
Mattias Johannesson et al., “Five Contributions to the Art of Sheet-of-light Range Imaging on MAPP2200,” Report No. LiTH-ISY-R-1611, Image Processing Group, Department of Electrical Engineering, Linköping University (Apr. 14, 1994). |
Federal Register, vol. 73 (70695-70696). |
Newman et al., “A Survey of Automated Visual Inspection,” Computer Vision an Image Understanding vol. 61, No. 2, March, pp. 231-262, 1995. |
J. Velten et al., “Application of a Brightness-Adapted Edge Detector for Real-Time Railroad Tie Detection in Video Images,” Institute of Electrical and Electronics Engineers (1999). |
R. Gordon Kennedy, “Problems of Cartographic Design in Geographic Information Systems for Transportation,” Cartographic Perspectives (Jul. 20, 1999). |
Richard Reiff, “An Evaluation of Remediation Techniques for Concrete Tie Rail Seat Abrasion in the Fast Environment,” American Railway Engineering Association, Bulletin 753 (1995). |
Russell H. Lutch et al., “Causes and Preventative Methods for Rail Seat Abrasion in North America's Railroads,” Conference Paper (Oct. 2014). |
Nigel Peters and Steven R. Mattson, “CN 60E Concrete Tie Development,” AREMA: 25 (2003). |
Federal Register, vol. 76, No. 175, pp. 55819-55825. |
National Transportation Safety Board, “Railroad Accident Brief” (NTSB/RAB-06/03). |
Arthur L. Clouse et al. “Track Inspection Into the 21st Century” (Sep. 19, 2006). |
Federal Register, vol. 76, No. 63, pp. 18001-18346 (18073). |
Railroad Safety Advisory Committee (RSAC), Minutes of Meeting, Dec. 10, 2008, Washington, D.C. |
Dennis P. Curtin, “An Extension to the Textbook of Digital Photography, Pixels and Images” (2007). |
Holland L.P.'s Combined Motion for Early Markman Claim Construction and Summary Judgment of Non-Infringement in Georgetown Rail Equipment Company v. Holland L.P., (E.D. Tex.) (Tyler) (6:13-cv-366). |
Georgetown Rail Equipment Company's Response to Holland L.P.'s Combined Motion for Early Markman Claim Construction and Summary Judgment of Non-Infringement in Georgetown Rail Equipment Company v. Holland L.P., (E.D. Tex.) (Tyler) (6:13-cv-366). |
Georgetown Rail Equipment Company's P.R. 4-5(a) Opening Markman Claim Construction Brief in Georgetown Rail Equipment Company v. Holland L.P., (E.D. Tex.) (Tyler) (6:13-cv-366). |
Holland L.P.'s Responsive Markman Claim Construction Brief Under P.R. 4-5 in Georgetown Rail Equipment Company v. Holland L.P., (E.D. Tex.) (Tyler) (6:13-cv-366). |
Claim Construction Memorandum Opinion and Order in Georgetown Rail Equipment Company v. Holland L.P., (E.D. Tex) (Tyler) (6:13-cv-366). |
Public Judgment and Reasons in Georgetown Rail Equipment Company v. Rail Radar Inc. and Tetra Tech EBA Inc. (T-896-15) (2018 FC 70). |
US Patent and Trademark Office, Final Office Action for U.S. Appl. No. 16/255,928 dated Apr. 27, 2020. |
US Patent and Trademark Office, Non-Final Office Action for U.S. Appl. No. 16/742,057 dated May 26, 2020. |
Invitation to Pay Additional Fees, PCT App. Ser. No. PCT/US2020/033449 dated Jul. 9, 2020. |
International Report on Patentability, PCT App. Ser. No. PCT/IB2018/058574 dated Aug. 6, 2020. |
T. Kanade, ed., Three-Dimensional Machine Vision, Kluwer Academic Publishers (1987) [Part 1]. |
T. Kanade, ed., Three-Dimensional Machine Vision, Kluwer Academic Publishers (1987) [Part 2]. |
D.D. Davis et al., “Tie Condition Inspection a Case Study of Tie Failure Rate, Mods, and Clustering,” Report No. R-714, Association of American Railroads Research and Test Department (Jul. 1989). |
John Choros et al., “Prevention of Derailments due to Concrete Tie Rail Seat Deterioration,” Proceedings of ASME/IEEE Joint Rail Conference & Internal Combustion Engine Spring Technical Conference. No. 40096 (2007). |
International Search Report and Written Opinion of the International Searching Authority, PCT App. Ser. No. PCT/US2020/033449 dated Sep. 14, 2020 (including copy of Kovalev et al. “Freight car models and their computer-aided dynamic analysis”, Multibody System Dynamics, Nov. 2009). |
Julio Molleda et al., “A Profile Measurement System for Rail Manufacturing using Multiple Laser Range Finders” (2015). |
Number | Date | Country | |
---|---|---|---|
20200363532 A1 | Nov 2020 | US |
Number | Date | Country | |
---|---|---|---|
63016661 | Apr 2020 | US | |
62988630 | Mar 2020 | US | |
62848630 | May 2019 | US |