Systems and methods for monitoring vehicles with tires

Information

  • Patent Grant
  • 12090795
  • Patent Number
    12,090,795
  • Date Filed
    Friday, August 30, 2019
    5 years ago
  • Date Issued
    Tuesday, September 17, 2024
    2 months ago
Abstract
A vehicle comprising wheels having tires can be monitored to obtain information regarding the vehicle, including information regarding the wheels, such as an indication of a physical state of a tire and/or other component of the wheel based on at least 3D recognition and/or 2D recognition of image data of the tire and/or other wheel component.
Description
FIELD

This disclosure relates to tires, including pneumatic and non-pneumatic tires (NPTs), for vehicles, including road vehicles and off-road vehicles and, more particularly, to monitoring wheels comprising tires of such vehicles.


BACKGROUND

Wheels for vehicles comprise tires, which may be pneumatic tires or non-pneumatic tires. Tires are subject to various forces and environments that cause them to wear and sometimes fail. Road vehicles and off-road vehicles equipped with tires are used on soft, slippery and/or irregular grounds (e.g., soil, mud, sand, ice, snow, etc.) for work and/or other purposes. In some cases, off-road vehicles may also be operable on paved roads.


Numerous factors affect performance of road vehicles and off-road vehicles, including their components (e.g., tires) and their environments (e.g., grounds on which they operate). While some of these factors may be managed by users (e.g., operator) of the vehicles, this may lead to suboptimal work (e.g., construction work), greater wear or other deterioration of components of the vehicles, and/or other issues in some cases.


Similar issues may arise because of wear or other deterioration of a wheel (e.g., the outer rim) around which a tire is disposed. For these and other reasons, there is a need to improve monitoring tires of road vehicles and off-road vehicles.


SUMMARY

In accordance with various aspects of this disclosure, a vehicle (e.g., a road vehicle or an off-road vehicle) comprising wheels can be monitored to obtain information regarding the vehicle, including information regarding a given one of the wheels, such as an indication of deterioration of a tire and/or another component of the given one of the wheels (e.g., an indication of a level of wear, a rupture like a break, a puncture, chunking, de-bonding, etc. of the tire and/or other component of that wheel), an identifier of the tire and/or another component of the given one of the wheels, and/or other parameters of the tire and/or another component of the given one of the wheels, which can be used for various purposes, such as, for example, to: convey the information to a user (e.g., an operator of the vehicle); control the vehicle (e.g., a speed of the vehicle); transmit the information to a remote party (e.g., a provider such as a manufacturer or distributor of the tire and/or another component of the given one of the wheels, and/or of the vehicle; etc.).


In accordance with an aspect, this disclosure relates to a system for monitoring a tire for traction of a vehicle. The system comprises an interface configured to receive data regarding at least one image of the tire. The system also comprises a processor configured to process the data regarding the at least one image of the tire to obtain an indication of a physical state of the tire, and to generate a signal based on the indication of the physical state of the tire.


In accordance with another aspect, this disclosure relates to a method of monitoring a tire for traction of a vehicle. The method comprises receiving data regarding at least one image of the tire. The method also comprises processing the data regarding the at least one image of the tire to obtain an indication of a physical state of the tire. The method also comprises generating a signal based on the indication of the physical state of the tire.


In accordance with yet another aspect, this disclosure relates to a wheel monitoring system. The system comprises an image data capture device configured to capture image data relating to a wheel component (e.g. tire, rim, etc.). The system also comprises an image processing device, in data communication with the image capture device. The image processing device is configured to receive captured image data from the image data capture device and process the captured image data to determine at least one physical characteristic of the wheel component.


In accordance with yet another aspect, this disclosure relates to wheel monitoring system. The system comprises a 3D scanning device configured to generate a 3D scan relating to a wheel component. The system also comprises a processing device, in data communication with the 3D scanning device. The processing device is configured to receive the 3D scan from the 3D scanning device and process the 3D scan to determine at least one physical characteristic of the wheel component.





BRIEF DESCRIPTION OF THE DRAWINGS

A detailed description of embodiments is provided below, by way of example only, with reference to accompanying drawings, in which:



FIG. 1 shows an example of an embodiment of a vehicle comprising a wheel including a pneumatic tire in accordance with an embodiment;



FIG. 2 shows a perspective view of a pneumatic tire in accordance with an embodiment;



FIG. 3 shows a cutaway view of the pneumatic tire of FIG. 2;



FIG. 4 shows a diagram of an image processing system according to one embodiment;



FIG. 5 shows a schematic network diagram of a system for monitoring road vehicles and off road vehicles, according to one embodiment;



FIG. 6 shows a diagram of a computer of a vehicle according to one embodiment;



FIG. 7 shows a schematic network diagram of a system for monitoring road vehicles and off road vehicles, according to another embodiment;



FIG. 8 A to C illustrate representations of different databases according to various embodiments;



FIGS. 9 to 15 show flowcharts of the use of a monitoring system, according to various embodiments;



FIGS. 16 to 19 show data capture for a system according to various embodiments;



FIG. 20 shows the communication of a result in a system according to one embodiment;



FIG. 21 shows scheduling of maintenance according to one embodiment;



FIG. 22 shows a system capable of facilitating the purchase of parts according to one embodiment;



FIG. 23 shows a system able schedule a maintenance request according to one embodiment;



FIG. 24 shows a system able to connect two electronic devices, according to one embodiment;



FIGS. 25 to 27 show a system that can determine that the vehicle has a critical malfunction according to various embodiments;



FIG. 28 shows an example of a camera station for inspecting tires;



FIG. 29 shows an example of a laser line scanner station for inspecting tires;



FIG. 30 shows an embodiment of a drone device for inspecting the tire;



FIG. 31 shows the drone device for inspecting the tire;



FIG. 32 shows an example of a vehicle-mounted inspection device for inspecting tires;



FIG. 33 shows an example of an embodiment of a computing apparatus;



FIG. 34 shows an example 3D tire model generated using the system described herein;



FIGS. 35 to 38 show example 3D tire models of used and/or damaged tires generated using the system described herein, overlaid by 3D tire models of unused and/or undamaged tires;



FIGS. 39 and 40 show examples of wear/damage being detected and characterized using 2D recognition techniques; and



FIGS. 41 to 43 shows flowcharts of the use of a monitoring system, according to various embodiments.





DETAILED DESCRIPTION OF EMBODIMENTS


FIG. 1 shows an example of an embodiment of a vehicle 10 comprising wheels 201-204 on a ground surface 11. Each of the wheels 201-204 comprises a tire 34 for contacting the ground surface 11.


As further discussed later, in this embodiment, the vehicle 10, including the wheels 201-204, can be monitored (e.g., during operation of the vehicle 10) to obtain information regarding the vehicle 10, including information regarding the wheels 201-204, such as an indication of deterioration of the tire 34 and/or another component of a given one of the wheels 201-204 (e.g., an indication of a level of wear, a rupture like a break, a puncture, chunking, de-bonding, etc. of the tire 34 and/or other component of that wheel), an identifier of the tire 34 and/or another component of the given one of the wheels 201-204, and/or other parameters of the tire 34 and/or another component of the given one of the wheels 201-204, which can be used for various purposes, such as, for example, to: convey the information to a user (e.g., an operator of the vehicle 10); control the vehicle 10 (e.g., a speed of the vehicle 10); transmit the information to a remote party (e.g., a provider such as a manufacturer or distributor of the tire 34 and/or another component of the given one of the wheels 201-204, and/or of the vehicle 10; etc.); etc. This may be useful, for example, to gain knowledge about the vehicle 10, including the wheels 201-204, to enhance efficiency of the vehicle 10, help prevent rapid wear or other deterioration of the wheels 201-204, facilitate maintenance (e.g., replacement or repair) of the tire 34 and/or another component of each of the wheels 201-204, and/or for various other reasons.


In this embodiment, the ground surface 11 is a road and the vehicle 10 is a road vehicle that is designed to legally carry people or cargo on the road 11, which is part of a public road infrastructure (e.g., public streets, highways, etc.). More particularly, in this embodiment, the road vehicle 10 is a truck. In this example, the vehicle 10 is a light truck for cargo transportation (e.g., having a gross vehicle weight rating (GVWR) greater than 6,001 lbs or 2,722 kg, such as in class 2 or higher according to the U.S. Department of Transportation's Federal Highway Administration (FHWA)). As will be appreciated, other examples can relate to trucks of any class or size, as well as any other vehicle requiring tires.


In addition to the wheels 201-204, in this embodiment, the vehicle 10 comprises a frame, a powertrain, a steering system, a suspension, a cabin, and a control system. The vehicle 10 has a longitudinal direction, a widthwise direction, and a heightwise direction.


The powertrain is configured to generate power for the vehicle 10, including motive power for respective ones of the wheels 201-204 to propel the vehicle 10 on the ground surface 11. To that end, the powertrain comprises a power source (e.g., a primer mover) that includes one or more motors. For example, in some embodiments, the power source may comprise an internal combustion engine, an electric motor (e.g., powered by a battery), or a combination of different types of motor (e.g., an internal combustion engine and an electric motor). The powertrain can transmit power from the power source to one or more of the wheels 201-204 in any suitable way (e.g., via a transmission, a differential, a shaft engaging (i.e., directly connecting) a motor and a given one of the wheels 201-204, etc.).


The steering system is configured to steer the vehicle 10 on the ground surface 11. In this embodiment, the steering system is configured to turn front ones of the wheels 201-204 to change their orientation relative to the frame of the vehicle 10 in order to cause the vehicle 10 to move in a desired direction.


The suspension is connected between the frame and the wheels 201-204 to allow relative motion between the frame and the wheels 201-204 as the vehicle 10 travels on the ground surface 11. For example, the suspension may enhance handling of the vehicle 10 on the ground surface 11 by absorbing shocks and helping to maintain traction between the wheels 201-204 and the ground surface 11. The suspension may comprise an arrangement of springs and dampers. A spring may be a coil spring, a leaf spring, a gas spring (e.g., an air spring), or any other elastic object used to store mechanical energy. A damper (also sometimes referred to as a “shock absorber”) may be a fluidic damper (e.g., a pneumatic damper, a hydraulic damper, etc.), a magnetic damper, or any other object which absorbs or dissipates kinetic energy to decrease oscillations. In some cases, a single device may itself constitute both a spring and a damper (e.g., a hydropneumatic device).


The cabin is configured to be occupied by one or more occupants of the vehicle 10. In this embodiment, the cabin comprises a user interface configured to interact with one or more occupants of the vehicle 10, including, in this example, the operator (e.g., a driver) of the vehicle 10. The user interface comprises an input portion including one or more input devices (e.g., a set of buttons, levers, dials, etc., a touchscreen, a microphone, etc.) allowing an occupant of the vehicle 10 to input commands and/or other information into the vehicle 10 and an output portion including one or more output devices (e.g., a display, a speaker, etc.) to provide information to an occupant of the vehicle 10. The output portion of the user interface which may comprise an instrument panel (e.g., a dashboard) which provides indicators (e.g., a speedometer indicator, a tachometer indicator, etc.) related to operation of the vehicle 10.


The wheels 201-204 engage the ground surface 11 for traction of the vehicle 10. Each wheel 20i comprises its tire 34 for contacting the ground surface 11 and a hub 32 for connecting the wheel 20i to an axle.


With additional reference to FIGS. 2 and 3, the wheel 20i has an axis of rotation 35, which defines an axial direction (also referred to as a “Y” direction) parallel to the axis of rotation 35 of the wheel 20i, a vertical direction (also referred to as a “Z” direction) that is normal to the axis of rotation 35 of the wheel 20i, and a horizontal direction (also referred to as a “X” direction) that is normal to the axis of rotation 35 of the wheel 20i and the vertical direction and can be viewed as corresponding to a heading direction of the wheel 20i. The axial direction of the wheel 20i can also be referred to as a lateral or widthwise direction of the wheel 20i, while each of the vertical direction and the horizontal direction of the wheel 20i can also be referred to as radial direction of the wheel 20i (also referred to as a “R” direction). The wheel 20i also has a circumferential direction (also referred to as a “C” direction). The wheel 20i has an outer diameter DW and a width WW. It comprises an inboard lateral side 47 for facing towards a center of the vehicle 10 in the widthwise direction of the vehicle 10 and an outboard lateral side 49 opposite its inboard lateral side 47.


Similarly, the tire 34 has an axial direction, a vertical direction and a horizontal direction that each are a radial direction, and a circumferential direction, which respectively correspond to the axial direction, the vertical direction and the horizontal direction that each are the radial direction, and the circumferential direction of the wheel 20i, has an inner diameter DTI, an outer diameter DT, and a width WT, and comprises an inboard lateral side 53 and an outboard lateral side 57, which are respectively part of the inboard lateral side 47 and the outboard lateral side 49 of the wheel 20i.


When it is in contact with the ground surface 11, the tire 34 has an area of contact with the ground surface 11, which may be referred to as a “contact patch” of the tire 34 with the ground surface 11.


In this embodiment, the tire 34 is a pneumatic tire, which comprises a body 40 to define a cavity 42 containing pressurized gas (e.g., air) to support loading on the tire 34 and allow the tire 34 to be resiliently deformable (i.e., changeable in configuration) as it contacts the ground surface 11. The tire 34 is configured to be mounted to a rim 44 of the hub 32 to form the cavity 42 containing the pressurized gas. Inflation pressure of the tire 34 is suitable for use of the vehicle 10.


More particularly, in this embodiment, the tire 34 comprises a tread 50, a shoulder 52, a sidewall 54, and a bead 56. The tread 50 is configured to contact the ground surface 11 and enhance traction. The tread 50 may comprise a plurality of tread recesses 231-23R and a plurality of tread projections 271-27P such that each of the tread recesses 231-23R is disposed between adjacent ones of the tread projections 271-27P. The tread 50 may be implemented in any suitable way in other embodiments (e.g., may have a smooth outer surface without tread recesses or projections). The bead 56 is configured to engage the rim 44. The sidewall 54 extends between the tread 50 and the bead 56 and contains the pressurized gas within the cavity 42. The shoulder 52 is a transition between the tread 50 and the sidewall 54.


The tire 34 comprises elastomeric material 45 to allow the tire 34 to be resiliently deformable. The elastomeric material 45 can include any polymeric material with suitable elasticity. In this embodiment, the elastomeric material 45 includes rubber. Various rubber compounds may be used and, in some cases, different rubber compounds may be present in different areas of the tire 34. In other embodiments, the elastomeric material 45 may include another elastomer in addition to or instead of rubber (e.g., polyurethane elastomer).


Also, the tire 34 comprises reinforcement 40 disposed within (e.g., embedded in) the elastomeric material 45 to reinforce the tire 34. In this embodiment, the reinforcement 40 comprises a plurality of reinforcing members 461-46R each of which can be stiffer and stronger than the elastomeric material 45 to reinforce the tire 34 in one or more directions. For example, a given one of the reinforcement members 461-46R may be metallic in that it is at least mainly (i.e., mainly or entirely) made of metal. As another example, a given one of the reinforcing members 461-46R may be polymeric but non-elastomeric in that it is at least mainly made of polymeric but non-elastomeric material (e.g., nylon, polyester, aramid, etc.).


More particularly, in this embodiment, each of the reinforcing members 461, 462 is a belt running in the circumferential direction of the tire 34. In this example, each of the belts 461, 462 comprises a layer of reinforcing cables 371-37M that extend generally parallel to one another. In this example, the reinforcing cables 371-37M of the belt 462 extend diagonally to the circumferential direction tire, and in the general direction of outboard lateral side 49 of the tire 34 to reinforce the tire 34 in that direction, whereas the reinforcing cables 371-37M of the belt 461 extend diagonally to the circumferential direction tire, and in the general direction of the inboard lateral side 47 of tire 34 to reinforce the tire 34 in that direction. In other examples, the reinforcing cables 371-37M of the belt 461 extend in the circumferential direction of the tire 34 to reinforce the tire 34 in that direction, whereas the reinforcing cables 371-37M of the belt 462 extend transversally to the circumferential direction of the tire 34 to reinforce the tire 34 in that direction. In this embodiment, each of the reinforcing cables 371-37M of the each of belts 461, 462 is a cord including a plurality of strands (e.g., metallic fibers or wires). Specifically, in this embodiment, each of the reinforcing members 461, 462 is a metallic (e.g., steel) belt in which the reinforcing cables 371-37M are metallic.


In some embodiments, the belts 461, 462 are separated by belt edge wedges 551 and 552 extending circumferentially around the tire 34 between the edges of the belts 461, 462. The belt edge wedges 551 and 552 are configured to suppress the formation of cracks at the edges of the belts 461, 462.


Also, in this embodiment, each of the reinforcing members 463, 464 is a layer of reinforcing fabric. Each of the layers of reinforcing fabric 463, 464 comprises thin pliable material made usually by weaving, felting, knitting, interlacing, or otherwise crossing natural or synthetic elongated fabric elements, such as fibers, filaments, strands and/or others, such that some elongated fabric elements extend transversally others. For instance, each of the layers of reinforcing fabric 463, 464 may comprise a ply of reinforcing woven fibers (e.g., nylon, polyester, aramid, and/or other synthetic fibers).



FIG. 4 shows a schematic block diagram of an image processing system 500 for use with a system 100 for monitoring road vehicles and off-road vehicles such as one or more vehicles like the vehicle 10. In some embodiments, one or more images captured by an electronic device 501 can be processed using the image processing system 500. For example, in some embodiments, the electronic device 501 may transmit image information relating to a tire 34 of a vehicle, such as the tires 341 to 344 of the vehicle 10, through a communication network 502, to an image processing entity 505 over a communication link, which may be implemented over a cellular network, a WiFi network or other wireless LAN, a WiMAX network or other wireless WAN, etc.


In some examples, the electronic device 501 can be a smartphone, a tablet, a smartwatch, a computer, etc., of a user, who may be the operator of the vehicle or another person having access to the vehicle. In other examples, the electronic device 501 may be integrated with the vehicle.


In some embodiments, the image processing entity 505 can be an application running on a server. In other embodiments, the image processing entity 505 can be a dedicated network appliance. In yet other embodiments, the image processing entity 505 may be an application running on the electronic device 501. In the embodiment of FIG. 4, the image processing entity 505 comprises a memory 506 for storing image information and instructions for processing images, a processor 507 implementing a plurality of computing modules 508x (for example, Artificial Intelligence, or “AI”, modules) for performing image recognition, pattern recognition and 3D model matching in order to assess a level and nature of wear, degradation and/or other deterioration of the tire 34. In some embodiments, the computing modules 508x can be implemented using a processor 507. In some embodiments, the computing modules 508x may be implemented by way of an Application Program Interface (API) that results in the computing modules 508x being implemented on a separate device or system.


Computing modules 508x may for example be implemented using known computer vision products, such as, AutoML Vision™ and/or Vision API™, each provided by Google™. In other embodiments, computing modules 508x may comprise standalone AI or machine-learning solutions forming part of image processing entity 505. As herein defined, AI refers to some implementation of artificial intelligence and/or machine learning (e.g., heuristics, support vector machines, artificial neural networks, convolutional neural networks, any types of deep neural networks, etc.) in software, hardware or some combination of both.


In some embodiments, complex algorithms, like artificial intelligence, are used to categorize what may be considered uncategorizable data. For example, the system 100 can be configured for generating conclusions about a physical state of a tire based on one or more images of the tire itself. This analysis can include whether or not there is a defect in the tire, according to some embodiments. In some embodiments, this can include indications as to the physical state of the tire and/or useful life remaining. As will be described below, a machine learning algorithm may be trained to identify a defect or other characteristic in a tire by way of image analysis.


In some embodiments, computing modules 508x are first taught how to identify parameters in a training mode (sometimes referred to as supervised learning mode). This is done by analyzing a given set of values, making quantitative comparisons, and cross-referencing conclusions with known results. Iterative refinement of these analyses and comparisons allows an algorithm to achieve greater predictive certainty. This process is continued iteratively until the solution converges or reaches a desired accuracy.


In this embodiment, computing modules 508x can compare image data for a given tire to a previously-analyzed mass of known data. When placed in a supervised learning mode, information can be generated from already populated tire data provided to the computing modules 508x. For example, this data could contain images of tires, along with determinations of the remaining life of the tires. In other words, in the supervised learning mode, both the inputs and the outputs are provided to the system 100. The system 100 can process the given inputs and compare the calculated outputs according to its algorithm to the provided outputs. Based on this comparison, the system 100 can determine a metric to represent the percentage of error between calculated and provided outputs. Using this error metric, the system 100 can adjust its method of calculating an output. During training, the system 100 can continuously repeat analysis of different inputs and provided outputs in order to fine-tune its method of determining tire information.


In some embodiments, while the computing modules 508x may require initial supervised learning, as the computing modules 508x continue to gain access to data, they may be able to further refine their predictive analytics based on new inputs. For example, if a user is able to confirm that an assessment (e.g. wheel imbalance) or prediction (e.g. 6 months of use left in a given tire) made by the system 100 is/was incorrect, the user can upload to the system 100 what the correct conclusion/prediction was. This allows the computing modules 508x to continue to improve accuracy in their analysis.


In some embodiments, multiple computing modules 508x can be configured to determine different characteristics of a given tire. Each of these modules can offer a different analysis for a given input. The processor may direct these modules to be used independently or concurrently based on an operational parameter determined by a given user. For example, the system 100 may use a different analytical technique to determine tires life compared to wheel imbalance. Based on an image communicated to the system 100 from an electronic device, the system 100 may analyze a given for tires life, wheel imbalance, or other forms of wear and/or damage.


In some embodiments, the computing modules 508x are configured to assess a level of wear, damage and/or other deterioration of the tire 34. For example, a computing module 508x can be configured to determine that the tread projections 271-27P are worn to 30% of the level of wear that would require replacement of the tires. In some embodiments, the computing modules 508x are configured to assess the nature of damage to the tire 34. For example, a computing module 508x can be configured to determine that any other component of the given one of wheels 201-204 is damaged or missing.


In some embodiments, the computing modules 508x are further configured to predict the cause of the wear and/or damage to the tire 34. In one specific example, a computing module 5081 is configured to predict whether a specific wear pattern of the elastomeric material of a tire 34i is caused by imbalanced wheel. In another specific example, a computing module 5082 is configured to predict whether a specific wear pattern of the elastomeric material of a tread projections 271-27P is caused by incorrectly installed tire. In another specific example, another computing module 5083 is configured to predict whether a specific wear pattern of the tires is caused by a categorized or uncategorized event. As will be appreciated, each computing module 508x can be implemented using a combination of deep learning, supervised or unsupervised machine learning, image recognition and/or machine vision.


In some embodiments, the system 100 is configured to capture one or more 2D images to detect specific patterns of wear and/or damage. For example, the system 100 may be configured to implement one or more computer vision (CV) models to detect specific visible wear/damage features. Examples of such visible wear/damage features include, but are not limited to, the presents of nails, or other road debris capable of piercing tires, as well as cracks in tire sidewalls.


In some embodiments, the image processing system 500 may produce a three-dimensional (3D) scan to generate a 3D model of at least part of the tire 34. For example, in some embodiments, the image data received by the electronic device 501 or any other image capture means are processed by way of photogrammetry in order to create the 3D model of the tire 34 and/or another component of a given one of the wheels 201-204. In some embodiments, as described in more detail below, laser line scanners are instead used to generate the 3D model of the tire 34 and/or another component of a given one of the wheels 201-204.


Such precise 3D models can be compared to 3D models of unworn and/or undamaged tires in order to precisely measure wear, damage and/or other deterioration. For example, by comparing the 3D model of a worn tire 34i to the 3D model of a new, unworn tire, it is possible to precisely measure a volumetric loss of material of the worn tire 34i and thereby assess the wear and/or other deterioration of the worn tire 34i, very precisely.


With reference to FIG. 34, in some embodiments, the system 100 may generate a 3D model 55 of a tire 34i, using any of the above methods, or a combination thereof. In some embodiments, the system 100 can then be superimposed onto an image of the tire 34i captured by electronic device 501. Such superimposition may be achieved using known augmented reality (AR) techniques and processes.


As described above, in some embodiments, the system 100 can implement a 2D recognition technique. In some embodiments, the system 100 can implement a 3D recognition technique. In some embodiments, the system 100 can implement a combination of a 2D recognition technique and a 3D recognition technique.


In some embodiments, the 3D recognition technique used is based on generating a 3D model using a point cloud. For example, as shown in FIG. 41, method 4000 can be used to identify tire wear/damage and/or the extent thereof. At step 4001, a plurality of images of the tire can be acquired using the electronic device 501, before sending the images to the image processing entity 505 at step 4002. At step 4003, the system 100 generates a 3D point cloud using the plurality of images. This can be accomplished by system 100 using, for example, open source algorithms, such as those available from Point Cloud Library (PCL). Alternatively, the point cloud can be generated a third party, through use of an Application Program Interface (API) by system 100. At step 4004, the system 100 uses the generated 3D point cloud to generate a 3D model of the tire. Once generated, the 3D model is matched to known 3D models of tires in a tire database at step 4005. Once matched, at step 4006, wear, damage and/or the extent thereof can be identified by comparing the generated 3D model to the known 3D model, as described in more detail below.


2D recognition techniques include four basic steps, namely image acquisition, image processing, feature extraction and classification. Such techniques include, but are not limited to, Optical Character Recognition (OCR), feature detection, image gradient analysis, pattern recognition algorithms and feature/pattern classification algorithms.


In some embodiments, the system 100 can be configured to implement the method of FIG. 42. In particular, at step 4101, the electronic device 501 can acquire one or more images of a tire, before sending the images to the image processing entity 505 at step 4102. In some embodiments, the image processing entity 505 can perform image processing steps prior to feature extraction. For example, in some embodiments, the image processing entity 505 can be configured to perform image processing including the use of fiducial markers. Then, at step 4104, the image processing entity 505 can perform feature extraction in order to detect and isolate various portions or shapes (features) of the image or images. Feature extraction can include, but is not limited to, edge detection, corner detection, blob detection, ridge detection, scale-invariant feature transform, thresholding, blob extraction, Haar-like feature extraction, template matching, Hough transforms and generalized Hough transforms.


Then at step 4105, the system 100 can perform feature classification. In some embodiments, feature classification can include, but is not limited to, the use of nearest neighbor classification, cascading classifiers, neural networks, statistical classification techniques and/or Bayesian classification techniques. Once the features have been classified, it is possible to separate, at step 4106, features which represent undamaged/unused parts of the tire, and features (e.g. cracks, exposed cables, etc.) which represent patterns of wear or damage. Once features relating to patterns of wear or damage have been detected, it is possible for the system 100 to perform further feature classification on the wear or damage pattern.


As shown in FIGS. 39 and 40, in some embodiments, system 100 is configured to use the system of FIG. 42 in order to detect damage or wear patterns in tires. For example, as shown in FIG. 39, system 100 can be configured to detect partially embedded nails 55A using the 2D analysis method described with reference to FIG. 42. Similarly, as shown in FIG. 40, system 100 can be configured to recognize narrow (though potentially deep) cracks 55B in sidewalls of a tire. As will be appreciated by the skilled reader, such patterns are difficult to detect using volumetric analysis alone. As such, the 3D recognition techniques of the present disclosure can be combined with any of the 2D recognition techniques in order to facilitate tire matching, as well as wear and/or damage recognition and characterization. Moreover, the 2D recognition techniques of the present disclosure can be used on images generated by the system 100 of various views of the 3D model generated using the 3D recognition techniques of the present disclosure.


As shown in the method of FIG. 43, once a plurality of images are acquired at step 4201, the system 100 can sequentially use 3D recognition at step 4202 and then 2D recognition at step 4203 in order to detect patterns of wear and/or damage on a tire. In some embodiments, 2D recognition may be performed before 3D recognition.


Advantageously however, 3D recognition is performed first, as in such an arrangement, the system 100 may be configured to superimpose 2D features onto 3D models, thereby allowing a more precise classification of the type of wear and/or damage.


As shown in FIGS. 35 to 38, in some embodiments, the system 100 is configured to generate a 3D model 55 of a used and/or damaged tire and compare it to a 3D model 77 of an unused and undamaged tire. The 3D model 77 of an unused and undamaged tire may be generated by the system 100 based on a previously-scanned tire, may be acquired by the system 100 from a database of 3D models of tires, or may be acquired by the system in any other suitable way. Once the 3D model 77 of an unused and undamaged tire is acquired or generated by the system 100, it can be compared to the 3D model 55 of a used and/or damaged tire generated by the system 100 using various volumetric comparison techniques. For example, the system 100 may compare the models by calculating the amount of missing material of a given tire feature (e.g. tread projections 271-27P). For example, volumetric comparison of the 3D model 55 of a used and/or damaged tire and a 3D model 77 of an unused and undamaged tire can establish that a given tread projection 27x has been worn to 78% of its original volume.


In some embodiment, the cause and/or nature of the wear and/or damage of the tire 34, can be established by the system 100 performing a volumetric comparison of the 3D model 55 of a used and/or damaged tire and a 3D model 77 of an unused and undamaged tire.


For example, as shown in FIG. 35, based on a comparison of the 3D model 55 of a damaged tire and a 3D model 77 of an undamaged tire, the system 100 can determine a pattern of damage that is indicative of the cause and/or nature of the damage. In particular, the chunking detected by the system 100 in the tread projections 271-27P of FIG. 35 is typically caused by using the tire on abrasive ground and/or on ground containing oil or some other form of contaminant.


As shown in FIG. 36, based on a comparison of the 3D model 55 of a used tire and a 3D model 77 of an unused tire, the system 100 can determine a pattern of tread wear that is indicative of the cause and/or nature of the tread wear. In particular, the uneven tread wear detected by the system 100 in the tread projections 271-27P of FIG. 36 is typically caused by an abnormal camber angle and/or a misaligned axle.


As shown in FIGS. 37 and 38, based on a comparison of the 3D model 55 of a damaged tire and a 3D model 77 of an undamaged tire, the system 100 can determine a pattern of damage that is indicative of the cause and/or nature of the damage. In particular, the radial cracks detected by the system 100 between the tread projections 271-27P and the sidewall 54 of FIG. 37 is typically caused by inadequate installation of the tire on the wheel and/or overloading of the vehicle 10. The concentric crack damage detected by the system 100 on the side wall of the tire above the rim of FIG. 38 is typically caused by an overloading of the vehicle 10.


As described above, and as shown in FIG. 39, system 100 can use the 2D recognition technique described above to recognize and characterize the presence of partially embedded nails 55A. Also, as shown in FIG. 40, the system 100 can use the 2D recognition technique described above to recognize and characterize the presence of cracks 55B in the sidewall of a tire.


Once the computing modules 508, has determined the cause, level and/or nature of the wear and/or damage of the tire 34, the image processing entity 505 may send data relating to the cause, level and/or nature of the wear and/or damage of the tire 34 back to electronic device 501 for further processing and/or notification to a user. By using this information, electronic device 501 may determine that an event arising from usage of the tire 34, such as a usage threshold event (e.g. an amount of tread wear, an amount of time such as a number of hours the tire 34 has been used), wear threshold event (e.g. the number of exposed radial cracks between the projection treads 271-27P and the sidewall 54) and/or damage event (e.g. minor delamination damage and/or “chunking” damage), has occurred.


According to some embodiments, the computing modules 508x may have access to information stored elsewhere on the internet. For example, the computing modules 508x may be configured to query databases stored on external servers by sending requests over the network in order to analyze the image based on pertinent cross-referential data. This may include weather, humidity, or information about the vehicle or tires that can be periodically updated.



FIG. 5 illustrates a schematic network diagram of a system 100 for monitoring vehicles such as one or more road vehicles and off-road vehicles like the vehicle 10, according to one embodiment. In the embodiment of FIG. 5, the system 100 includes an electronic device 501, a network 124, and a system server 1142 that can implement the image processing entity 505 of FIG. 4. The server includes a memory 1146, processor 1144, and network interface 1148.


The electronic device 501 may include elements such as a processor, a memory, a display, a data input module, and a network interface. The electronic device 501 may include other components, but these have been omitted for the sake of brevity. In operation, the electronic device 501 is configured to perform the operations described herein. The electronic device 501 processor may be configured to execute instructions stored in memory. The instructions, when executed, cause the electronic device 501 to perform the operations described herein. In some embodiments, the instructions may be part of a software application downloaded into memory by the electronic device 501. Alternatively, some or all of the functionality described herein may be implemented using dedicated circuitry, such as an ASIC, a GPU, or a programmed FPGA for performing the operations of the processor.


In some embodiments, an application (“app”, i.e., software) may be installed on the electronic device 501 to interact with the system server 1142 and or the vehicle 10. For example, in some embodiments, such as where the electronic device 501 is a smartphone, a tablet, a computer, etc., the user (e.g., the operator) may download the app from a repository (e.g., Apple's App Store, iTunes, Google Play, Android Market, etc.) or any other website onto the electronic device 501. Upon activation of the app on the electronic device 501, the user may access certain features relating to the system server 1142 and/or the vehicle 10 locally on the electronic device 501.


In operation, a user can use the electronic device 501 to generate data about the vehicle 10. For example, for embodiments where the electronic device is a smart phone equipped with a camera, the user can take one or more images of a tire 34 of the vehicle 10. The system 100 may then take the image data captured by the electronic device 501 and transmit the image data over a network 124 to a system server 1142.


According to some embodiments, the electronic device 501 may be a portable electronic device with multiple uses such as a mobile phone, tablet or laptop. According to other embodiments, the electronic device may be a single-use electronic device, such that the device is designed to only be used in operation with the system 100. Further, the electronic device 501 may also be capable of establishing a communicable link with an accessory device. This communicable link be may be wireless, wired, or partly wireless and partly wired (e.g., Bluetooth or other short-range or near-field wireless connection, WiFi or other wireless LAN, WiMAX or other wireless WAN, cellular, Universal Serial Bus (USB), etc.).


According to other embodiments, the electronic device 501 may integrated into an internal computer 1342 in the vehicle (as shown in FIG. 6). The internal computer 1342 may have a vehicle memory 1346, processor 1344, network interface 1348, and internal sensor network 1350. In some embodiments, vehicle internal computer 1342 can communicate and upload images to system server 1142 independently.


The internal sensor network 1350 can include sensors to provide information about the vehicle or the tires of the vehicle. For example, this may include a camera positioned to take images of the tires. In some embodiments where the electronic device is integrated into an internal computer in the vehicle, the system 100 may be configured to continuously monitor the tires. This can be achieved by continuously capturing data, for example, images of the vehicle tires, at various intervals. The electronic device 501 can then automatically upload the data over the network 124 to the system server 1142 for image processing. After processing, the image processing entity 505 can automatically communicate over the network 124 if a fault state has been determined.


The electronic device 501 can also send additional data to the image processing entity 505 over the network 124. For example, this can include (but is not limited to) GPS location, date and time, or any information from an onboard computer within the vehicle. This data can be cross-referenced and analyzed within the computing modules 508x.


For example, given GPS and date and time data, the computing modules 508x can access the specific weather and weather history for the vehicle location. In some embodiments, such information may be used in, for example, determining the end-of-life of a tire (i.e. the amount of time until a tire is expected to fail or until the likelihood of tires failure rises above a predetermined threshold).


This may be achieved by a separate electronic device 501 being communicably linked to an internal computer 1342 of a vehicle 10. The internal computer 1342 may periodically receive and record information relating to the vehicle 10 and/or the tires 341-344 determined by the internal sensor network 1350. For example, the information received from the internal sensor network 1350 may include an image taken of the tires or information about the vehicle 10, such as the speed of the vehicle 10.


According to some embodiments, the electronic device 501 may communicate a unique identifier for a specific tire under inspection. In some embodiments, the unique identifier can be a serial number of the tire. This allows the server 1146 and/or internal computer 1342 to catalog the inspection and produce a history of a given tire. According to some embodiments, the internal computer 1342 and/or the server 1146 may store data about the serial numbers of the tires installed on the vehicle 10.


According to some embodiments, the electronic device 501 may be capable of determining a serial number from a tire based on an image of the tire. This can be done by the electronic device 501 capturing an image of an embossed serial number on a surface of the tires, and using the image processing entity 505 to determine the specific characters of the serial number. This can be cross-referenced with a database stored in server memory 1146 (or otherwise accessible by system server 1142) to determine elements such as the model and date of manufacture of the tire.


Serial number analysis may be performed using AI techniques employed by the computing modules 508x, may be performed using techniques such as optical character recognition (OCR), or a combination thereof. These techniques may include preprocessing of an image in order to improve the ability to analyze the target components, such as de-skewing, layout analysis, and binarization. In some embodiments, a tire and/or another component of a given one of the wheels 201-204 (e.g., outer rim) can be identified by way of another marking or tag suitable for communicating information relating to the tires. Such markings or tags can include, but are not limited to, barcodes, Quick Response (QR) codes or other matrix barcodes and Radio Frequency Identification (RFID) tags.


Another method of tire identification that can be performed by the electronic device 501 is tire pattern recognition. The electronic device 501 may be configured to analyze the tread pattern and measure tire width to determine a number of characteristics about the tire. The electronic device 501 may then send this data and information to the system server 1142 for further data analysis to identify the type of tire. The type of tire may be a tire brand, model number, or any other suitable information capable of identifying a tire.


According to some embodiments, the vehicle may be capable of communicating all the necessary data over the network without the use of an external electronic device 501 such as a mobile phone. For example, the vehicle 10 may be equipped with a network interface capable of independently communicating with the system server 1142 over the network 124.


According to some embodiments, a system server 1142 hosts the image processing entity 505. Server processor 1144 is able to access instructions stored in the memory 1146 that can initialize the image processing entity 505. This initialization can include operational parameters that include which computing module 508x to use.


Image processing entity 505 can store instructions relating to a specific computing module 508x within the memory 1146. The processor 1144 may instruct the server to save the data received from the network via the network interface 1148 in memory 1146. The processor 1144 may analyze the data in memory 1146 and determine information about the tire 34. Based on the data analysis, the processor 1144 may send a communication to the electronic device 501 over the network 124 via the network interface 1148.



FIG. 7 illustrates a schematic network diagram of a system 100 for monitoring road vehicles and off-road vehicles, according to another embodiment. According to this embodiment, the system 100 can communicate with multiple vehicles 10A-10N. While this figure shows the vehicles 10A-10N communicating independently with the system server 1142 over the network 124, the vehicles may alternatively each be communicably linked with an electronic device 501 as described with reference to FIG. 5. According to this embodiment, the system server 1142 may communicate with an electronic device 501 located at a dispatch center 1102, a service center 1104, or a parts supplier 1107.


In operation, based on the analysis determined by the image processing entity 505, the system server 1142 may communicate with the user via an electronic device 501 or the vehicle 10, a dispatch center 1102, a service center 1104, or a parts supplier 1106. The system 100 may also communicate with any combination of these, or any other suitable device registered within the system 100. This communication can contain information such as that indicating the determination of tire wear and/or damage concluded by the image processing entity 505. Based on this information, the dispatch center 1502 or user may schedule maintenance with the service center 1104. Based on the conclusion on tire wear and/or damage (for example, that the tire needs to be replaced) and vehicle information (tire type, vehicle type) available, the system 100 can determine the amount of time required or parts available at the service center 1104 and facilitate scheduling a maintenance appointment or a shipment from the parts supplier 1106. This can be done by maintaining a database of inventory at the service center, along with a calendar.



FIG. 8 A to C illustrate representations of different databases that may be generated by the server processor 1144 based on information stored in memory 1146. The server memory 1146 can store a history of all information necessary for performance of the system 100, including a record of all inspections and conclusions made. These databases, or the information stored within them, may be accessible to users and administrators of the system 100, or to software able to interact with the system 100 through the use of an application programming interface (API).



FIG. 8A shows an example of a visual representation of a database that can be generated by the system 100 according to an embodiment directed towards a specific tire manufacturer. This includes an indication of tire model, a serial number for the tire, the date of an inspection, the type of inspection, along with the registered owner. This database representation gives the manufacturer access to all registered tires sold and registered within the system 100, and allows access to information on tire wear and damage.



FIG. 8B shows an example of a visual representation of a database that can be generated by the system 100 according to an embodiment directed towards a vehicle fleet manager. The database includes an indication of tire model, a unique identifier for the vehicle itself, the date of an inspection, tire status, and an additional field for manager notes. This database representation gives the fleet manager access to all vehicles registered within the system 100, and allows them to access a history of information on tire wear and/or damage.



FIG. 8C shows an example of a visual representation of a database that can be generated by the system 100 according to an embodiment directed towards a specific vehicle manufacturer. This includes an indication of vehicle model, tire model, a date of an inspection, tire status, and an additional field for manager notes. This database representation gives the vehicle manufacturer access to all of their vehicles registered within the system 100, and allows them to access a history of information on tire wear and/or damage.


The disclosed embodiments of database representations are structured merely by way of example for illustrative purposes, and a skilled reader would know that these visual representations can be changed to include more or less information available to the system 100.



FIG. 9 shows an example flowchart of the use of the system 100 which could be used (e.g., by the operator of the vehicle 10, in a rental market, etc.) to monitor usage of tires.


In operation, a user can use the electronic device 501 to generate image data relating to the tire 34 of the vehicle 10. According to some embodiments, the electronic device 501 may also access internal information stored on the vehicle onboard computer 1342. The electronic device 501 may then communicate both the data captured and the information retrieved by the electronic device 501 over the network 124 to the system server 1142 to be stored in memory 1146. Using both the data captured and the information retrieved the processor 1144 may determine information about the tire 34. Based on the data analysis, the processor 1144 may send a communication to the electronic device 501 over the network 124 via the network interface 1148.


At step 1301, the system 100 determines that an event arising from use of a tire 34, such as a usage threshold event (e.g. an amount of time such as a number of hours the tire 34 has been used) or a deterioration threshold event (e.g. the number of exposed radial cracks between the projection treads 271-27P and the sidewall 54) and/or deterioration event (e.g. minor delamination damage and the “chunking” damage), has occurred. As described above, the system 100 can make these determinations by analysis of the images taken by the image capture devices described above.


At step 1302, the system 100 identifies the tire for which the usage threshold event or deterioration threshold event has occurred. In some embodiments, the tire information and information relating to the usage threshold event and deterioration threshold event is conveyed to the operator of the vehicle by the system 100 in order to facilitate scheduling of tire servicing and/or other maintenance.


For purposes of this example, it is assumed that the usage threshold event or deterioration threshold event is for the tire 34.


For example, the system 100 may issue a notification conveying this information to the operator via the user interface of the operator of the vehicle 10 and/or the electronic device 501. According to embodiments wherein the electronic device 501 is a mobile phone, this could be in the form of a push notification sent to the app over the network 124. In other embodiments, the system 100 conveys the tire information and information relating to the usage threshold event and deterioration threshold event to an organization providing maintenance services. For example, the system 100 may issue a notification conveying this information to a system server 1142 associated with the organization via a network 124 (e.g. which may be implemented by the Internet, a cellular connection, and/or any other network infrastructure). Once the information is received, the organization can schedule maintenance of the vehicle at step 1303, and subsequently replace or repair the tire. Accordingly, tire maintenance operations can be initiated and scheduled without the need for input from the vehicle operator.


As shown in FIG. 10, the system 100 may allow organizations to provide tire-as-a-service type payment/usage models, in which tires are not purchased, but are rather provided as a service to vehicle operators in exchange for a subscription fee. For example, for a monthly fee, an organization may provide vehicle operators with tires, as well as usage rights to the system 100 described herein which will allow the organization to ensure that the vehicle operator is never without an operable/functional tires, regardless of how much and how (i.e. under what circumstances) the vehicle operator uses the tires.


This can lead to significant savings in term of vehicle downtime and logistics. For example, at step 1401, the system 100 determines that an event arising from usage of a tire, such as a usage threshold event (e.g. an amount of tread wear, an amount of time such as a number of hours the tire 34 has been used), deterioration threshold event (e.g. the number of exposed radial cracks between the projection treads 271-27P and the sidewall 54) and/or deterioration event (e.g. minor delamination damage and the “chunking” damage), has occurred. At step 1402, the system 100 identifies the tire for which the usage threshold event, deterioration threshold event and/or deterioration event has occurred. At step 1403, vehicle location information relating to the geographic location of the vehicle is determined. This can be achieved by any suitable means including, but not limited to, Global Positioning System (GPS) receivers. In some embodiment, the system 100 conveys the tire information, vehicle location information and information relating to the usage threshold event, deterioration threshold event and/or deterioration event to the tire-as-a-service organization.


As shown in above, the system 100 may communicate with the system server 1142 of the tire-as-a-service organization over a network 124 (e.g. which may be implemented by the Internet, a cellular connection, and/or any other network infrastructure). Then, at step 1404, the tire-as-a-service organization ships a replacement tire to a location related to the geographic location of the vehicle. For example, the tire-as-a-service location could ship the replacement tire to the nearest maintenance service dispatch location or third party maintenance organization. At step 1405, the tire-as-a-service organization can schedule a maintenance of the tires. In some embodiments, the tire-as-a-service organization schedules a third party mobile maintenance team to perform onsite maintenance based on the geographic location of the vehicle. Finally, at step 1406, the tire-as-a-service organization, or an agent thereof, replaces the tire. In some embodiments, this can be performed onsite, based at least in part on the vehicle location information received from the tire-as-a-service organization.



FIG. 11 shows an example flowchart of the use of the system 100 which could be used, for example, by a fleet manager to monitor usage of tires. In this system 100, the preferences of a given fleet manager can be included in any part purchase or system maintenance request. For example, a fleet manager may consider a specific tire to be superior to all other on the market. The fleet manager may want to only purchase that specific brand of tire. Another example of purchase preferences may include only to purchase a specific tire if the supplier inventory and price database indicates that the part is available with a discount. Further, if there is no supply of a first preferred tire in the inventory, the user may store a preference for an alternate tire to be purchased. In this embodiment, steps 1501 and 1502 are the same as those described in steps 1301, 1401, and 1302, 1402 respectively.


At step 1503, the system 100 will query the memory to determine if the specific user has a purchase preference stored in the system 100. If the system 100 has a purchase preference stored for the given user, the system 100 will order the tire for replacement based on the saved preference at step 1506. If the system 100 does not find a purchase preference for the given user, the system 100 may send a communication to the user's electronic device 501 with information indicating the part purchase options and information about the parts (for example the various options of price and part characteristics). The system 100 may also send a communication instructing the electronic device 501 to prompt the user to store a purchase preference. Based on this information, the system 100 will order the tire at step 1507.


At step 1509, the system 100 may schedule maintenance with a given service center or technician. At this step user preferences may also be considered. For example, a user may be able to store in their profile a preference for scheduling. This may include a preference for the first available time to service the vehicle. Alternatively, a fleet manager may try and coordinate scheduling of maintenance with other vehicles within a fleet. This could include wanting all vehicles to be serviced at the same time, or to stagger vehicle services. Scheduling preferences may also include a time of day preference for the user to have maintenance scheduled. Based on these preferences, the user may be automatically scheduled for maintenance.


According to other embodiments, the system 100 may prompt the user via a date and time entry interface, such as a calendar interface, on the electronic device 501 to input a date and time for maintenance. Based on this input data, the system 100 can schedule maintenance with a technician or service center.


Finally, at step 1520, the tire-as-a-service organization, or an agent thereof, replaces the tire. In some embodiments, this can be performed onsite, based at least in part on the vehicle 10 location information received from the tire-as-a-service organization.



FIG. 16 shows an example flowchart of the use of the system 100 which could be used, for example, by a fleet manager to monitor usage of tires. According to this embodiment, inventory of the tires at a given service center can be monitored. The system 100 allows organizations managing large fleets (e.g. vehicle rental companies, construction companies, forestry companies, etc.) to ensure that maintenance operations can be scheduled and carried out effectively and efficiently. For example, by monitoring the wear of tires, it is possible to more precisely predict when a tire will fail and/or when a replacement tire should be ordered and/or shipped.


Moreover, for an organization managing a fleet of vehicles, knowing which vehicles will shortly require maintenance and/or replacement parts contributes to efficient and effective deployment of vehicles and maintenance resources. For example, at step 1601, the system 100 determines that an event arising from usage of a tire 34, such as a usage threshold event (e.g. an amount of tread wear, an amount of time such as a number of hours the tire 34 has been used), deterioration threshold event (e.g. the number of exposed radial cracks between the projection treads 271-27P and the sidewall 54) and/or deterioration event (e.g. minor delamination damage and the “chunking” damage), has occurred. At step 1602, the system 100 identifies the tire for which the usage threshold event, deterioration threshold event and/or deterioration event has occurred. In some embodiments, as shown in FIG. 12, the system 100 conveys the tire information and information relating to the usage threshold event, deterioration threshold event and/or deterioration event to an automated fleet management system. The system 100 may communicate with the automated fleet management system over a network 124 (e.g. which may be implemented by the Internet, a cellular connection, and/or any other network infrastructure). At step 1603, the automated feet management system queries a tire supply database to determine whether the identified tire is available or needs to be ordered.


The tire supply database can be managed by the fleet management system, or can be managed by a third-party tire supplier. If the identified tire is available, the vehicle can be scheduled for maintenance. If, on the other hand, the tire is not available, the fleet management system can cause the tire to be ordered at step 1604, before scheduling maintenance of the vehicle at step 1605. This system may also include ordering based on stored user preference as previously described.


In some embodiments, the scheduling of the vehicle maintenance is at least in part based on the estimated delivery time for an ordered tire. In other embodiments, the dispatching of the vehicle relating to the identified tire can, at least partially, be based on a pre-scheduled maintenance. This system 100 may also include scheduling based on stored user preference as previously described. Finally, at step 1606, the maintenance operation is carried out and the tire is replaced or repaired.



FIG. 13 shows an example flowchart of the use of the system 100 which could be used, for example, by a vehicle operator to monitor usage of tires. According to this embodiment, the system 100 has determined a critical error to have taken place or imminent. In this embodiment, steps 1701 and 1702 are similar to those described in steps 1301, 1401, and 1302, 1402 respectively.


If the system 100 has determined that a critical error has taken place or is imminent, it can prompt the user to establish an audiovisual and/or textual connection with a technician at 1703. This could be achieved by using a Voice Over IP (VoIP) system, a phone call over a cellular network, or any other means of text, audio or video communication. This will allow the vehicle operator to communicate with the technician and get or receive pertinent information to vehicle maintenance. For example, the technician may instruct the user to drive the vehicle to a safe location and wait for the technician to arrive. In the case of a video call, the technician may be able to instruct the user to point the camera of the electronic device at a specific component of the vehicle 10 in order to provide the technician with more information about the vehicle status.



FIG. 14 shows an example flowchart of the use of the system 100 which could be used, for example, by a vehicle operator to monitor usage of tires. According to this embodiment, the system 100 has determined a critical status of the tire and/or another component of a given one of the wheels 201-204. In this embodiment, steps 1801 and 1802 are similar to those described in steps 1301, 1401, and 1302, 1402 respectively.


At step 1803, the system 100 alerts relevant parties of the critical status. This can include fleet managers, technicians or other operators. For example, the system 100 may send a text message, email or app push notification to any interested party that the status and operability of a given vehicle with a unique identifier has reached a certain threshold of wear or damage. Based on the information determined by the system 100, the vehicle operator or fleet manager may override the decision determined by the system 100 and continue to operate the vehicle. Alternatively, the system 100 may have the capability to safely disable the vehicle given specific parameters. For example, the system 100 may only allow the vehicle to operate for another specific distance or time, or may not allow the vehicle to restart after it has switched off without an appointment with a technician.



FIG. 15 shows an example flowchart of the use of the system 100 by, for example, a vehicle operator to monitor usage of tires. According to this embodiment, the system 100 is able to determine a specific tire brand or type, and cross-reference this brand or type with a database of compatible brands stored in a memory. In this embodiment, steps 1901 and 1902 are similar to those described in steps 1301, 1401, and 1302, 1402 respectively.


According to this embodiment, the system 100 is able to identify the tire characteristics 1903. These characteristics may include thickness, length, weight, width, tread pattern, etc. Based on an analysis of the vehicle's tires, the system 100 can determine tire alternatives at step 1904. This can be done using a pre-populated database stored on a server of all major available tire brands and products, along with compatible alternatives. Once the system 100 has determined the tire and tire characteristics, it can query the database to find all other products that could be used for the vehicle.


The system 100 can then communicate the tires to the user at step 1905. This can be done by sending the information over the network to the electronic device. The user may determine that an alternative tire could be used for the vehicle. If the user selects the alternative tire, the system 100 will send that message back to the server over the network and proceed to organize any part replacement using the user's selection.


As shown in FIGS. 16 to 19, image data capture is shown according to different embodiments. According to some embodiments, the electronic device 501 can display an instruction to the user to position and/or move the electronic device 501 in order to optimally capture the image.



FIG. 16 shows an embodiment in which the system 100 may instruct the user to take an image of the vehicle 10. The electronic device 501 will communicate this image along with any other information to be communicated to the system server 1142 for analysis, as described above.


As shown in FIG. 17, the system 100 may also or instead instruct the user to take a video of the tire 34. The electronic device 501 may then communicate this video along with any other information to be communicated to the system server 1142 for data analysis, as described above.


As shown in FIG. 18, the system 100 can instruct the user to use an accessory device 2202 in conjunction with the vehicle 10 in order to generate data about the tire 34. According to this embodiment, the accessory device 2202 can be an optical sensor communicatively linked to the electronic device 501. The accessory device 2202 can communicate the image data captured to the electronic device 501. The electronic device 501 may communicate this data along with any other information to be communicated to the system server 1142 for analysis, as described above.


As shown in FIG. 19, the electronic device 501 can be communicably linked to the vehicle, according to some embodiments. According to this embodiment, the electronic device 501 communicates with an onboard computer 1342 in the vehicle 10 in order to generate data about the vehicle. The electronic device 501 may communicate this data along with any other information to be communicated to the system server 1142 for analysis, as described above.


As shown in FIG. 20, the information determined about the vehicle based on the analysis conducted by the system 100 is communicated to the electronic device 501 over the network 124. According to this embodiment, the information was a length of time before the tires needed to be replaced.


As shown in FIG. 21, the system 100 has already communicated to the electronic device that based on the data analysis, vehicle maintenance is required. The electronic device 501 can then prompt the user to schedule the maintenance. If the user decides to schedule the maintenance, the electronic device 501 can communicate directly with a service center 2604 in order to schedule the maintenance over the network 124. For example, the user may have access to a booking calendar for the service center and select a time. Based on this selection, the parties will be notified that maintenance has been booked. According to some embodiments, the system 100 may have access to information about the service center 2604, such as parts inventory. Based on this inventory, the system 100 can calculate any lead time if required that can be factored into the booking span.


According to other embodiments such as those shown in FIG. 22, the system 100 can order new parts through the network 124 by creating a request to a retailer or parts center 2704. In this embodiment, if the service center requires a unique part that they do not have, the system 100 may create a request to the parts center to ship the part to the service center in advance of the booked maintenance time.


According to another embodiment, the system 100 may have access to pricing information or alternative replacement parts available at the parts center 2704. The system 100 may present the user with pricing options, sale information for different components they may require ordering for replacement. The user may then inform the system 100 of their preference and the system 100 will submit the order to the parts center accordingly.


As shown in FIG. 23, according to some embodiments, the system 100 is configured to schedule a maintenance request over the network 124 without requiring a user to select a time. This time may be based on a user preference saved in the server memory for a given vehicle owner. For example, an owner may have a preference that all vehicles are scheduled for maintenance one month before the system 100 determined date. Accordingly, the system 100 can notify the user of scheduled maintenance as it is automatically scheduled.


Similarly, according to some embodiments, the system 100 is able to make purchase requests over the network 124 without requiring the user to select a part component. This choice may be based on a user preference saved in the server memory for a given vehicle owner. For example, an owner may have a preference for a specific brand of vehicle parts. Accordingly the system 100 can notify the user of the part purchase as it is automatically scheduled.


As shown in FIG. 24, the electronic device may be communicably linked to a technician 3204. Alongside the user of the electronic device, the technician can also be notified over the network 124 of any determined vehicle information, scheduled maintenance, parts purchased, location of maintenance etc. Based on the user selection the user can be connected to a technician 3204 via the network 124. This connection could be by way of a telephone call, wherein the system 100 communicates a phone number over the network for the electronic device. Alternatively, the system 100 may use a Voice Over IP (VoIP) connection between the user and the technician. According to other embodiments, the communication between user and technician established could be a video call, wherein the technician is able to view a feed coming from a camera module within the user's electronic device.


According to the embodiments disclosed in FIGS. 25 to 27, the system 100 may determine that the vehicle has a critical malfunction. This could be determined through information captured form the onboard computer's internal sensor network or through data captured via the electronic device. For example, the vehicle tires may have been damaged to the point where further driving would cause greater permanent damage to the vehicle and may endanger the safety of the driver. Using the communication link between the electronic device and the vehicle, the system 100 can instruct the electronic device to prompt a user with a notification of the critical malfunction and request instruction for whether or not the vehicle should be allowed to continue to operate. Based on this decision, the electronic device can instruct an onboard computer in the vehicle that the vehicle should not be operated again until the system 100 has determined the vehicle is no longer in a critical malfunction state.


According to another embodiment and shown in FIG. 25, the electronic device may offer the user a choice to immediately disable the vehicle. Based on this decision, the electronic device can instruct an onboard computer in the vehicle that the vehicle should not be operated again until the system 100 has determined that the vehicle and/or the tires is no longer in a critical state.


According to another embodiment and shown in FIG. 26, the electronic device may not offer the user a choice and immediately disable the vehicle. Based on this decision, the electronic device can instruct an onboard computer in the vehicle that the vehicle should not be operated again until the system 100 has determined the vehicle and/or the tires is no longer in a critical state.


According to yet another embodiment and shown in FIG. 27, the electronic device may not offer the user a choice and may disable the vehicle once the vehicle has been returned to a specific location. This can be done by using a location coordinate determined by either the electronic device 501 or in the vehicle itself. While the vehicle may be continued to be used to complete the current job, when the location coordinate of the vehicle is determined to be the same as a specific location such as a storage facility, the electronic device can instruct an onboard computer in the vehicle that the vehicle should not be operated again until the system 100 has determined that the vehicle and/or the tires is no longer in a critical state.


In some embodiments, with additional reference to FIGS. 28 and 29, in addition to or instead of the electronic device 501, the system server 1142 may receive image data from an inspection station for inspecting vehicles such as the vehicle 10 when they are in proximity.


For example, in some embodiments, as shown in FIG. 28, the system 100 may include a imaging inspection station 463 for inspecting tires of vehicles 461x. In some embodiments, the imaging inspection station 463 comprises camera systems 462x arranged to capture images of each of the tires 341 to 344 and their environment. The captured images can then be optionally processed and analyzed locally or remotely in system 100. The camera systems 462x can include directional cameras having any configuration of lenses suitable for inspecting the tires 341 to 344 and their environment.


In other embodiments, with additional reference to FIG. 29, the system server 1142 may receive image data from a scanning inspection station 473 for inspecting tires of vehicles 471x. In some embodiments, the inspection station 473 comprises laser line scanner and/or laser area scanner systems 472x arranged to scan each of the tires 341 to 344 and their environment as each vehicle 471, moves past the inspection station 473. The information generated by the laser line scanner and/or laser area scanner systems 472x can then be optionally processed and analyzed locally or remotely by system server 1142. This embodiment is particularly advantageous for producing 3D scanning data suitable for subsequent volumetric analysis, as described in more detail above.


In some embodiments, with additional reference to FIGS. 30 and 31, the system server 1142 may receive image data from a drone 3201 for inspecting the tire 34 and/or other components of each of the wheels 201 to 204 and/or their environment (e.g., detecting the presence of debris, etc.), so that information derived from the drone 3210 may be relayed to the operator of the vehicle 10 and/or another remote device or person. The vehicle 10 may comprise a drone mount 3220 configured to mount the drone 3220 to the vehicle 10 and release the drone 3201 when the drone 3201 is to monitor the vehicle 10 by moving around it.


In some embodiments, the drone 3201 is arranged to follow the vehicle, capture and analyze images of each of the tires 341 to 344 and their environment. In other embodiments, the drone 3201 is equipped with a laser line scanner for scanning the tires 341 to 344 and their environment. Communication between the drone 3201 and the vehicle 10 can be provided for by any suitable means, including but not limited to any combination of Global Positioning System (GPS) signals, Radio Frequency (RF) signals, Bluetooth signals, LIDAR, and RADAR signals. This embodiment is particularly advantageous for producing 3D scanning data suitable for subsequent volumetric analysis, as described in more detail above.


In this embodiment, the drone 3210 is an aerial drone configured to fly about the vehicle 10. While the drone 3201 shown in FIG. 30 is a multi-rotor flying drone, other drones are possible, including but not limited to fixed-wing drones, or any other type of unmanned aerial vehicle. Also, in other embodiments, the drone 3210 may be a land drone configured to travel on the ground about the vehicle 10 (e.g., on wheels or on tracks).


In some embodiments, with additional reference to FIG. 32, in addition to or instead of the electronic device 501, the system server 1142 may receive image data from a vehicle-mounted inspection device 4801 for inspecting the tire 34 of the vehicle 10. In particular, the system 100 may include one or more vehicle-mounted inspection device 4801 for inspecting tire 34 of vehicles by way of image data. In some embodiments, each tire 341 to 344 is provided with a vehicle-mounted inspection device 4801.


In some embodiments, the vehicle-mounted inspection device 4801 comprises a camera system arranged to capture images of the tire 34 and its environment as the tire 34 rotates. The information generated by the camera system can then be optionally processed and analyzed locally or remotely by the system server 1142.


In some embodiments, the vehicle-mounted inspection device 4801 comprises a laser line scanner system and/or a laser area scanner system arranged to scan the tire 34 and its environment as the tire 34 rotates. The information generated by the laser line scanner and/or laser area scanner systems can then be optionally processed and analyzed locally or remotely by system server 1142. This embodiment is particularly advantageous for producing 3D scanning data suitable for subsequent volumetric analysis, as described in more detail above.


In some embodiments, as shown in FIG. 33, a given component mentioned herein (e.g., the electronic device 501, the image processing entity 505, the server 1142, etc.) may comprise a computing system 1500 comprising suitable hardware and/or software (e.g., firmware) configured to implement functionality of that given component. The computing system 1500 comprises an interface 1520, a processor 1540, and a memory 1560.


The interface 1520 comprises one or more inputs and outputs allowing the computing system 1500 to receive signals from and send signals to other components to which the computing system 1500 is connected (i.e., directly or indirectly connected).


The processor 1540 comprises one or more processing devices for performing processing operations that implement functionality of the computing system 1500. A processing device of the processor 1540 may be a general-purpose processor executing program code stored in the memory 1560. Alternatively, a processing device of the processor 1540 may be a specific-purpose processor comprising one or more preprogrammed hardware or firmware elements (e.g., application-specific integrated circuits (ASICs), electrically erasable programmable read-only memories (EEPROMs), etc.) or other related elements).


The memory 1560 comprises one or more memory elements for storing program code executed by the processor 1540 and/or data used during operation of the processor 1540. A memory element of the memory portion 1560 may be a semiconductor medium (including, e.g., a solid state memory), a magnetic storage medium, an optical storage medium, and/or any other suitable type of memory element. A memory element of the memory portion 1560 may be read-only memory (ROM) and/or random-access memory (RAM), for example.


In some embodiments, two or more elements of the computing system 1500 may be implemented by devices that are physically distinct from one another (e.g., located in a common site or in remote sites) and may be connected to one another via a bus (e.g., one or more electrical conductors or any other suitable bus) or via a communication link which may be wired, wireless, or both and which may traverse one or more networks (e.g., the Internet or any other computer network such as a local-area network (LAN) or wide-area network (WAN), a cellular network, etc.). In other embodiments, two or more elements of the computing system 1500 may be implemented by a single device.


While in embodiments considered above the vehicle 10 is a passenger vehicle, in other embodiments, the vehicle 10 may be another type of work vehicle (such as a tractor, etc.) for performing forestry work, or a military vehicle (e.g., reconnaissance vehicle, military light utility vehicle, etc.) for performing military work, a carrier (e.g. carrying a boom, a rig, and/or other equipment), or may be any other type of vehicle operable on paved road. Also, while in embodiments considered above the road vehicle 10 is driven by a human operator in the vehicle 10, in other embodiments, the vehicle 10 may be an unmanned ground vehicle (e.g., a teleoperated or autonomous unmanned ground vehicle).


Any feature of any embodiment discussed herein may be combined with any feature of any other embodiment discussed herein in some examples of implementation.


Certain additional elements that may be needed for operation of certain embodiments have not been described or illustrated as they are assumed to be within the purview of those of ordinary skill in the art. Moreover, certain embodiments may be free of, may lack and/or may function without any element that is not specifically disclosed herein.


Although various embodiments and examples have been presented, this was for purposes of description, but should not be limiting. Various modifications and enhancements will become apparent to those of ordinary skill in the art.

Claims
  • 1. A system for monitoring a tire of a vehicle on a ground, the tire including elastomeric material embedded with a reinforcement and defining sidewalls, the system comprising: an interface configured to receive data regarding at least one image of the tire; anda processor configured to: perform a two-dimensional (2D) image analysis based on the data regarding the at least one image of the tire to obtain a result of the 2D image analysis, the result of the 2D image analysis being an indication of a defect of the tire;create a three-dimensional (3D) model of the tire based on the data regarding the at least one image of the tire;generate information on a state of the tire based on at least one of the result of the 2D image analysis and the 3D model of the tire; anddetermine that a deterioration threshold event relating to at least one of a crack in a sidewall of the tire, exposure of the reinforcement and a partially embedded foreign object in the elastomeric material, has occurred based on the information on the state of the tire;wherein the processor is further configured to issue a signal based on the determination to inform a user of the system of an action to be performed in respect of the tire.
  • 2. The system of claim 1, wherein the information on the state of the tire includes an indication of a volumetric loss of the elastomeric material of the tire.
  • 3. The system of claim 1, wherein the action to be performed in respect of the tire is replacement of the tire by a new tire.
  • 4. The system of claim 1, wherein the processor is configured to issue a signal based on the information on the state of the tire that is directed to the vehicle.
  • 5. The system of claim 1, wherein the at least one image of the tire is captured by a wireless communication device.
  • 6. The system of claim 1, wherein the 3D model of the tire is generated using a 3D point cloud.
  • 7. The system of claim 1, wherein the defect of the tire comprises a breakage of the reinforcement.
  • 8. The system of claim 1, wherein the reinforcement is a reinforcing cable extending around the tire.
  • 9. The system of claim 1, wherein the processor comprises an artificial intelligence module configured to perform the 2D image analysis.
  • 10. The system of claim 1, wherein the processor is configured to compare the 3D model of the tire to a 3D model of a reference tire to generate the information on the state of the tire.
  • 11. The system of claim 1, wherein the processor is configured to generate the information on the state of the tire based on both the result of the 2D image analysis and the 3D model of the tire.
  • 12. The system of claim 1, wherein the information on the state of the tire comprises an indication of a level of wear of the tire.
  • 13. The system of claim 1, wherein the information on the state of the tire comprises an indication of a remaining useful life of the tire.
  • 14. The system of claim 4, wherein the signal is issued to control the vehicle.
  • 15. The system of claim 9, wherein the artificial intelligence module comprises a neural network.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application is a National Stage of international PCT Patent Application No. PCT/CA2019/051219 filed on Aug. 30, 2019, which claims the benefit of U.S. Provisional Patent Application Ser. No. 62/724,853, filed on Aug. 30, 2018 and U.S. Provisional Patent Application Ser. No. 62/861,684, filed on Jun. 14, 2019. The contents of the aforementioned applications are incorporated by reference herein.

PCT Information
Filing Document Filing Date Country Kind
PCT/CA2019/051219 8/30/2019 WO
Publishing Document Publishing Date Country Kind
WO2020/041899 3/5/2020 WO A
US Referenced Citations (330)
Number Name Date Kind
1539582 Andry May 1925 A
1673541 Wilson Jun 1928 A
2025999 Myers Dec 1935 A
2040696 Johnston May 1936 A
2369130 Benson Feb 1945 A
2461150 Flynn Feb 1949 A
2523182 Battaglia Sep 1950 A
2562264 Ford Jul 1951 A
2596919 Smith May 1952 A
2854294 Bannister Sep 1958 A
3019061 Schomers Jan 1962 A
3118709 Case Jan 1964 A
3464476 Scheuba Sep 1969 A
3582154 Russ, Sr. Jun 1971 A
3612626 Fuchs Oct 1971 A
3747995 Russ, Sr. Jul 1973 A
3747996 Huber Jul 1973 A
3767275 Russ, Sr. Oct 1973 A
3781067 Dodson et al. Dec 1973 A
3858948 Johnson et al. Jan 1975 A
3887244 Haslett et al. Jun 1975 A
3914990 Borg Oct 1975 A
3944006 Lassanske Mar 1976 A
3981943 Fujio et al. Sep 1976 A
4059313 Beyers et al. Nov 1977 A
RE29718 Reinsma et al. Aug 1978 E
4150858 Fox et al. Apr 1979 A
4218101 Thompson Aug 1980 A
4279449 Martin et al. Jul 1981 A
4538860 Edwards et al. Sep 1985 A
4583791 Nagata et al. Apr 1986 A
4586757 Bloechlinger May 1986 A
4587280 Guha et al. May 1986 A
4607892 Payne et al. Aug 1986 A
4614508 Kerivan Sep 1986 A
4696520 Henke et al. Sep 1987 A
4721498 Grob Jan 1988 A
D298018 Cartwright Oct 1988 S
4843114 Touchet et al. Jun 1989 A
4844561 Savage et al. Jul 1989 A
4880283 Savage et al. Nov 1989 A
4953919 Langford Sep 1990 A
4953921 Burns Sep 1990 A
4981188 Kadela Jan 1991 A
5018591 Price May 1991 A
5050710 Bargfrede Sep 1991 A
5117930 Argouarc'h Jun 1992 A
5145242 Togashi Sep 1992 A
5190363 Brittain et al. Mar 1993 A
5299860 Anderson Apr 1994 A
5320585 Kato Jun 1994 A
5352029 Nagorcka Oct 1994 A
5362142 Katoh Nov 1994 A
5368115 Crabb Nov 1994 A
5380076 Hori Jan 1995 A
5447365 Muramatsu et al. Sep 1995 A
5452949 Kelderman Sep 1995 A
5482364 Edwards et al. Jan 1996 A
5498188 Deahr Mar 1996 A
5511869 Edwards et al. Apr 1996 A
5513683 Causa et al. May 1996 A
5529267 Giras et al. Jun 1996 A
5540489 Muramatsu et al. Jul 1996 A
5632537 Yoshimura et al. May 1997 A
5707123 Grob Jan 1998 A
5722745 Courtemanche et al. Mar 1998 A
5813733 Hori et al. Sep 1998 A
5866265 Reilly et al. Feb 1999 A
5894900 Yamamoto et al. Apr 1999 A
5904217 Yamamoto et al. May 1999 A
5984438 Tsunoda et al. Nov 1999 A
5997109 Kautsch Dec 1999 A
6000766 Takeuchi et al. Dec 1999 A
6024183 Dietz et al. Feb 2000 A
6030057 Fikse Feb 2000 A
6056656 Kitano et al. May 2000 A
6065818 Fischer May 2000 A
6068354 Akiyama et al. May 2000 A
6074025 Juncker et al. Jun 2000 A
6079802 Nishimura et al. Jun 2000 A
6095275 Shaw Aug 2000 A
6120405 Oertley et al. Sep 2000 A
6129426 Tucker Oct 2000 A
6139121 Muramatsu Oct 2000 A
6153686 Granatowicz et al. Nov 2000 A
6170925 Ono Jan 2001 B1
6176557 Ono Jan 2001 B1
6186604 Fikse Feb 2001 B1
6193335 Edwards Feb 2001 B1
6206492 Moser Mar 2001 B1
6224172 Goodwin May 2001 B1
6241327 Gleasman et al. Jun 2001 B1
6259361 Robillard et al. Jul 2001 B1
6296329 Rodgers et al. Oct 2001 B1
6299264 Kautsch et al. Oct 2001 B1
6300396 Tsunoda et al. Oct 2001 B1
6352320 Bonko et al. Mar 2002 B1
6386652 Bonko May 2002 B1
6386653 Brandenburger May 2002 B1
6386654 Singer et al. May 2002 B1
6416142 Oertley Jul 2002 B1
6474756 Hori et al. Nov 2002 B2
6494548 Courtemanche Dec 2002 B2
6536852 Katayama et al. Mar 2003 B2
6536853 Egle et al. Mar 2003 B2
6568769 Watanabe et al. May 2003 B1
6581449 Brown et al. Jun 2003 B1
D476599 Whittington Jul 2003 S
6588862 Pringiers Jul 2003 B1
6637276 Adderton et al. Oct 2003 B2
6652043 Oertley Nov 2003 B2
6671609 Nantz et al. Dec 2003 B2
6698850 Ueno Mar 2004 B2
D488171 Juncker et al. Apr 2004 S
6716012 Yovichin et al. Apr 2004 B2
6733091 Deland et al. May 2004 B2
6733093 Deland et al. May 2004 B2
6769746 Rodgers et al. Aug 2004 B2
6800236 Kurata et al. Oct 2004 B1
6848757 Ueno Feb 2005 B2
6874586 Boivin et al. Apr 2005 B2
D505136 Brazier May 2005 S
6904986 Brazier Jun 2005 B2
6913329 Rodgers et al. Jul 2005 B1
6921197 Aubel et al. Jul 2005 B2
6923515 Konickson et al. Aug 2005 B2
6932442 Hori Aug 2005 B2
6935708 Courtemanche Aug 2005 B2
6948784 Wodrich et al. Sep 2005 B2
6962222 Kirihata Nov 2005 B2
6964462 Katoh et al. Nov 2005 B2
6974196 Gagne et al. Dec 2005 B2
7001294 Fukuda Feb 2006 B2
7032636 Salakari Apr 2006 B2
7077216 Juncker Jul 2006 B2
D528133 Brazier Sep 2006 S
7114788 Deland et al. Oct 2006 B2
7131508 Brazier Nov 2006 B2
7137675 Simula et al. Nov 2006 B1
7197922 Rimkus et al. Apr 2007 B2
7202777 Tsuji et al. Apr 2007 B2
7222924 Christianson May 2007 B2
7229141 Dandurand et al. Jun 2007 B2
7252348 Gingras Aug 2007 B2
7293844 Uchiyama Nov 2007 B2
D556791 Brazier Dec 2007 S
7316251 Kogure et al. Jan 2008 B2
7325888 Fujita et al. Feb 2008 B2
7367637 Gleasman May 2008 B2
7407236 Fukushima Aug 2008 B2
7413268 Kato et al. Aug 2008 B2
7416266 Soucy et al. Aug 2008 B2
7497530 Bessette Mar 2009 B2
7567171 Dufournier Jul 2009 B2
7597161 Brazier Oct 2009 B2
D603880 Brazier Nov 2009 S
7625050 Bair Dec 2009 B2
7676307 Schmitt et al. Mar 2010 B2
7708092 Despres May 2010 B2
7729823 Ruoppolo Jun 2010 B2
7740094 Pelletier Jun 2010 B2
7778741 Rao et al. Aug 2010 B2
7779947 Stratton Aug 2010 B2
7784884 Soucy et al. Aug 2010 B2
7798260 Albright et al. Sep 2010 B2
7823987 Dandurand et al. Nov 2010 B2
7866766 Berg Jan 2011 B2
7914088 Bair Mar 2011 B2
7914089 Bair Mar 2011 B2
D644670 Barrelmeyer Sep 2011 S
8122581 Hurst et al. Feb 2012 B1
8327960 Couture et al. Dec 2012 B2
8342257 Rosenboom Jan 2013 B2
D680561 Zuchoski et al. Apr 2013 S
D681071 Zuchoski et al. Apr 2013 S
D683371 Aube May 2013 S
D683769 Aube Jun 2013 S
8567876 Wellman Oct 2013 B2
8628152 Delisle et al. Jan 2014 B2
8736147 Wang May 2014 B1
D711928 Brazier Aug 2014 S
8985250 Lussier et al. Mar 2015 B1
9033431 Zuchoski et al. May 2015 B1
9067631 Lussier et al. Jun 2015 B1
9334001 Lussier et al. May 2016 B2
9805697 Dorrance et al. Oct 2017 B1
9855843 Vik et al. Jan 2018 B2
9878750 Leblanc et al. Jan 2018 B2
9880075 Finch et al. Jan 2018 B2
9975554 Lussier et al. May 2018 B2
10006836 Monti et al. Jun 2018 B2
10267710 Regoli Apr 2019 B2
10272959 Zuchoski et al. Apr 2019 B2
10328982 Lussier et al. Jun 2019 B2
10745069 Gustafson Aug 2020 B2
10783723 Richard et al. Sep 2020 B2
10933877 Lussier et al. Mar 2021 B2
11046377 Lussier et al. Jun 2021 B2
11167810 Boily et al. Nov 2021 B2
11186330 Zuchoski et al. Nov 2021 B2
20020070607 Edwards Jun 2002 A1
20020140288 Robert, Sr. et al. Oct 2002 A1
20020145335 Soucy et al. Oct 2002 A1
20030019133 Hori Jan 2003 A1
20030034189 Lemke et al. Feb 2003 A1
20030034690 Hori et al. Feb 2003 A1
20030080618 Krishnan et al. May 2003 A1
20030089534 Kanzler et al. May 2003 A1
20030161687 Kirihata Aug 2003 A1
20040004395 Soucy et al. Jan 2004 A1
20040070273 Safe et al. Apr 2004 A1
20040084962 Courtemanche May 2004 A1
20040130212 Ishibashi Jul 2004 A1
20040135433 Inaoka et al. Jul 2004 A1
20050035654 Tamaru et al. Feb 2005 A1
20050056468 Tucker Mar 2005 A1
20050103540 Lavoie May 2005 A1
20050104449 Lavoie et al. May 2005 A1
20050104450 Gagne et al. May 2005 A1
20050168069 Ueno Aug 2005 A1
20060060395 Boivin et al. Mar 2006 A1
20060090558 Raskas May 2006 A1
20060103236 Soucy et al. May 2006 A1
20060124366 Le Masne De Chermont Jun 2006 A1
20060144480 Takayama et al. Jul 2006 A1
20060175108 Kubota Aug 2006 A1
20060220456 Sugahara Oct 2006 A1
20060248484 Baumgartner et al. Nov 2006 A1
20060273660 Dandurand et al. Dec 2006 A1
20070046100 McGilvrey et al. Mar 2007 A1
20070075456 Feldmann Apr 2007 A1
20070126286 Feldmann et al. Jun 2007 A1
20070159004 St-Amant Jul 2007 A1
20070251621 Prost Nov 2007 A1
20070252433 Fujita Nov 2007 A1
20080007118 Fujita Jan 2008 A1
20080073971 Paradis et al. Mar 2008 A1
20080084111 Rainer Apr 2008 A1
20080100134 Soucy et al. May 2008 A1
20080136255 Feldmann et al. Jun 2008 A1
20080169147 Brazier Jul 2008 A1
20080179124 Stratton Jul 2008 A1
20080203813 Doyle Aug 2008 A1
20080211300 Matsuo et al. Sep 2008 A1
20080211301 Jee et al. Sep 2008 A1
20090085398 Maltais Apr 2009 A1
20090102283 Choi Apr 2009 A1
20090166101 Wenger et al. Jul 2009 A1
20090195062 Uchida Aug 2009 A1
20090302676 Brazier Dec 2009 A1
20090302677 Sugihara Dec 2009 A1
20090309415 Shimozono Dec 2009 A1
20090326109 Kameda et al. Dec 2009 A1
20100012399 Hansen Jan 2010 A1
20100033010 Shimozono Feb 2010 A1
20100079238 Gravelle et al. Apr 2010 A1
20100095506 Bair Apr 2010 A1
20100096915 Hagio Apr 2010 A1
20100096917 Bair Apr 2010 A1
20100121644 Glasser May 2010 A1
20100133019 Muemken Jun 2010 A1
20100139994 Hansen Jun 2010 A1
20100191417 Murahashi et al. Jul 2010 A1
20100194553 Mizutani et al. Aug 2010 A1
20100230185 Mallette et al. Sep 2010 A1
20100253138 Despres Oct 2010 A1
20100256946 Carresjo et al. Oct 2010 A1
20100283317 Soucy et al. Nov 2010 A1
20110068620 Delisle et al. Mar 2011 A1
20110121644 Wellman May 2011 A1
20110148189 Courtemanche et al. Jun 2011 A1
20110301825 Grajkowski et al. Dec 2011 A1
20110315459 Zuchoski et al. Dec 2011 A1
20120001478 Zuchoski et al. Jan 2012 A1
20120056473 Hashimoto et al. Mar 2012 A1
20120104840 Zuchoski et al. May 2012 A1
20120242142 Kautsch et al. Sep 2012 A1
20120242143 Feldmann Sep 2012 A1
20120253590 Fink Oct 2012 A1
20120306916 Marumoto et al. Dec 2012 A1
20130033271 Woodard Feb 2013 A1
20130073157 Person et al. Mar 2013 A1
20130082846 McKinley et al. Apr 2013 A1
20130126196 Rosenboom May 2013 A1
20130134772 Dandurand et al. May 2013 A1
20130134773 Dandurand et al. May 2013 A1
20130162016 Lajoie et al. Jun 2013 A1
20130245911 Nakajima et al. Sep 2013 A1
20130255354 Hawkins et al. Oct 2013 A1
20130325266 Padilla et al. Dec 2013 A1
20140105481 Hasselbusch et al. Apr 2014 A1
20140125117 Weeks et al. May 2014 A1
20140180534 Son Jun 2014 A1
20140182960 Bedard et al. Jul 2014 A1
20140288763 Bennett et al. Sep 2014 A1
20140324301 Rebinsky Oct 2014 A1
20150042152 Lussier et al. Feb 2015 A1
20150042153 Lussier et al. Feb 2015 A1
20150107975 Chen et al. Apr 2015 A1
20150129329 Cox May 2015 A1
20150191173 Lussier et al. Jul 2015 A1
20150321710 Zuchoski et al. Nov 2015 A1
20160052572 McKinley et al. Feb 2016 A1
20160059779 Vandendriessche Mar 2016 A1
20160121945 Rust et al. May 2016 A1
20160129954 Hasselbusch et al. May 2016 A1
20160236733 Tiede et al. Aug 2016 A1
20160332682 Ussier et al. Nov 2016 A1
20170036714 Lunkenbein Feb 2017 A1
20170087987 Vik et al. Mar 2017 A1
20170094894 Heim et al. Apr 2017 A1
20170174277 Zuchoski et al. Jun 2017 A1
20170177011 Garvin et al. Jun 2017 A1
20180043949 Boily Feb 2018 A1
20180093724 Boily Apr 2018 A1
20180172557 Ghidotti Piovan et al. Jun 2018 A1
20180190045 Richard et al. Jul 2018 A1
20180237020 Lussier et al. Aug 2018 A1
20180265145 Todd et al. Sep 2018 A1
20180364744 Garvin et al. Dec 2018 A1
20190351957 Zuchoski et al. Nov 2019 A1
20190359270 Lussier et al. Nov 2019 A1
20200070906 Laperle et al. Mar 2020 A1
20210024150 Boily et al. Jan 2021 A1
20210074089 Boily et al. Mar 2021 A1
20210173399 Richard et al. Jun 2021 A1
20210213954 Lussier et al. Jul 2021 A1
20210339758 Laperle et al. Nov 2021 A1
20220055699 Desmarais et al. Feb 2022 A1
20220063740 Boily et al. Mar 2022 A1
Foreign Referenced Citations (70)
Number Date Country
2606039 Apr 2009 CA
2838935 Jan 2013 CA
2978482 Sep 2016 CA
3064100 Sep 2016 CA
2991072 Jan 2017 CA
2947575 May 2017 CA
3110606 Mar 2020 CA
104981685 Oct 2015 CN
110588813 Dec 2019 CN
20207342 Aug 2002 DE
0578504 Jan 1994 EP
1008509 Jun 2000 EP
1211169 Jun 2002 EP
1260429 Nov 2002 EP
1273504 Jan 2003 EP
1283152 Feb 2003 EP
1325820 Sep 2004 EP
1506913 Feb 2005 EP
1882627 Jan 2008 EP
1953070 Aug 2008 EP
2014542 Jan 2009 EP
2119620 Nov 2009 EP
3185097 Jun 2017 EP
3265368 Jan 2018 EP
3313717 May 2018 EP
3721200 Oct 2020 EP
3721200 Oct 2020 EP
3747746 Dec 2020 EP
3812249 Apr 2021 EP
3844473 Jul 2021 EP
3844474 Jul 2021 EP
3960593 Feb 2022 EP
2229410 Sep 1990 GB
2360501 Sep 2001 GB
H02074476 Mar 1990 JP
H11129946 May 1999 JP
2000053037 Feb 2000 JP
2003011860 Jan 2003 JP
2003089366 Mar 2003 JP
2004330830 Nov 2004 JP
2006103482 Apr 2006 JP
2007022304 Feb 2007 JP
2009061829 Mar 2009 JP
2009248924 Oct 2009 JP
2010018091 Jan 2010 JP
2010047040 Mar 2010 JP
2010089729 Apr 2010 JP
100829059 May 2008 KR
20120055071 May 2012 KR
20130043954 May 2013 KR
20150004969 Jan 2015 KR
201500004969 Jan 2015 KR
1446017 Dec 1988 SU
2008088355 Jul 2008 WO
2008108439 Sep 2008 WO
2009105892 Sep 2009 WO
2009106617 Sep 2009 WO
2013002781 Jan 2013 WO
2014056089 Apr 2014 WO
2014168851 Oct 2014 WO
2015059457 Apr 2015 WO
2016138592 Sep 2016 WO
2017000068 Jan 2017 WO
2017049393 Mar 2017 WO
2019109191 Jun 2019 WO
2019225752 Nov 2019 WO
2020041897 Mar 2020 WO
2020041899 Mar 2020 WO
2020049532 Mar 2020 WO
2021189154 Sep 2021 WO
Non-Patent Literature Citations (100)
Entry
Communication pursuant to Article 94 (3) EPC issued on Aug. 22, 2019, in connection with European Patent Application No. 16816886.2, 4 pages.
Communication pursuant to Rule 63(1) EPC issued on Nov. 8, 2018 in connection with European Patent Application No. 16758408.5, 4 pages.
Communication under Rule 71(3) EPC issued on Mar. 16, 2020, in connection with European Patent Application 16758408.5, 7 pages.
Communication under Rule 71(3) EPC issued on May 6, 2020, in connection with European Patent Application No. 16816886.2, 7 pages.
European Extended Search Report issued on Nov. 6, 2020, in connection with European Patent Application No. 20187079.7, 15 pages.
Ex Parte Quayle issued on Sep. 11, 2018, in connection with U.S. Appl. No. 15/139,572, 7 pages.
Examiners Report issued on Apr. 7, 2020, in connection with Canadian Patent Application 3,064,100, 3 pages.
Examiners report issued on Aug. 28, 2018, in connection with Canadian Patent application No. 2,991,072, 4 pages.
Examiners Report issued on Feb. 1, 2018 in connection with Canadian Patent Application No. 2,991,072, 3 pages.
Examiners Report issued on Jan. 11, 2018 in connection with Canadian Patent Application No. 2,978,482, 3 pages.
Examiners report issued on Jan. 16, 2019 in connection with Canadian Patent Application No. 2,991,072, 3 pages.
Examiners Report issued on Jul. 9, 2020, in connection with Canadian Patent Application 3,064,100, 3 pages.
Examiners Report issued on Oct. 14, 2020, in connection with Canadian Patent Application 3,042,780, 4 pages.
Extended European Search Report and Written Opinion issued on Sep. 13, 2013 in connection, with European Patent Application No. 10835318.6, 8 pages.
Extended European Search Report issued on Jan. 9, 2019 in connection with European Patent Application No. 16816886.2, 7 pages.
Extended European Search Report issued on Mar. 20, 2019 in connection with European Patent Application No. 16758408.5, 8 pages.
Final Office Action issued on Aug. 19, 2019, in connection with U.S. Appl. No. 15/724,733, 18 pages.
Final Office Action issued on Aug. 24, 2018 in connection with U.S. Appl. No. 15/454,881, 30 pages.
Final Office Action issued on Dec. 23, 2014 in connection with U.S. Appl. No. 13/326,132, 15 pages.
Final Office Action issued on Dec. 31, 2015 in connection with U.S. Appl. No. 13/326,132, 15 pages.
Final Office Action issued on Jan. 12, 2017 in connection with U.S. Appl. No. 14/665,075, 11 pages.
Final Office Action issued on Jul. 6, 2020, in connection with U.S. Appl. No. 15/555,148, 20 pages.
Final Office Action issued on Jun. 26, 2015 in connection with U.S. Appl. No. 13/326,278, 15 pages.
Final Office Action issued on Mar. 27, 2014 in connection with U.S. Appl. No. 13/170,753, 16 pages.
Final Office Action issued on Mar. 5, 2015 in connection with U.S. Appl. No. 13/325,796, 33 pages.
Final Office Action issued on Sep. 8, 2017 in connection with U.S. Appl. No. 15/139,572, 14 pages.
International Search Report and Written Opinion issued on Sep. 29, 2016 in connection with PCT/CA2016/050760, 8 pages.
International Search Report and Written Opinion issued on Nov. 12, 2019, in connection with International PCT Application No. PCT/CA2019/051217, 8 pages.
International Search Report and Written Opinion issued on Oct. 29, 2019 in connection with International PCT Application No. PCT/CA2019/051219, 9 pages.
International Search Report issued on Mar. 5, 2019, in connection with International PCT application No. PCT/CA2018/051567, 4 pages.
Interview Summary issued on Dec. 10, 2018 in connection with U.S. Appl. No. 15/139,572, 3 pages.
Interview Summary issued on Nov. 2, 2018 in connection with U.S. Appl. No. 15/454,881, 3 pages.
Interview Summary Report issued on Sep. 12, 2015 in connection with U.S. Appl. No. 13/326,278, 3 pages.
Non-Final Office Action issued on Apr. 5, 2016 in connection with U.S. Appl. No. 14/665,075, 5 pages.
Non-Final Office Action issued on Apr. 9, 2019 in connection with U.S. Appl. No. 15/724,733, 56 pages.
Non-Final Office Action issued on Dec. 1, 2016 in connection with U.S. Appl. No. 15/139,572, 5 pages.
Non-Final Office Action issued on Dec. 30, 2019 in connection with U.S. Appl. No. 15/740,976, 16 pages.
Non-Final Office Action issued on Feb. 21, 2017 in connection with U.S. Appl. No. 14/721,326, 12 pages.
Non-Final Office Action issued on Jan. 11, 2017 in connection with U.S. Appl. No. 14/886,327, 11 pages.
Non-Final Office Action issued on Jul. 10, 2014 in connection with U.S. Appl. No. 13/326,110, 16 pages.
Non-Final Office Action issued on Jul. 10, 2014 in connection with U.S. Appl. No. 13/424,459, 11 pages.
Non-Final Office Action issued on Jul. 14, 2014 in connection with U.S. Appl. No. 13/112,840, 16 pages.
Non-Final Office Action issued on Jun. 12, 2017 in connection with U.S. Appl. No. 14/665,075, 9 pages.
Non-Final Office Action issued on Jun. 2, 2014 in connection with U.S. Appl. No. 13/326,010, 6 pages.
Non-Final Office Action issued on Jun. 24, 2015 in connection with U.S. Appl. No. 13/326,132, 17 pages.
Non-Final Office Action issued on Mar. 20, 2018 in connection with U.S. Appl. No. 15/139,572, 7 pages.
Non-Final Office Action issued on Mar. 6, 2015 in connection with U.S. Appl. No. 13/326,278, 16 pages.
Non-Final Office Action issued on May 17, 2018 in connection with U.S. Appl. No. 15/454,881, 8 pages.
Non-Final Office Action issued on Oct. 11, 2017 in connection with U.S. Appl. No. 14/665,075, 6 pages.
Non-Final Office Action issued on Oct. 2, 2013 in connection with U.S. Appl. No. 13/170,753, 15 pages.
European Extended Search Report issued on Aug. 18, 2021, in connection with European Patent Application No. 18887057.0, 8 pages.
Examiner's report issued on Jan. 22, 2021, in connection with Canadian Patent Application No. 2,991,072, 3 pages.
Non-Final Office Action issued on Dec. 14, 2020, in connection with U.S. Appl. No. 15/555,148, 9 pages.
Notice of Allowance issued on Feb. 23, 2021 in connection with U.S. Appl. No. 15/724,733, 18 pages.
Final Office Action issued on Mar. 9, 2021, in connection with U.S. Appl. No. 15/555,148, 14 pages.
Notice of Allowance issued on Mar. 4, 2015 in connection with U.S. Appl. No. 13/325,783, 13 pages.
Notice of Allowance issued on May 20, 2020 in connection with U.S. Appl. No. 15/740,976, 12 pages.
Notice of Allowance issued on May 30, 2018 in connection with U.S. Appl. No. 15/139,572, 10 pages.
Notice of Allowance issued on Nov. 14, 2014 in connection with U.S. Appl. No. 13/326,010, 7 pages.
Notice of Allowance issued on Oct. 30, 2020, in connection with U.S. Appl. No. 15/958,156, 30 pages.
Notice of Allowance issued on Oct. 24, 2014 in connection with U.S. Appl. No. 13/170,753, 5 pages.
Non-Final Office Action issued on Feb. 2, 2021, in connection with U.S. Appl. No. 16/360,060, 55 pages.
Non-Final Office Action issued on Jan. 24, 2020 in connection with U.S. Appl. No. 15/555,148, 57 pages.
Non-Final Office Action issued on May 29, 2020, in connection with U.S. Appl. No. 15/958,156, 45 pages.
Non-Final Office Action issued on Nov. 9, 2012, in connection with U.S. Appl. No. 29/405,414, 9 pages.
Restriction Requirement issued on Jan. 27, 2014 in connection with U.S. Appl. No. 13/326,010, 5 pages.
Restriction Requirement issued on Jun. 9, 2014 in connection with U.S. Appl. No. 13/326,278, 7 pages.
Restriction Requirement issued on Sep. 26, 2013 in connection with U.S. Appl. No. 13/326,010, 6 pages.
Supplementary Search Report issued on Jan. 2, 2019 in connection with the European Patent Application No. 16816886, 1 page.
Written Opinion issued on Mar. 5, 2019 in connection with International PCT application No. PCT/CA2018/051567, 6 pages.
International Search Report and Written Opinion issued on Jun. 21, 2021 in connection with International PCT application No. PCT/CA2021/050404, 8 pages.
Notice of Allowance issued on Jul. 9, 2021, in connection with U.S. Appl. No. 15/555,148, 6 pages.
Notice of Allowance issued on Jul. 26, 2021, in connection with U.S. Appl. No. 16/360,060, 14 pages.
Extended European Search Report issued on Mar. 30, 2021, in connection with European Patent Application No. 20201638.2, 10 pages.
Non-Final Office Action issued on Mar. 11, 2021 in connection with U.S. Appl. No. 16/419,760, 28 pages.
Final Office Action issued on Nov. 9, 2021 in connection with U.S. Appl. No. 16/419,760, 57 pages.
International Search Report and Written Opinion issued on Nov. 3, 2021 in connection with International PCT application No. PCT/CA2021/051143, 17 pages.
Non-Final Office Action issued on Nov. 9, 2012, in connection with U.S. Appl. No. 29/405,417, 10 pages.
Extended European Search Report issued on Feb. 1, 2022, in connection with European Patent Application No. 21191893.3, 11 pages.
Bair Products, Inc. “Larry Lugs—Patented Bolt-On-Replacement Drive Lugs”, http://www.bairproductsinc.com/products/larry_lugs.html, Jan. 1, 2011, 2 pages.
Bridgestone Industrial Products America Inc. “Stay on the Right Track. New Generation Features Rubber Tracks for Excavators”, Brochure 2009, 12 pages.
Bridgestone Industrial Products America Inc., “Stay on the Right Track. Rubber Tracks for Track Loaders”, Brochure 2009, 6 pages.
Camoplast Inc. “ATV/UTV Track Systems”, 2009-2010 Catalog, 8 pages.
CAN-AM BRP, “Parts & Accessories—Track Systems”, Parts, Accessories & Riding Gear Catalogue, p. 66 (2011).
Red Roo Solutions Pty Ltd—World Class Solutions for the Earth Moving Industry, “Save thousands of dollars and add thousands of hours to your tracks with Larry Lugs”, http://www.redroosolutions.com.au/larrylugs.html, Jun. 8, 2009, 3 pages.
Story by Staff, BRP upgrades Apache ATV track system for Outlander—Apache system features larger footprint and power steering mode, atv.com, http://www.atc.com/news/brp-upgrades-apache-atv-track-system-for-outlander-1481.html 2 pages (Nov. 24, 2009).
Non-Final Office Action issued on Nov. 9, 2012, in connection with U.S. Appl. No. 29/405,416, 9 pages.
Restriction Requirement issued on Oct. 25, 2019 in connection with U.S. Appl. No. 15/555,148, 7 pages.
Non-final Office Action issued on Oct. 5, 2020, in connection with U.S. Appl. No. 15/724,733, 7 pages.
Non-Final Office Action issued on Sep. 22, 2014 in connection with U.S. Appl. No. 13/325,796, 25 pages.
Non-Final Office Action issued on Sep. 29, 2014 in connection with U.S. Appl. No. 13/325,783, 27 pages.
Notice of Allowance issued on Apr. 2, 2020, in connection with U.S. Appl. No. 15/724,733, 5 pages.
Notice of Allowance issued on Dec. 17, 2018 in connection with U.S. Appl. No. 15/454,881, 5 pages.
Notice of Allowance issued on Feb. 15, 2019 in connection with U.S. Appl. No. 15/139,572, 7 pages.
Notice of Allowance issued on Jan. 11, 2016 in connection with U.S. Appl. No. 13/326,278, 7 pages.
Notice of Allowance issued on Jan. 19, 2018 in connection with U.S. Appl. No. 14/665,075, 8 pages.
Notice of Allowance issued on Jan. 21, 2015 in connection with U.S. Appl. No. 13/326,110, 6 pages.
European Extended Search Report issued on Mar. 20, 2019 in connection with European Patent Application No. 16758408.5, 8 pages.
International Search Report issued on Jun. 29, 2016 in connection with International Patent Application No. PCT/CA2016/050234, 3 pages.
Written Opinion issued on Jun. 29, 2016 in connection with International Patent Application No. PCT/CA2016/050234, 3 pages.
Related Publications (1)
Number Date Country
20210197625 A1 Jul 2021 US
Provisional Applications (2)
Number Date Country
62861684 Jun 2019 US
62724853 Aug 2018 US