This disclosure relates generally to tracked vehicles (e.g., agricultural vehicles or other industrial vehicles, etc.) and, more particularly, to monitoring track systems of such vehicles.
Off-road vehicles, including industrial vehicles such as agricultural vehicles (e.g., tractors, harvesters, combines, etc.), construction vehicles (e.g., loaders, excavators, bulldozers, etc.), and forestry vehicles (e.g., feller-bunchers, tree chippers, knuckleboom loaders, etc.), military vehicles (e.g., combat engineering vehicles (CEVs), etc.), snowmobiles, and all-terrain vehicles (ATVs), are used on soft, slippery and/or irregular grounds (e.g., soil, mud, sand, ice, snow, etc.) for work and/or other purposes. To enhance their traction and floatation on such grounds, certain off-road vehicles are equipped with track systems. In some cases, off-road vehicles may also be operable on paved roads.
For example, agricultural vehicles can travel in agricultural fields to perform agricultural work and possibly on paved roads (e.g., to travel between agricultural fields). Numerous factors affect performance of the agricultural vehicles and efficiency of agricultural work they do, including their components (e.g., track systems) and their environments (e.g., grounds on which they operate). While some of these factors may be managed by users (e.g., operators) of the agricultural vehicles, this may lead to suboptimal agricultural work, greater wear or other deterioration of components of the agricultural vehicles, and/or other issues in some cases.
Similar considerations may arise in relation to other off-road vehicles (e.g., construction vehicles, snowmobiles, ATVs, etc.) in some cases.
For these and other reasons, there is a need to improve monitoring tracks and track systems of off-road vehicles.
In accordance with various aspects of this disclosure, a vehicle (e.g., an agricultural vehicle or other off-road vehicle) comprising a track system can be monitored to obtain information regarding the vehicle, including information regarding the track system, such as an indication of a physical state (e.g., wear, damage and/or other deterioration) of a track and/or other component of the track system based on at least one image of the track and/or other component of the track system, respectively, which can be used for various purposes, such as, for example, to: convey the information to a user (e.g., an operator of the vehicle); control the vehicle (e.g., a speed of the vehicle, operation of a work implement, etc.); transmit the information to a remote party (e.g., a provider such as a manufacturer or distributor of the track system, the track and/or another component thereof, and/or of the vehicle; a service provider for servicing (e.g., maintenance or repair of) the track system, the track and/or another component thereof, etc.).
In accordance with an aspect, this disclosure relates to a system for monitoring a track for traction of a vehicle. The system comprises an interface configured to receive data regarding at least one image of the track. The system also comprises a processor configured to process the data regarding the at least one image of the track to obtain an indication of a physical state of the track, and to generate a signal based on the indication of the physical state of the track.
In accordance with another aspect, this disclosure relates to a method of monitoring a track for traction of a vehicle. The method comprises receiving data regarding at least one image of the track. The method also comprises processing the data regarding the at least one image of the track to obtain an indication of a physical state of the track. The method also comprises generating a signal based on the indication of the physical state of the track.
In accordance with yet another aspect, this disclosure relates to a system for monitoring a component of a track system for traction of a vehicle. The system comprises an interface configured to receive data regarding at least one image of the component of the track system. The system also comprises a processor configured to process the data regarding the at least one image of the component of the track system to obtain an indication of a physical state of the component of the track system. The processor is also configured to generate a signal based on the indication of the physical state of the component of the track system.
In accordance with yet another aspect, this disclosure relates to method of monitoring a component of a track system for traction of a vehicle. The method comprises receiving data regarding at least one image of the component of the track system. The method also comprises processing the data regarding the at least one image of the component of the track system to obtain an indication of a physical state of the component of the track system. The method also comprises generating a signal based on the indication of the physical state of the component of the track system.
In accordance with yet another aspect, this disclosure relates to a track system monitoring system. The system comprises an image data capture device configured to capture image data relating to a track system component. The system also comprises an image processing device, in data communication with the image capture device. The image processing device is configured to receive captured image data from the image data capture device and process the captured image data to determine at least one physical characteristic of the track system component.
In accordance with yet another aspect, this disclosure relates to track system monitoring system. The system comprises a 3D scanning device configured to generate a 3D scan relating to a track system component. The system also comprises a processing device, in data communication with the 3D scanning device. The processing device is configured to receive the 3D scan from the 3D scanning device and process the 3D scan to determine at least one physical characteristic of the track system component.
In accordance with yet another aspect, this disclosure relates to a track system monitoring system. The system comprises an image data capture device configured to capture image data relating to a track system component. The system also comprises an image processing device in data communication with the image capture device. The image processing device is configured to receive captured image data from the image data capture device and generate a 3D model of at least a portion of the track system component based on the image data. The image processing device is also configured to compare the 3D model to at least one known 3D model of a track system component to determine at least one aspect of the physical state of the track system component.
A detailed description of embodiments is provided below, by way of example only, with reference to accompanying drawings, in which:
In this embodiment, the vehicle 10 comprises a frame 11, a powertrain 15, a steering mechanism 18, a suspension 24, and an operator cabin 20 that enable a user to move the vehicle 10 on the ground, including on an agricultural field and possibly on a paved road (e.g., between agricultural fields), using the track systems 161-164 and perform work using a work implement 13.
As further discussed later, in this embodiment, the agricultural vehicle 10, including the track systems 161-164, can be monitored (e.g., while the agricultural vehicle 10 is parked, inspected or otherwise at rest and/or during operation of the agricultural vehicle 10) to obtain information regarding the agricultural vehicle 10, including information regarding the track systems 161-164, such as indications of physical states of tracks and/or other components of the track systems 161-164 (e.g., information indicative of wear or other degradation thereof) that is derivable from one or more images of the tracks and/or other components of the track systems 161-164, which can be used for various purposes, such as, for example, to: convey the information to a user (e.g., the operator); control the agricultural vehicle 10 (e.g., a speed of the agricultural vehicle 10, operation of the work implement 13, etc.); transmit the information to a remote party (e.g., a provider such as a manufacturer or distributor of the track systems 161-164 or their tracks or other components and/or of the agricultural vehicle 10; a service provider for servicing (e.g., maintenance or repair of) the track system, the track and/or another component thereof, etc.); etc. This may be useful, for example, to gain knowledge about the agricultural vehicle 10, the track systems 161-164, and/or their environment to enhance efficiency of agricultural work performed by the agricultural vehicle 10 and to help prevent excessive wear or other deterioration of the track systems 161-164, to schedule maintenance or replacement of the track system 161-164 or individual components thereof, to effectively manage the wear of the track system 161-164 or individual components thereof, for the agricultural vehicle 10 or a fleet of such agricultural vehicles, to achieve any of various other outcomes herein described, and/or for various other reasons.
The powertrain 15 is configured to generate power for the agricultural vehicle 10, including motive power for the track systems 161-164 to propel the vehicle 10 on the ground. To that end, the powertrain 15 comprises a power source 14 (e.g., a primer mover) that includes one or more motors. For example, in this embodiment, the power source 14 comprises an internal combustion engine. In other embodiments, the power source 14 may comprise another type of motor (e.g., an electric motor) or a combination of different types of motor (e.g., an internal combustion engine and an electric motor). The powertrain 15 can transmit power from the power source 14 to one or more of the track systems 161-164 in any suitable way (e.g., via a transmission, a differential, a direct connection, and/or any other suitable mechanism). In some embodiments, at least part of the powertrain 15 (e.g., a motor and/or a transmission) may be part of one or more of the track systems 161-164.
The operator cabin 20 is where the user sits and controls the vehicle 10. More particularly, the operator cabin 20 comprises a user interface 70 allowing the user to steer the vehicle 10 on the ground, operate the work implement 13, and control other aspects of the vehicle 10. In this embodiment, the user interface 70 comprises input devices, such as an accelerator, a brake control, and a steering device (e.g., a steering wheel, a stick, etc.) that are operated by the user to control motion of the vehicle 10 on the ground. The user interface 70 also comprises output devices such as an instrument panel (e.g., a dashboard) which provides indicators (e.g., a speedometer indicator, a tachometer indicator, etc.) to convey information to the user.
The work implement 13 is used to perform agricultural work. For example, in some embodiments, the work implement 13 may include a combine head, a cutter, a scraper pan, a tool bar, a planter, or any other type of agricultural work implement.
The track systems 161-164 engage the ground to provide traction to the vehicle 10. More particularly, in this embodiment, front ones of the track systems 161-164 provide front traction to the vehicle 10, while rear ones of the track systems 161-164 provide rear traction to the vehicle 10.
In this embodiment, each of the front ones of the track systems 161-164 is pivotable relative to the frame 11 of the vehicle 10 about a steering axis 19 by the steering mechanism 18 (e.g., in response to input of the user at the steering device of the user interface 70) to change the orientation of that track system relative to the frame 11 in order to steer the vehicle 10 on the ground. The orientation of each of the front ones of the track systems 161-164 relative to a longitudinal axis 33 of the vehicle 10, which defines a steering angle θ of that track system, is thus changeable. In this example, the steering mechanism 18 includes a steering unit 34 (e.g., comprising a steering knuckle) on each side of the vehicle 10 dedicated to each of the front ones of the track systems 161-164 and defining the steering axis 19 for that track system. Each of the front ones of the track systems 161-164 is therefore steerable.
With additional reference to
The track 41 engages the ground to provide traction to the vehicle 10. A length of the track 41 allows the track 41 to be mounted around the track-engaging assembly 17. In view of its closed configuration without ends that allows it to be disposed and moved around the track-engaging assembly 17, the track 41 can be referred to as an “endless” track. Referring additionally to
The track 41 is elastomeric, i.e., comprises elastomeric material, allowing it to flex around the wheels 42, 501-508. The elastomeric material of the track 41 can include any polymeric material with suitable elasticity. In this embodiment, the elastomeric material includes rubber. Various rubber compounds may be used and, in some cases, different rubber compounds may be present in different areas of the track 41. In other embodiments, the elastomeric material of the track 41 may include another elastomer in addition to or instead of rubber (e.g., polyurethane elastomer). The track 41 can be molded into shape in a mold by a molding process during which its elastomeric material is cured.
More particularly, the track 41 comprises an elastomeric belt-shaped body 36 underlying its inner side 45 and its ground-engaging outer side 47. In view of its underlying nature, the body 36 can be referred to as a “carcass”. The carcass 36 comprises elastomeric material 37 which allows the track 41 to flex around the wheels 42, 501-508.
In this embodiment, the carcass 36 comprises a plurality of reinforcements embedded in its elastomeric material 37. One example of a reinforcement is a layer of reinforcing cables 381-38C that are adjacent to one another and that extend in the longitudinal direction of the track 41 to enhance strength in tension of the track 41 along its longitudinal direction. In some cases, a reinforcing cable may be a cord or wire rope including a plurality of strands or wires. In other cases, a reinforcing cable may be another type of cable and may be made of any material suitably flexible longitudinally (e.g., fibers or wires of metal, plastic or composite material). Another example of a reinforcement is a layer of reinforcing fabric 40. Reinforcing fabric comprises pliable material made usually by weaving, felting, or knitting natural or synthetic fibers. For instance, a layer of reinforcing fabric may comprise a ply of reinforcing woven fibers (e.g., nylon fibers or other synthetic fibers). Various other types of reinforcements may be provided in the carcass 36 in other embodiments.
The carcass 36 may be molded into shape in the track's molding process during which its elastomeric material 37 is cured. For example, in this embodiment, layers of elastomeric material providing the elastomeric material 37 of the carcass 36, the reinforcing cables 381-38C and the layer of reinforcing fabric 40 may be placed into the mold and consolidated during molding.
In this embodiment, the inner side 45 of the track 41 comprises an inner surface 32 of the carcass 36 and a plurality of wheel-contacting projections 481-48N that project from the inner surface 32 to contact at least some of the wheels 42, 501-508 and that are used to do at least one of driving (i.e., imparting motion to) the track 41 and guiding the track 41. In that sense, the wheel-contacting projections 481-48N can be referred to as “drive/guide projections”, meaning that each drive/guide projection is used to do at least one of driving the track 41 and guiding the track 41. Also, such drive/guide projections are sometimes referred to as “drive/guide lugs” and will thus be referred to as such herein. More particularly, in this embodiment, the drive/guide lugs 481-48N interact with the drive wheel 42 in order to cause the track 41 to be driven, and also interact with the idler wheels 501-508 in order to guide the track 41 as it is driven by the drive wheel 42. The drive/guide lugs 481-48N are thus used to both drive the track 41 and guide the track 41 in this embodiment.
The drive/guide lugs 481-48N are spaced apart along the longitudinal direction of the track 41. In this case, the drive/guide lugs 481-48N are arranged in a plurality of rows that are spaced apart along the widthwise direction of the track 41. The drive/guide lugs 481-48N may be arranged in other manners in other embodiments (e.g., a single row or more than two rows). Each of the drive/guide lugs 481-48N is an elastomeric drive/guide lug in that it comprises elastomeric material 68. The drive/guide lugs 481-48N can be provided and connected to the carcass 36 in the mold during the track's molding process.
The ground-engaging outer side 47 of the track 41 comprises a ground-engaging outer surface 31 of the carcass 36 and a plurality of traction projections 611-61M that project from the outer surface 31 and engage and may penetrate into the ground to enhance traction. The traction projections 611-61M, which can sometimes be referred to as “traction lugs”, are spaced apart in the longitudinal direction of the track system 16i. The ground-engaging outer side 47 comprises a plurality of traction-projection-free areas 711-71F (i.e., areas free of traction projections) between successive ones of the traction projections 611-61M. In this example, each of the traction projections 611-61M is an elastomeric traction projection in that it comprises elastomeric material 69. The traction projections 611-61M can be provided and connected to the carcass 36 in the mold during the track's molding process.
The track 41 may be constructed in various other ways in other embodiments. For example, in some embodiments, the track 41 may comprise a plurality of parts (e.g., rubber sections) interconnected to one another in a closed configuration, the track 41 may have recesses or holes that interact with the drive wheel 42 in order to cause the track 41 to be driven (e.g., in which case the drive/guide lugs 481-48N may be used only to guide the track 41 without being used to drive the track 41), and/or the ground-engaging outer side 47 of the track 41 may comprise various patterns of traction projections.
The drive wheel 42 is rotatable about an axis of rotation 49 for driving the track 41 in response to rotation of an axle of the vehicle 10. In this example, the axis of rotation 49 corresponds to the axle of the vehicle 10. More particularly, in this example, the drive wheel 42 has a hub which is mounted to the axle of the vehicle 10 such that power generated by the power source 14 and delivered over the powertrain 15 of the vehicle 10 rotates the axle, which rotates the drive wheel 42, which imparts motion of the track 41.
In this embodiment, the drive wheel 42 comprises a drive sprocket engaging the drive/guide lugs 481-48N of the inner side 45 of the track 41 in order to drive the track 41. In this case, the drive sprocket 42 comprises a plurality of drive members 461-46T (e.g., bars, teeth, etc.) distributed circumferentially of the drive sprocket 42 to define a plurality of lug-receiving spaces therebetween that receive the drive/guide lugs 481-48N of the track 41. The drive wheel 42 may be configured in various other ways in other embodiments. For example, in embodiments where the track 41 comprises recesses or holes, the drive wheel 42 may have teeth that enter these recesses or holes in order to drive the track 41. As yet another example, in some embodiments, the drive wheel 42 may frictionally engage the inner side 45 of the track 41 in order to frictionally drive the track 41.
The idler wheels 501-508 are not driven by power supplied by the powertrain 15, but are rather used to do at least one of supporting part of a weight of the vehicle 10 on the ground via the track 41, guiding the track 41 as it is driven by the drive wheel 42, and tensioning the track 41. More particularly, in this embodiment, the leading and trailing idler wheels 501, 502, 507, 508 maintain the track 41 in tension, and can help to support part of the weight of the vehicle 10 on the ground via the track 41. The roller wheels 503-506 roll on the inner side 45 of the track 41 along the bottom run 66 of the track 41 to apply the bottom run 66 on the ground. The idler wheels 501-508 may be arranged in other configurations and/or the track system 16i may comprise more or less idler wheels in other embodiments.
The frame 44 of the track system 16i supports components of the track system 16i, including the idler wheels 501-508. More particularly, in this embodiment, the front idler wheels 501, 502 are mounted to the frame 44 in a front longitudinal end region of the frame 44 proximate the front longitudinal end 57 of the track system 16i, while the rear idler wheels 507, 508 are mounted to the frame 44 in a rear longitudinal end region of the frame 44 proximate the rear longitudinal end 59 of the track system 16i. The roller wheels 503-506 are mounted to the frame 44 in a central region of the frame 44 between the front idler wheels 501, 502 and the rear idler wheels 507, 508. Each of the roller wheels 503-506 may be rotatably mounted directly to the frame 44 or may be rotatably mounted to a link which is pivotally mounted to the frame 44 to which is rotatably mounted an adjacent one of the roller wheels 503-506 (e.g., forming a “tandem”).
The frame 44 of the track system 16i is supported at a support area 39. More specifically, in this embodiment, the frame 44 is supported by the axle of the vehicle 10 to which is coupled the drive wheel 42, such that the support area 39 is intersected by the axis of rotation 49 of the drive wheel 42.
In this example of implementation, the track system 16i comprises a tensioner 93 for tensioning the track 41. For instance, in this embodiment, the tensioner 93 comprises an actuator (e.g., a hydraulic actuator) mounted at one end to the frame 44 of the track system 16i and at another end to a hub of the leading idler wheels 501, 502. This allows the tensioner 93 to modify a distance between the front idler wheels 501, 502 and the rear idler wheels 507, 508 in the longitudinal direction of the track system 16i.
In some examples, the electronic device 501 can be a smartphone, a tablet, a smartwatch, a computer, etc., of a user, who may be the operator of the vehicle or another person having access to the vehicle. In other examples, the electronic device 501 may be integrated with the vehicle.
In some embodiments, the image processing entity 505 can be an application running on a server. In other embodiments, the image processing entity 505 can be a dedicated network appliance. In yet other embodiments, the image processing entity 505 may be an application running on the electronic device 501. In the embodiment of
Computing modules 508x may for example be implemented using known computer vision products, such as, AutoML Vision™ and/or Vision API™, each provided by Google™. In other embodiments, computing modules 508x may comprise standalone AI or machine-learning solutions forming part of image processing entity 505. As defined herein, AI refers to some implementation of artificial intelligence and/or machine learning (e.g., heuristics, support vector machines, artificial neural networks, convolutional neural networks, etc.) in software, hardware or some combination of both.
In some embodiments, complex algorithms, like artificial intelligence, are used to categorize what may be considered uncategorizable data. For example, the system 100 can be configured for generating conclusions about a physical state of a track based on one or more images of the track itself. This analysis can include whether or not there is a defect in the track, according to some embodiments. In some embodiments, this can include indications as to the physical state of the track and/or useful life remaining. As will be described below, a machine learning algorithm may be trained to identify a defect or other characteristic in a track by way of image analysis.
In some embodiments, computing modules 508x are first taught how to identify parameters in a training mode (sometimes referred to as supervised learning mode). This is done by analyzing a given set of values, making quantitative comparisons, and cross-referencing conclusions with known results. Iterative refinement of these analyses and comparisons allows an algorithm to achieve greater predictive certainty. This process is continued iteratively until the solution converges or reaches a desired accuracy.
In this embodiment, computing modules 508x can compare image data for a given track to a previously-analyzed mass of known data. When placed in a supervised learning mode, information can be generated from already populated track data provided to the computing modules 508x. For example, this data could contain images of tracks, along with determinations of the remaining life of the tracks. In other words, in the supervised learning mode, both the inputs and the outputs are provided to the system 100. The system 100 can process the given inputs and compare the calculated outputs according to its algorithm to the provided outputs. Based on this comparison, the system 100 can determine a metric to represent the percentage of error between calculated and provided outputs. Using this error metric, the system 100 can adjust its method of calculating an output. During training, the system 100 can continuously repeat analysis of different inputs and provided outputs in order to fine-tune its method of determining track information.
In some embodiments, while the computing modules 508x may require initial supervised learning, as the computing modules 508x continue to gain access to data, they may be able to further refine their predictive analytics based on new inputs. For example, if a user is able to confirm that an assessment (e.g. broken/exposed reinforcing cables 381-38C) or prediction (e.g. 6 months of use left in a given track) made by the system 100 is/was incorrect, the user can upload to the system 100 what the correct conclusion/prediction was. This allows the computing modules 508x to continue to improve accuracy in their analysis.
In some embodiments, multiple computing modules 508x can be configured to determine different characteristics of a given track. Each of these modules can offer a different analysis for a given input. The processor may direct these modules to be used independently or concurrently based on an operational parameter determined by a given user. For example, the system 100 may use a different analytical technique to determine track life compared to drive wheel misalignment. Based on an image communicated to the system 100 from an electronic device, the system 100 may analyze a given for track life, drive wheel misalignment, or other forms of wear and/or damage.
In some embodiments, the computing modules 508x are configured to assess a level of wear, damage and/or other deterioration of the track 41 or other track system component. For example, a computing module 508x can be configured to determine that the traction projections 611-61M are worn to 30% of the level of wear that would require replacement of the track. In some embodiments, the computing modules 508x are configured to assess the nature of damage to the track 41 or other track system component. For example, a computing module 508x can be configured to determine that a midroller (or any other track system component, such as a sprocket) is damaged or missing.
In some embodiments, the computing modules 508x are further configured to predict the cause of the wear and/or damage to the track 41 or other track system component. In one specific example, a computing module 5081 is configured to predict whether a specific wear pattern of the elastomeric material of a track 41 is caused by a misaligned drive wheel. In another specific example, a computing module 5082 is configured to predict whether a specific wear pattern of the elastomeric material of a traction projections 611-61M is caused by excessive roading (i.e. traversing a paved road). In another specific example, another computing module 5083 is configured to predict whether a specific wear pattern of the track (e.g. the abnormal relative position of two adjoining track links) is caused by a broken reinforcing cable 381-38C. As will be appreciated, each computing module 508x can be implemented using a combination of deep learning, supervised or unsupervised machine learning, image recognition and/or machine vision.
In some embodiments, the system 100 is configured to capture one or more 2D images to detect specific patterns of wear and/or damage. For example, the system 100 may be configured to implement one or more computer vision (CV) models to detect specific visible wear/damage features. Examples of such visible wear/damage features include, but are not limited to, broken and/or exposed reinforcing cables 381-38C, linear recesses in the carcass 36 caused by delamination and changes in the shape of drive wheel 42 (sprocket) teeth, evidencing sprocket tooth wear caused by debris and/or normal engagement with drive/guide lugs 481-48N.
In some embodiments, the image processing system 500 may produce a three-dimensional (3D) scan to generate a 3D model of at least part of the track 41 or other track system component. For example, in some embodiments, the image data received by the electronic device 501 or any other image capture means are processed by way of photogrammetry in order to create the 3D model of the track 41 and/or track component. In some embodiments, as described in more detail below, laser line scanners are instead used to generate the 3D model of the track 41 and/or track component.
Such precise 3D models can be compared to 3D models of unworn and/or undamaged tracks in order to precisely measure wear, damage and/or other deterioration. For example, by comparing the 3D model of a worn track 41 to the 3D model of a new, unworn track, it is possible to precisely measure a volumetric loss of material of the worn track 41, and thereby assess the wear and/or other deterioration of the worn track 41, very precisely.
With reference to
As described above, in some embodiments, the system 100 can implement a 2D recognition technique. In some embodiments, the system 100 can implement a 3D recognition technique. In some embodiments, the system 100 can implement a combination of a 2D recognition technique and a 3D recognition technique.
In some embodiments, the 3D recognition technique used is based on generating a 3D model using a point cloud. For example, as shown in
2D recognition techniques include four basic steps, namely image acquisition, image processing, feature extraction and classification. Such techniques include, but are not limited to, Optical Character Recognition (OCR), feature detection, image gradient analysis, pattern recognition algorithms and feature/pattern classification algorithms.
In some embodiments, the system 100 can be configured to implement the method of
Then at step 5205, the system 100 can perform feature classification. In some embodiments, feature classification can include, but is not limited to, the use of nearest neighbor classification, cascading classifiers, neural networks, statistical classification techniques and/or Bayesian classification techniques. Once the features have been classified, it is possible to separate, at step 5206, features which represent undamaged/unused parts of the track system component, and features (e.g. cracks, exposed cables, etc.) which represent patterns of wear or damage. Once features relating to patterns of wear or damage have been detected, it is possible for the system 100 to perform further feature classification on the wear or damage pattern.
As shown in
As shown in the method of
As shown in
In some embodiment, the cause and/or nature of the wear and/or damage of the track 41, or other track system component, can be established by the system 100 performing a volumetric comparison of the 3D model 55 of a used and/or damaged track and a 3D model 77 of an unused and undamaged track.
For example, as shown in
As shown in
As shown in
As shown in
For example, as shown in
As shown in
As shown in
As described above, and as shown in
Once the computing modules 508x has determined the cause, level and/or nature of the wear and/or damage of the track 41 or other track system component, the image processing entity 505 may send data relating to the cause, level and/or nature of the wear and/or damage of the track 41 or other track system component back to electronic device 501 for further processing and/or notification to a user. By using this information, electronic device 501 may determine that an event arising from usage of a track system 16x, such as a usage threshold event (e.g. an amount of tread wear, an amount of time such as a number of hours the track 41 has been used), wear threshold event (e.g. the number of exposed reinforcing cables caused by chunking) and/or damage event (e.g. one or more severed reinforcing cables), has occurred.
According to some embodiments, the computing modules 508x may have access to information stored elsewhere on the internet. For example, the computing modules 508x may be configured to query databases stored on external servers by sending requests over the network in order to analyze the image based on pertinent cross-referential data. This may include weather, humidity, or information about the vehicle or track that can be periodically updated.
The electronic device 501 may include elements such as a processor, a memory, a display, a data input module, and a network interface. The electronic device 501 may include other components, but these have been omitted for the sake of brevity. In operation, the electronic device 501 is configured to perform the operations described herein. The electronic device 501 processor may be configured to execute instructions stored in memory. The instructions, when executed, cause the electronic device 501 to perform the operations described herein. In some embodiments, the instructions may be part of a software application downloaded into memory by the electronic device 501. Alternatively, some or all of the functionality described herein may be implemented using dedicated circuitry, such as an ASIC, a GPU, or a programmed FPGA for performing the operations of the processor.
In some embodiments, an application (“app”, i.e., software) may be installed on the electronic device 501 to interact with the system server 1142 and or the vehicle 10. For example, in some embodiments, such as where the electronic device 501 is a smartphone, a tablet, a computer, etc., the user (e.g., the operator) may download the app from a repository (e.g., Apple's App Store, iTunes, Google Play, Android Market, etc.) or any other website onto the electronic device 501. Upon activation of the app on the electronic device 501, the user may access certain features relating to the system server 1142 and/or the vehicle 10 locally on the electronic device 501.
In operation, a user can use the electronic device 501 to generate data about the vehicle 10. For example, for embodiments where the electronic device is a smart phone equipped with a camera, the user can take one or more images of a track 41 of the vehicle 10. The system 100 may then take the image data captured by the electronic device 501 and transmit the image data over a network 124 to a system server 1142.
According to some embodiments, the electronic device 501 may be a portable electronic device with multiple uses such as a mobile phone, tablet or laptop. According to other embodiments, the electronic device may be a single-use electronic device, such that the device is designed to only be used in operation with the system 100. Further, the electronic device 501 may also be capable of establishing a communicable link with an accessory device. This communicable link be may be wireless, wired, or partly wireless and partly wired (e.g., Bluetooth or other short-range or near-field wireless connection, WiFi or other wireless LAN, WiMAX or other wireless WAN, cellular, Universal Serial Bus (USB), etc.).
According to other embodiments, the electronic device 501 may integrated into an internal computer 1342 in the off-road vehicle (as shown in
The internal sensor network 1350 can include sensors to provide information about the vehicle or the track of the vehicle. For example, this may include a camera positioned to take images of the track. In some embodiments where the electronic device is integrated into an internal computer in the off-road vehicle, the system 100 may be configured to continuously monitor the track. This can be achieved by continuously capturing data, for example, images of the vehicle track, at various intervals. The electronic device 501 can then automatically upload the data over the network 124 to the system server 1142 for image processing. After processing, the image processing entity 505 can automatically communicate over the network 124 if a fault state has been determined.
The electronic device 501 can also send additional data to the image processing entity 505 over the network 124. For example, this can include (but is not limited to) GPS location, date and time, or any information from an onboard computer within the vehicle. This data can be cross-referenced and analyzed within the computing modules 508x. For example, given GPS and date and time data, the AI module can access the specific weather and weather history for the vehicle location. In some embodiments, such information may be used in, for example, determining the end-of-life of a track (i.e. the amount of time until a track is expected to fail or until the likelihood of track failure rises above a predetermined threshold).
This may be achieved by a separate electronic device 501 being communicably linked to an internal computer 1342 of a vehicle 10. The internal computer 1342 may periodically receive and record information relating to the vehicle 10 and/or track systems 161-164 determined by the internal sensor network 1350. For example, the internal sensor network 1350 may include an image taken of the track or information about the vehicle 10, such as the speed of the vehicle 10.
According to some embodiments, the electronic device 501 may communicate a unique identifier for a specific track under inspection. In some embodiments, the unique identifier can be a serial number of the track. This allows the server 1146 and/or internal computer 1342 to catalog the inspection and produce a history of a given track. According to some embodiments, the internal computer 1342 and/or the server 1146 may store data about the serial numbers of the tracks installed on the vehicle 10.
According to some embodiments, the electronic device 501 may be capable of determining a serial number from a track based on an image of the track. This can be done by the electronic device 501 capturing an image of an embossed serial number on a surface of the track, and using the image processing entity 505 to determine the specific characters of the serial number. This can be cross-referenced with a database stored in server memory 1146 (or otherwise accessible by system server 1142) to determine elements such as the model and date of manufacture of the track.
Serial number analysis may be performed using AI techniques employed by the computing modules 508x, may be performed using techniques such as optical character recognition (OCR), or a combination thereof. These techniques may include preprocessing of an image in order to improve the ability to analyze the target components, such as de-skewing, layout analysis, and binarization. In some embodiments, a track system and/or track system component (such as a track) can be identified by way of another marking or tag suitable for communicating information relating to the track system and/or track system component. Such markings or tags can include, but are not limited to, barcodes, Quick Response (QR) codes or other matrix barcodes and Radio Frequency Identification (RFID) tags.
Another method of track identification that can be performed by the electronic device 501 is track pattern recognition. The electronic device 501 may be configured to analyze the tread pattern and measure track width to determine a number of characteristics about the track. The electronic device 501 may then send this data and information to the system server 1142 for further data analysis to identify the type of track. The type of track may be a track brand, model number, or any other suitable information capable of identifying a track.
According to some embodiments, the vehicle may be capable of communicating all the necessary data over the network without the use of an external electronic device 501 such as a mobile phone. For example, the vehicle 10 may be equipped with a network interface capable of independently communicating with the system server 1142 over the network 124.
According to some embodiments, a system server 1142 hosts the image processing entity 505. Server processor 1144 is able to access instructions stored in the memory 1146 that can initialize the image processing entity 505. This initialization can include operational parameters that include which AI module 508x to use.
Image processing entity 505 can store instructions relating to a specific AI module 508x within the memory 1146. The processor 1144 may instruct the server to save the data received from the network via the network interface 1148 in memory 1146. The processor 1144 may analyze the data in memory 1146 and determine information about the track 41. Based on the data analysis, the processor 1144 may send a communication to the electronic device 501 over the network 124 via the network interface 1148.
In operation, based on the analysis determined by the image processing entity 505, the system server 1142 may communicate with the user via an electronic device 501 or the vehicle 10, a dispatch center 1102, a service center 1104, or a parts supplier 1106. The system 100 may also communicate with any combination of these, or any other suitable device registered within the system 100. This communication can contain information such as that indicating the determination of track wear and/or damage concluded by the image processing entity 505. Based on this information, the dispatch center 1502 or user may schedule maintenance with the service center 1104. Based on the conclusion on track wear and/or damage (for example, that the track needs to be replaced) and vehicle information (track type, vehicle type) available, the system 100 can determine the amount of time required or parts available at the service center 1104 and facilitate scheduling a maintenance appointment or a shipment from the parts supplier 1106. This can be done by maintaining a database of inventory at the service center, along with a calendar.
The disclosed embodiments of database representations are structured merely by way of example for illustrative purposes, and a skilled reader would know that these visual representations can be changed to include more or less information available to the system 100.
In operation, a user can use the electronic device 501 to generate image data relating to the track 41 and/or track system 16x of the vehicle 10. According to some embodiments, the electronic device 501 may also access internal information stored on the vehicle onboard computer 1342. The electronic device 501 may then communicate both the data captured and the information retrieved by the electronic device 501 over the network 124 to the system server 1142 to be stored in memory 1146. Using both the data captured and the information retrieved the processor 1144 may determine information about the track 41. Based on the data analysis, the processor 1144 may send a communication to the electronic device 501 over the network 124 via the network interface 1148.
At step 1301, the system 100 determines that an event arising from use of a track system 16x, such as a usage threshold event (e.g. an amount of time such as a number of hours the track 41 has been used) or a deterioration threshold event (e.g. chunking or other loss of elastomeric material of the track, the number of exposed reinforcing cables, one or more severed reinforcing cables, etc.), has occurred. As described above, the system 100 can make these determinations by analysis of the images taken by the image capture devices described above.
At step 1302, the system 100 identifies the track system component for which the usage threshold event or deterioration threshold event has occurred. In some embodiments, the track system component information and information relating to the usage threshold event and deterioration threshold event is conveyed to the operator of the vehicle by the system 100 in order to facilitate scheduling of track system component servicing and/or other maintenance.
For purposes of this example, it is assumed that the usage threshold event or deterioration threshold event is for the track 41.
For example, the system 100 may issue a notification conveying this information to the operator via the user interface of the operator cabin 20 of the vehicle 10 and/or the electronic device 501. According to embodiments wherein the electronic device 501 is a mobile phone, this could be in the form of a push notification sent to the app over the network 124. In other embodiments, the system 100 conveys the track system component information and information relating to the usage threshold event and deterioration threshold event to an organization providing maintenance services. For example, the system 100 may issue a notification conveying this information to a system server 1142 associated with the organization via a network 124 (e.g. which may be implemented by the Internet, a cellular connection, and/or any other network infrastructure). Once the information is received, the organization can schedule maintenance of the vehicle at step 1303, and subsequently replace or repair the track system component. Accordingly, track system component maintenance operations can be initiated and scheduled without the need for input from the vehicle operator.
As shown in
This can lead to significant savings in term of vehicle downtime and logistics. For example, at step 1401, the system 100 determines that an event arising from usage of a track system 16x, such as a usage threshold event (e.g. an amount of tread wear, an amount of time such as a number of hours the track 41 has been used), deterioration threshold event (e.g. the number of exposed reinforcing cables) and/or deterioration event (e.g. one or more severed reinforcing cables), has occurred. At step 1402, the system 100 identifies the track system component for which the usage threshold event, deterioration threshold event and/or deterioration event has occurred. At step 1403, vehicle location information relating to the geographic location of the vehicle is determined. This can be achieved by any suitable means including, but not limited to, Global Positioning System (GPS) receivers. In some embodiment, the system 100 conveys the track system component information, vehicle location information and information relating to the usage threshold event, deterioration threshold event and/or deterioration event to the track-as-a-service organization.
As shown in above, the system 100 may communicate with the system server 1142 of the track-as-a-service organization over a network 124 (e.g. which may be implemented by the Internet, a cellular connection, and/or any other network infrastructure). Then, at step 1404, the track-as-a-service organization ships a replacement track system component to a location related to the geographic location of the vehicle. For example, the track-as-a-service location could ship the replacement track system component to the nearest maintenance service dispatch location or third party maintenance organization. At step 1405, the track-as-a-service organization can schedule a maintenance of the track system. In some embodiments, the track-as-a-service organization schedules a third party mobile maintenance team to perform onsite maintenance based on the geographic location of the vehicle. Finally, at step 1406, the track-as-a-service organization, or an agent thereof, replaces the track system component. In some embodiments, this can be performed onsite, based at least in part on the vehicle location information received from the track-as-a-service organization.
At step 1503, the system 100 will query the memory to determine if the specific user has a purchase preference stored in the system 100. If the system 100 has a purchase preference stored for the given user, the system 100 will order the track system component for replacement based on the saved preference at step 1506. If the system 100 does not find a purchase preference for the given user, the system 100 may send a communication to the user's electronic device 501 with information indicating the part purchase options and information about the parts (for example the various options of price and part characteristics). The system 100 may also send a communication instructing the electronic device 501 to prompt the user to store a purchase preference. Based on this information, the system 100 will order the track system component at step 1507.
At step 1509, the system 100 may schedule maintenance with a given service center or technician. At this step user preferences may also be considered. For example, a user may be able to store in their profile a preference for scheduling. This may include a preference for the first available time to service the vehicle. Alternatively, a fleet manager may try and coordinate scheduling of maintenance with other vehicles within a fleet. This could include wanting all vehicles to be serviced at the same time, or to stagger vehicle services. Scheduling preferences may also include a time of day preference for the user to have maintenance scheduled. Based on these preferences, the user may be automatically scheduled for maintenance.
According to other embodiments, the system 100 may prompt the user via a date and time entry interface, such as a calendar interface, on the electronic device 501 to input a date and time for maintenance. Based on this input data, the system 100 can schedule maintenance with a technician or service center.
Finally, at step 1520, the track-as-a-service organization, or an agent thereof, replaces the track system component. In some embodiments, this can be performed onsite, based at least in part on the vehicle location information received from the track-as-a-service organization.
Moreover, for an organization managing a fleet of vehicles, knowing which vehicles will shortly require maintenance and/or replacement parts contributes to efficient and effective deployment of vehicles and maintenance resources. For example, at step 1601, the system 100 determines that an event arising from usage of a track system 16x, such as a usage threshold event (e.g. an amount of tread wear, an amount of time such as a number of hours the track 41 has been used), deterioration threshold event (e.g. the number of exposed reinforcing cables) and/or deterioration event (e.g. one or more snapped or broken reinforcing cables), has occurred. At step 1602, the system 100 identifies the track system component for which the usage threshold event, deterioration threshold event and/or deterioration event has occurred. In some embodiments, as shown in
The track system component supply database can be managed by the fleet management system, or can be managed by a third-party track system component supplier. If the identified track system component is available, the vehicle can be scheduled for maintenance. If, on the other hand, the track system component is not available, the fleet management system can cause the track system component to be ordered at step 1604, before scheduling maintenance of the vehicle at step 1605. This system may also include ordering based on stored user preference as previously described.
In some embodiments, the scheduling of the vehicle maintenance is at least in part based on the estimated delivery time for an ordered track system component. In other embodiments, the dispatching of the vehicle relating to the identified track system component can, at least partially, be based on a pre-scheduled maintenance. This system 100 may also include scheduling based on stored user preference as previously described. Finally, at step 1606, the maintenance operation is carried out and the track system component is replaced or repaired.
If the system 100 has determined that a critical error has taken place or is imminent, it can prompt the user to establish an audiovisual and/or textual connection with a technician at 1703. This could be achieved by using a Voice Over IP (VoIP) system, a phone call over a cellular network, or any other means of text, audio or video communication. This will allow the vehicle operator to communicate with the technician and get or receive pertinent information to vehicle maintenance. For example, the technician may instruct the user to drive the vehicle to a safe location and wait for the technician to arrive. In the case of a video call, the technician may be able to instruct the user to point the camera of the electronic device at a specific component of the vehicle 10 in order to provide the technician with more information about the vehicle status.
At step 1803, the system 100 alerts relevant parties of the critical status. This can include fleet managers, technicians or other operators. For example, the system 100 may send a text message, email or app push notification to any interested party that the status and operability of a given vehicle with a unique identifier has reached a certain threshold of wear or damage. Based on the information determined by the system 100, the vehicle operator or fleet manager may override the decision determined by the system 100 and continue to operate the vehicle. Alternatively, the system 100 may have the capability to safely disable the vehicle given specific parameters. For example, the system 100 may only allow the vehicle to operate for another specific distance or time, or may not allow the vehicle to restart after it has switched off without an appointment with a technician.
According to this embodiment, the system 100 is able to identify the track characteristics 1903. These characteristics may include thickness, length, weight, width, tread pattern, internal cable strength, etc. Based on an analysis of the vehicle's track, the system 100 can determine track alternatives at step 1904. This can be done using a pre-populated database stored on a server of all major available track brands and products, along with compatible alternatives. Once the system 100 has determined the track and track characteristics, it can query the database to find all other products that could be used for the vehicle.
The system 100 can then communicate the tracks to the user at step 1905. This can be done by sending the information over the network to the electronic device. The user may determine that an alternative track could be used for the vehicle. If the user selects the alternative track, the system 100 will send that message back to the server over the network and proceed to organize any part replacement using the user's selection.
As shown in
As shown in
As shown in
The electronic device 501 may communicate this data along with any other information to be communicated to the system server 1142 for analysis, as described above.
As shown in
As shown in
As shown in
According to other embodiments such as those shown in
According to another embodiment, the system 100 may have access to pricing information or alternative replacement parts available at the parts center 2704. The system 100 may present the user with pricing options, sale information for different components they may require ordering for replacement. The user may then inform the system 100 of their preference and the system 100 will submit the order to the parts center accordingly.
As shown in
Similarly, according to some embodiments, the system 100 is able to make purchase requests over the network 124 without requiring the user to select a part component. This choice may be based on a user preference saved in the server memory for a given vehicle owner. For example, an owner may have a preference for a specific brand of vehicle parts. Accordingly the system 100 can notify the user of the part purchase as it is automatically scheduled.
As shown in
According to the embodiments disclosed in
According to another embodiment and shown in
According to another embodiment and shown in
According to yet another embodiment and shown in
In some embodiments, with additional reference to
For example, in some embodiments, as shown in
In other embodiments, with additional reference to
In some embodiments, with additional reference to
In some embodiments, the drone 3201 is arranged to follow the vehicle, capture and analyze images of each of the track system 161, 162 and their environment. In other embodiments, the drone 3201 is equipped with a laser line scanner for scanning the track system 161, 162 and their environment. Communication between the drone 3201 and the vehicle 10 (e.g., between the drone 3201 and the processing entity 88) can be provided for by any suitable means, including but not limited to any combination of Global Positioning System (GPS) signals, Radio Frequency (RF) signals, Bluetooth signals, LIDAR, and RADAR signals. This embodiment is particularly advantageous for producing 3D scanning data suitable for subsequent volumetric analysis, as described in more detail above.
In this embodiment, the drone 3210 is an aerial drone configured to fly about the vehicle 10. While the drone 3201 shown in
In some embodiments, with additional reference to
In some embodiments, the vehicle-mounted inspection device 4801 comprises a camera system arranged to capture images of the track system 161, 162 and its environment as the track 22 moves around the track-engaging assembly 21. The information generated by the camera system can then be optionally processed and analyzed locally or remotely by the system server 1142.
In some embodiments, the vehicle-mounted inspection device 4801 comprises a laser line scanner system and/or a laser area scanner system arranged to scan the track system 161, 162 and its environment as the track 22 move around the track-engaging assembly 21. The information generated by the laser line scanner and/or laser area scanner systems can then be optionally processed and analyzed locally or remotely by system server 1142. This embodiment is particularly advantageous for producing 3D scanning data suitable for subsequent volumetric analysis, as described in more detail above.
In some embodiments, as shown in
The interface 1520 comprises one or more inputs and outputs allowing the computing system 1500 to receive signals from and send signals to other components to which the computing system 1500 is connected (i.e., directly or indirectly connected).
The processor 1540 comprises one or more processing devices for performing processing operations that implement functionality of the computing system 1500. A processing device of the processor 1540 may be a general-purpose processor executing program code stored in the memory 1560. Alternatively, a processing device of the processor 1540 may be a specific-purpose processor comprising one or more preprogrammed hardware or firmware elements (e.g., application-specific integrated circuits (ASICs), electrically erasable programmable read-only memories (EEPROMs), etc.) or other related elements).
The memory 1560 comprises one or more memory elements for storing program code executed by the processor 1540 and/or data used during operation of the processor 1540. A memory element of the memory portion 1560 may be a semiconductor medium (including, e.g., a solid state memory), a magnetic storage medium, an optical storage medium, and/or any other suitable type of memory element. A memory element of the memory portion 1560 may be read-only memory (ROM) and/or random-access memory (RAM), for example.
In some embodiments, two or more elements of the computing system 1500 may be implemented by devices that are physically distinct from one another (e.g., located in a common site or in remote sites) and may be connected to one another via a bus (e.g., one or more electrical conductors or any other suitable bus) or via a communication link which may be wired, wireless, or both and which may traverse one or more networks (e.g., the Internet or any other computer network such as a local-area network (LAN) or wide-area network (WAN), a cellular network, etc.). In other embodiments, two or more elements of the computing system 1500 may be implemented by a single device.
While in embodiments considered above the off-road vehicle 10 is a construction or agricultural vehicle, in other embodiments, the vehicle 10 may be another type of work vehicle such as a knuckleboom loader, etc.) for performing forestry work, or a military vehicle (e.g., a combat engineering vehicle (CEV), etc.) for performing military work, a carrier (e.g. carrying a boom, a rig, and/or other equipment t), or may be any other type of vehicle operable off paved road. Although operable off paved roads, the vehicle 10 may also be operable on paved roads in some cases. Also, while in embodiments considered above the off-road vehicle 10 is driven by a human operator in the vehicle 10, in other embodiments, the vehicle 10 may be an unmanned ground vehicle (e.g., a teleoperated or autonomous unmanned ground vehicle).
Any feature of any embodiment discussed herein may be combined with any feature of any other embodiment discussed herein in some examples of implementation.
Certain additional elements that may be needed for operation of certain embodiments have not been described or illustrated as they are assumed to be within the purview of those of ordinary skill in the art. Moreover, certain embodiments may be free of, may lack and/or may function without any element that is not specifically disclosed herein.
Although various embodiments and examples have been presented, this was for purposes of description, but should not be limiting. Various modifications and enhancements will become apparent to those of ordinary skill in the art.
This application is a National Stage of international PCT Patent Application No. PCT/CA2019/051217 filed on Aug. 30, 2019, which claims the benefit of U.S. Provisional Patent Application Ser. No. 62/724,846, filed on Aug. 30, 2018 and U.S. Provisional Patent Application Ser. No. 62/861,677, filed on Jun. 14, 2019. The contents of the aforementioned applications are incorporated by reference herein.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/CA2019/051217 | 8/30/2019 | WO |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2020/041897 | 3/5/2020 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
1539582 | Landry | May 1925 | A |
1673541 | Wilson | Jun 1928 | A |
2025999 | Myers | Dec 1935 | A |
2040696 | Johnston | May 1936 | A |
2369130 | Benson | Feb 1945 | A |
2461150 | Flynn | Feb 1949 | A |
2523182 | Battaglia | Sep 1950 | A |
2562264 | Ford | Jul 1951 | A |
2596919 | Smith | May 1952 | A |
2854294 | Bannister | Sep 1958 | A |
3019061 | Schomers | Jan 1962 | A |
3118709 | Case | Jan 1964 | A |
3464476 | Scheuba | Sep 1969 | A |
3582154 | Russ, Sr. | Jun 1971 | A |
3612626 | Fuchs | Oct 1971 | A |
3747995 | Russ, Sr. | Jul 1973 | A |
3747996 | Huber | Jul 1973 | A |
3767275 | Russ, Sr. | Oct 1973 | A |
3781067 | Dodson et al. | Dec 1973 | A |
3858948 | Johnson et al. | Jan 1975 | A |
3887244 | Haslett et al. | Jun 1975 | A |
3914990 | Borg | Oct 1975 | A |
3944006 | Lassanske | Mar 1976 | A |
3981943 | Fujio et al. | Sep 1976 | A |
4059313 | Beyers et al. | Nov 1977 | A |
RE29718 | Reinsma et al. | Aug 1978 | E |
4150858 | Fox et al. | Apr 1979 | A |
4218101 | Thompson | Aug 1980 | A |
4279449 | Martin et al. | Jul 1981 | A |
4538860 | Edwards et al. | Sep 1985 | A |
4583791 | Nagata et al. | Apr 1986 | A |
4586757 | Bloechlinger | May 1986 | A |
4587280 | Guha et al. | May 1986 | A |
4607892 | Payne et al. | Aug 1986 | A |
4614508 | Kerivan | Sep 1986 | A |
4696520 | Henke et al. | Sep 1987 | A |
4721498 | Grob | Jan 1988 | A |
D298018 | Cartwright | Oct 1988 | S |
4843114 | Touchet et al. | Jun 1989 | A |
4844561 | Savage et al. | Jul 1989 | A |
4880283 | Savage et al. | Nov 1989 | A |
4953919 | Langford | Sep 1990 | A |
4953921 | Burns | Sep 1990 | A |
4981188 | Kadela | Jan 1991 | A |
5018591 | Price | May 1991 | A |
5050710 | Bargfrede | Sep 1991 | A |
5117930 | Argouarc'h | Jun 1992 | A |
5145242 | Togashi | Sep 1992 | A |
5190363 | Brittain et al. | Mar 1993 | A |
5299860 | Anderson | Apr 1994 | A |
5320585 | Kato | Jun 1994 | A |
5352029 | Nagorcka | Oct 1994 | A |
5362142 | Katoh | Nov 1994 | A |
5368115 | Crabb | Nov 1994 | A |
5380076 | Hori | Jan 1995 | A |
5447365 | Muramatsu et al. | Sep 1995 | A |
5452949 | Kelderman | Sep 1995 | A |
5482364 | Edwards et al. | Jan 1996 | A |
5498188 | Deahr | Mar 1996 | A |
5511869 | Edwards et al. | Apr 1996 | A |
5513683 | Causa et al. | May 1996 | A |
5529267 | Giras et al. | Jun 1996 | A |
5540489 | Muramatsu et al. | Jul 1996 | A |
5632537 | Yoshimura et al. | May 1997 | A |
5707123 | Grob | Jan 1998 | A |
5722745 | Courtemanche et al. | Mar 1998 | A |
5813733 | Hori et al. | Sep 1998 | A |
5866265 | Reilly et al. | Feb 1999 | A |
5894900 | Yamamoto et al. | Apr 1999 | A |
5904217 | Yamamoto et al. | May 1999 | A |
5984438 | Tsunoda et al. | Nov 1999 | A |
5997109 | Kautsch | Dec 1999 | A |
6000766 | Takeuchi et al. | Dec 1999 | A |
6024183 | Dietz et al. | Feb 2000 | A |
6030057 | Fikse | Feb 2000 | A |
6056656 | Kitano et al. | May 2000 | A |
6065818 | Fischer | May 2000 | A |
6068354 | Akiyama et al. | May 2000 | A |
6074025 | Juncker et al. | Jun 2000 | A |
6079802 | Nishimura et al. | Jun 2000 | A |
6095275 | Shaw | Aug 2000 | A |
6120405 | Oertley et al. | Sep 2000 | A |
6129426 | Tucker | Oct 2000 | A |
6139121 | Muramatsu | Oct 2000 | A |
6153686 | Granatowicz et al. | Nov 2000 | A |
6170925 | Ono | Jan 2001 | B1 |
6176557 | Ono | Jan 2001 | B1 |
6186604 | Fikse | Feb 2001 | B1 |
6193335 | Edwards | Feb 2001 | B1 |
6206492 | Moser | Mar 2001 | B1 |
6224172 | Goodwin | May 2001 | B1 |
6241327 | Gleasman et al. | Jun 2001 | B1 |
6259361 | Robillard et al. | Jul 2001 | B1 |
6296329 | Rodgers et al. | Oct 2001 | B1 |
6299264 | Kautsch et al. | Oct 2001 | B1 |
6300396 | Tsunoda et al. | Oct 2001 | B1 |
6352320 | Bonko et al. | Mar 2002 | B1 |
6386652 | Bonko | May 2002 | B1 |
6386653 | Brandenburger | May 2002 | B1 |
6386654 | Singer et al. | May 2002 | B1 |
6416142 | Oertley | Jul 2002 | B1 |
6474756 | Hori et al. | Nov 2002 | B2 |
6494548 | Courtemanche | Dec 2002 | B2 |
6536852 | Katayama et al. | Mar 2003 | B2 |
6536853 | Egle et al. | Mar 2003 | B2 |
6568769 | Watanabe et al. | May 2003 | B1 |
6581449 | Brown et al. | Jun 2003 | B1 |
D476599 | Whittington | Jul 2003 | S |
6588862 | Pringiers | Jul 2003 | B1 |
6637276 | Adderton et al. | Oct 2003 | B2 |
6652043 | Oertley | Nov 2003 | B2 |
6671609 | Nantz et al. | Dec 2003 | B2 |
6698850 | Ueno | Mar 2004 | B2 |
D488171 | Juncker et al. | Apr 2004 | S |
6716012 | Yovichin et al. | Apr 2004 | B2 |
6733091 | Deland et al. | May 2004 | B2 |
6733093 | Deland et al. | May 2004 | B2 |
6769746 | Rodgers et al. | Aug 2004 | B2 |
6800236 | Kurata et al. | Oct 2004 | B1 |
6848757 | Ueno | Feb 2005 | B2 |
6874586 | Boivin et al. | Apr 2005 | B2 |
D505136 | Brazier | May 2005 | S |
6904986 | Brazier | Jun 2005 | B2 |
6913329 | Rodgers et al. | Jul 2005 | B1 |
6921197 | Aubel et al. | Jul 2005 | B2 |
6923515 | Konickson et al. | Aug 2005 | B2 |
6932442 | Hori | Aug 2005 | B2 |
6935708 | Courtemanche | Aug 2005 | B2 |
6948784 | Wodrich et al. | Sep 2005 | B2 |
6962222 | Kirihata | Nov 2005 | B2 |
6964462 | Katoh et al. | Nov 2005 | B2 |
6974196 | Gagne et al. | Dec 2005 | B2 |
7001294 | Fukuda | Feb 2006 | B2 |
7032636 | Salakari | Apr 2006 | B2 |
7077216 | Juncker | Jul 2006 | B2 |
D528133 | Brazier | Sep 2006 | S |
7114788 | Deland et al. | Oct 2006 | B2 |
7131508 | Brazier | Nov 2006 | B2 |
7137675 | Simula et al. | Nov 2006 | B1 |
7197922 | Rimkus et al. | Apr 2007 | B2 |
7202777 | Tsuji et al. | Apr 2007 | B2 |
7222924 | Christianson | May 2007 | B2 |
7229141 | Dandurand et al. | Jun 2007 | B2 |
7252348 | Gingras | Aug 2007 | B2 |
7293844 | Uchiyama | Nov 2007 | B2 |
D556791 | Brazier | Dec 2007 | S |
7316251 | Kogure et al. | Jan 2008 | B2 |
7325888 | Fujita et al. | Feb 2008 | B2 |
7367637 | Gleasman | May 2008 | B2 |
7407236 | Fukushima | Aug 2008 | B2 |
7413268 | Kato et al. | Aug 2008 | B2 |
7416266 | Soucy et al. | Aug 2008 | B2 |
7497530 | Bessette | Mar 2009 | B2 |
7567171 | Dufournier | Jul 2009 | B2 |
7597161 | Brazier | Oct 2009 | B2 |
D603880 | Brazier | Nov 2009 | S |
7625050 | Bair | Dec 2009 | B2 |
7676307 | Schmitt et al. | Mar 2010 | B2 |
7708092 | Despres | May 2010 | B2 |
7729823 | Ruoppolo | Jun 2010 | B2 |
7740094 | Pelletier | Jun 2010 | B2 |
7778741 | Rao et al. | Aug 2010 | B2 |
7779947 | Stratton | Aug 2010 | B2 |
7784884 | Soucy et al. | Aug 2010 | B2 |
7798260 | Albright et al. | Sep 2010 | B2 |
7823987 | Dandurand et al. | Nov 2010 | B2 |
7866766 | Berg | Jan 2011 | B2 |
7914088 | Bair | Mar 2011 | B2 |
7914089 | Bair | Mar 2011 | B2 |
D644670 | Barrelmeyer | Sep 2011 | S |
8122581 | Hurst et al. | Feb 2012 | B1 |
8327960 | Couture et al. | Dec 2012 | B2 |
8342257 | Rosenboom | Jan 2013 | B2 |
D680561 | Zuchoski et al. | Apr 2013 | S |
D681071 | Zuchoski et al. | Apr 2013 | S |
D683371 | Aube | May 2013 | S |
D683769 | Aube | Jun 2013 | S |
8567876 | Wellman | Oct 2013 | B2 |
8628152 | Delisle et al. | Jan 2014 | B2 |
8736147 | Wang | May 2014 | B1 |
D711928 | Brazier | Aug 2014 | S |
8985250 | Lussier et al. | Mar 2015 | B1 |
9033431 | Zuchoski et al. | May 2015 | B1 |
9067631 | Lussier et al. | Jun 2015 | B1 |
9334001 | Lussier et al. | May 2016 | B2 |
9805697 | Dorrance | Oct 2017 | B1 |
9855843 | Vik et al. | Jan 2018 | B2 |
9878750 | Leblanc et al. | Jan 2018 | B2 |
9880075 | Finch et al. | Jan 2018 | B2 |
9975554 | Lussier et al. | May 2018 | B2 |
10006836 | Monti et al. | Jun 2018 | B2 |
10272959 | Zuchoski et al. | Apr 2019 | B2 |
10328982 | Lussier et al. | Jun 2019 | B2 |
10745069 | Gustafson | Aug 2020 | B2 |
10783723 | Richard et al. | Sep 2020 | B2 |
10933877 | Lussier et al. | Mar 2021 | B2 |
11046377 | Lussier et al. | Jun 2021 | B2 |
11167810 | Boily et al. | Nov 2021 | B2 |
11186330 | Zuchoski et al. | Nov 2021 | B2 |
20020070607 | Edwards | Jun 2002 | A1 |
20020140288 | Herberger, Sr. et al. | Oct 2002 | A1 |
20020145335 | Soucy et al. | Oct 2002 | A1 |
20030019133 | Hori | Jan 2003 | A1 |
20030034189 | Lemke et al. | Feb 2003 | A1 |
20030034690 | Hori et al. | Feb 2003 | A1 |
20030080618 | Krishnan et al. | May 2003 | A1 |
20030089534 | Kanzler et al. | May 2003 | A1 |
20030161687 | Kirihata | Aug 2003 | A1 |
20040004395 | Soucy et al. | Jan 2004 | A1 |
20040070273 | Safe et al. | Apr 2004 | A1 |
20040084962 | Courtemanche | May 2004 | A1 |
20040130212 | Ishibashi | Jul 2004 | A1 |
20040135433 | Inaoka et al. | Jul 2004 | A1 |
20050035654 | Tamaru et al. | Feb 2005 | A1 |
20050056468 | Tucker | Mar 2005 | A1 |
20050103540 | Lavoie | May 2005 | A1 |
20050104449 | Lavoie et al. | May 2005 | A1 |
20050104450 | Gagne et al. | May 2005 | A1 |
20050168069 | Ueno | Aug 2005 | A1 |
20060060395 | Boivin et al. | Mar 2006 | A1 |
20060090558 | Raskas | May 2006 | A1 |
20060103236 | Soucy et al. | May 2006 | A1 |
20060124366 | Le Masne De Chermont | Jun 2006 | A1 |
20060144480 | Takayama et al. | Jul 2006 | A1 |
20060175108 | Kubota | Aug 2006 | A1 |
20060220456 | Sugahara | Oct 2006 | A1 |
20060248484 | Baumgartner et al. | Nov 2006 | A1 |
20060273660 | Dandurand et al. | Dec 2006 | A1 |
20070046100 | McGilvrey et al. | Mar 2007 | A1 |
20070075456 | Feldmann | Apr 2007 | A1 |
20070126286 | Feldmann et al. | Jun 2007 | A1 |
20070159004 | St-Amant | Jul 2007 | A1 |
20070251621 | Prost | Nov 2007 | A1 |
20070252433 | Fujita | Nov 2007 | A1 |
20080007118 | Fujita | Jan 2008 | A1 |
20080073971 | Paradis et al. | Mar 2008 | A1 |
20080084111 | Rainer | Apr 2008 | A1 |
20080100134 | Soucy et al. | May 2008 | A1 |
20080136255 | Feldmann et al. | Jun 2008 | A1 |
20080169147 | Brazier | Jul 2008 | A1 |
20080179124 | Stratton | Jul 2008 | A1 |
20080203813 | Doyle | Aug 2008 | A1 |
20080211300 | Matsuo et al. | Sep 2008 | A1 |
20080211301 | Jee et al. | Sep 2008 | A1 |
20090085398 | Maltais | Apr 2009 | A1 |
20090102283 | Choi | Apr 2009 | A1 |
20090166101 | Wenger et al. | Jul 2009 | A1 |
20090195062 | Uchida | Aug 2009 | A1 |
20090302676 | Brazier | Dec 2009 | A1 |
20090302677 | Sugihara | Dec 2009 | A1 |
20090309415 | Shimozono | Dec 2009 | A1 |
20090326109 | Kameda et al. | Dec 2009 | A1 |
20100012399 | Hansen | Jan 2010 | A1 |
20100033010 | Shimozono | Feb 2010 | A1 |
20100079238 | Gravelle et al. | Apr 2010 | A1 |
20100095506 | Bair | Apr 2010 | A1 |
20100096915 | Hagio | Apr 2010 | A1 |
20100096917 | Bair | Apr 2010 | A1 |
20100121644 | Glasser | May 2010 | A1 |
20100133019 | Muemken | Jun 2010 | A1 |
20100139994 | Hansen | Jun 2010 | A1 |
20100191417 | Murahashi et al. | Jul 2010 | A1 |
20100194553 | Mizutani et al. | Aug 2010 | A1 |
20100230185 | Mallette et al. | Sep 2010 | A1 |
20100253138 | Despres | Oct 2010 | A1 |
20100256946 | Carresjo et al. | Oct 2010 | A1 |
20100283317 | Soucy et al. | Nov 2010 | A1 |
20110068620 | Delisle et al. | Mar 2011 | A1 |
20110121644 | Wellman | May 2011 | A1 |
20110148189 | Courtemanche et al. | Jun 2011 | A1 |
20110301825 | Grajkowski et al. | Dec 2011 | A1 |
20110315459 | Zuchoski et al. | Dec 2011 | A1 |
20120001478 | Zuchoski et al. | Jan 2012 | A1 |
20120056473 | Hashimoto et al. | Mar 2012 | A1 |
20120104840 | Zuchoski et al. | May 2012 | A1 |
20120242142 | Kautsch et al. | Sep 2012 | A1 |
20120242143 | Feldmann | Sep 2012 | A1 |
20120253590 | Fink | Oct 2012 | A1 |
20120306916 | Marumoto et al. | Dec 2012 | A1 |
20130033271 | Woodard | Feb 2013 | A1 |
20130073157 | Person et al. | Mar 2013 | A1 |
20130082846 | McKinley et al. | Apr 2013 | A1 |
20130126196 | Rosenboom | May 2013 | A1 |
20130134772 | Dandurand et al. | May 2013 | A1 |
20130134773 | Dandurand et al. | May 2013 | A1 |
20130162016 | Lajoie et al. | Jun 2013 | A1 |
20130245911 | Nakajima et al. | Sep 2013 | A1 |
20130255354 | Hawkins et al. | Oct 2013 | A1 |
20130325266 | Padilla et al. | Dec 2013 | A1 |
20140105481 | Hasselbusch | Apr 2014 | A1 |
20140125117 | Weeks et al. | May 2014 | A1 |
20140180534 | Son | Jun 2014 | A1 |
20140182960 | Bedard et al. | Jul 2014 | A1 |
20140288763 | Bennett et al. | Sep 2014 | A1 |
20140324301 | Rebinsky | Oct 2014 | A1 |
20150042152 | Lussier et al. | Feb 2015 | A1 |
20150042153 | Lussier et al. | Feb 2015 | A1 |
20150107975 | Chen et al. | Apr 2015 | A1 |
20150129329 | Cox | May 2015 | A1 |
20150191173 | Lussier et al. | Jul 2015 | A1 |
20150321710 | Zuchoski et al. | Nov 2015 | A1 |
20160052572 | McKinley et al. | Feb 2016 | A1 |
20160059779 | Vandendriessche | Mar 2016 | A1 |
20160121945 | Rust et al. | May 2016 | A1 |
20160129954 | Hasselbusch et al. | May 2016 | A1 |
20160236733 | Tiede et al. | Aug 2016 | A1 |
20160332682 | Lussier et al. | Nov 2016 | A1 |
20170036714 | Lunkenbein | Feb 2017 | A1 |
20170087987 | Vik et al. | Mar 2017 | A1 |
20170094894 | Heim et al. | Apr 2017 | A1 |
20170174277 | Zuchoski et al. | Jun 2017 | A1 |
20170177011 | Garvin et al. | Jun 2017 | A1 |
20180043949 | Boily | Feb 2018 | A1 |
20180093724 | Boily | Apr 2018 | A1 |
20180172557 | Ghidotti Piovan et al. | Jun 2018 | A1 |
20180190045 | Richard et al. | Jul 2018 | A1 |
20180237020 | Lussier et al. | Aug 2018 | A1 |
20180265145 | Todd et al. | Sep 2018 | A1 |
20180364744 | Garvin et al. | Dec 2018 | A1 |
20190066283 | Gros | Feb 2019 | A1 |
20190351957 | Zuchoski et al. | Nov 2019 | A1 |
20190359270 | Lussier et al. | Nov 2019 | A1 |
20190392569 | Finch | Dec 2019 | A1 |
20200070906 | Laperle et al. | Mar 2020 | A1 |
20210024150 | Boily et al. | Jan 2021 | A1 |
20210074089 | Boily et al. | Mar 2021 | A1 |
20210173399 | Richard et al. | Jun 2021 | A1 |
20210197625 | Laperle et al. | Jul 2021 | A1 |
20210213954 | Lussier et al. | Jul 2021 | A1 |
20220055699 | Desmarais et al. | Feb 2022 | A1 |
20220063740 | Boily et al. | Mar 2022 | A1 |
Number | Date | Country |
---|---|---|
2606039 | Apr 2009 | CA |
2838935 | Jan 2013 | CA |
2978482 | Sep 2016 | CA |
3064100 | Sep 2016 | CA |
2991072 | Jan 2017 | CA |
2947575 | May 2017 | CA |
3110606 | Mar 2020 | CA |
108061735 | May 2018 | CN |
108061735 | May 2018 | CN |
110588813 | Dec 2019 | CN |
20207342 | Aug 2002 | DE |
102010017531 | Dec 2011 | DE |
0578504 | Jan 1994 | EP |
1008509 | Jun 2000 | EP |
1211169 | Jun 2002 | EP |
1260429 | Nov 2002 | EP |
1273504 | Jan 2003 | EP |
1283152 | Feb 2003 | EP |
1325820 | Sep 2004 | EP |
1506913 | Feb 2005 | EP |
1882627 | Jan 2008 | EP |
1953070 | Aug 2008 | EP |
2014542 | Jan 2009 | EP |
2119620 | Nov 2009 | EP |
3185097 | Jun 2017 | EP |
3265368 | Jan 2018 | EP |
3313717 | May 2018 | EP |
3721200 | Oct 2020 | EP |
3721200 | Oct 2020 | EP |
3747746 | Dec 2020 | EP |
3812249 | Apr 2021 | EP |
3844473 | Jul 2021 | EP |
3844474 | Jul 2021 | EP |
3960593 | Feb 2022 | EP |
2229410 | Sep 1990 | GB |
2360501 | Sep 2001 | GB |
H02074476 | Mar 1990 | JP |
H11129946 | May 1999 | JP |
2000053037 | Feb 2000 | JP |
2003011860 | Jan 2003 | JP |
2003089366 | Mar 2003 | JP |
2004330830 | Nov 2004 | JP |
2006103482 | Apr 2006 | JP |
2007022304 | Feb 2007 | JP |
2009061829 | Mar 2009 | JP |
2009248924 | Oct 2009 | JP |
2010018091 | Jan 2010 | JP |
2010047040 | Mar 2010 | JP |
2010089729 | Apr 2010 | JP |
2014126370 | Jul 2014 | JP |
100829059 | May 2008 | KR |
20120055071 | May 2012 | KR |
20130043954 | May 2013 | KR |
20150004969 | Jan 2015 | KR |
1446017 | Dec 1988 | SU |
2008088355 | Jul 2008 | WO |
2008108439 | Sep 2008 | WO |
2009105892 | Sep 2009 | WO |
2009106617 | Sep 2009 | WO |
2013002781 | Jan 2013 | WO |
2014056089 | Apr 2014 | WO |
2014168851 | Oct 2014 | WO |
2015059457 | Apr 2015 | WO |
2016138592 | Sep 2016 | WO |
WO-2017000068 | Jan 2017 | WO |
2017049393 | Mar 2017 | WO |
WO2018165753 | Sep 2018 | WO |
2019109191 | Jun 2019 | WO |
2019225752 | Nov 2019 | WO |
2020041897 | Mar 2020 | WO |
2020041899 | Mar 2020 | WO |
2020049532 | Mar 2020 | WO |
2021189154 | Sep 2021 | WO |
Entry |
---|
Perelmuter et al., Classification of Bidimensional Images Using Artificial Intelligence Techniques, 1996, IEEE Conferences, Proceedings II Workshop on Cybernetic Vision (pp. 39-44) (Year: 1996). |
Perelmuter et al., Classification of Bidimensional Images Using Artificial Intelligence Techniques, 1996,IEEE Conferences, ProceedingsWorkshoponCyberneticVision(pp. 39-44)(Year: 1996) (Year: 1996). |
JP-2014126370 (“Saka”) translation and original (Year: 2014). |
CN-108061735 (“Bian”) translation and original (Year: 2018). |
WO-2018165753 (“Cha”) translation and original (Year: 2018). |
Communication pursuant to Article 94 (3) EPC issued on Aug. 22, 2019, in connection with European Patent Application No. 16816886.2, 4 pages. |
Communication pursuant to Rule 63(1) EPC issued on Nov. 8, 2018 in connection with European Patent Application No. 16758408.5, 4 pages. |
Communication under Rule 71(3) EPC issued on Mar. 16, 2020, in connection with European Patent Application 16758408.5, 7 pages. |
Communication under Rule 71(3) EPC issued on May 6, 2020, in connection with European Patent Application No. 16816886.2, 7 pages. |
European Extended Search Report issued on Nov. 6, 2020, in connection with European Patent Application No. 20187079.7, 15 pages. |
Examiners Report issued on Apr. 7, 2020, in connection with Canadian Patent Application 3,064,100, 3 pages. |
Examiners report issued on Aug. 28, 2018, in connection with Canadian Patent application No. 2,991,072, 4 pages. |
Examiners Report issued on Feb. 1, 2018 in connection with Canadian Patent Application No. 2,991,072, 3 pages. |
Examiners Report issued on Jan. 11, 2018 in connection with Canadian Patent Application No. 2,978,482, 3 pages. |
Examiners report issued on Jan. 16, 2019 in connection with Canadian Patent Application No. 2,991,072, 3 pages. |
Examiners Report issued on Jul. 9, 2020, in connection with Canadian Patent Application 3,064,100, 3 pages. |
Examiners Report issued on Oct. 14, 2020, in connection with Canadian Patent Application 3,042,780, 4 pages. |
Extended European Search Report and Written Opinion issued on Sep. 13, 2013 in connection, with European Patent Application No. 10835318.6, 8 pages. |
Extended European Search Report issued on Jan. 9, 2019 in connection with European Patent Application No. 16816886.2, 7 pages. |
Extended European Search Report issued on Mar. 20, 2019 in connection with European Patent Application No. 16758408.5, 8 pages. |
Final Office Action issued on Aug. 19, 2019, in connection with U.S. Appl. No. 15/724,733, 18 pages. |
Final Office Action issued on Aug. 24, 2018 in connection with U.S. Appl. No. 15/454,881, 30 pages. |
Final Office Action issued on Dec. 23, 2014 in connection with U.S. Appl. No. 13/326,132, 15 pages. |
Final Office Action issued on Dec. 31, 2015 in connection with U.S. Appl. No. 13/326,132, 15 pages. |
Final Office Action issued on Jan. 12, 2017 in connection with U.S. Appl. No. 14/665,075, 11 pages. |
Final Office Action issued on Jul. 6, 2020, in connection with U.S. Appl. No. 15/555,148, 20 pages. |
Final Office Action issued on Jun. 26, 2015 in connection with U.S. Appl. No. 13/326,278, 15 pages. |
Final Office Action issued on Mar. 27, 2014 in connection with U.S. Appl. No. 13/170,753, 16 pages. |
Final Office Action issued on Mar. 5, 2015 in connection with U.S. Appl. No. 13/325,796, 33 pages. |
Final Office Action issued on Sep. 8, 2017 in connection with U.S. Appl. No. 15/139,572, 14 pages. |
International Search Report and Written Opinion issued on Sep. 29, 2016 in connection with PCT/CA2016/050760, 8 pages. |
International Search Report and Written Opinion issued on Nov. 12, 2019, in connection with International PCT Application No. PCT/CA2019/051217, 8 pages. |
International Search Report and Written Opinion issued on Oct. 29, 2019 in connection with International PCT Application No. PCT/CA2019/051219, 9 pages. |
International Search Report issued on Mar. 5, 2019, in connection with International PCT application No. PCT/CA2018/051567, 4 pages. |
Interview Summary issued on Dec. 10, 2018 in connection with U.S. Appl. No. 15/139,572, 3 pages. |
Interview Summary issued on Nov. 2, 2018 in connection with U.S. Appl. No. 15/454,881, 3 pages. |
Interview Summary Report issued on Sep. 12, 2015 in connection with U.S. Appl. No. 13/326,278, 3 pages. |
Non-Final Office Action issued on Apr. 5, 2016 in connection with U.S. Appl. No. 14/665,075, 5 pages. |
Non-Final Office Action issued on Apr. 9, 2019 in connection with U.S. Appl. No. 15/724,733, 56 pages. |
Non-Final Office Action issued on Dec. 1, 2016 in connection with U.S. Appl. No. 15/139,572, 5 pages. |
Non-Final Office Action issued on Dec. 30, 2019 in connection with U.S. Appl. No. 15/740,976, 16 pages. |
Non-Final Office Action issued on Feb. 21, 2017 in connection with U.S. Appl. No. 14/721,326, 12 pages. |
Non-Final Office Action issued on Jan. 11, 2017 in connection with U.S. Appl. No. 14/886,327, 11 pages. |
Non-Final Office Action issued on Jul. 10, 2014 in connection with U.S. Appl. No. 13/326,110, 16 pages. |
Non-Final Office Action issued on Jul. 10, 2014 in connection with U.S. Appl. No. 13/424,459, 11 pages. |
Non-Final Office Action issued on Jul. 14, 2014 in connection with U.S. Appl. No. 13/112,840, 16 pages. |
Non-Final Office Action issued on Jun. 12, 2017 in connection with U.S. Appl. No. 14/665,075, 9 pages. |
Non-Final Office Action issued on Jun. 2, 2014 in connection with U.S. Appl. No. 13/326,010, 6 pages. |
Non-Final Office Action issued on Jun. 24, 2015 in connection with U.S. Appl. No. 13/326,132, 17 pages. |
Non-Final Office Action issued on Mar. 20, 2018 in connection with U.S. Appl. No. 15/139,572, 7 pages. |
Non-Final Office Action issued on Mar. 6, 2015 in connection with U.S. Appl. No. 13/326,278, 16 pages. |
Non-Final Office Action issued on May 17, 2018 in connection with U.S. Appl. No. 15/454,881, 8 pages. |
Non-Final Office Action issued on Oct. 11, 2017 in connection with U.S. Appl. No. 14/665,075, 6 pages. |
Non-Final Office Action issued on Oct. 2, 2013 in connection with U.S. Appl. No. 13/170,753, 15 pages. |
European Extended Search Report issued on Aug. 18, 2021, in connection with European Patent Application No. 18887057.0, 8 pages. |
Examiner's report issued on Jan. 22, 2021, in connection with Canadian Patent Application No. 2,991,072, 3 pages. |
Non-Final Office Action issued on Dec. 14, 2020, in connection with U.S. Appl. No. 15/555,148, 9 pages. |
Notice of Allowance issued on Feb. 23, 2021 in connection with U.S. Appl. No. 15/724,733, 18 pages. |
Final Office Action issued on Mar. 9, 2021, in connection with U.S. Appl. No. 15/555,148, 14 pages. |
Notice of Allowance issued on Mar. 4, 2015 in connection with U.S. Appl. No. 13/325,783, 13 pages. |
Notice of Allowance issued on May 20, 2020 in connection with U.S. Appl. No. 15/740,976, 12 pages. |
Notice of Allowance issued on May 30, 2018 in connection with U.S. Appl. No. 15/139,572, 10 pages. |
Notice of Allowance issued on Nov. 14, 2014 in connection with U.S. Appl. No. 13/326,010, 7 pages. |
Notice of Allowance issued on Oct. 30, 2020, in connection with U.S. Appl. No. 15/958,156, 30 pages. |
Notice of Allowance issued on Oct. 24, 2014 in connection with U.S. Appl. No. 13/170,753, 5 pages. |
Non-Final Office Action issued on Feb. 2, 2021, in connection with U.S. Appl. No. 16/360,060, 55 pages. |
Non-Final Office Action issued on Jan. 24, 2020 in connection with U.S. Appl. No. 15/555,148, 57 pages. |
Non-Final Office Action issued on May 29, 2020, in connection with U.S. Appl. No. 15/958,156, 45 pages. |
Non-Final Office Action issued on Nov. 9, 2012, in connection with U.S. Appl. No. 29/405,414, 9 pages. |
Ex parte Quayle issued on Sep. 11, 2018 in connection with U.S. Appl. No. 15/139,572, 7 pages. |
Restriction Requirement issued on Jan. 27, 2014 in connection with U.S. Appl. No. 13/326,010, 5 pages. |
Restriction Requirement issued on Jun. 9, 2014 in connection with U.S. Appl. No. 13/326,278, 7 pages. |
Restriction Requirement issued on Sep. 26, 2013 in connection with U.S. Appl. No. 13/326,010, 6 pages. |
Supplementary Search Report issued on Jan. 2, 2019 in connection with the European Patent Application No. 16816886, 1 page. |
Written Opinion issued on Mar. 5, 2019 in connection with International PCT application No. PCT/CA2018/051567, 6 pages. |
International Search Report and Written Opinion issued on Jun. 21, 2021 in connection with International PCT application No. PCT/CA2021/050404, 8 pages. |
Notice of Allowance issued on Jul. 9, 2021, in connection with U.S. Appl. No. 15/555,148, 6 pages. |
Notice of Allowance issued on Jul. 26, 2021, in connection with U.S. Appl. No. 16/360,060, 14 pages. |
Extended European Search Report issued on Mar. 30, 2021, in connection with European Patent Application No. 20201638.2, 10 pages. |
Non-Final Office Action issued on Mar. 11, 2021 in connection with U.S. Appl. No. 16/419,760, 28 pages. |
Final Office Action issued on Nov. 9, 2021 in connection with U.S. Appl. No. 16/419,760, 57 pages. |
International Search Report and Written Opinion issued on Nov. 3, 2021 in connection with International PCT application No. PCT/CA2021/051143, 17 pages. |
Non-Final Office Action issued on Nov. 9, 2012, in connection with U.S. Appl. No. 29/405,417, 10 pages. |
Extended European Search Report issued on Feb. 1, 2022, in connection with European Patent Application No. 21191893.3, 11 pages. |
Bair Products, Inc. “Larry Lugs—Patented Bolt-On-Replacement Drive Lugs”, http://www.bairproductsinc.com/products/larry_lugs.html, Jan. 1, 2011, 2 pages. |
Bridgestone Industrial Products America Inc. “Stay on the Right Track. New Generation Features Rubber Tracks for Excavators”, Brochure 2009, 12 pages. |
Bridgestone Industrial Products America Inc., “Stay on the Right Track. Rubber Tracks for Track Loaders”, Brochure 2009, 6 pages. |
Camoplast Inc. “ATV/UTV Track Systems”, 2009-2010 Catalog, 8 pages. |
CAN-AM BRP, “Parts & Accessories—Track Systems”, Parts, Accessories & Riding Gear Catalogue, p. 66 (2011). |
Red Roo Solutions Pty Ltd—World Class Solutions for the Earth Moving Industry, “Save thousands of dollars and add thousands of hours to your tracks with Larry Lugs”, http://www.redroosolutions.com.au/larrylugs.html, Jun. 8, 2009, 3 pages. |
Story by Staff, BRP upgrades Apache ATV track system for Outlander—Apache system features larger footprint and power steering mode, atv.com, http://www.atc.com/news/brp-upgrades-apache-atv-track-system-for-outlander-1481.html 2 pages (Nov. 24, 2009). |
Non-Final Office Action issued on Nov. 9, 2012, in connection with U.S. Appl. No. 29/405,416, 9 pages. |
Restriction Requirement issued on Oct. 25, 2019 in connection with U.S. Appl. No. 15/555,148, 7 pages. |
Non-final Office Action issued on Oct. 5, 2020, in connection with U.S. Appl. No. 15/724,733, 7 pages. |
Non-Final Office Action issued on Sep. 22, 2014 in connection with U.S. Appl. No. 13/325,796, 25 pages. |
Non-Final Office Action issued on Sep. 29, 2014 in connection with U.S. Appl. No. 13/325,783, 27 pages. |
Notice of Allowance issued on Apr. 2, 2020, in connection with U.S. Appl. No. 15/724,733, 5 pages. |
Notice of Allowance issued on Dec. 17, 2018 in connection with U.S. Appl. No. 15/454,881, 5 pages. |
Notice of Allowance issued on Feb. 15, 2019 in connection with U.S. Appl. No. 15/139,572, 7 pages. |
Notice of Allowance issued on Jan. 11, 2016 in connection with U.S. Appl. No. 13/326,278, 7 pages. |
Notice of Allowance issued on Jan. 19, 2018 in connection with U.S. Appl. No. 14/665,075, 8 pages. |
Notice of Allowance issued on Jan. 21, 2015 in connection with U.S. Appl. No. 13/326,110, 6 pages. |
European Extended Search Report issued on Mar. 20, 2019 in connection with European Patent Application No. 16758408.5, 8 pages. |
International Search Report issued on Jun. 29, 2016 in connection with International Patent Application No. PCT/CA2016/050234, 3 pages. |
Written Opinion issued on Jun. 29, 2016 in connection with International Patent Application No. PCT/CA2016/050234, 3 pages. |
Number | Date | Country | |
---|---|---|---|
20210339758 A1 | Nov 2021 | US |
Number | Date | Country | |
---|---|---|---|
62861677 | Jun 2019 | US | |
62724846 | Aug 2018 | US |