The present disclosure is directed generally to methods and systems for evaluating mobile phones and other consumer electronic devices and, more particularly, to methods and systems associated with detecting cracks in screens of such devices.
It is often necessary to visually evaluate a screen of a mobile device (e.g., a smartphone or tablet) to identify cracks or other defects in the mobile device. For example, pricing the mobile device, assessing the mobile device for possible repair, and evaluating the mobile device for warranty coverage all may require identification of any cracks in the mobile device's screen and/or in non-screen portions of the device. Individualized manual inspection of mobile device screens for cracks is slow, cumbersome, and can yield inconsistent results. Current automated methods for detecting cracks in other contexts are often over-inclusive resulting in high rates of false-positive crack indications. Accordingly, there is a need for improved methods and systems for automatically detecting cracks in mobile device screens.
Overview
The following disclosure describes systems and methods for automated visual inspection and evaluation of electronic device screens. In particular, at least some embodiments of the present technology enable automatic crack detection in electronic device screens. In one embodiment, an image of an electronic device screen is automatically analyzed to detect cracks therein. First, edges can be automatically identified in the image (e.g., using a Canny edge detector or another edge-detection algorithm). Next, close proximity edge pairings can be identified. For example, for each identified edge, the routine can determine whether a corresponding adjacent edge is present within a predetermined distance. With this approach, edges with nearby adjacent edges can be identified. Such edge-nonedge-edge groupings, or edge pairings, are indicative of a genuine crack rather than an artifact such as a reflection or fingerprint. Further processing can be performed, for example assigning various weights to the identified edges, identifying straight line segments from the identified edges (e.g., using a Hough line transform or another straight-line detector), etc. Based on these analyses, a crack detection score can be calculated.
Certain details are set forth in the following description and in
The accompanying Figures depict embodiments of the present technology and are not intended to be limiting of the scope of the present technology. The sizes of various depicted elements are not necessarily drawn to scale, and these various elements may be arbitrarily enlarged to improve legibility. Component details may be abstracted in the Figures to exclude details such as the position of components and certain precise connections between such components when such details are unnecessary for a complete understanding of how to make and use the invention.
In the Figures, identical reference numbers may identify identical, or at least generally similar, elements. To facilitate the discussion of any particular element, the most significant digit or digits of any reference number may refer to the Figure in which that element is first introduced. For example, element 110 is first introduced and discussed with reference to
Crack Detection in Electronic Devices
The imaging device 101 includes a processing component 111, a memory 113, input and output components 115 and 117, and a power component 125. The imaging device 101 further includes an image sensor 127, associated optics 129 and an illumination source 131. A communication component 119 of the imaging device 101 includes a wired connection 123 and a wireless transceiver 121. The computing device 105 can include several components similar to components of the imaging device 101. For example, the computing device 105 can include a processing component 133, memory 135 (which can store an operating system 137, applications 139, and data 141), along with input 143 and output 145 components and a power component 153. A communication component 147 of the computing device 105 includes a wired connection 151 and a wireless transceiver 147. These features of the imaging device 101 and the computing device 105 are described in more detail below in the context of a routine for detecting cracks in device screens in accordance with an embodiment of the present technology.
In block 203, the routine 200 finds the four corners of the screen region and applies a mask to the image. For example, the mask can separate the image into a screen region and a non-screen region. In some embodiments, the screen region can be dilated with respect to the actually identified four corners of the screen so as to include at least a portion of the area outside of the device screen. The corners of the screen region can be identified using edge detection or other suitable techniques.
In block 205, the routine 200 determines the phone color. For example, the routine 200 can evaluate the average brightness of pixels in the non-screen region of the phone to determine if they exceed a predetermined threshold. Higher average brightness in this region can indicate a white phone, while lower average brightness can indicate a black phone. Other suitable approaches can be used to determine the phone color, for example calculating a color histogram of the non-screen portion of the phone or other such technique.
In block 207, the routine 200 creates an “enhanced black image” in which the image is modified so as to render cracks more visible in black phones or in the screen portion of white phones (which are generally black). This enhanced black image can be created, for example, by calculating the average brightness of the screen portion and determining how much that average brightness must be multiplied by to get some set value (e.g., 140 brightness values). That multiplier factor can be clipped to some maximum, for example 6.3. Other set values of brightness values and maximum multiplier factors can be used. The entire image can then be multiplied by this determined multiplier factor such that both the screen region portion and the non-screen portion have increased brightness levels.
In block 209, the routine 200 determines whether the phone color is white. If not, then the routine 200 continues to block 211 and retains the black enhanced image generated in block 207. This enhanced black image can be used to detect cracks in black phones, as the enhancement applied to both the screen portion and non-screen portion can aid in detecting cracks in those regions. Although this description refers to “white” and “black” devices, the routine can be equally applied to a variety of dark-colored and light-colored devices, and the devices need not be substantially white or substantially black.
If, in block 209, the phone color is white, then the routine 200 continues to block 213 to create an “enhanced white image.” This enhanced white image can be generated by computing a histogram of the entire image and setting all pixels less than some shadow cutoff (e.g., approximately 100 brightness values, or the 25th percentile of brightness in the image). In some embodiments, the shadow cutoff can be approximately 100 brightness values. Next, the routine 200 can set all pixels greater than some highlight cutoff (e.g., approximately 185 brightness values, or the 75th percentile of brightness in the image) to the maximum brightness value of 255. The routine 200 can then stretch all remaining values between the shadow cutoff and the highlight cutoff (e.g., between 100 and 185, or between the 25th and 75th percentiles of brightness in the image) to a range between 0 and 255 brightness values. This produces an enhanced white image in which cracks in the non-screen portions of white phones are more easily visible.
In block 215, the routine 200 merges the enhanced black image and the enhanced white image. For example, the routine 200 can retain the screen portion of the enhanced black image (as the screen portion of the white phone is generally black) and retain the non-screen portion of the enhanced white image (since this non-screen portion of the white phone is white). This merged image includes high visibility of cracks in the screen (black) portion of the image as well as high visibility of cracks in the non-screen (white) portion of the image. After block 215, the routine ends.
In block 253, the routine 250 applies an edge detector (e.g., a Canny edge detector) to the image to identify edges. In a particular example, the edge detector identifies points in the image at which the image brightness changes sharply or has discontinuities. The identified points can then be organized into a set of line segments termed “edges.” A Gaussian or other suitable filter can then be applied to smooth the image and reduce noise. The edge detector can identify intensity gradients in the image. A non-maximum suppression can be applied to eliminate spurious response to edge detection. A threshold (e.g., a double threshold) can then be applied to determine potential edges. Finally, a hysteresis approach can be used to suppress potential edges that are weak and/or not connected to strong edges.
The routine 250 can perform any number of pre-processing steps prior to and/or after applying the edge detector to the image. For example, the image can be enhanced, inverted, cropped, dilated, merged with other images (e.g., merging a high contrast image with a low contrast image), etc.
Although a Canny edge detector is described above, various other edge detection algorithms can be used in its place. For example, the routine can include applying other gradient operator methods such as Extended Prewitt 7×7 and Abdou, thresholding and linking, edge thinning, differential edge detection, phase congruency-based edge detection, etc.
In block 255, the routine 250 determines, for each identified edge, whether a corresponding adjacent edge is present. For example, for each identified edge, the routine can determine whether another identified edge is located within a predetermined distance (e.g., within 10 pixels, within 9, 8, 7, 6, 5, 4, 3, 2, or 1 pixel, etc.). In some embodiments, the predetermined distance can be determined as a number of pixels that correspond to a predetermined physical distance, for example given a resolution and imaging distance such that 7 pixels correspond to approximately 1 millimeter in physical distance on the device, the predetermined distance can be 10 pixels in one example, which correspond to a physical distance of approximately 1.3 mm. In various embodiments, the predetermined distance can correspond to a range of physical distances, e.g., within 2 mm, within 1.5 mm, within 1 mm, within 0.5 mm, etc. In some embodiments, the routine 250 determines, at each point along an identified edge, whether another identified edge falls within the predetermined distance by analyzing the spacing along two or more axes. For example, for a given pixel that has been found to be on one of the identified edges, the routine 250 can determine whether the next pixel to the right is also on an identified edge. If not, but the next pixel to the right (i.e., two pixels to the right of the original pixel) does fall along an identified edge, then the original pixel can be retained. This procedure can be repeated in several directions (e.g., up, down, left, and right), and at a pre-selected proximity threshold (e.g., determining whether another edge is located within 5 pixels, within 7 pixels, etc.). With this approach, the routine 250 can identify edges associated with nearby edges. Such edge-nonedge-edge groupings, or edge pairings, are indicative of a genuine crack rather than an artifact such as a reflection or fingerprint. In block 257, the routine 250 discards those edges for which no adjacent edge was identified. Alternatively, the edges for which no adjacent edge was identified can be assigned a lower weight rather than being discarded entirely.
In the illustrated embodiment, the routine 250 next applies a straight-line detector (e.g., a suitable Hough transform) to the retained edges to identify line segments (block 259). For example, a Hough transform can be applied and varied based on three tuning parameters: (1) the minimum number of intersections to detect a line, (2) the minimum number of points that can form a line, and (3) the maximum gap between two points to be considered in the same line. These three parameters can be varied to maximize the number of cracks identified using such a Hough transform. In one example, the first parameter can have a value of between 10 and 20, e.g., approximately 15; the second parameter can have a value of between 15 and 20, e.g., approximately 18, and the third parameter can have a value between 5 and 15, e.g. approximately 10. Solely by way of theory, and without wishing to be limited to any particular theory, cracks in mobile device screens may tend to be straight or at least capable of being represented by a series of straight lines, whereas extraneous features (e.g., fingerprints and reflections) may be more likely to follow curved paths. In other embodiments, applying the straight-line detector is omitted. For example the routine 250 can further process all of the retained edges from block 257.
With reference again to
As described above with respect to blocks 255 and 257, the output of the edge detector can be further filtered by determining, for each edge, whether a corresponding adjacent edge is present within a predetermined distance. If no adjacent edge is found within the predetermined distance, then the edge is determined not to be a crack and is discarded, otherwise it is retained. Image 305 is the output of this step on the cracked mobile device and image 313 is the output of this step on the non-cracked mobile device. Image 305 shows a reduced number of edges compared to image 303. Similarly, image 313 shows a reduced number of edges compared to image 311. As illustrated, the cracks 304 have been identified and retained in image 305, indicating the presence of edge-nonedge-edge pairings. This step results in a more accurate assessment of the degree of crack damage in the screen, by removing, for example, edged features that are wider than the predetermined distance (e.g., wider than 10 pixels). These relatively wide features are less likely to be genuine cracks and more likely to be reflections, smudges, or other artifacts in the image.
Although the illustration in
Referring again to
Computing Environment
Referring again to
The electronic device 103 can be, for example, a smartphone, a tablet, a laptop, a handheld gaming device, a media player, or any such device that has a screen or other surface that may suffer cracks or similar defects. Although many embodiments of the present technology are described herein in the context of mobile phones, aspects of the present technology are not limited to mobile phones and generally apply to other consumer electronic devices. Such devices include, as non-limiting examples, all manner of mobile phones; smartphones; handheld devices; personal digital assistants (PDAs); MP3 or other digital music players; tablet, notebook, ultrabook and laptop computers; e-readers; all types of cameras; GPS devices; set-top boxes and other media players; VoIP phones; universal remote controls; wearable computers; and larger consumer electronic devices, such as desktop computers, TVs, projectors, DVRs, game consoles, etc.
The computing device 105 can be a desktop computer or another suitable device. The computing device 105 is configured to receive images of the electronic device 103 from the imaging device 101 and to automatically analyze the images to detect the presence of cracks or other defects. In some embodiments, the computing device 105 is remote from the imaging device 101 and can be in communication via the communications link 109. In other embodiments, the computing device 105 is connected to the imaging device 101 via a hardwire connection, or in certain embodiments the imaging device 101 and the computing device 105 are integrated into the same machine. The computing device 105 is also in communication with the database 107 which can store data used in automatically analyzing the images of the electronic device 103. The database 107 may also store the results of the automatic analysis of the images, other data about the electronic device 103, etc.
In the illustrated embodiment, various devices, including the imaging device 101 and the computing device 105, can exchange information with one another via the communication link 109. Although the communication link 109 can include a publicly available network (e.g., the Internet with a web interface), a private communication link, such as an intranet or other network can also be used. Moreover, in various embodiments the imaging device 101 is connected to a host computer (not shown) that facilitates the exchange of information between the imaging device 101, the computing device 105, remote computers, mobile devices, etc.
In the illustrated embodiment, the imaging device 101 includes the processing component 111 that controls operation of the imaging device 101 in accordance with computer-readable instructions stored in memory 113. The processing component 111 may include any logic processing unit, such as one or more central processing units (CPUs), graphics processing units (GPUs), digital signal processors (DSPs), application-specific integrated circuits (ASICs), etc. The processing component 111 may be a single processing unit or multiple processing units in an electronic device or distributed across multiple devices. Aspects of the present technology can be embodied in a special purpose computing device or data processor that is specifically programmed, configured, or constructed to perform one or more of the computer-executable instructions explained in detail herein. Aspects of the present technology can also be practiced in distributed computing environments in which functions or modules are performed by remote processing devices that are linked through a communications network, such as a local area network (LAN), wide area network (WAN), or the Internet. In a distributed computing environment, modules can be located in both local and remote memory storage devices.
The processing component 111 is connected to memory 113, which can include a combination of temporary and/or permanent storage, and both read-only memory (ROM) and writable memory (e.g., random access memory or RAM), writable non-volatile memory such as flash memory or other solid-state memory, hard drives, removable media, magnetically or optically readable discs, nanotechnology memory, biological memory, and so forth. As used herein, memory does not include a transitory propagating signal per se. The memory 213 includes data storage that contains programs, software, and information, such as an operating system and data. Imaging device 101 operating system and data can include software and databases configured to control imaging device 101 components, process images, communicate and exchange data and information with remote computers and other devices, etc.
The imaging device 101 further includes input components 115 that can receive input from user interactions and provide input to the processing component 111, typically mediated by a hardware controller that interprets the raw signals received from the input device and communicates the information to the processing component 111 using a known communication protocol. Examples of an input component 115 include touchpad, a keyboard (with physical or virtual keys), a pointing device (such as a mouse, dial, or eye tracking device), a touchscreen that detects contact events when it is touched by a user, a microphone that receives audio input, etc. The imaging device 101 can also include various other input components 115 such as GPS or other location determination sensors, motion sensors, wearable input devices with accelerometers (e.g. wearable glove-type input devices), biometric sensors (e.g., fingerprint sensors), light sensors, card readers (e.g., magnetic stripe readers or memory card readers) or the like.
The processing component 111 is also connected to one or more various output components 117, e.g., directly or via a hardware controller. The output devices can include a display on which text and graphics are displayed. The display can be, for example, an LCD, LED, or OLED display screen, an e-ink display, a projected display (such as a heads-up display device), and/or a display integrated with a touchscreen that serves as an input device as well as an output device that provides graphical and textual visual feedback to a user. The output components 117 can also include a speaker for playing audio signals, haptic feedback devices for tactile output such as vibration, etc. In some implementations, a speaker and microphone are implemented by a combined audio input-output device.
In the illustrated embodiment, the imaging device 101 further includes one or more communication components 119. The communication components can include, for example, a wireless transceiver 121 (e.g., one or more of a Wi-Fi transceiver; Bluetooth transceiver; near-field communication (NFC) device; wireless modem or cellular radio utilizing GSM, CDMA, 3G and/or 4G technologies; etc.) and/or a wired network connection 123 (e.g., one or more of an Ethernet port, cable modem, FireWire cable, Lightning connector, universal serial bus (USB) port, etc.). The communication components 119 are suitable for communication between the imaging device 101 and other local and/or remote devices, e.g., the computing device 105, directly via a wired or wireless peer-to-peer connection and/or indirectly via the communication link 109 (which can include the Internet, a public or private intranet, a local or extended Wi-Fi network, cell towers, the plain old telephone system (POTS), etc.). For example, the wireless transceiver 121 of the imaging device 101 can connect to a wireless transceiver 149 of the computing device via the wireless connection. The imaging device 101 further includes power 125, which can include battery power and/or facility power for operation of the various electrical components associated with the imaging device 101.
The imaging device 101 further includes the image sensor 127, optics 129, and illumination source 131. The image sensor 127 can be, for example, a CCD sensor, a CMOS sensor, or any other type of image sensor or array of sensors. The image sensor 127 can be aligned with optics 129, for example one or more lenses, filters, or other optical elements configured to orient and modulate incoming light before it reaches the image sensor 127. The illumination source 131 can be configured to direct illumination towards the field of view of the imaging device 101, and can be any type of light source, for example LEDs, fluorescent bulbs, etc. In some embodiments, the illumination source 131 includes multiple different types of light sources which can be individually activated, for example infrared, ultraviolet, broadband, etc.
The computing device 105 includes several components similar to those in the imaging device 101. In the illustrated embodiment, the computing device 105 includes a processing component 133 that controls operation of the computing device 105 in accordance with computer-readable instructions stored in memory 135. The processing component 133 may be any logic processing unit, such as one or more central processing units (CPUs), graphics processing units (GPUs), digital signal processors (DSPs), application-specific integrated circuits (ASICs), etc. The processing component 133 may be a single processing unit or multiple processing units in an electronic device or distributed across multiple devices. The processing component 133 is connected to memory 135, which includes data storage that contains programs, software, and information, such as an operating system 137, application programs 139, and data 141. The operating system 137 can include, for example, Windows®, Linux®, Android™, iOS®, and/or an embedded real-time operating system. The application programs 139 and data 141 can include software and databases configured to control computing device 105 components, process and evaluate images received from the imaging device 101, communicate and exchange data and information with remote computers and other devices, etc.
The computing device 105 can include input components 143, such as a keyboard (with physical or virtual keys), a pointing device (such as a mouse, joystick, dial, or eye tracking device), a touchscreen, a microphone, and a camera for still photograph and/or video capture. The computing device 105 can also include various other input components 143 such as GPS or other location determination sensors, motion sensors, wearable input devices with accelerometers (e.g. wearable glove-type input devices), biometric sensors (e.g., fingerprint sensors), light sensors, card readers (e.g., magnetic stripe readers or memory card readers) and the like.
The processing component 133 can also be connected to one or more various output components 145, e.g., directly or via a hardware controller. The output devices can include a display such as an LCD, LED, or OLED display screen (such as a desktop computer screen, handheld device screen, or television screen), an e-ink display, a projected display (such as a heads-up display device), and/or a display integrated with a touchscreen that serves as an input device as well as an output device that provides graphical and textual visual feedback to the user. The output devices can also include a speaker for playing audio signals, haptic feedback devices for tactile output such as vibration, etc.
In the illustrated embodiment, computing device 105 further includes one or more communication components 147. The communication components can include, for example, a wireless transceiver 149 (e.g., one or more of a Wi-Fi transceiver; Bluetooth transceiver; near-field communication (NFC) device; wireless modem or cellular radio utilizing GSM, CDMA, 3G and/or 4G technologies; etc.) and/or a wired network connector port 251 (e.g., one or more of an Ethernet port, cable modem, FireWire cable, Lightning connector, universal serial bus (USB) port, etc.). The communication components 147 are suitable for communication between the computing device 105 and other local and/or remote computing devices, e.g., the imaging device 101 a via a wired or wireless peer-to-peer connection and/or indirectly via the communication link 109. For example, the wireless transceiver 149 of the computing device 105 can connect to the wireless transceiver 121 of imaging device 101, and/or the wired connector port 151 of the computing device 105 can connect to the wired connector port 123 of the imaging device 101. The computing device 105 further includes power 153, which can include battery power and/or facility power for operation of the various electrical components associated with the computing device 105.
Unless described otherwise, the construction and operation of the various components shown in
Kiosk Environment
In some embodiments, the routines described herein can be carried out using a kiosk that includes an imaging device (e.g., the imaging device 101) therein. In some embodiments, the kiosk can perform some or all of the functions performed by the computing device 105 described above, for example processing and evaluating images received from the imaging device 101. The kiosk can include, for example, a processing component and memory storing instructions that, when executed by the processing component, perform operations such as the routine 250 described above.
In the illustrated embodiment, the kiosk 400 is a floor-standing self-service kiosk configured for use by a user 401 (e.g., a consumer, customer, etc.) to recycle, sell, and/or perform other operations with a mobile phone or other consumer electronic device. In other embodiments, the kiosk 400 can be configured for use on a countertop or a similar raised surface. Although the kiosk 400 is configured for use by consumers, in various embodiments the kiosk 400 and/or various portions thereof can also be used by other operators, such as a retail clerk or kiosk assistant to facilitate the selling or other processing of mobile phones and other electronic devices.
In the illustrated embodiment, the kiosk 400 includes a housing 402 that is approximately the size of a conventional vending machine. The housing 402 can be of conventional manufacture from, for example, sheet metal, plastic panels, etc. A plurality of user interface devices are provided on a front portion of the housing 402 for providing instructions and other information to users, and/or for receiving user inputs and other information from users. For example, the kiosk 400 can include a display screen 404 (e.g., a liquid crystal display (“LCD”) or light emitting diode (“LED”) display screen, a projected display (such as a heads-up display or a head-mounted device), and so on) for providing information, prompts, etc., to users. The display screen 404 can include a touch screen for receiving user input and responses to displayed prompts. In addition or alternatively, the kiosk 400 can include a separate keyboard or keypad for this purpose. The kiosk 400 can also include an ID reader or scanner 412 (e.g., a driver's license scanner), a fingerprint reader 414, and one or more cameras 416 (e.g., digital still and/or video cameras, identified individually as cameras 416a-c). The kiosk 400 can additionally include output devices such as a label printer having an outlet 410, and a cash dispenser having an outlet 418. Although not identified in
A sidewall portion of the housing 402 can include a number of conveniences to help users recycle or otherwise process their mobile phones. For example, in the illustrated embodiment the kiosk 400 includes an accessory bin 428 that is configured to receive mobile device accessories that the user wishes to recycle or otherwise dispose of. Additionally, the kiosk 400 can provide a free charging station 426 with a plurality of electrical connectors 424 for charging a wide variety of mobile phones and other consumer electronic devices.
In the illustrated embodiment, the inspection plate 544 is configured to translate back and forth (on, e.g., parallel mounting tracks) to move an electronic device, such as the mobile phone 550, between a first position directly behind the access door 406 and a second position between an upper chamber 530 and an opposing lower chamber 532. Moreover, in this embodiment the inspection plate 544 is transparent, or at least partially transparent (e.g., formed of glass, Plexiglas, etc.) to enable the mobile phone 550 to be photographed and/or otherwise optically evaluated from all, or at least most viewing angles (e.g., top, bottom, sides, etc.) using, e.g., one or more cameras, mirrors, etc. mounted to or otherwise associated with the upper and lower chambers 530 and 532. When the mobile phone 550 is in the second position, the upper chamber 530 can translate downwardly to generally enclose the mobile phone 550 between the upper chamber 530 and the lower chamber 532. The upper chamber 530 is operably coupled to a gate 538 that moves up and down in unison with the upper chamber 530.
In some embodiments, the kiosk 400 includes the imaging device 101 disposed within the upper hood 530. The imaging device 101 can be used as described above to facilitate visual inspection of the mobile phone 550 in order to detect cracks in the screen. The upper chamber 530 and/or the lower chamber 532 can also include one or more magnification tools, scanners (e.g., bar code scanners, infrared scanners, etc.) or other imaging components (not shown) and an arrangement of mirrors (also not shown) to view, photograph and/or otherwise visually evaluate the mobile phone 550 from multiple perspectives. In some embodiments, one or more of the cameras and/or other imaging components discussed above can be movable to facilitate device evaluation. For example, as noted above with respect to
The kiosk 400 can be used in a number of different ways to efficiently facilitate the recycling, selling and/or other processing of mobile phones and other consumer electronic devices. Referring to
Referring next to
After the visual and electronic analysis of the mobile phone 550, the user is presented with a phone purchase price via the display screen 404. If the user declines the price (via, e.g., the touch screen), a retraction mechanism (not shown) automatically disconnects the connector 542 from the mobile phone 550, the door 406 opens, and the user can reach in and retrieve the mobile phone 550. If the user accepts the price, the door 406 remains closed and the user may be prompted to place his or her identification (e.g., a driver's license) in the ID scanner 412 and provide a thumbprint via the fingerprint reader 414. As a fraud prevention measure, the kiosk 400 can be configured to transmit an image of the driver's license to a remote computer screen, and an operator at the remote computer can visually compare the picture (and/or other information) on the driver's license to an image of the person standing in front of the kiosk 400 as viewed by one or more of the cameras 416a-c (
As those of ordinary skill in the art will appreciate, the foregoing routines are but some examples of ways in which the kiosk 400 can be used to recycle or otherwise process consumer electronic devices such as mobile phones. Although the foregoing example is described in the context of mobile phones, it should be understood that the kiosk 400 and various embodiments thereof can also be used in a similar manner for recycling virtually any consumer electronic device, such as MP3 players, tablet computers, PDAs, and other portable devices, as well as other relatively non-portable electronic devices such as desktop computers, printers, devices for implementing games, entertainment or other digital media on CDs, DVDs, Blu-ray, etc. Moreover, although the foregoing example is described in the context of use by a consumer, the kiosk 400 in various embodiments thereof can similarly be used by others, such as a store clerk, to assist consumers in recycling, selling, exchanging, etc. their electronic devices.
The disclosed technology also includes the disclosures of U.S. patent application Ser. No. 14/498,763, titled “METHODS AND SYSTEMS FOR PRICING AND PERFORMING OTHER PROCESSES ASSOCIATED WITH RECYCLING MOBILE PHONES AND OTHER ELECTRONIC DEVICES,” filed by the applicant on Sep. 26, 2014; U.S. patent application Ser. No. 14/500,739, titled “MAINTAINING SETS OF CABLE COMPONENTS USED FOR WIRED ANALYSIS, CHARGING, OR OTHER INTERACTION WITH PORTABLE ELECTRONIC DEVICES,” filed by the applicant on Sep. 29, 2014; U.S. patent application Ser. No. 14/873,158, titled “WIRELESS-ENABLED KIOSK FOR RECYCLING CONSUMER DEVICES,” filed by the applicant on Oct. 1, 2015; U.S. patent application Ser. No. 14/873,145, titled “APPLICATION FOR DEVICE EVALUATION AND OTHER PROCESSES ASSOCIATED WITH DEVICE RECYCLING,” filed by the applicant on Oct. 1, 2015; U.S. patent application Ser. No. 14/506,449, titled “SYSTEM FOR ELECTRICALLY TESTING MOBILE DEVICES AT A CONSUMER-OPERATED KIOSK, AND ASSOCIATED DEVICES AND METHODS,” filed by the applicant on Oct. 3, 2014; U.S. patent application Ser. No. 14/925,357, titled “SYSTEMS AND METHODS FOR RECYCLING CONSUMER ELECTRONIC DEVICES,” U.S. patent application Ser. No. 14/925,375, titled “METHODS AND SYSTEMS FOR FACILITATING PROCESSES ASSOCIATED WITH INSURANCE SERVICES AND/OR OTHER SERVICES FOR ELECTRONIC DEVICES,” filed by the applicant on Oct. 28, 2015; U.S. patent application Ser. No. 14/964,963, titled “METHODS AND SYSTEMS FOR PROVIDING INFORMATION REGARDING COUPONS/PROMOTIONS AT KIOSKS FOR RECYCLING MOBILE PHONES AND OTHER ELECTRONIC DEVICES,” filed by the applicant on Dec. 10, 2015; U.S. patent application Ser. No. 14/568,051, titled “METHODS AND SYSTEMS FOR IDENTIFYING MOBILE PHONES AND OTHER ELECTRONIC DEVICES,” filed by the applicant on Dec. 11, 2014; U.S. patent application Ser. No. 14/966,346, titled “SYSTEMS AND METHODS FOR RECYCLING CONSUMER ELECTRONIC DEVICES,” filed by the applicant on Dec. 11, 2015; U.S. patent application Ser. No. 14/598,469, titled “METHODS AND SYSTEMS FOR DYNAMIC PRICING AND PERFORMING OTHER PROCESSES ASSOCIATED WITH RECYCLING MOBILE PHONES AND OTHER ELECTRONIC DEVICES,” filed by the applicant on Jan. 16, 2015; U.S. patent application Ser. No. 14/660,768, titled “SYSTEMS AND METHODS FOR INSPECTING MOBILE DEVICES AND OTHER CONSUMER ELECTRONIC DEVICES WITH A LASER,” filed by the applicant on Mar. 17, 2015; U.S. patent application Ser. No. 14/663,331, titled “DEVICE RECYCLING SYSTEMS WITH FACIAL RECOGNITION,” filed by the applicant on Mar. 19, 2015; U.S. provisional application No. 62/169,072, titled “METHODS AND SYSTEMS FOR VISUALLY EVALUATING ELECTRONIC DEVICES,” filed by the applicant on Jun. 1, 2015; U.S. provisional application No. 62/202,330, titled “METHODS AND SYSTEMS FOR INSPECTING MOBILE DEVICES AND OTHER CONSUMER ELECTRONIC DEVICES WITH ROBOTIC ACTUATION,” filed by the applicant on Aug. 7, 2015; U.S. patent application Ser. No. 15/057,707, titled “METHODS AND SYSTEMS FOR RECORDING INTERACTIONS WITH A SYSTEM FOR PURCHASING MOBILE PHONES AND OTHER ELECTRONIC DEVICES,” filed by the applicant on Mar. 1, 2016; U.S. patent application Ser. No. 14/873,158, titled “WIRELESS-ENABLED KIOSK FOR RECYCLING CONSUMER DEVICES,” filed by the applicant on Oct. 1, 2015; U.S. patent application Ser. No. 14/873,145, titled “APPLICATION FOR DEVICE EVALUATION AND OTHER PROCESSES ASSOCIATED WITH DEVICE RECYCLING,” filed by the applicant on Oct. 1, 2015; U.S. patent application Ser. No. 14/925,357, titled “SYSTEMS AND METHODS FOR RECYCLING CONSUMER ELECTRONIC DEVICES,” filed by the applicant on Oct. 28, 2015; U.S. patent application Ser. No. 14/925,375, titled “METHODS AND SYSTEMS FOR FACILITATING PROCESSES ASSOCIATED WITH INSURANCE SERVICES AND/OR OTHER SERVICES FOR ELECTRONIC DEVICES,” filed by the applicant on Oct. 28, 2015; U.S. patent application Ser. No. 14/934,134, titled “METHODS AND SYSTEMS FOR EVALUATING AND RECYCLING ELECTRONIC DEVICES,” and U.S. patent application Ser. No. 14/967,183, titled “SYSTEMS AND METHODS FOR RECYCLING CONSUMER ELECTRONIC DEVICES,” filed Dec. 11, 2015, each of which is incorporated herein by reference in its entirety. All of the patents and patent applications listed above are commonly owned by the applicant of the present application, and they along with any other patents or patent applications identified herein are incorporated herein by reference in their entireties.
While the Internet is shown, a private network, such as an intranet may likewise be used herein. The network may have a client-server architecture, in which a computer is dedicated to serving other client computers, or it may have other architectures such as peer-to-peer, in which one or more computers serve simultaneously as servers and clients. A database or databases, coupled to the server computer(s), stores much of the web pages and content exchanged between the user computers. The server computer(s), including the database(s), may employ security measures to inhibit malicious attacks on the system and preserve the integrity of the messages and data stored therein (e.g., firewall systems, message encryption and/or authentication (e.g., using transport layer security (TLS) or secure socket layers (SSL)), password protection schemes, encryption of stored data (e.g., using trusted computing hardware), and the like).
One skilled in the relevant art will appreciate that the concepts of the invention can be used in various environments other than location based or the Internet. In general, a display description may be in HTML, XML or WAP format, email format or any other format suitable for displaying information (including character/code-based formats, algorithm-based formats (e.g., vector generated), and bitmapped formats). Also, various communication channels, such as local area networks, wide area networks, or point-to-point dial-up connections, may be used instead of the Internet. The system may be conducted within a single computer environment, rather than a client/server environment. Also, the user computers may comprise any combination of hardware or software that interacts with the server computer, such as television-based systems and various other consumer products through which commercial or noncommercial transactions can be conducted. The various aspects of the invention described herein can be implemented in or for any e-mail environment.
Although not required, aspects of the invention are described in the general context of computer-executable instructions, such as routines executed by a general-purpose data processing device, e.g., a server computer, wireless device or personal computer. Those skilled in the relevant art will appreciate that aspects of the invention can be practiced with other communications, data processing, or computer system configurations, including Internet appliances, hand-held devices (including personal digital assistants (PDAs)), wearable computers, all manner of cellular or mobile phones (including Voice over IP (VoIP) phones), dumb terminals, media players, gaming devices, multi-processor systems, microprocessor-based or programmable consumer electronics, set-top boxes, network PCs, mini-computers, mainframe computers, and the like. Indeed, the terms “computer,” “server,” “host,” “host system,” and the like, are generally used interchangeably herein, and refer to any of the above devices and systems, as well as any data processor. Input devices may include a touchpad, keyboard and/or a pointing device such as a mouse. Other input devices are possible such as a microphone, joystick, pen, game pad, scanner, digital camera, video camera, and the like. The data storage devices may include any type of computer-readable media that can store data accessible by a computer, such as magnetic hard and floppy disk drives, optical disk drives, magnetic cassettes, tape drives, flash memory cards, digital video disks (DVDs), Bernoulli cartridges, RAMs, ROMs, smart cards, etc. Indeed, any medium for storing or transmitting computer-readable instructions and data may be employed, including a connection port to a network such as a local area network (LAN), wide area network (WAN) or the Internet.
Aspects of the invention can be embodied in a special purpose computer or data processor that is specifically programmed, configured, or constructed to perform one or more of the computer-executable instructions explained in detail herein. While aspects of the invention, such as certain functions, are described as being performed exclusively on a single device, the invention can also be practiced in distributed environments where functions or modules are shared among disparate processing devices, which are linked through a communications network, such as a Local Area Network (LAN), Wide Area Network (WAN), or the Internet. In a distributed computing environment, program modules may be located in both local and remote memory storage devices.
Aspects of the invention may be stored or distributed on tangible computer-readable media, including magnetically or optically readable computer discs, hard-wired or preprogrammed chips (e.g., EEPROM semiconductor chips), nanotechnology memory, biological memory, or other data storage media. The data storage devices may include any type of computer-readable media that can store data accessible by a computer, such as magnetic hard and floppy disk drives, optical disk drives, magnetic cassettes, tape drives, flash memory cards, DVDs, Bernoulli cartridges, RAM, ROMs, smart cards, etc. Indeed, any medium for storing or transmitting computer-readable instructions and data may be employed, including a connection port to a network such as a LAN, WAN, or the Internet. Alternatively, computer implemented instructions, data structures, screen displays, and other data under aspects of the invention may be distributed over the Internet or over other networks (including wireless networks), on a propagated signal on a propagation medium (e.g., an electromagnetic wave(s), a sound wave, etc.) over a period of time, or they may be provided on any analog or digital network (packet switched, circuit switched, or other scheme). The terms “memory” and “computer-readable storage medium” include any combination of temporary, persistent, and/or permanent storage, e.g., ROM, writable memory such as RAM, writable non-volatile memory such as flash memory, hard drives, solid state drives, removable media, and so forth, but do not include a propagating signal per se.
The above Detailed Description of examples and embodiments of the invention is not intended to be exhaustive or to limit the invention to the precise form disclosed above. While specific examples for the invention are described above for illustrative purposes, various equivalent modifications are possible within the scope of the invention, as those skilled in the relevant art will recognize. References throughout the foregoing description to features, advantages, or similar language do not imply that all of the features and advantages that may be realized with the present technology should be or are in any single embodiment of the invention. Rather, language referring to the features and advantages is understood to mean that a specific feature, advantage, or characteristic described in connection with an embodiment is included in at least one embodiment of the present technology. Thus, discussion of the features and advantages, and similar language, throughout this specification may, but do not necessarily, refer to the same embodiment. Furthermore, the described features, advantages, and characteristics of the present technology may be combined in any suitable manner in one or more embodiments. One skilled in the relevant art will recognize that the present technology can be practiced without one or more of the specific features or advantages of a particular embodiment. In other instances, additional features and advantages may be recognized in certain embodiments that may not be present in all embodiments of the present technology.
Any patents and applications and other references noted above, including any that may be listed in accompanying filing papers, are incorporated herein by reference. Aspects of the invention can be modified, if necessary, to employ the systems, functions, and concepts of the various references described above to provide yet further implementations of the invention.
Unless the context clearly requires otherwise, throughout the description and the claims, the words “comprise,” “comprising,” and the like, are to be construed in an inclusive sense, as opposed to an exclusive or exhaustive sense; that is to say, in the sense of “including, but not limited to.” As used herein, the terms “connected,” “coupled,” or any variant thereof, means any connection or coupling, either direct or indirect, between two or more elements; the coupling or connection between the elements can be physical, logical, or a combination thereof. Additionally, the words “herein,” “above,” “below,” and words of similar import, when used in this application, refer to this application as a whole and not to any particular portions of this application. Where the context permits, words in the above Detailed Description using the singular or plural number may also include the plural or singular number respectively. The word “or,” in reference to a list of two or more items, covers all of the following interpretations of the word: any of the items in the list, all of the items in the list, and any combination of the items in the list.
The teachings of the invention provided herein can be applied to other systems, not necessarily the system described above. The elements and acts of the various examples described above can be combined to provide further implementations of the invention. Some alternative implementations of the invention may include not only additional elements to those implementations noted above, but also may include fewer elements. Further any specific numbers noted herein are only examples—alternative implementations may employ differing values or ranges.
While the above description describes various embodiments of the invention and the best mode contemplated, regardless how detailed the above text, the invention can be practiced in many ways. Details of the system may vary considerably in its specific implementation, while still being encompassed by the present disclosure. As noted above, particular terminology used when describing certain features or aspects of the invention should not be taken to imply that the terminology is being redefined herein to be restricted to any specific characteristics, features, or aspects of the invention with which that terminology is associated. In general, the terms used in the following claims should not be construed to limit the invention to the specific examples disclosed in the specification, unless the above Detailed Description section explicitly defines such terms. Accordingly, the actual scope of the invention encompasses not only the disclosed examples, but also all equivalent ways of practicing or implementing the invention under the claims.
From the foregoing, it will be appreciated that specific embodiments of the invention have been described herein for purposes of illustration, but that various modifications may be made without deviating from the spirit and scope of the various embodiments of the invention. Further, while various advantages associated with certain embodiments of the invention have been described above in the context of those embodiments, other embodiments may also exhibit such advantages, and not all embodiments need necessarily exhibit such advantages to fall within the scope of the invention. Accordingly, the invention is not limited, except as by the appended claims. Although certain aspects of the invention are presented below in certain claim forms, the applicant contemplates the various aspects of the invention in any number of claim forms. Accordingly, the applicant reserves the right to pursue additional claims after filing this application to pursue such additional claim forms, in either this application or in a continuing application.
Number | Name | Date | Kind |
---|---|---|---|
3808439 | Renius | Apr 1974 | A |
4248334 | Hanley et al. | Feb 1981 | A |
4519522 | McElwee | May 1985 | A |
4715709 | Sekine et al. | Dec 1987 | A |
4821118 | Lafreniere | Apr 1989 | A |
4870357 | Young et al. | Sep 1989 | A |
4878736 | Hekker et al. | Nov 1989 | A |
4927051 | Falk et al. | May 1990 | A |
4951308 | Bishop et al. | Aug 1990 | A |
5027074 | Haferstat | Jun 1991 | A |
5077462 | Newell et al. | Dec 1991 | A |
5091773 | Fouche et al. | Feb 1992 | A |
5105149 | Tokura | Apr 1992 | A |
5216502 | Katz | Jun 1993 | A |
5280170 | Baldwin | Jan 1994 | A |
5319459 | Mochizuki et al. | Jun 1994 | A |
5339096 | Beaufort et al. | Aug 1994 | A |
5419438 | Squyres et al. | May 1995 | A |
5436554 | Decker | Jul 1995 | A |
5570920 | Crisman et al. | Nov 1996 | A |
5572444 | Lentz et al. | Nov 1996 | A |
5610710 | Canfield et al. | Mar 1997 | A |
5717780 | Mitsumune et al. | Feb 1998 | A |
5747784 | Walter et al. | May 1998 | A |
5775806 | Allred | Jul 1998 | A |
5839058 | Phillips et al. | Nov 1998 | A |
5920338 | Katz | Jul 1999 | A |
5949901 | Nichani et al. | Sep 1999 | A |
5965858 | Suzuki et al. | Oct 1999 | A |
5966654 | Croughwell et al. | Oct 1999 | A |
5987159 | Nichani | Nov 1999 | A |
5988431 | Roe | Nov 1999 | A |
6029851 | Jenkins et al. | Feb 2000 | A |
6041229 | Turner | Mar 2000 | A |
6181805 | Koike et al. | Jan 2001 | B1 |
6228008 | Pollington et al. | May 2001 | B1 |
6259827 | Nichani | Jul 2001 | B1 |
6264104 | Jenkins et al. | Jul 2001 | B1 |
6330354 | Companion et al. | Dec 2001 | B1 |
6330958 | Ruskin et al. | Dec 2001 | B1 |
6393095 | Robinson | May 2002 | B1 |
6462644 | Howell et al. | Oct 2002 | B1 |
6529837 | Kang | Mar 2003 | B1 |
6535637 | Wootton et al. | Mar 2003 | B1 |
6573886 | Lehtinen et al. | Jun 2003 | B1 |
6587581 | Matsuyama et al. | Jul 2003 | B1 |
6596684 | Casagrande et al. | Jul 2003 | B2 |
6633377 | Weiss et al. | Oct 2003 | B1 |
6667800 | Larsson et al. | Dec 2003 | B1 |
6754637 | Stenz | Jun 2004 | B1 |
6758370 | Cooke et al. | Jul 2004 | B2 |
6798528 | Hartman | Sep 2004 | B1 |
6822422 | Sagawa | Nov 2004 | B2 |
6842596 | Morii et al. | Jan 2005 | B2 |
6854656 | Matsumori | Feb 2005 | B2 |
7069236 | Tsunenari | Jun 2006 | B1 |
7076449 | Tsunenari et al. | Jul 2006 | B2 |
7178720 | Strubbe et al. | Feb 2007 | B1 |
7234609 | DeLazzer et al. | Jun 2007 | B2 |
7251458 | O'Connell | Jul 2007 | B2 |
7268345 | Schultz | Sep 2007 | B2 |
7334729 | Brewington | Feb 2008 | B2 |
7520666 | Pevzner et al. | Apr 2009 | B2 |
7567344 | LeBlanc et al. | Jul 2009 | B2 |
7646193 | Suzuki et al. | Jan 2010 | B2 |
7649450 | Campion et al. | Jan 2010 | B2 |
7702108 | Amon et al. | Apr 2010 | B2 |
7735125 | Alvarez et al. | Jun 2010 | B1 |
7761331 | Low et al. | Jul 2010 | B2 |
7783379 | Beane et al. | Aug 2010 | B2 |
7881965 | Bowles et al. | Feb 2011 | B2 |
8010402 | Sharma et al. | Aug 2011 | B1 |
8019588 | Wohlberg et al. | Sep 2011 | B1 |
8025229 | Hammond et al. | Sep 2011 | B2 |
8031930 | Wang et al. | Oct 2011 | B2 |
8107243 | Guccione et al. | Jan 2012 | B2 |
8112325 | Foy et al. | Feb 2012 | B2 |
8195511 | Bowles et al. | Jun 2012 | B2 |
8200533 | Librizzi et al. | Jun 2012 | B2 |
8254883 | Uchida | Aug 2012 | B2 |
8266008 | Siegel et al. | Sep 2012 | B1 |
8423404 | Bowles et al. | Apr 2013 | B2 |
8463646 | Bowles et al. | Jun 2013 | B2 |
8718717 | Vaknin et al. | May 2014 | B2 |
8743215 | Lee | Jun 2014 | B1 |
8824136 | Interian et al. | Sep 2014 | B1 |
9043026 | Lien et al. | May 2015 | B2 |
9582101 | Chang et al. | Feb 2017 | B2 |
9595238 | Won | Mar 2017 | B2 |
20010039531 | Aoki | Nov 2001 | A1 |
20020014577 | Ulrich et al. | Feb 2002 | A1 |
20020035515 | Moreno | Mar 2002 | A1 |
20020067184 | Smith et al. | Jun 2002 | A1 |
20020087413 | Mahaffy et al. | Jul 2002 | A1 |
20020129170 | Moore et al. | Sep 2002 | A1 |
20020157033 | Cox | Oct 2002 | A1 |
20020162966 | Yoder | Nov 2002 | A1 |
20020186878 | Hoon et al. | Dec 2002 | A1 |
20030006277 | Maskatiya et al. | Jan 2003 | A1 |
20030036866 | Nair et al. | Feb 2003 | A1 |
20030061150 | Kocher | Mar 2003 | A1 |
20030146898 | Kawasaki et al. | Aug 2003 | A1 |
20030170529 | Sagawa | Sep 2003 | A1 |
20030197782 | Ashe | Oct 2003 | A1 |
20030204289 | Banerjee et al. | Oct 2003 | A1 |
20040012825 | Tesavis | Jan 2004 | A1 |
20040114153 | Andersen et al. | Jun 2004 | A1 |
20040141320 | Bock et al. | Jul 2004 | A1 |
20040150815 | Sones et al. | Aug 2004 | A1 |
20040156557 | Van Der Weij | Aug 2004 | A1 |
20040156667 | Berger et al. | Aug 2004 | A1 |
20040186744 | Lux | Sep 2004 | A1 |
20040205015 | DeLaCruz | Oct 2004 | A1 |
20040235513 | O'Connell | Nov 2004 | A1 |
20040242216 | Boutsikakis | Dec 2004 | A1 |
20040262521 | Devitt et al. | Dec 2004 | A1 |
20050027622 | Walker et al. | Feb 2005 | A1 |
20050128551 | Yang | Jun 2005 | A1 |
20050139661 | Eglen et al. | Jun 2005 | A1 |
20050143149 | Becker et al. | Jun 2005 | A1 |
20050167620 | Cho et al. | Aug 2005 | A1 |
20050187657 | Hashimoto et al. | Aug 2005 | A1 |
20050216120 | Rosenberg et al. | Sep 2005 | A1 |
20050222690 | Wang et al. | Oct 2005 | A1 |
20050231595 | Wang et al. | Oct 2005 | A1 |
20050240958 | Nguyen et al. | Oct 2005 | A1 |
20060167580 | Whittier | Jul 2006 | A1 |
20060184379 | Tan et al. | Aug 2006 | A1 |
20060195384 | Bauer et al. | Aug 2006 | A1 |
20060217152 | Fok et al. | Sep 2006 | A1 |
20060229108 | Cehelnik | Oct 2006 | A1 |
20060235747 | Hammond et al. | Oct 2006 | A1 |
20060261931 | Cheng | Nov 2006 | A1 |
20060271431 | Wehr et al. | Nov 2006 | A1 |
20060279307 | Wang et al. | Dec 2006 | A1 |
20060280356 | Yamagashi | Dec 2006 | A1 |
20070013124 | Graef et al. | Jan 2007 | A1 |
20070057815 | Foy et al. | Mar 2007 | A1 |
20070129906 | Stoecker et al. | Jun 2007 | A1 |
20070133844 | Waehner et al. | Jun 2007 | A1 |
20070140310 | Rolton et al. | Jun 2007 | A1 |
20070150403 | Mock et al. | Jun 2007 | A1 |
20070205751 | Suzuki et al. | Sep 2007 | A1 |
20070263099 | Motta et al. | Nov 2007 | A1 |
20070269099 | Nishino | Nov 2007 | A1 |
20070281734 | Mizrachi | Dec 2007 | A1 |
20080004828 | Mizrachi | Jan 2008 | A1 |
20080027581 | Saether et al. | Jan 2008 | A1 |
20080033596 | Fausak et al. | Feb 2008 | A1 |
20080097770 | Low et al. | Apr 2008 | A1 |
20080109746 | Mayer | May 2008 | A1 |
20080111989 | Dufour et al. | May 2008 | A1 |
20080149720 | Colville | Jun 2008 | A1 |
20080177598 | Davie | Jul 2008 | A1 |
20080207198 | Juric | Aug 2008 | A1 |
20080231113 | Guccione et al. | Sep 2008 | A1 |
20080255901 | Carroll et al. | Oct 2008 | A1 |
20080256008 | Kwok | Oct 2008 | A1 |
20080281691 | Pearson et al. | Nov 2008 | A1 |
20080296374 | Gonen et al. | Dec 2008 | A1 |
20080303915 | Omi | Dec 2008 | A1 |
20080306701 | Zhong et al. | Dec 2008 | A1 |
20090051907 | Li et al. | Feb 2009 | A1 |
20090078775 | Giebel et al. | Mar 2009 | A1 |
20090079388 | Reddy | Mar 2009 | A1 |
20090095047 | Patel et al. | Apr 2009 | A1 |
20090184865 | Valo et al. | Jul 2009 | A1 |
20090190142 | Taylor et al. | Jul 2009 | A1 |
20090207743 | Huq et al. | Aug 2009 | A1 |
20090244285 | Chathukutty | Oct 2009 | A1 |
20090247133 | Holmen et al. | Oct 2009 | A1 |
20090251815 | Wang et al. | Oct 2009 | A1 |
20090262341 | Konopa et al. | Oct 2009 | A1 |
20090265035 | Jenkinson et al. | Oct 2009 | A1 |
20090299543 | Cox et al. | Dec 2009 | A1 |
20090312009 | Fishel | Dec 2009 | A1 |
20090321511 | Browne | Dec 2009 | A1 |
20100005004 | Hudak et al. | Jan 2010 | A1 |
20100063894 | Lundy | Mar 2010 | A1 |
20100088192 | Bowles et al. | Apr 2010 | A1 |
20100110174 | Leconte | May 2010 | A1 |
20100115887 | Schroeder et al. | May 2010 | A1 |
20100147953 | Barkan | Jun 2010 | A1 |
20100157280 | Kusevic et al. | Jun 2010 | A1 |
20100161397 | Gauthier et al. | Jun 2010 | A1 |
20100169231 | Bowles et al. | Jul 2010 | A1 |
20100185506 | Wolff | Jul 2010 | A1 |
20100219234 | Forbes | Sep 2010 | A1 |
20100228676 | Librizzi et al. | Sep 2010 | A1 |
20100235198 | Fini et al. | Sep 2010 | A1 |
20100237854 | Kumhyr et al. | Sep 2010 | A1 |
20100262481 | Baker et al. | Oct 2010 | A1 |
20110035322 | Lively | Feb 2011 | A1 |
20110043628 | Yun | Feb 2011 | A1 |
20110060641 | Grossman et al. | Mar 2011 | A1 |
20110067520 | Ihrke et al. | Mar 2011 | A1 |
20110235853 | Bowles et al. | Sep 2011 | A1 |
20110313840 | Mason et al. | Dec 2011 | A1 |
20120016518 | Saario et al. | Jan 2012 | A1 |
20120029985 | Wilson et al. | Feb 2012 | A1 |
20120030097 | Hagan et al. | Feb 2012 | A1 |
20120054113 | Jayaraman et al. | Mar 2012 | A1 |
20120063501 | Aguren | Mar 2012 | A1 |
20120078413 | Baker, Jr. | Mar 2012 | A1 |
20120116928 | Gventer et al. | May 2012 | A1 |
20120116929 | Gventer et al. | May 2012 | A1 |
20120117001 | Gventer et al. | May 2012 | A1 |
20120127307 | Hassenzahl | May 2012 | A1 |
20120146956 | Jenkinson | Jun 2012 | A1 |
20120191562 | Bowles et al. | Jul 2012 | A1 |
20120235812 | De Mello et al. | Sep 2012 | A1 |
20120254046 | Librizzi et al. | Oct 2012 | A1 |
20130006713 | Haake et al. | Jan 2013 | A1 |
20130034305 | Jahanshahi | Feb 2013 | A1 |
20130046611 | Bowles et al. | Feb 2013 | A1 |
20130046699 | Bowles et al. | Feb 2013 | A1 |
20130124426 | Bowles et al. | May 2013 | A1 |
20130126741 | Srivastava et al. | May 2013 | A1 |
20130144797 | Bowles et al. | Jun 2013 | A1 |
20130155061 | Jahanshahi | Jun 2013 | A1 |
20130173434 | Hartman | Jul 2013 | A1 |
20130181935 | McKenzie et al. | Jul 2013 | A1 |
20130191236 | Bowles | Jul 2013 | A1 |
20130198089 | Bowles | Aug 2013 | A1 |
20130198144 | Bowles | Aug 2013 | A1 |
20130226679 | Bowles | Aug 2013 | A1 |
20130246212 | Sullivan | Sep 2013 | A1 |
20130253700 | Carson et al. | Sep 2013 | A1 |
20130275314 | Bowles | Oct 2013 | A1 |
20130284805 | Kraft et al. | Oct 2013 | A1 |
20130290146 | West et al. | Oct 2013 | A1 |
20130297388 | Kyle, Jr. et al. | Nov 2013 | A1 |
20140012643 | Behrisch | Jan 2014 | A1 |
20140038556 | De Sousa | Feb 2014 | A1 |
20140067710 | Gventer et al. | Mar 2014 | A1 |
20140143161 | Ahn | May 2014 | A1 |
20140150100 | Gupta et al. | May 2014 | A1 |
20140156883 | Bowles | Jun 2014 | A1 |
20140214505 | Shuster-Arechiga et al. | Jul 2014 | A1 |
20140244315 | Cahill et al. | Aug 2014 | A1 |
20140347473 | Wolff et al. | Nov 2014 | A1 |
20150006281 | Takahashi | Jan 2015 | A1 |
20150066677 | Bowles et al. | Mar 2015 | A1 |
20150120485 | Nash | Apr 2015 | A1 |
20150193797 | Gerrity | Jul 2015 | A1 |
20150278529 | Cho et al. | Oct 2015 | A1 |
20160019685 | Nguyen | Jan 2016 | A1 |
20160055392 | Nakano | Feb 2016 | A1 |
20160091549 | Snook et al. | Mar 2016 | A1 |
20160092849 | Cirannek et al. | Mar 2016 | A1 |
20160098688 | Hunt et al. | Apr 2016 | A1 |
20160098689 | Bowles et al. | Apr 2016 | A1 |
20160098690 | Silva et al. | Apr 2016 | A1 |
20160125367 | Bowles et al. | May 2016 | A1 |
20160125548 | Bowles et al. | May 2016 | A1 |
20160125612 | Seki | May 2016 | A1 |
20160132840 | Bowles et al. | May 2016 | A1 |
20160171544 | Heminger et al. | Jun 2016 | A1 |
20160171575 | Bowles et al. | Jun 2016 | A1 |
20160210648 | Cirannek | Jul 2016 | A1 |
20160269401 | Saito et al. | Sep 2016 | A1 |
20160275460 | Ploetner et al. | Sep 2016 | A1 |
20160275518 | Bowles et al. | Sep 2016 | A1 |
20170083886 | Silva et al. | Mar 2017 | A1 |
20170091823 | Adinarayan et al. | Mar 2017 | A1 |
20170169401 | Beane et al. | Jun 2017 | A1 |
20170278191 | Tassone et al. | Sep 2017 | A1 |
Number | Date | Country |
---|---|---|
1365479 | Aug 2002 | CN |
2708415 | Jul 2005 | CN |
1864088 | Nov 2006 | CN |
1957320 | May 2007 | CN |
200965706 | Oct 2007 | CN |
102246384 | Nov 2011 | CN |
202351953 | Jul 2012 | CN |
202394296 | Aug 2012 | CN |
102654927 | Sep 2012 | CN |
102812500 | Dec 2012 | CN |
102930642 | Feb 2013 | CN |
102976004 | Mar 2013 | CN |
103198562 | Jul 2013 | CN |
103226870 | Jul 2013 | CN |
203242065 | Oct 2013 | CN |
103440607 | Dec 2013 | CN |
103544772 | Jan 2014 | CN |
203408902 | Jan 2014 | CN |
103662541 | Mar 2014 | CN |
103679147 | Mar 2014 | CN |
203520502 | Apr 2014 | CN |
203588366 | May 2014 | CN |
103954626 | Jul 2014 | CN |
105513201 | Apr 2016 | CN |
1168253 | Jan 2002 | EP |
1703436 | Sep 2006 | EP |
2167553 | May 1986 | GB |
07112801 | May 1995 | JP |
7334583 | Dec 1995 | JP |
2000121564 | Apr 2000 | JP |
3123095 | Jan 2001 | JP |
2002019147 | Jan 2002 | JP |
2002183286 | Jun 2002 | JP |
2002259528 | Sep 2002 | JP |
2002302252 | Oct 2002 | JP |
2002324264 | Nov 2002 | JP |
2002358354 | Dec 2002 | JP |
2003139516 | May 2003 | JP |
2003242243 | Aug 2003 | JP |
2003264007 | Sep 2003 | JP |
2003267509 | Sep 2003 | JP |
2004021569 | Jan 2004 | JP |
2004288143 | Oct 2004 | JP |
2004303102 | Oct 2004 | JP |
2004341681 | Dec 2004 | JP |
2006127308 | May 2006 | JP |
2006195814 | Jul 2006 | JP |
2006227764 | Aug 2006 | JP |
2006260246 | Sep 2006 | JP |
2007141266 | Jun 2007 | JP |
2007179516 | Jul 2007 | JP |
2007265340 | Oct 2007 | JP |
2008522299 | Jun 2008 | JP |
2008293391 | Dec 2008 | JP |
2007086725 | Apr 2009 | JP |
2009245058 | Oct 2009 | JP |
2009250971 | Oct 2009 | JP |
2010177720 | Aug 2010 | JP |
2012058932 | Mar 2012 | JP |
2013033361 | Feb 2013 | JP |
2013037441 | Feb 2013 | JP |
2013531823 | Aug 2013 | JP |
20000064168 | Nov 2000 | KR |
20130085255 | Jul 2013 | KR |
20140037543 | Mar 2014 | KR |
115096 | Mar 2001 | WO |
205176 | Jan 2002 | WO |
WO-225613 | Mar 2002 | WO |
239357 | May 2002 | WO |
3012717 | Feb 2003 | WO |
3014994 | Feb 2003 | WO |
2004021114 | Mar 2004 | WO |
2004114490 | Dec 2004 | WO |
2005008566 | Jan 2005 | WO |
2005101346 | Oct 2005 | WO |
2006058601 | Jun 2006 | WO |
2006080851 | Aug 2006 | WO |
WO-2007066166 | Jun 2007 | WO |
9128176 | Oct 2009 | WO |
2009128173 | Oct 2009 | WO |
2009129526 | Oct 2009 | WO |
2010040116 | Apr 2010 | WO |
2010128267 | Nov 2010 | WO |
2010128315 | Nov 2010 | WO |
2011131016 | Oct 2011 | WO |
2012138679 | Oct 2012 | WO |
2013074819 | May 2013 | WO |
WO-2013063042 | May 2013 | WO |
WO-2014075055 | May 2014 | WO |
2015022409 | Feb 2015 | WO |
Entry |
---|
International Search Report and Written Opinion dated Oct. 3, 2017, in International Application No. PCT/US2017/027476, 16 pages. |
Kanter, James Max, “Color Crack:Identifying Cracks in Glass,” dated Dec. 9, 2014; retrieved from the internet http://www.jmaxkanter.com/static/papers/color_crack.pdf on Sep. 22, 2017. |
Oliveira, et al., “Automatic crack detection on road imagery using anisotropic diffusion and region linkage,” 18th European Signal Processing Conference (EUSIPCO-2010), Aug. 23, 2010, pp. 274-278. |
Zhang, Yiyang, “The design of glass crack detection system based on image preprocessing technology,” 2014 IEEE 7th Joint International Information Technology and Artificial Intelligence Conference, IEEE, Dec. 20, 2014; pp. 39-42. |
2006 Florida Statutes Title XXXIII, Chapter 538, Sections 538.03 and 538.04, 7 pages. |
Aftermarket Cellular Accessories, “Cellular Phone Model Identification,” retrieved from http://web/archive.org/web/20060328064957/http://aftermarketcellular.com/ic/identification.html on Mar. 16, 2014, published Mar. 28, 2006, 3 pages. |
Altec Lansing User's Guide 2007, 8 pages. |
Bussiness Wire, “The World's First Office Photography Machine” at CES 2008 Launched by Ortery Technologies, Jan. 7, 2008, 3 pages. |
CNET, “Tackling LCD “burn ins”, and dead/stick Pixels”, published Sep. 2, 2009, retrieved from http://www.cnet.com/news/tackling-lcd-burn-ins-and-deadstuck-pixels/. |
Evgenii Masunov, Mar. 25, 2010, http://www.appleinsider.ru/news/ipone-obladaet-luchshim-tachskrinom-provereno_robotom.html, 4 pages. |
Geekanoids, You Tube Video, “Apple iPhone 3GS Unboxing and Review”, uploaded on Jun. 19, 2009, retrieved from http://www.youtube.com/watch?v=GCEi9QAeDqk on Sep. 2, 2009. |
GSM Arena Glossary, “LCD (Liquid Crystal Display”, retrieved from http://www.gsmarena.com/glossary.php3?term=lcd on Apr. 28, 2016, 1 page. |
Lambert, Emily, “Use It Up, Wear It Out”, Forbes 175.5 (2005): 77-78. Business Source Complete. Web. Jan. 6, 2015, 3 pages. |
Littleton Partners with Donations Ink (Jan. 19, 2006) US Fed News Service, Including US State News. Web. Jan. 6, 2015, 1 page. |
MobileGazette.com, “2006 in Review: The Good, The Bad and the Ugly”, published Dec. 2006, retrieved from http://www.mobilegazette.com/2006-review-06x12x22.htm on Nov. 11, 2015. |
PC World, “Wipe Your Cell Phone's Memory Before Giving it Away”, published Jan. 2006, retrieved from http://www.washingtonpost.com/wp-dyn/content/article/2006/01/30/AR2006013001144.html on Nov. 10, 2015. |
Perng et al., “A Novel Vision System for CRT Panel Auto-Inspection”, Proceedings of the 2005 IEEE International Conference on Mechatronics, Jul. 10-12, 2005, pp. 4. |
Perng et al., “A Novel Vision System for CRT PaNnel Auto-Inspection”, Journal of the Chinese Institute of Industrial Engineers, vol. 24, No. 5, pp. 341-350 (2007). |
Rawson, Chris, “TUAW: 25 Ways to Check the Hardware on Your iPhone 4”, published Aug. 12, 2010, retrieved at http://www.tuaw.com/2010/08/13/hardware-test-your-iphone-4/ on Feb. 28, 2014. |
Rehg et al. “Vision for a Smart Kiosk” IEEE, Computer Society Conference on Computer Vision and Pattern Recognition (1997). |
Rolf Steinhilper “Remanufacturing: The Ultimate Form of Recycling”, Fraunhofer IRBVerlag, 1998, parts 1-3, http://www.reman.org/Publications_main.htm. |
SimplySellular, “Get Cash for your Old Cell Phone”, published Apr. 2, 2010, retrieved from http://simplysellular.com/conditions.php on Jan. 6, 2015, 2 pages. |
Wilson, Doug, “Liquid Crystal Display (LCD) Inspection System”, National Instruments Case Study, available May 10, 2009, retrieved from http://sine.ni.com/cs/app/cod/p/id/cs-345 on Jan. 5, 2015, 2 pages. |
Yahoo Answers, “What is a Clean ESN?” published Jun. 23, 2009, retrieved from http://web.archive.org/web/20090623215042/http://answers.yahoo.com/question/inde,8020US?qid=20080318061012AANFRco on Apr. 3, 2014. |
Co-Pending U.S. Appl. No. 15/130,851 of Forutanpour, B. et al., filed Apr. 15, 2016. |
Co-Pending U.S. Appl. No. 15/630,508 of Silva, J. et al., filed Jun. 22, 2017. |
Co-Pending U.S. Appl. No. 15/630,539 of Bowles, M. et al., filed Jun. 22, 2017. |
Investopedia: What's the difference between weighted average accounting and FIFO/LILO accounting methods? Aug. 19, 2010. Accessed via archive.org [https://web.archive.org/web/20100819200402/http://www.investopedia.com/ask/answers/09/weighted-average-fifo-lilo-accounting.asp]. |
Dennis Bournique: “Mobile Karma Shuts Down As iCloud and Blacklists Challenge Used Phone Buyers”, Prepaid Phone News, Jul. 23, 2014 (Jul. 23, 2014), XP055229747, Retrieved from the Internet <URL:http://www.prepaidphonenews.com/2014/07/mobile-karma-shuts-down-as-icloud-and.html>; accessed Nov. 27, 2017; 2 pages. |
Tecace Software: “Your phone appraisal—Movaluate—Android Apps on Google Play”, Android Apps on Google Play, Aug. 12, 2013 (Aug. 12, 2013), XP055230264, Retrieved from the Internet <URL:https://play.google.com/store/apps/details?id=com.tecace.android.app.movaluate&hl=en>; accessed Nov. 27, 2017; 2 pages. |
Co-Pending U.S. Appl. No. 15/855,320 of Forutanpour et al., filed Dec. 27, 2017. |
Number | Date | Country | |
---|---|---|---|
20170301078 A1 | Oct 2017 | US |