Handheld apparatus and method for the automated application of cosmetics and other substances

Information

  • Patent Grant
  • 8942775
  • Patent Number
    8,942,775
  • Date Filed
    Wednesday, February 2, 2011
    14 years ago
  • Date Issued
    Tuesday, January 27, 2015
    10 years ago
Abstract
An applicator head is provided for a reflectance modifying agent (RMA) applicator is moved across the skin by means of a floating ring having dispersed raised contact points to maintain a proper distance from the surface to be treated, reduce the influence of outside light during scanning, and limit smudging during deposition. During an application session, software on the computer uses a camera to sense aspects of color and texture on human features, calculates cosmetic enhancements, and uses the printer head to apply RMA precisely to the features to create those enhancements. Skin landmarks are used for registration. The head uses differential lighting by providing a sequence of directional lighting, with some exposures left dark to adjust for ambient light leakage. The exposures are co-synchronized in stacks, where each stack is a grouping of data about a particular instant of time during the scanning.
Description
FIELD OF THE INVENTION

The current invention relates to automated computer-controlled methods to identify skin texture and to selectively and precisely apply one or more reflectance modifying agent, such as a dye or pigment, to human skin to improve its visual attractiveness.


BACKGROUND OF THE INVENTION
Prior Cosmetic Techniques and their Disadvantages

Prior art techniques for modifying the appearance of skin include natural tanning, artificial tanning, and the deliberate application of cosmetics. Each of these prior art techniques has limitations.


Typically, the applications of cosmetic substances to skin are largely manual, for example through the use of brushes, application tubes, pencils, pads, and fingers. The application methods make prior art cosmetics imprecise, labor intensive, expensive, and sometimes harmful, when compared to the computerized techniques of the present invention.


Most prior art cosmetic approaches are based on the application of opaque substances. As explained in the cross-referenced application U.S. Ser. No. 11/503,806, there is a need for the precise computer-controlled application of reflectance modifying agents (RMAs), such as transparent dyes, to provide a more effective modification of appearance.


In this specification, the terms “reflectance modifying agent” or “RMA” refer to any compound useful for altering the reflectance of another material, and are explained in further detail below. Some examples of RMA are inks, dyes, pigments, bleaching agents, chemically altering agents, and other substances that can alter the reflectance of human skin and other features. The terms “dye” and “transparent dyes” are used for brevity in this specification to represent any RMA.


Manual cosmetic applications are imprecise compared to computer-controlled techniques, and this imprecision may make them less effective. For example, the heavy application of a foundation base for makeup may cause an unattractive, caked-on appearance. Manual techniques also typically take a long time to employ, as can be seen in any morning commute on a highway, where people frantically take advantage of stops to finish applying their makeup. In addition, manually applied makeup is not cheap, and when the help of professionals such as beauticians is required, is even more expensive. Moreover, often the materials applied to the skin in manual techniques are themselves potentially harmful. For example, a foundation base for makeup may cause skin to dry out and may inhibit the skin's breathing. Sunlight or artificial light used for tanning may cause cancer.


Therefore, there is a need for the precise application of reflectance modifying agents (RMAs) to provide a more effective, more automated, faster, less expensive, and less dangerous modification of the appearance of skin. The cross-referenced patent application cited above presents a system and method for this need.


One problem that an automated system and method of applying RMAs must solve is the design of an applicator with an efficient head. In an embodiment, a useful applicator would be small enough to be held in the hand, would be easy to clean, and would be inexpensive to produce. In addition, it would maintain the scanner and RMA application system at an appropriate distance from the surface to be treated, to ensure accurate scanning and deposition. If the scanner is located too far from or too close to the surface, for example, the results of scanning may not be not be accurate enough to provide a basis for pixel-level cosmetic enhancements. In the same way, a printer head that is not maintained at a proper distance from the surface, for example, will not be able to apply the RMAs with pixel-level precision.


An additional challenge in designing an automated RMA system is preventing outside light from entering around the base of the applicator and scanner and distorting the accuracy of the scanning.


Moreover, the design of the applicator must limit smudging of the RMAs on the surface treated, which may result from contact with hardware elements of the scanner or inkjet printer head. If the rim of an inkjet printer head used for applying RMAs drags across the skin during deposition, for example, it may smudge the effect of the RMAs on the skin. This is especially a problem when applications involve making multiple passes over the surface, because the freshly deposited RMAs may be easily smudged by too much contact with hard surfaces.


Therefore, there is also a need for an RMA applicator head designed so that the applicator is small enough to be handheld, easy to clean, and inexpensive, and that maintains a proper distance between the scanner and RMA printer head and the surface to be treated, while reducing the influence of outside light during scanning and limiting smudging during deposition.


An important element of a cosmetic enhancement system is the ability to separate a scanned image of an area of skin or other human feature into two components, color and surface texture. Color refers to an area's light value, such as lightness and darkness, as well as hue, such as pinkness or yellowness. Surface texture refers to the area's topography, such as the contours of pores, wrinkles, and bumps, both large and small. For example, the system's software uses strategies to accentuate, alter, or camouflage color effects and different strategies for surface texture effects, to make a woman look both young and real.


BRIEF SUMMARY OF THE INVENTION

These and other needs are addressed by the present invention. The following explanation describes the present invention by way of example and not by way of limitation.


It is aspect of the present invention to provide an RMA applicator head that is small enough for a handheld applicator.


It is another aspect of the present invention to provide an RMA applicator head that is easy to clean.


It is still another aspect of the present invention to provide an RMA applicator head that is inexpensive.


It is another aspect of the present invention to provide an RMA applicator head that maintains a proper distance between its scanner and printer head and the surface to treated, while reducing the influence of outside light during scanning and limiting smudging during deposition.


In accordance with the present invention, a computer-controlled system determines attributes of a frexel, an area of human skin, and applies a reflectance modifying agent (RMA) at the pixel level, to make the skin appear more attractive. The system's scanner and RMA applicator are moved across the skin by means of elements with dispersed raised contact points, for example pounce wheels, which are wheels with points around their outer rims. These contact points maintain a proper distance from the surface to be treated, reduce the influence of outside light during scanning, and limit smudging during deposition. Different motion means with dispersed raise contact points may also used, such as a ball, a comb-like walker, or other geometrical shapes. For example, a square configuration of motion means may be used or a circular one.


In one embodiment, the applicator head further comprises a thin inkjet printer head, a telecentric field lens, a camera, and an RMA reservoir and is attached via a power and data cable to a computer. During an application session, software on the computer uses a camera to sense aspects of color and texture on human features, calculates cosmetic enhancements, and uses the printer head to apply RMA precisely to the features to create those enhancements. Skin landmarks are used for registration.


It is aspect of the present invention to provide an effective method to determine surface texture from scanned data about an area of skin or other human feature.


This and other aspects, features, and advantages are achieved according to the system and method of the present invention. In accordance with the present invention, a software method of differential lighting automatically determines aspects of surface texture from scanned data about an area of skin or other human feature, using a computerized system for scanning that area, calculating enhancements, and applying cosmetics. Scanning with varying configurations of applied lighting captures images in a cycle, with the lighting for some exposures left dark. The exposures are co-synchronized in stacks, where each stack is a grouping of data about a particular instant of time during the scanning. This data may be obtained from directly captured exposures or from interpolated data. Data from an exposure made without applied lighting is subtracted from each stack to remove edge leakage from ambient light. The remaining exposures are used to generate the albedo (color), the north-south tilt, and the east-west tilt of the scanned area. The exposures are then stacked in their visual position in computer memory, and the data is updated using a noise filtering average.





BRIEF DESCRIPTION OF THE DRAWINGS

The following embodiment of the present invention is described by way of example only, with reference to the accompanying drawings, in which:



FIG. 1 is a block diagram that illustrates the relative size an RMA applicator head;



FIG. 2 is a block diagram that illustrates elements of an RMA applicator head;



FIG. 3 is a flow chart illustrating the general steps for a process of applying RMA with the applicator head;



FIG. 4 is a representative diagram illustrating a path of movement of the applicator head over an area of skin whereby multiple overlapping images may be captured;



FIG. 5 is a representative diagram illustrating aspects associated with the present invention that require registration;



FIG. 6A is a representational diagram illustrating a side view of a pounce wheel;



FIG. 6B is a representational diagram illustrating a view of a pounce wheel at an angle;



FIG. 7 is a representational diagram illustrating a ball with dispersed contact points;



FIG. 8 is a representational diagram illustrating a comb-like walker with dispersed contact points;



FIG. 9 is a representational diagram illustrating a square configuration of motion means;



FIG. 10 is a representational diagram illustrating a circular configuration of motion means;



FIG. 11 is a block diagram showing an operating environment in which embodiments of the present invention may be employed for applying RMAs onto skin, using motion means with dispersed contact points;



FIG. 12 is a block diagram illustrating an operating environment in which embodiments of the present invention may be employed for applying RMAs onto skin through communications over a network, using motion means with dispersed contact points; and



FIG. 13 is a block diagram illustrating an operating environment in which embodiments of the present invention may be employed for applying RMAs onto skin through communications over a network and a portable application device, using motion means with dispersed contact points.



FIG. 14 is a flow chart illustrating the general steps for a process to determine surface texture from scanned data about an area of skin or other human feature;



FIG. 15 is a chart illustrating a configuration of exposures for determining textures.





DETAILED DESCRIPTION
Applicator Head with Raised Contact Points

In this embodiment, the present invention comprises an applicator head for an applicator used with a computer-controlled system and method that scans an area of human skin, identifies unattractive attributes, and applies the RMA, typically with an inkjet printer, to improve the appearance of that area of skin. U.S. application Ser. No. 11/503,806 filed Aug. 14, 2006 by the present applicants describes a computer-controlled system and method.


An example applicator head 2, shown in FIG. 1, covers an area of skin about equal to a single electric razor head. Such a size is proven daily to fit in intimate contact across a human face.


In an embodiment for speed of application, multiple applicator heads 2 may be assembled in a floating mount, just as multiple floating heads are combined in a single electric razor.


Applicator Head


In one embodiment, the applicator head 2 comprises the following elements, as illustrated in FIG. 2.


Plastic Case


The molded case 4A and 4B has rubber “O” type rings for waterproofing, so that the applicator head 2 can be run under the faucet for cleaning, like a razor. The inkjet printer head 8 can be maintained this way, which is not an option in normal printers. In an embodiment, the applicator head 2 may “park” for storage on a stand that would cap the applicator head 2.


Floating Ring


The applicator head 2 is moved across the skin by means of a floating ring 6 with elements with dispersed raised contact points. These contact points maintain a proper distance from the surface to be treated, reduce the influence of outside light during scanning, and limit smudging during deposition. One example of dispersed raised contact points are pounce wheels which are discussed below.


Inkjet Head


A very thin inkjet head 8, illustrated in FIG. 2, fits perpendicularly to the skin into case groove 10.


Field Lens


A field lens 12 assembly with LED assembly 13 provides telecentric viewing so that size is independent of distance and the view fits around the inkjet head. It fits into case groove 14 and helps protect the electronics behind the lens from water and dirt.


Camera


A camera module 16 with electronics fits into case groove 18.


In an embodiment, the camera module 16 may be a module made for mobile devices such as cell phones. The newer of these modules have 3 megapixels and above. In covering an area half an inch across, just a 1 megapixel camera would have 4 times the resolution of the human eye at 10 inches.


Cosmetic Reservoir


A replaceable cosmetics reservoir 20 and ink is shown only as a block, but it should have a visually appealing and protectable design because it is what consumers would actually buy repeatedly, like razor blades. In an embodiment, the cosmetics reservoir 20 may contain multiple separate RMA colors that may be mixed to achieve desired effects. In another embodiment, it may contain a single RMA color premixed to achieve a desired aim color or effect.


Cable and Computer


In one embodiment, the applicator head 2 is attached to a computer with a cable. In this example, a data and power cable 22 is required. In an embodiment, a USB 2.0 cable may be used. In this example, a consumer computer 24 is required. Almost any newer computer configured correctly with enough disk memory, good display, and a USB port may be used.


Software


Software 26 is required that runs on the computer 24 and provides the functionality for scanning an area of a human feature, such as skin, calculating cosmetic enhancements, tracking registration, and applying the RMA, explained in detail in the cross-referenced application and outlined below.


Method of Operation


The applicator head 2 enables RMA applications that are like conventional cosmetic applications in the sense that the user actually chooses an aim color and “brushes” it on the desired area. This allows a user to select an “aim” skin color, and then deposit to that density. By optical feedback on each frexel (area of human skin), RMA, such as ink or dye, is deposited on each pass until that density is reached. Then no more dye is deposited on subsequent passes. The user may choose to change the aim color manually while applying to different parts of the skin, just as current art requires different colors of blush to be manually selected and applied to different parts of the face to achieve a shaded effect.


In this patent application, the phrase “area of skin” is used to represent any human feature to be enhanced.


The general steps of this process are illustrated in FIG. 3.


Step 1000 in FIG. 3—Choosing an Aim Color.


In an embodiment, a user employs an interface on the computer 24, shown in FIG. 2, to select an “aim” skin color to be achieved through an appropriate mix of separate RMA colors contained in the cosmetic reservoir 20. In another embodiment, controls on the applicator may be used to select the aim color.


In yet another embodiment, an applicator may contain premixed RMA for a single aim color, and different applicators may contain different single aim colors. For example, one user might buy an applicator with a light RMA aim color, and another user might buy a different applicator with a darker aim color.


Step 1002 in FIG. 3—Moving the Applicator Containing the Applicator Head 2 Over the Area to be Enhanced.


The user moves the applicator containing the applicator head 2, shown in FIG. 2, over an area of skin 302, shown in FIG. 4, to be enhanced. As the applicator head 2, shown in FIG. 2, is placed on the skin, data from the skin under the applicator is immediately seen as a “current” view.


Step 1004 in FIG. 3—Capturing Images of the Area to be Enhanced.


As the applicator is moved in a pattern of movement 30, for example the pattern 30 shown in FIG. 4, overlapping images 32 are captured by the camera module 16, shown in FIG. 2, at least 10 per second. Most of the image at each capture is redundant with the previous capture.


Step 1006 in FIG. 3—Using Landmarks on the Area to be Enhanced to Provide a Current View of the Area.


Using landmarks, or “skinmarks” on the area of skin 302, shown in FIG. 4, software 26, shown in FIG. 2, tracks relative movement of the applicator and combines images in computer memory to give an expanded current view of the skin, using only views from the current application session. For example, pores, moles, scars, lines, wrinkles, age spots, sun damage, freckles, color variations, contours of features, textural variations such as bumps, and many other aspects of human features may be used as landmarks.


An application session is defined to start when the applicator touches the skin and to end when it is retracted. When the applicator is retracted from the skin, in this mode, all knowledge is erased, and the next application session starts fresh by again placing the applicator some place on the skin and manually sweeping it around an area.


As the applicator is swept back and forth in a path of movement 30, shown in FIG. 4, in a single application session, this current view can cover a large area of skin 302 and develop momentary strategies for partial and overlapping deposition on each sweep. Whole-body makeup may be accomplished through multiple sweeps without retracting the applicator from the skin.


Positional Data


The software 26, shown in FIG. 2, must be able to determine that a skin defect is underneath the inkjet head 8 at the moment the inkjet head 8 needs to be fired, even though at that precise moment the inkjet head 8 is covering the skin defect from view. As shown in FIG. 5, this requires knowledge of applicator head 2 position relative to real skin 36, and a mapping from real skin 36 to abstract layers 38 in computer memory that model that skin, describe aesthetic choices, guide execution strategies, and track long term changes. The positional information provided by the skinmarks described above enables the software 26, shown in FIG. 2, to keep the applicator head 2, the area of skin 302, shown in FIG. 4, and computer models in register.


Step 1008 in FIG. 3—Calculating Enhancements for the Area Represented by the Current View.


The software 26, shown in FIG. 2, calculates cosmetic enhancements to the area of skin 302, shown in FIG. 4, using the methods described in the cross-referenced patent application.


Step 1010 in FIG. 3—Applying RMA to the Area Represented by the Current View.


RMA, such as ink or dye, contained in the cosmetic reservoir 20, shown in FIG. 2, is deposited by the inkjet head 8 onto the area of skin shown in the current view, for example the area of skin 302 shown in FIG. 4, to achieve desired cosmetic enhancement. The RMA is deposited on each pass of the applicator over the area of skin 302 until the chosen aim color is reached. Then no more dye is deposited on subsequent passes.


Application of Other Substances than RMAs


The applicator of the present invention may be used to apply other substances than RMAs, for example medically beneficial compounds or live skin.


Pounce Wheels and Other Motion Means with Dispersed Contact Points


Example elements with dispersed raised contact points are shown in FIG. 6A, FIG. 6B, FIG. 7, and FIG. 8. FIG. 6A and FIG. 6B illustrate a pounce wheel 7, which is a wheel with points around its outer rim. Typically pounce wheels 7 have been used to transfer a design onto a surface. A pounce wheel 7 on a handle is rolled over an object and leaves holes in the object's surface, and chalk or another marker is then rubbed over the surface to reveal the holes. For example, pounce wheels 7 are used for tracing regular designs onto wood for wood-burning or carving. FIG. 6B illustrates a view of a pounce wheel 7 at an angle. FIG. 7 illustrates a ball 704 with dispersed points, and FIG. 8 a comb-like walker 706.


In an embodiment, the floating ring 6 shown in FIG. 2 has multiple micro pounce wheels 7, shown in FIG. 6A and FIG. 6B, around the rim. The height of the points maintains a proper distance from the surface for both scanning and inkjet deposition. The pounce wheels 7, shown in FIG. 2, also reduce the amount of outside light entering around the base of the applicator to prevent distorting the accuracy of the scanning. In addition, the points on the pounce wheels 7 limit contact of the applicator head 2 with the cosmetics being deposited, to prevent smudging. Thus, they will typically leave behind minimal deposits of the RMA as they are moved over surfaces.


The pounce wheels 7 should be made of durable non-absorptive and hydrophobic material, for example silicon rubber or Teflon, so that they last and do not absorb the RMA. Their heights should also be low, for example 3/16 of an inch (4.8 mm). The use of low heights keeps the system close to the surface so that too much light does not come in underneath the system. The pounce wheels 7 may further be colored black to help absorb light. Their widths should be narrow to further reduce the area that comes into contact with the RMA. Their points should not be very sharp, so that they will not easily puncture surfaces such as skin.


In an embodiment, the pounce wheels 7 may be mounted on thin wires serving as axles.


In an embodiment, twelve pounce wheels may be mounted on each side of the floating ring 6.


In an embodiment, a non-contact, electrostatic wipe (not shown) may be used to blow off the RMA from the pounce wheels 7.


Still other geometrical shapes with dispersed contact points may also be used.


Advantages


Motion means with dispersed contact points are useful for the present invention in several ways. First, the height of the contact points can be used to maintain proper distance from the surface. During scanning, they may be used to maintain a proper distance from the surface to be scanned to ensure effective scanning During deposition, they may also be used to maintain a proper distance from the surface to be enhanced by application with the RMA and so ensure proper application.


Second, the configuration of the dispersed contact points, explained below, can be used to reduce the amount of outside light entering around the base of the applicator and scanner, to prevent distorting the accuracy of the scanning. In this aspect, the contact points serve a baffle to block outside light.


Third, the dispersion and sharpness of the contact points can be used to limit contact with the RMA on the surface being enhanced, reducing smudging. Other motion means, such as a shroud that simply drags across the surface or wheel with flat rims without raised points, would typically cause greater smudging. Motion means with dispersed contact points are especially useful during multiple-pass enhancement, when the applicator must be moved more than once over a freshly applied RMA that would smudge easily.


Configurations of Motion Means


The motion means described above may be mounted in different configurations on an appropriate housing side of elements of the present invention. As shown in FIG. 9, motion means 700 with dispersed contact points may be mounted in a square pattern on a housing side 710. Alternately, these motion means 700 may be mounted in a circular pattern, as shown in FIG. 10, as well as in other useful patterns, for example a rectangle.


The motion means 700 should be made of durable non-absorptive and hydrophobic material, for example silicon rubber or Teflon, so that they last do not absorb the RMA. Thus, they will typically leave behind minimal deposits of the RMA as they are moved over surfaces.


Their heights should also be low, for example 3/16 of an inch. The use of low heights keeps elements of the system close to the surface so that too much light doesn't come in underneath the system. Their widths should be narrow to further reduce the area that comes into contact with the RMA


Their points should not be very sharp, so that they will not easily puncture surfaces such as skin.


Multiple contact points may be used to achieve the advantages explained above, including baffling light. In an embodiment, twelve pounce wheels or balls may be mounted on a side, as shown in FIG. 9 and FIG. 10. In another embodiment, a hundred comb-like walkers may be mounted in a circle on a side.


The motion means 700 is preferably colored black to help absorb light.


Examples of Embodiments


FIG. 11 shows an operating environment in which embodiments of the present invention may be employed for applying RMAs onto skin, using motion means 700 with dispersed contact points. The motion means 700 may be used on

    • the side of the means of application 240 that comes into contact with the surface to be treated, such as an area of skin 302, and
    • the side of the scanner 220 that comes into contact with the surface to be treated, such as an area of skin 302.



FIG. 12 shows an operating environment in which embodiments of the present invention may be employed for applying RMAs onto skin through communications over a network, using motion means 700 with dispersed contact points. Again, the motion means 700 may be used on

    • the side of a printer 241 that comes into contact with the surface to be treated, such as an area of skin 302, and
    • the side of the scanner 220 that comes into contact with the surface to be treated, such as an area of skin 302.



FIG. 13 shows an operating environment in which embodiments of the present invention may be employed for applying RMAs onto skin through communications over a network and a portable application device, using motion means 700 with dispersed contact points. In this embodiment, the motion means 700 may be used on the side of the application device 246 that comes into contact with the surface to be treated, such as an area of skin 302. The application device 246 further comprises both an inkjet printer 242 and a scanner 220.


DETAILED DESCRIPTION OF EMBODIMENT
Differential Lighting

The present invention comprises a method of differential lighting that can be used to determine surface texture from scanned data about an area of skin or other human feature. In an embodiment, this method may be used with a computer-controlled system and method that scans an area of human skin, identifies unattractive attributes, and applies RMA, typically with an inkjet printer, to improve the appearance of that area of skin. U.S. application Ser. No. 11/503,806 by the present applicants describes computer-controlled systems and methods.


The present invention is an innovation comprising a method of differential lighting that, in an embodiment, may be employed using this computer-controlled system and method and applicator head.


In this patent application, the phrase “area of skin” is used to represent any human feature to be enhanced cosmetically.


Light Sources


In this embodiment, a plurality of light sources, such as LEDs, are provided, such that each light source represents a directional orientation with respect to the housing. For example, one directional orientation is a North-South-East-West orientation where one or more lights represents each of those directions. Another directional orientation is a circular alignment of light sources such as three light sources arranged about 120 degrees apart. These three sources may be one or more LED.


Field Lens


A field lens 12 with LED assembly 13 provides telecentric viewing so that size is independent of distance and the view fits around the inkjet head. It fits into case groove 14 and helps protect the electronics behind the lens from water and dirt. In one example as described below, the LEDs are configured on a North/South and East/West alignment so that one or more LEDs representing each of those directions may be cycled on an off.


Using Differential Lighting to Determine Surface Texture


The general steps of the present invention's method of using differential lighting to determine surface texture from scanned data about an area of skin are illustrated in FIG. 14.


Step 2010 in FIG. 14—Using Varying Lighting Configurations to Capture Images in a Cycle


Images are scanned in a cycle by the applicator head 2, shown in FIG. 2, which comprises lighting means and one or more cameras 16. In an embodiment, the lighting means comprise a field lens 12 and LEDs 13.


In an embodiment, six images are captured in a cycle of lighting modes. Each image represents an exposure 40, shown in FIG. 15, that is captured in 1/60 second in the US to null out the 120 Hz flicker of ambient fluorescent lighting leaking under the edges of the floating ring 6, shown in FIG. 2, during scanning. As shown in FIG. 15, each exposure 40 in one example lighting cycle represents a different configuration or lighting mode of flashed and unflashed LEDs 13, shown in FIG. 2. As shown in FIG. 15, there is thus an approximate visual positional fix P (Position) every 1/60 second for the exposures 40.


The exposures 40 in a cycle are arbitrarily labeled as follows:

    • N (North),
    • S (South),
    • D (Dark),
    • E (East),
    • W (West), and
    • D (Dark).


There is a hard fix on P, shown in FIG. 15, every 1/20 second, which is fast enough for the software 26, shown in FIG. 2, to track normal movement of the applicator head 2. That is, every third exposure 40, shown in FIG. 15, is D (Dark), an exposure captured with the LEDs 13, shown in FIG. 2, or other lighting means, turned off. This is done so that the software 26, shown in FIG. 2, can identify edge leakage in the D (Dark) exposure 40, shown in FIG. 15, and subtract that edge leakage from calculations.


Step 2020 in FIG. 14—Stacking Channels of Exposures


The exposures 40 are co-synchronized in stacks, where each stack is a grouping of data about a P representing a particular instant of time during the scanning. This data may be obtained from directly captured exposures or from interpolated data.


In the embodiment shown in FIG. 15, five channels of exposures 40, N, S, E, W, and D are stacked in a stack 42 every 1/60 second frame. In other embodiments, the stacking may be calculated at other time periods, for example 1/15 sec.; 1/20 sec.; or 1/30 sec.


One channel 44 in each stack 4 may be directly captured by the camera 16, shown in FIG. 2. The other four channels are then interpolated from data temporally adjacent exposures 40, shown in FIG. 15, based on image movement calculated in software 26, shown in FIG. 2. Such movement may be calculated at video rates by, for example, a hardware motion detection engine in an MPEG encoder, such as the one found in NVIDIA GPUs for mobile devices, or other techniques in CPU software. The use of one or more accelerometers associated with the applicator head 2 may provide additional data showing micro-variations in positioning.


If channel 44, shown in FIG. 15, does not coincide with a real frame captured by the camera 16, shown in FIG. 2, during the camera's 16 capture cycle, channel 44, shown in FIG. 15, may be interpolated from data from one or more temporally adjacent direct exposures 40. That is, data from either a frame captured before channel 44 or after it, or from an average of the two, may be used for interpolation of channel 44.


Interpolations may be derived by simple averaging of data from one or more temporally adjacent frames. A weighted average may also be used, where the weight in the average may be higher for a directly captured frame that is closer in temporal position to the frame that is to be derived.


However, simple averaging may be inaccurate for moving images. Thus, for moving images it may be desirable to “push,” or adjust, data about an image in software 26, shown in FIG. 2, according to known position changes and to feature recognition based on a stored model of features, as is done in images for gaming.


Step 2040 in FIG. 14—Subtracting the Dark Exposure from Each Stack to Remove Edge Leakage


In an embodiment, after the five channels are co-synchronized in a stack 42, shown in FIG. 15, the D (Dark) signal is subtracted from the other four. This gives a precise black level, and eliminates the effect of light leaking around the pounce wheels 7, shown in FIG. 2, so long as the total light is not so bright as to saturate the camera 16 or the ratio so low as to amplify sensor noise excessively.


Step 2060 in FIG. 14—Using a Set of Exposures to Generate Albedo and Tilt


In an embodiment, the remaining N, S, E, and W exposures 40, shown in FIG. 15, in a stack 42 are an overspecified set of three exposures 40 each, which are used to generate three numbers, namely the albedo (color), the north-south tilt, and the east-west tilt of each area of skin. The overspecification allows some options to detect and remove specular reflections when the fourth dimension is in disagreement. In one example, the color is actually three numbers for R, G, and B, while the captured image is reduced to monochrome using the lowest noise combination before calculating tilt.


The tilt is a complex equation derived from the differential brightness as seen when illuminated by each LED 13, shown in FIG. 2, as a function of the four LED configurations. Because the angle struck by each LED 13 varies across the field, the equation changes for each pixel. The practical approach to refine the equation is to first reduce the image to a N/S and E/W differential, and then calibrate gains in a matrix empirically by passing the applicator head 2 over a dimpled surface with known dimple depth.


Step 2080 in FIG. 14—Storing the Derived Data in Computer Memory


Again based on the known visual position “P” of each exposure 40, shown in FIG. 15, the derived data is stored in the computer memory of computer 24, shown in FIG. 2.


Step 2090 in FIG. 14—Using a Noise Filtering Average


The stored data is updated using a noise filtering average.


Other Applications


The method explained above may be adapted for extrapolating the position of the applicator for actual application of the RMA, for example for firing the inkjet head 8 shown in FIG. 2.


In practice, the view from the leading part of the applicator head 2 would confirm and refine position of the applicator head 2 prior to deposition. The “P” value, shown in FIG. 15, would be extrapolated to determine the exact moment to fire the inkjet head 8, shown in FIG. 2, because the skin is not visible when directly under the inkjet head 8. The differential of P, shown in FIG. 15, indicating velocity, would be used to determine how rapidly to fire the inkjet head 8, shown in FIG. 2. For example, the inkjet head 8 would deposit more ink to attain an aim density if the applicator head 2 were moving faster, or deposition would stop if the applicator head 2 were held motionless at the end of a stroke. The trailing part of the applicator head 2 would then be used to update the data about color and texture with the effects of the immediately preceding deposition.


In an embodiment, the applicator head 2 is bidirectional, and can be passed back and forth over the skin with continuous deposition. The sensing of the leading and trailing part of the applicator head 2 would alternate according to software's 26 determination of position and direction of movement.


Reducing Reflections and Glare


In an embodiment, the molded plastic case 4, shown in FIG. 2, may be colored black to reduce retro reflections and reverse reflections off the opposite wall relative to a particular LED configuration.


In an embodiment, several LEDs 13, like those shown in FIG. 2, may be stacked in a group, or other diffusion techniques may be used to reduce specular glare effects. In addition, polarization may be highly beneficial to reduce specular glare effects, although this may complicate assembly and require several fold more power and light from the LEDs 13.


Other Hardware and Software


It will also be apparent to those skilled in the art that different embodiments of the present invention may employ a wide range of possible hardware and of software techniques. For example the communication between a Web service provider and client business computers could take place through any number of links, including wired, wireless, infrared, or radio ones, and through other communication networks beside those cited, including any not yet in existence.


Also, the term computer is used here in its broadest sense to include personal computers, laptops, telephones with computer capabilities, personal data assistants (PDAs) and servers, and it should be recognized that it could include multiple servers, with storage and software functions divided among the servers. A wide array of operating systems, compatible e-mail services, Web browsers and other communications systems can be used to transmit messages among client applications and Web services.

Claims
  • 1. A handheld device, comprising: an applicator for applying one or more reflectance modifying agent (RMAs), the applicator being configured to reduce outside light from entering around a base of the applicator that comprises edges in contact with an area of skin;a plurality of light sources, each light source being positioned within the applicator and above the base of the applicator, in a respective direction to selectively illuminate the area of skin;a camera for generating images of the area of skin; anda computing environment coupled to the applicator and the camera, the computing environment being configured to: select an aim color,selectively activate light sources of the plurality of light sources to provide a lighting cycle comprising a plurality of lighting modes, the plurality of lighting modes comprising a first lighting mode comprising deactivation of all of the plurality of light sources and a second-lighting mode comprising activation of at least one of the plurality of light sources,selectively activate the camera to generate a plurality of images of the area of skin, each image corresponding to reflective properties of the area of skin associated with a respective lighting mode of the plurality of lighting modes while the applicator reduces outside light from illuminating the area of skin,process the plurality of images to determine an amount of ambient light leaking under edges of the applicator during the first lighting mode and subtracting the amount of ambient light from a subset of the plurality of images corresponding to the second lighting mode,calculate cosmetic enhancements to the area of skin based on the processing,determine a position of the handheld device relative to the area of skin based on the processing, andtransmit a signal to the applicator for applying the one or more RMAs to the area of skin based on the cosmetic enhancements to achieve the aim color.
  • 2. The handheld device of claim 1, wherein the computing environment is configured such that the plurality of lighting modes comprises: a first lighting mode, in which one or more light sources of the plurality of light sources corresponding to a first directional orientation are turned on,a second lighting mode, in which one or more light sources of the plurality of light sources corresponding to a second directional orientation are turned on, anda third lighting mode, in which one or more light sources of the plurality of light sources corresponding to a third directional orientation are turned on.
  • 3. The handheld device of claim 1, wherein the first lighting mode is used to identify at least one of a camera saturation and a level of a sensor noise.
  • 4. The handheld device of claim 1, further comprising an applicator head housing the applicator.
  • 5. The handheld device of claim 4, wherein the computing environment is configured to process the plurality of images, calculate the cosmetic enhancements, determine the position, and transmit the signal to the applicator during an application session, the application session beginning when the applicator head touches the area of skin and ending when the applicator head is retracted from the area of skin.
  • 6. The handheld device of claim 1, further comprising a spacer that maintains a distance between the applicator and the area of skin.
  • 7. The handheld device of claim 6, wherein the spacer comprises a floating ring.
  • 8. The handheld device of claim 1, wherein the one or more light sources each comprise one or more light-emitting diodes (LEDs).
  • 9. The handheld device of claim 1, further comprising a case housing the applicator, the camera, the one or more light sources, and the computing environment.
  • 10. The handheld device of claim 2, wherein each of the first directional orientation, the second directional orientation and the third directional orientation is a directional orientation relative to the case.
  • 11. The handheld device of claim 9, wherein processing the plurality of images comprises: analyzing the plurality of images to compensate for image movement; andadjusting images to account for applicator movement between images.
  • 12. The handheld device of claim 9, wherein processing the plurality of images comprises analyzing the plurality of images to compensate for ambient lighting.
  • 13. The handheld device of claim 1, wherein calculating cosmetic enhancements comprises determining albedo and tilt, the cosmetic enhancements being determined based on the albedo and tilt.
  • 14. A method of providing cosmetic enhancements to an area of skin, the method comprising: selectively activating a plurality of light sources provided in a handheld device, the handheld device being configured to reduce outside light entering around a base of the handheld device that comprises edges in contact with the area of skin, each light source being positioned within the handheld device and above the base of the handheld device, in a respective direction to selectively illuminate an area of skin and being selectively activated to provide a lighting cycle comprising a plurality of lighting modes, the plurality of lighting modes comprising a first lighting mode comprising deactivation of all of the plurality of light sources and a second lighting mode-comprising activation of at least one of the plurality of light sources, while the outside light to illuminate the area of skin is reduced;generating a plurality of images of the area of skin using a camera provided in the handheld device, each image corresponding to reflective properties of the area of skin associated with a respective lighting mode of the plurality of lighting modes;selecting, using a computing environment, an aim color;processing, using the computing environment, the plurality of images to determine an amount of ambient lighting leaking under edges of the applicator during the first lighting mode and subtracting the amount of ambient lighting from a subset of the plurality of images corresponding to the second lighting mode;calculating, using the computing environment, cosmetic enhancements to the area of skin based on the processing;determining, using the computing environment, a position of the handheld device relative to the area of skin based on the processing; andtransmitting a signal to an applicator for applying one or more reflectance modifying agents (RMAs) to the area of skin based on the cosmetic enhancements to achieve the aim color.
  • 15. The method of claim 14, wherein the computing environment is provided in the handheld device.
  • 16. The method of claim 14, wherein the lighting cycle comprises a plurality of lighting modes.
  • 17. The method of claim 16, wherein the plurality of lighting modes comprises: a first lighting mode, in which one or more light sources of the plurality of light sources corresponding to a first directional orientation are turned on,a second lighting mode, in which one or more light sources of the plurality of light sources corresponding to a second directional orientation are turned on, anda third lighting mode, in which one or more light sources of the plurality of light sources corresponding to a third directional orientation are turned on.
  • 18. The method of claim 16, wherein the plurality of lighting modes comprises a dark mode, in which none of the one or more light sources are turned on.
  • 19. The method of claim 14, wherein the handheld device comprises a case housing the applicator, the camera, the one or more light sources, and the computing environment.
  • 20. The method of claim 17, wherein the directional orientation is a directional orientation relative to the case.
  • 21. The method of claim 14, wherein processing the plurality of images comprises: analyzing the plurality of images to compensate for image movement; andadjusting images to account for applicator movement between images.
  • 22. The method of claim 14, wherein processing the plurality of images comprises analyzing the plurality of images to compensate for ambient lighting.
  • 23. The method of claim 14, wherein calculating cosmetic enhancements comprises determining albedo and tilt, the cosmetic enhancements being determined based on the albedo and tilt.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application is a continuation of U.S. application Ser. No. 12/028,835, filed on Feb. 11, 2008, the disclosure of which is expressly incorporated herein by reference in its entirety, and which is related to U.S. Provisional Patent Application No. 60/889, 291 filed Feb. 11, 2007 by the present inventors for “HANDHELD APPARATUS AND METHOD FOR THE AUTOMATED APPLICATION OF COSMETICS AND OTHER SUBSTANCES;” and is related to U.S. Provisional Patent Application No. 60/889,299 filed Feb. 12, 2007 by the present inventors for “SYSTEM AND METHOD FOR APPLYING A REFLECTANCE MODIFYING AGENT TO IMPROVE THE VISUAL ATTRACTIVENESS OF HUMAN SKINWITH MOTION MEANS WITH DISPERSED CONTACT POINTS” and is related to U.S. Provisional Patent Application No. 60/889, 288 for “DIFFERENTIAL LIGHTING FOR IDENTIFYING SURFACE TEXTURE” filed Feb. 11, 2007 by the present inventors. This patent application incorporates by reference the specification, drawings, and claims of U.S. patent application Ser. No. 11/503,806 filed Aug. 14, 2006 by the present inventors for “SYSTEM AND METHOD FOR APPLYING A REFLECTANCE MODIFYING AGENT TO IMPROVE THE VISUAL ATTRACTIVENESS OF HUMAN SKIN”.

US Referenced Citations (97)
Number Name Date Kind
4190056 Tapper et al. Feb 1980 A
4401122 Clark Aug 1983 A
4628356 Spillman et al. Dec 1986 A
4807991 Carew Feb 1989 A
4882492 Schlager Nov 1989 A
5027817 John Jul 1991 A
5156479 Iizuka Oct 1992 A
5241468 Kenet Aug 1993 A
5268166 Barnett et al. Dec 1993 A
5836872 Kenet et al. Nov 1998 A
5931166 Weber et al. Aug 1999 A
6021344 Lui et al. Feb 2000 A
6067996 Weber et al. May 2000 A
6111653 Bucknell et al. Aug 2000 A
6122042 Wunderman et al. Sep 2000 A
6151031 Atkins et al. Nov 2000 A
6208749 Gutkowicz et al. Mar 2001 B1
6250927 Narlo Jun 2001 B1
6286517 Weber et al. Sep 2001 B1
6292277 Kikinis Sep 2001 B1
6293284 Rigg Sep 2001 B1
6295737 Patton et al. Oct 2001 B2
6312124 Desormeaux Nov 2001 B1
6341831 Weber et al. Jan 2002 B1
6436127 Anderson et al. Aug 2002 B1
6487440 Deckert et al. Nov 2002 B2
6502583 Utsugi Jan 2003 B1
6543893 Desormeaux Apr 2003 B2
6554452 Bourn et al. Apr 2003 B1
6575751 Lehmann et al. Jun 2003 B1
6578276 Patton et al. Jun 2003 B2
6641578 Mukai Nov 2003 B2
6706035 Cense et al. Mar 2004 B2
6719467 Hess et al. Apr 2004 B2
6810130 Aubert et al. Oct 2004 B1
7027619 Pavlidis et al. Apr 2006 B2
7233693 Momma Jun 2007 B2
7382400 Sablak Jun 2008 B2
7602942 Bazakos et al. Oct 2009 B2
7890152 Edgar et al. Feb 2011 B2
8027505 Edgar et al. Sep 2011 B2
8582830 Edgar et al. Nov 2013 B2
8610767 Uzenbajakava et al. Dec 2013 B2
20010040982 Kim et al. Nov 2001 A1
20020054714 Hawkins et al. May 2002 A1
20020064302 Massengill May 2002 A1
20020081003 Sobol Jun 2002 A1
20020105662 Patton et al. Aug 2002 A1
20020107456 Leveque Aug 2002 A1
20020155069 Pruche et al. Oct 2002 A1
20020172419 Lin et al. Nov 2002 A1
20020176926 Pletcher et al. Nov 2002 A1
20030045799 Bazin et al. Mar 2003 A1
20030050561 Bazin et al. Mar 2003 A1
20030053685 Lestideau Mar 2003 A1
20030060810 Syrowicz et al. Mar 2003 A1
20030062058 Utsugi Apr 2003 A1
20030063102 Rubinstenn et al. Apr 2003 A1
20030067545 Giron et al. Apr 2003 A1
20030100837 Lys et al. May 2003 A1
20030108228 Garnier Jun 2003 A1
20030130575 Desai Jul 2003 A1
20030208190 Roberts et al. Nov 2003 A1
20030223622 Simon et al. Dec 2003 A1
20030229514 Brown Dec 2003 A2
20040005086 Wolff et al. Jan 2004 A1
20040078278 Dauga et al. Apr 2004 A1
20040125996 Eddowes et al. Jul 2004 A1
20040170337 Simon et al. Sep 2004 A1
20040174525 Mullani Sep 2004 A1
20040179101 Bodnar et al. Sep 2004 A1
20040201694 Gartstein et al. Oct 2004 A1
20040236229 Freeman et al. Nov 2004 A1
20040254546 Lefebvre Dec 2004 A1
20040267189 Mavor et al. Dec 2004 A1
20050004475 Giron Jan 2005 A1
20050010102 Marchesini et al. Jan 2005 A1
20050053628 Montanari et al. Mar 2005 A1
20050053637 Ma Or et al. Mar 2005 A1
20050154382 Altshuler et al. Jul 2005 A1
20060153470 Simon et al. Jul 2006 A1
20060228037 Simon et al. Oct 2006 A1
20060228038 Simon et al. Oct 2006 A1
20060228039 Simon et al. Oct 2006 A1
20060228040 Simon et al. Oct 2006 A1
20070016078 Hoyt et al. Jan 2007 A1
20070035815 Edgar et al. Feb 2007 A1
20070047761 Wasilunas et al. Mar 2007 A1
20070049832 Edgar et al. Mar 2007 A1
20070203413 Frangioni Aug 2007 A1
20080192999 Edgar et al. Aug 2008 A1
20080193195 Edgar et al. Aug 2008 A1
20080194971 Edgar et al. Aug 2008 A1
20080219528 Edgar et al. Sep 2008 A1
20090025747 Edgar et al. Jan 2009 A1
20110270200 Edgar et al. Nov 2011 A1
20140050377 Edgar et al. Feb 2014 A1
Foreign Referenced Citations (44)
Number Date Country
101287607 Sep 2010 CN
202004003148 Apr 2005 DE
1184663 Mar 2002 EP
1210909 Jun 2002 EP
1304056 Apr 2003 EP
1495781 Jan 2005 EP
1677254 Jul 2006 EP
1763380 Mar 2007 EP
2810761 Dec 2001 FR
59171280 Sep 1984 JP
5281041 Oct 1993 JP
6201468 Jul 1994 JP
11019050 Jan 1999 JP
11019051 Jan 1999 JP
2000139846 May 2000 JP
2000331167 Nov 2000 JP
2001112722 Apr 2001 JP
2002017689 Jan 2002 JP
2002263084 Sep 2002 JP
2003052642 Feb 2003 JP
2003057169 Feb 2003 JP
2003057170 Feb 2003 JP
2003513735 X Apr 2003 JP
2003519019 Jun 2003 JP
2003210248 Jul 2003 JP
2004501707 Jan 2004 JP
2004105748 Apr 2004 JP
2004315426 Nov 2004 JP
2006271654 Oct 2006 JP
2007231883 Sep 2007 JP
2008526241 Jul 2008 JP
2008526284 Jul 2008 JP
2336866 Oct 2008 RU
WO0126735 Apr 2001 WO
WO0149360 Jul 2001 WO
WO0177976 Oct 2001 WO
WO2004028420 Apr 2004 WO
WO2004091590 Oct 2004 WO
WO2004095372 Nov 2004 WO
WO2005123172 Dec 2005 WO
W02006008414 Jan 2006 WO
WO2006008414 Jan 2006 WO
WO2006074881 Jul 2006 WO
WO2007022095 Feb 2007 WO
Non-Patent Literature Citations (51)
Entry
Second Examiner's Report in Application No. 2006279652, mailed Nov. 3, 2011, 2 pages.
Railan et al; Laser Treatment of Acne, Psoriasis, Leukoderma and Scars; Seminars in Cutaneous Medicine and Surgery; Dec. 2008; 285-291.
Chiu et al; Fractionated Photothermolysis: The Fraxel 1550-nm Glass Fiber Laser Treatment; Facial Plastic Surgery Clinics of North America (2007), vol. 15, Issue 2; May 2007, 229-237.
Cula et al; Bidirectional Imaging and Modeling of Skin Texture; IEEE Engineering of Medicine and Biology Society; Nov. 2004; 1-6.
Laubach et al; Effects of Skin Temperature on Lesion Size in Fractional Photothermolysis; Lasers in Surgery and Medicine; Jan. 2007; 14-18.
Bon et al; Quantitative and Kinetic Evolution of Wound Healing through Image Analysis; 2000 IEEE Transactions on Medical Imaging, vol. 19, No. 7; Jul. 2000; 767-772.
Examiner's First Report in Application No. 2008260040, mailed Apr. 13, 2012, 2 pages.
Notice to File a Response in Application No. 10-2008-7006079, dated Aug. 6, 2012, 10 pages.
Notice of Reasons for Rejection for Application No. 2008-526284, dated Apr. 18, 2012, 10 pages.
Notification of the Second Office Action for Application No. 200880009579.0, dated Mar. 10, 2012, 4 pages.
Office Action for Application No. 2009148819, mailed May 30, 2012, 7 pages.
“Lehrstuhl für Optik 2004 Annual Report” Jun. 2005 (2005-2006), Lehrstuhl Für Optik, Institute Für Optik, Information and Photonik, Max-Planck-Forschungsgruppe, Universität Erlangen-Nu{umlaut over (r)}nberg, Erlangen, Germany, XP002460048, 2 pages.
EPO Office Action in App. No. 06 801 295.4, mailed Feb. 3, 2010, 3 pages.
Authorized Officer Nora Lindner, International Preliminary Report on Patentability and Written Opinion of the International Searching Authority for International Application No. PCT/US2006/031441, mailed Feb. 12, 2008, 9 pages.
Russian Official Action (including translation) for Application No. 2008109234, mailed Apr. 2, 2009, 7 pages.
EPO Office Action in Application No. 06 801 295.4, mailed Jun. 10, 2008, 3 pages.
Authorized Officer Moritz Knupling, International Search Report for International Application No. PCT/US2006/031441, mailed Dec. 7, 2007, 2 pages.
Authorized Officer Lars-Oliver Romich, International Search Report and the Written Opinion for International Application No. PCT/US2006/031441, mailed Dec. 7, 2007, 14 pages.
Notification of the First Office Action (including translation) in Application No. 200680037564.6, mailed Jul. 31, 2009, 7 pages.
Examiner's First Report in Application No. 2006279800, mailed Feb. 2, 2011, 2 pages.
Russian Deputy Chief S.V. Artamonov, Decision on Grant Patent for Invention (including translation) in Application 2008109235, dated Feb. 19, 2009.
Authorized Officer Dorothee Mulhausen, International Preliminary Report on Patentability for International Application No. PCT/US2006/031657, mailed Feb. 12, 2008, 7 pages.
Authorized Officer Laure Acquaviva, Invitation to Pay Additional Fees and, where applicable, Protest Fees International Application No. PCT/US2008/053527, mailed Jul. 7, 2008, 8 pages.
Examiner's First Report in Application No. 2006279652, mailed Jan. 28, 2011, 2 pages.
Notification of the First Office Action (including translation) in Application No. 200680037560.8, mailed Jul. 17, 2009, 8 pages.
EPO Office Action in Application No. 06 789 746.2, mailed Apr. 3, 2009, 3 pages.
International Search Report for International Application No. PCT/US2006/031657, mailed Dec. 20, 2006, 2 pages.
Authorized Officer Athina Nickitas-Etienne, International Preliminary Report on Patentability for International Application No. PCT/US2008/053640, mailed Aug. 19, 2009, 5 pages.
Authorized Officer Michael Eberwein, International Search Report and Written Opinion for International Application No. PCT/US2008/053640, mailed Jun. 3, 2008, 9 pages.
European Patent Office Action for Application No. 08 729 481.5, dated Aug. 23, 2010, 5 pages.
Authorized Officer Jens Clevorn, International Search Report for Application No. PCT/US2008/053528, dated Nov. 13, 2008, 4 pages.
Authorized Officer Jens Clevorn, International Preliminary Report on Patentability and Written Opinion of the International Searching Authority for Application No. PCT/US2008/053528, dated Aug. 11, 2009, 9 pages.
Notification of First Office Action for Application No. 200880009579.0, dated Jul. 14, 2010, 10 pages.
Authorized Officer Simin Baharlou, International Preliminary Report on Patentability and Written Opinion of the International Searching Authority for Application No. PCT/US2008/065168, mailed Dec. 1, 2009, 8 pages.
Anonymous, “Circular Polarizer Films,” Internet Article, [Online] 2005, http://www.optigrafix.com/circular.htm [retrieved on Sep. 5, 2008].
Authorized Officer Carlos Nicolas, International Search Report and Written Opinion for Application No. PCT/US2008/065168, mailed Sep. 19, 2008, 13 pages.
Mike Topping et al., “The Development of Handy 1, A Robotic System to Assist the Severely Disabled,” ICORR '99, Sixth International Conference of Rehabilitation Robotics, Stanford, CA, Jul. 1-2, 1999, pp. 244-249.
Robot News, “Handy1-Rehabilitation robot for the severely disabled; helping you to eat and drink and brush and even do make-up!”, posted on Apr. 3, 2006, http://robotnews.wordpress.com/2006/04/03/handy1-rehabiliation-robot-for-the-severely-disabledhelping-you-to-eat-and-drink-and-brush-and-even-do-make-up/, 6 pages.
Mike Topping, “An Overview of the Development of Handy 1, a Rehabilitation Robot to Assist the Severely Disabled” Journal of Intelligent and Robotic Systems, vol. 34, No. 3, 2002, pp. 253-263.
Notice of Reasons for Rejection for Application No. 2008-526241, dated Aug. 31, 2011, 7 pages.
Notification of the First Office Action (including translation) in Application No. 200880009069.3, mailed Jul. 1, 2011, 8 pages.
EPO Office Action in App. No. 06 801 295.4, mailed Oct. 10, 2011, 5 pages.
Cula O G et al., “Bidirectional Imaging and Modeling of Skin Texture,” IEEE Transactions on Biomedical Engineering, IEEE Service Center, Piscataway, NJ, USA, vol. 51, No. 12, Dec. 1, 2004, pp. 2148-2159.
Chujit Jeamsinkul, “MasqueArray Automatic Makeup Selector/Applicator”, Nov. 11, 1998, Rochester Institute of Technology, 79 pages.
Office Action for Japanese Patent Application No. 2009-549296, Apr. 30, 2013, 12 pages.
Notification of the Third Office Action for Application No. 200880009579.0, dated Jan. 7, 2013, 8 pages.
Notice to File a Response in Application No. 10-2008-7006041, dated Jan. 29, 2013, 10 pages.
Examination Report for European Application No. 08769826.2, dated Jul. 16, 2013, 6 pages.
Notice to File a Response in Application No. 10-2008-7006079, dated Jun. 25, 2013, 5 pages.
Office Action for Korean Patent Application No. 10-2009-7019063, Mar. 24, 2014, 8 pages.
Examination Report for Canadian Patent Application No. 2,619,706, Jul. 31, 2014, 3 pages.
Related Publications (1)
Number Date Country
20110124989 A1 May 2011 US
Continuations (1)
Number Date Country
Parent 12028835 Feb 2008 US
Child 13019462 US