AU2006279800A1 - System and method for medical monitoring and treatment through cosmetic monitoring and treatment - Google Patents

System and method for medical monitoring and treatment through cosmetic monitoring and treatment Download PDF

Info

Publication number
AU2006279800A1
AU2006279800A1 AU2006279800A AU2006279800A AU2006279800A1 AU 2006279800 A1 AU2006279800 A1 AU 2006279800A1 AU 2006279800 A AU2006279800 A AU 2006279800A AU 2006279800 A AU2006279800 A AU 2006279800A AU 2006279800 A1 AU2006279800 A1 AU 2006279800A1
Authority
AU
Australia
Prior art keywords
skin
reflectance
frexel
frexels
data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
AU2006279800A
Other versions
AU2006279800B2 (en
Inventor
Albert D. Edgar
David C. Iglehart
Rick B. Yeager
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
TCMS Transparent Beauty LLC
Original Assignee
RICK YEAGER
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by RICK YEAGER filed Critical RICK YEAGER
Publication of AU2006279800A1 publication Critical patent/AU2006279800A1/en
Assigned to TCMS TRANSPARENT BEAUTY LLC reassignment TCMS TRANSPARENT BEAUTY LLC Request for Assignment Assignors: YEAGER, Rick
Application granted granted Critical
Publication of AU2006279800B2 publication Critical patent/AU2006279800B2/en
Ceased legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • AHUMAN NECESSITIES
    • A45HAND OR TRAVELLING ARTICLES
    • A45DHAIRDRESSING OR SHAVING EQUIPMENT; EQUIPMENT FOR COSMETICS OR COSMETIC TREATMENTS, e.g. FOR MANICURING OR PEDICURING
    • A45D44/00Other cosmetic or toiletry articles, e.g. for hairdressers' rooms
    • A45D44/005Other cosmetic or toiletry articles, e.g. for hairdressers' rooms for selecting or displaying personal cosmetic colours or hairstyle
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/103Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
    • AHUMAN NECESSITIES
    • A45HAND OR TRAVELLING ARTICLES
    • A45DHAIRDRESSING OR SHAVING EQUIPMENT; EQUIPMENT FOR COSMETICS OR COSMETIC TREATMENTS, e.g. FOR MANICURING OR PEDICURING
    • A45D34/00Containers or accessories specially adapted for handling liquid toiletry or cosmetic substances, e.g. perfumes
    • A45D34/04Appliances specially adapted for applying liquid, e.g. using roller or ball
    • AHUMAN NECESSITIES
    • A45HAND OR TRAVELLING ARTICLES
    • A45DHAIRDRESSING OR SHAVING EQUIPMENT; EQUIPMENT FOR COSMETICS OR COSMETIC TREATMENTS, e.g. FOR MANICURING OR PEDICURING
    • A45D40/00Casings or accessories specially adapted for storing or handling solid or pasty toiletry or cosmetic substances, e.g. shaving soaps or lipsticks
    • A45D40/30Masks for marking lips or eyelashes
    • AHUMAN NECESSITIES
    • A45HAND OR TRAVELLING ARTICLES
    • A45DHAIRDRESSING OR SHAVING EQUIPMENT; EQUIPMENT FOR COSMETICS OR COSMETIC TREATMENTS, e.g. FOR MANICURING OR PEDICURING
    • A45D44/00Other cosmetic or toiletry articles, e.g. for hairdressers' rooms
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/0059Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence
    • A61B5/0062Arrangements for scanning
    • A61B5/0064Body surface scanning
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/44Detecting, measuring or recording for evaluating the integumentary system, e.g. skin, hair or nails
    • A61B5/441Skin evaluation, e.g. for skin disorder diagnosis
    • A61B5/442Evaluating skin mechanical properties, e.g. elasticity, hardness, texture, wrinkle assessment
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/44Detecting, measuring or recording for evaluating the integumentary system, e.g. skin, hair or nails
    • A61B5/441Skin evaluation, e.g. for skin disorder diagnosis
    • A61B5/444Evaluating skin marks, e.g. mole, nevi, tumour, scar
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/68Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient
    • A61B5/6887Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient mounted on external non-worn devices, e.g. non-medical devices
    • A61B5/6888Cabins
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B41PRINTING; LINING MACHINES; TYPEWRITERS; STAMPS
    • B41JTYPEWRITERS; SELECTIVE PRINTING MECHANISMS, i.e. MECHANISMS PRINTING OTHERWISE THAN FROM A FORME; CORRECTION OF TYPOGRAPHICAL ERRORS
    • B41J3/00Typewriters or selective printing or marking mechanisms characterised by the purpose for which they are constructed
    • B41J3/36Typewriters or selective printing or marking mechanisms characterised by the purpose for which they are constructed for portability, i.e. hand-held printers or laptop printers
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B41PRINTING; LINING MACHINES; TYPEWRITERS; STAMPS
    • B41JTYPEWRITERS; SELECTIVE PRINTING MECHANISMS, i.e. MECHANISMS PRINTING OTHERWISE THAN FROM A FORME; CORRECTION OF TYPOGRAPHICAL ERRORS
    • B41J3/00Typewriters or selective printing or marking mechanisms characterised by the purpose for which they are constructed
    • B41J3/407Typewriters or selective printing or marking mechanisms characterised by the purpose for which they are constructed for marking on special material
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B41PRINTING; LINING MACHINES; TYPEWRITERS; STAMPS
    • B41JTYPEWRITERS; SELECTIVE PRINTING MECHANISMS, i.e. MECHANISMS PRINTING OTHERWISE THAN FROM A FORME; CORRECTION OF TYPOGRAPHICAL ERRORS
    • B41J3/00Typewriters or selective printing or marking mechanisms characterised by the purpose for which they are constructed
    • B41J3/407Typewriters or selective printing or marking mechanisms characterised by the purpose for which they are constructed for marking on special material
    • B41J3/4073Printing on three-dimensional objects not being in sheet or web form, e.g. spherical or cubic objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/90Determination of colour characteristics
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/60Extraction of image or video features relating to illumination properties, e.g. using a reflectance or lighting model
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • G06V40/171Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N1/00Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
    • H04N1/46Colour picture communication systems
    • H04N1/56Processing of colour picture signals
    • H04N1/60Colour correction or control
    • H04N1/62Retouching, i.e. modification of isolated colours only or in isolated picture areas only
    • H04N1/628Memory colours, e.g. skin or sky
    • AHUMAN NECESSITIES
    • A45HAND OR TRAVELLING ARTICLES
    • A45DHAIRDRESSING OR SHAVING EQUIPMENT; EQUIPMENT FOR COSMETICS OR COSMETIC TREATMENTS, e.g. FOR MANICURING OR PEDICURING
    • A45D44/00Other cosmetic or toiletry articles, e.g. for hairdressers' rooms
    • A45D2044/007Devices for determining the condition of hair or skin or for selecting the appropriate cosmetic or hair treatment
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61MDEVICES FOR INTRODUCING MEDIA INTO, OR ONTO, THE BODY; DEVICES FOR TRANSDUCING BODY MEDIA OR FOR TAKING MEDIA FROM THE BODY; DEVICES FOR PRODUCING OR ENDING SLEEP OR STUPOR
    • A61M35/00Devices for applying media, e.g. remedies, on the human body
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61MDEVICES FOR INTRODUCING MEDIA INTO, OR ONTO, THE BODY; DEVICES FOR TRANSDUCING BODY MEDIA OR FOR TAKING MEDIA FROM THE BODY; DEVICES FOR PRODUCING OR ENDING SLEEP OR STUPOR
    • A61M35/00Devices for applying media, e.g. remedies, on the human body
    • A61M35/20Non-portable devices, e.g. spraying booths
    • A61M35/25Non-portable devices, e.g. spraying booths specially adapted for the application of sunscreen, tanning or self-tanning lotions

Landscapes

  • Health & Medical Sciences (AREA)
  • Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Public Health (AREA)
  • Animal Behavior & Ethology (AREA)
  • Veterinary Medicine (AREA)
  • Biophysics (AREA)
  • Pathology (AREA)
  • Surgery (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Medical Informatics (AREA)
  • Molecular Biology (AREA)
  • Multimedia (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Dermatology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Signal Processing (AREA)
  • Radiology & Medical Imaging (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Software Systems (AREA)
  • Manufacturing & Machinery (AREA)
  • Dentistry (AREA)
  • Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
  • Measuring And Recording Apparatus For Diagnosis (AREA)
  • Medicines Containing Material From Animals Or Micro-Organisms (AREA)
  • Cosmetics (AREA)
  • Investigating Or Analysing Materials By Optical Means (AREA)
  • Infusion, Injection, And Reservoir Apparatuses (AREA)

Abstract

A computer-controlled system determines attributes of a frexel, an area of human skin, and applies a modifying agent (RMA) at the pixel level, typically to make the skin appear more youthful and so more attractive. The system scans the frexel, identifies unattractive attributes, and applies the RMA, typically with an inkjet printer. The identified attributes relate to reflectance and may refer to features such as irregular-looking light and dark spots, age-spots, scars, and bruises. Identified attributes may also relate to the surface topology of the skin, for more precisely enhancing surface irregularities such as bumps and wrinkles. Feature mapping may be used, for example to make cheeks appear pinker and cheekbones more prominent. The RMA can be applied in agreement with identified patterns, such as adding red to a red frexel, or in opposition, such as adding green or blue to a red frexel, according to idealized models of attractiveness.

Description

WO 2007/021972 PCT/US2006/031441 Patent Application 5 of Albert D. Edgar, David C. Iglehart, and Rick B. Yeager for SYSTEM AND METHOD FOR MEDICAL MONITORING AND TREATMENT THROUGH COSMETIC MONITORING AND TREATMENT 10 CROSS-REFERENCE TO RELATED APPLICATIONS This application claims the benefit of U.S. Provisional Application Serial Number 60/708,118 filed August 12, 2005 by applicants. 15 FIELD OF THE DISCLOSURE The present invention relates to the monitoring of the human body to detect attributes that warrant preventative or curative medical attention or treatment and more particularly to an automated system and method to apply to medical monitoring and treatment of a region of skin. 20 BACKGROUND Monitoring the external surface of the human body is a well established method for detecting conditions indicating illness or an increased risk of illness. For example, doctors examine patients to discover melanoma and then quickly prescribe courses of treatment for that disease. In the same way, they may discover that a patient has sun-damaged skin and prescribe that the patient avoid excessive exposure to the sun to prevent the 25 future development of melanoma or other skin cancers. In addition, they may examine the surface topology of a woman's skin and discover a strong possibility that the woman has underlying breast cancer, so that they order further tests and appropriate treatment. Medical examinations may be performed visually or through many kinds of medical equipment. In some cases, 30 visual examinations may be more effective than those made by instruments, since the human eye sees both reflectance, for example color, and surface topology, for example bumps and depressions, but many mechanical instruments do not. This ability sometimes allows the eye to identify medical problems that instruments miss or cannot interpret. For example, many automated systems for medical monitoring lack the ability to analyze surface topology. 35 Through regular examinations, a base line of a patient's physical characteristics can be created, so that variations from that base line can be detected. One advantage of monitoring patients in this way is that early detection and treatment of illnesses and potentially dangerous conditions increases the success rates of treatments. For example, a detailed daily examination of patients' skin for signs of skin cancer could greatly 40 increase their ability to avoid or survive skin cancer. In addition, the detection of splotches of white in people WO 2007/021972 PCT/US2006/031441 with dark skin, such as many African Americans, can indicate internal cancer. Regular examinations of skin can aid in discovering for many other illnesses and health problems, including acne, blisters, bruises, scarring, jaundice, varicose veins, and infections, and in treating them successfully. 5 However, the high cost of manual examinations by doctors and of examinations made with expensive medical equipment typically limits the frequency of patient monitoring. The daily monitoring of patients is typically too expensive to be conducted, except in cases where serious illnesses have already occurred and the patients have been hospitalized. General practitioners usually ask their patients to undergo general examinations only yearly, or even less frequently. Specialists also usually only see patients at widely spaced intervals, such as yearly or 10 every six or three months. Although patients are asked to conduct self-exams more regularly, they often lack the expertise and discipline to detect medical conditions successfully. As a result, many diseases remain undetected until times when they are less susceptible to treatment. Therefore, there is a need for an automated system and method that provides frequent medical monitoring of the 15 external surface of the human body for characteristics of reflectance and surface topology, identifies and reports on medical problems, and, in an embodiment, treats those problems when appropriate. In this specification, the terms "reflectance modifying agent" or "RMA" refer to any compound useful for altering the reflectance of another material, and are explained in further detail below. Some examples of RMA 20 are inks, dyes, pigments, bleaching agents, chemically altering agents, and other substances that can alter the reflectance of human skin and other features. The terms "dye" and "transparent dyes" are used for brevity in this specification to represent any RMA. BRIEF SUMMARY OF THE INVENTION 25 These and other needs are addressed by the present invention. The following explanation describes the present invention by way of example and not by way of limitation. It is an aspect of the present invention to determine the visual attributes of an area of skin by electronically scanning the area and analyzing the scanned data in a computing environment. 30 In one embodiment, the scanning provides reflective data about the skin. The data is used to conduct feature identification and to monitor changes in reflectance over time. 35 In one embodiment, the scanning provides both reflective and surface profile data. The data is used to conduct feature identification and to monitor changes in reflectance or morphology over time. It is another aspect of the present invention to create a map of the area of skin, and to use that map at a later time to determine the location, relative to the skin, of a monitoring device. The map may also be used to compare 40 images from a first time and a second time in order to detect changes in reflectance or shape. 2 WO 2007/021972 PCT/US2006/031441 In this patent specification, the phrase "inkjet technology" refers generally to "drop control" technology, whereby each individual droplet of the substance being applied can be controlled by the applicator, as known to those skilled in the art. A particularly useful technique for the present invention is to employ "drop on demand" 5 technology, a subset of drop control technology. In this specification, the phrase "inkjet printer" is used for brevity represent any form of inkjet technology. It is another aspect of the present invention to precisely apply one or more medicinal agents to the skin in response to the local reflective properties of the skin. 10 The term "medicinal agent" is used broadly in this application to refer to any material or compound which is beneficial to a region of skin. The medicinal agent may be a pharmaceutical compound, a drug, a natural herb, a moisturizing agent, or any other material which provides a benefit to the skin. 15 These and other aspects, features, and advantages are achieved according to the system and method of the present invention. In accordance with the present invention, a computer-controlled system determines attributes of an area of human skin. The identified attributes may relate to reflectance and may refer to features such as irregular-looking light and dark spots, age-spots, scars, moles, and bruises, or other surface or subsurface features. Identified attributes may also relate to the surface topology of the skin, such as depth, for more 20 precisely enhancing surface irregularities such as bumps and wrinkles. The medicinal agents can be applied in agreement with identified patterns, such as adding red to a red area, or in opposition, such adding green or blue to a red area, according to idealized models of attractiveness. It is an aspect of the current invention to collect and analyze data at different wavelengths (color) in order to 25 provide a basis for detailed analysis of skin features. Some skin features may be identified from the characteristics that the features exhibit in different wavelengths. The application of medicinal agents at the pixel level allows much greater accuracy than with prior art methods, so that less of the applied material is used. In some cases, this precise delivery of medicinal agents to specific 30 areas provides an enhanced effect of the agent without side effects associated with a systematic application of the agent. For instance, acne medication can be applied effectively to skin legions rather than taken as an oral medicine. In one embodiment of the current invention, an application device comprising a scanner and an inkjet printer 35 makes a single pass over an area of skin. It scans the skin, identifies problems, calculates the amount and distribution of a medicinal agent and quickly applies the agent onto the skin. In another embodiment, the application device makes a first map of the features of the skin at a first time and identifies features of interest. It then makes a second map of the features of the skin at a second time and 3 WO 2007/021972 PCT/US2006/031441 identifies features of interest. The two maps, and the features of interest are compared to monitor changes in the features. In this embodiment, a detailed scan is made of a region of human skin such as a face, leg, or arm. The scan is 5 acquired by deliberately flashing multiple light sources arranged in a known configuration, and scanning a small area of skin as the light sources are turned on and off. By comparing readings from different light sources, both the reflectance and the surface profile of the skin can be determined. The data from the scan includes reflective characteristics of the skin. These characteristics can be used to 10 produce a detailed map of the skin which includes both reflectance and skin surface morphology. The detailed map can be used to develop a corrective plan to selectively apply a medicinal agent. It is an aspect of the present invention to provide an automated system and method that provides frequent medical monitoring of the external surface of the human body for characteristics of reflectance and surface 15 topology. It is another aspect of the present invention to provide an automated system and method that identifies medical problems in the external surface or subsurface of the human body through monitoring. 20 It is another aspect of the present invention to provide an automated system and method that reports of medical problems in the external surface or subsurface of the human body identified through monitoring. It is still another aspect of the present invention to provide an automated system and method that applies medicines to identified problems in the external surface of the human body when appropriate. 25 These and other aspects of the present invention will become readily apparent upon further review of the following specification and associated drawings. In accordance with one embodiment of the present invention, a system and method that scans areas of a human body to identify unattractive characteristics and make cosmetic enhancements is modified for medical monitoring and optionally for treatment. A 3-D model of the exterior 30 surface of the human body is created, and the scanned data is analyzed by characteristics of reflectance and surface topology to identify unhealthy characteristics. Because people will use a cosmetics system widely and frequently, base lines of patients' conditions can be created through frequent monitoring over a long time period, so that potentially dangerous changes from the base line can identified quickly and reported on. When appropriate, precise applications of medications to treat affected areas may be made automatically. Controlled 35 and precisely directed dosages of medications may be applied to reduce the risk of undesirable side effects. Medicines may also be applied over a large area of skin and during a long period of time to achieve desired treatments. 40 4 WO 2007/021972 PCT/US2006/031441 BRIEF DESCRIPTION OF THE DRAWINGS The following embodiment of the present invention is described by way of example only, with reference to the accompanying drawings, in which: 5 FIG. 1 is a block diagram showing an operating environment in which embodiments of the present invention may be employed for applying RMAs onto skin; FIG. 2 is a block diagram showing an operating environment in which embodiments of the present invention 10 may be employed for applying RMAs onto skin through communications over a network; FIG. 3 is a block diagram showing an operating environment in which embodiments of the present invention may be employed for applying RMAs onto skin through communications over a network and a portable application device; 15 FIG. 4 is a block diagram showing the use of magenta, yellow, and cyan RMAs. FIG. 5 is a block diagram showing an operating environment in which embodiments of the present invention may be employed through a self-contained portable application device for applying inks onto skin; 20 FIG. 6 is a flow chart that illustrates a process for employing an application system; FIG. 7 is a flow chart that illustrates a process for setting up an application system; 25 FIG. 8 is a flow chart that illustrates a process for the programming in an application algorithm in an embodiment for printing on skin; FIG. 9 is a flow chart that illustrates a process for creating a printable enhancement image; 30 FIG. 10 is a diagram that illustrates how a 3-D object maps to a 2-D surface in a computer model; FIG. 11 is a flow chart that illustrates a process defining principles of attractiveness; FIG. 12 is a block diagram that illustrates lighting from above on an area with surface texture variations; 35 FIG. 13 is a diagram that illustrates characteristics or features on a 3-D human face; FIG. 14 is a perspective diagram that illustrates features on a 2-D map of a human face; 40 FIG. 15 is a perspective diagram that illustrates characteristics or features on a 2-D map of a human face; 5 WO 2007/021972 PCT/US2006/031441 FIGS. 16A-E are charts of the reflectance, illuminance, and a printable enhancement image along line A-A' of the 2-D map of the human face of FIG. 14; 5 FIG. 17 is a diagram of a 3-D human face that has been enhanced through printing of RMAs according to a printable enhancement image; FIG. 18 is a diagram that illustrates characteristics of a 3-D human leg, with the corresponding reflectance per spectral base on a 2-D map, and a printable enhancement image; 10 FIG. 19 is a diagram of a 3-D human leg that has been enhanced through printing of RMAs according to a printable enhancement image; FIG. 20A-B are diagrams that illustrates characteristics of a 3-D human breast, with the corresponding 15 reflectance per spectral base on a 2-D map; FIG. 21 is a diagram that illustrates performing multiple passes of scanning and applications; FIG. 22A-C are diagrams that illustrate the effects of RMAs applied to improve the appearance of an age-related 20 freckle; FIG. 23 is a generalized graph of visual benefits versus resolution; FIG. 24 is a flow chart showing the general steps employed by the present invention; 25 FIG. 25 is a generalized graph of patterns of unattractive features in RGB bands; FIG. 26 is a block diagram showing a spacer cup on a sensor; 30 FIG. 27 is a block diagram showing an operating environment in which embodiments of the present invention may be employed for applying RMAs onto skin through communications over a network and an application device comprising a booth; FIG. 28 is a generalized graph of weaker and stronger middle frequencies; 35 FIG. 29 is a block diagram showing an operating environment in which embodiments of the present invention may be employed for applying RMAs onto skin through communications over a network and a blotter application device; 6 WO 2007/021972 PCT/US2006/031441 FIG. 30 is a block diagram showing an operating environment in which embodiments of the present invention may be employed for applying RMAs onto skin through communications over a network and a portable application device with a curved surface; 5 FIG. 31 is a flow chart for coordinating pixel-level mapping of skin; FIG. 32 flow chart for coordinating a pixel-level application of RMAs; FIG. 33 is a flow chart showing a process for employing enhancement techniques; 10 FIG. 34 is a flow chart showing a process for determining the aim depth of the scanned area; FIG. 35 is a flow chart showing a process for determining the aim illumination of the scanned area; 15 FIG. 36 is a block diagram showing an application device comprising a booth; FIG. 37 is a schematic for a simple skin smoothing example. FIG. 38 is a schematic for a multiple pass smoothing example. 20 FIG. 39 is a schematic for a facial map example. FIGS. 40A-B are sample layouts for LEDs and a sensors for acquiring reflectance and skin orientation data; 25 FIG. 41 is a schematic for feature recognition; FIG. 42 is a schematic of an example for feature recognition; FIG. 43 is a schematic of an artistic strategy for applying RMAs; 30 FIG. 44 is an example of a rotating printer for a blotter application device; FIG. 45 is an example of a text image showing apparent depth; and 35 FIG. 46 is flowchart illustrating a correction process. FIG. 47A is a side view of one embodiment of a handheld device for skin marks. FIG. 47B is a front view of the device of FIG. 47A. 40 7 WO 2007/021972 PCT/US2006/031441 FIG. 47C is a top cross sectional view along section AA' of FIG. 47B. FIG. 48 is a flow chart that illustrates a process for medical monitoring and optional treatment; 5 FIG. 49 is a flow chart that illustrates a process for identifying medical problems; FIG. 50 is a bar graph showing the number of samples collected for medical monitoring of a person in a year by different data collection methods; 10 FIG. 51 is a block diagram that illustrates an operating environment for the present inventions, with reporting elements; and FIG. 52 is a block diagram showing the use of a second reservoir with medical compounds and a second inkjet printer head for medical applications of a cosmetic system. 15 DETAILED DESCRIPTION OF EMBODIMENT - PROVIDING A COSMETIC SYSTEM FOR APPLYING REFLECTANCE MODIFYING AGENTS, AND USING THE COSMETICS SYSTEM FOR MEDICAL 20 MOITORING The details of the following explanation are offered to illustrate the present invention clearly. However, it will be apparent to those skilled in the art that the concepts of present invention are not limited to these specific details. Commonly known elements are also shown in block diagrams for clarity, as examples and not as limitations of the present invention. Furthermore, the order of processes, their numbered sequences, and their 25 labels are presented for clarity of illustration and not as limitations on the present invention. This embodiment describes a cosmetic method to improve the visual attractiveness of a region human skin as described in copending patent applications by applicants. As described in more detail below, in one embodiment of the current invention the cosmetic system is adapted for use in medical monitoring of the skin. 30 One advantage is that it provides a frequent monitoring of skin during the cosmetic sessions, without additional effort. In another embodiment, many of the scanning, mapping, and application techniques described for the cosmetic system are adapted for a medical monitoring device. As shown in FIG. 24, the cosmetic method comprises the general steps of 35 * Step 900 - allocating a region of skin into a plurality of frexels; * Step 910 - measuring at least one optical attribute of each of the plurality of frexels; * Step 920 - determining, from the optical attributes of the frexels, at least one measured skin characteristic affecting visual attractiveness; * Step 930 - determining a desired state of the skin characteristic; and 8 WO 2007/021972 PCT/US2006/031441 * Step 940 - applying at least one reflectance modifying agent to specific frexels in order to modify the measured skin characteristic to approach the desired state of the skin characteristic. This step may be modified to apply one or more medicinal agents in the current invention. 5 Allocating a Region of Skin into a Plurality of Frexels In this patent specification, the term "frexel" is defined as a small pixel-like region of the skin. In this patent application, the term "skin" is used not only to refer to skin as on the surface of the human body, but also to refer more broadly to any human feature that may be enhanced cosmetically, for example fingernails and hair. A frexel might correspond to a small portion of a freckle or other skin feature, or it may correspond to an area of 10 the skin that does not have special features. A frexel thus refers to skin rather than to an independent coordinate system. The term frexel is used to suggest that what is being measured is on a 3-D surface rather than a flat surface. A region of skin is comprised of a plurality of frexels. For instance, if a resolution of 300 dots per inch (11.8 pots 15 per mm or "dpmm") is used, a frexel may have a width and height of about 1/300th of an inch (0.085 mm) so that there are approximately 90,000 frexels per square inch (140 frexels per square mm). The surface of the human body may have millions of frexels. By allocating skin into frexels, the present invention can accomplish scanning and the application of RMAs for 20 enhancement at the higher end of the human visual ability to resolve detail. FIG. 23 is a generalized graph of relative visual benefits 450 versus dots per inch (DPI) 452 resolution. A target resolution in the range of 50 to 300 dpi (2-11.8 dpmm) provides much better resolution that existing cosmetic techniques, as well as advantages in making the adjustments in response to actual and desired skin 25 characteristics; and the further advantage of automatic application. Prior art techniques for applying makeup with brushes, tubes, and fingers have much coarser resolutions. For instance a fine brush has an approximate resolution of about 20 dpi (o.8 dpmm). Measuring at Least One Optical Attribute of Each of the Plurality of Frexels 30 Scanning As shown in FIG. 1, in one embodiment, an application device comprising a scanner 220 is moved across the area of skin 302 so that the scanner 220 can electronically record data about one optical attribute, such as the reflectance, of each of the plurality of frexels. For example, the area of skin 302 might be a face. 35 The scanning may acquire images under various frequencies to obtain useful data. For example, it may obtain data on reflectance in a particular color, for example red, to help determine a particular characteristic of skin for enhancement. The scanning may also provide data for determining other characteristics of skin, such as surface topology, based on reflectance angle from multiple light sources. 9 WO 2007/021972 PCT/US2006/031441 In an embodiment a two-dimensional array is used for the scanning. In other embodiments, a line array may be used. Alerting Sounds In an embodiment one or more alerting means, such as a sound, light, or vibration may be used to indicate when 5 sufficient scanning has been accomplished. Sensors In one embodiment, the scanner 220 comprises a sensor and four LED light sources arranged in a known configuration within a housing. The LED light sources are typically each turned on and off in a manner that allows the sensing of at least one optical characteristic for each light source. In one example, 120 captures may 10 be made per second, 30 from each light, quickly providing a large about of data about the skin. That data can then be used to determine both reflectance characteristics at various wavelengths, and the skin's surface profile. In an embodiment the captured images may be averaged for effectiveness. In an embodiment, the sensor comprises shading patterns on the LEDs useful for determining the relative 15 position of the sensor. In an embodiment a monochrome sensor with a Bayer array may be employed. Other arrangements of LEDs and sensors may be used. 20 Analyzing the Scanned Data The scanned data comprises information about * The reflectance from the skin, and * The location of the skin relative to the sensor, and the skin features. 25 In an embodiment, the application algorithm 230 puts the stored data into spatial frequency bands and uses pattern recognition to analyze them to determine the landscape of the area of skin 302 and the dimensions that require application of the RMAs 264. The process used to determine these dimensions will be explained in detail below. 30 The application algorithm 230 uses its analysis to create in software an application map 232 of the area of skin 302, which is stored in storage 250, for potential future use. Optical Attributes The reflectance, which is a measure of the reflection of the skin, is independent of its illuminance. Illuminance is a measure of how much light gets to the skin. The light reading is independent of the surface topology 35 reading. In an embodiment, certain optical attributes, such as the amount of reflectance of each frexel, may be determined directly from the scanned data. In another embodiment, the scanned data is translated into at least 10 WO 2007/021972 PCT/US2006/031441 one spatial frequency band for analysis. In still another embodiment, the scanned date may be translated into multiple spatial frequency bands, such as red, green, and blue (RGB) bands. FIGS. 16A-E represent the patterns of a 2-D face 232, shown in FIG. 14, after the data has been put into single 5 spatial frequency bands to determine the attributes of albedo 348 and illuminance 352. Albedo Albedo is the percentage of reflectivity of incident light from the surface of an object. In the case of electronic scanning, the albedo is the RBG values of the scanned area of skin. In this patent application, the term "actual albedo" means the observed albedo before correction and the term "aim albedo" refers to the desired reflectivity 10 of an area of skin in order to improve the appearance of that area of skin. In one example, the aim albedo is determined from one or more correction strategies, including general smoothing, specific feature enhancement, and artistic strategies. The top band in FIG. 16 represents the actual "albedo" along line A-A' in the 2-D surface map 232 of FIG. 14. 15 A rise in the actual albedo graph identifies the light spot 408. A deep, sharp drop in the graph identifies a non uniformity 412 such as a scar. And an irregular section identifies a freckle 410. Illuminance Illuminance is the incident light reaching a unit area of the surface of an object, and is a function of the angle of the incident light relative to the surface. 20 The spatial frequency bands also graph the actual illuminance or shading 352, shown in FIG. 16, of the 2-D surface map 232 shown in FIG. 14. Reflectance and Illuminance Data and Calculations 25 In one example, frexel data obtained from scanning a region of skin may be represented as [(xs, Ys, ZS, ats, PS, 7S), (xf, Yr, zf, P f, f), {(refl)A, (refl)N, (refl)s, (refl)E, (refl)w}]i 30 The term {(refl)A, (refl)N, (refl)s, (refl)E, (refl)w} represents reflective data for the frexel i under ambient lighting conditions, and for each of four light sources, such as LEDs, which are arbitrarily designated as north-south east-west for ease of discussion. Other numbers of light sources, such as three sources, can be used, but the mathematics is simplified with four light sources. The (refl) represents one or more data point for the reflectance measurement. The reflectance measurement for a wavelength is the product of a constant, the illuminance, and 35 the albedo for the wavelength: Reflectance = k * illuminance * albedo 11 WO 2007/021972 PCT/US2006/031441 For instance: Reflectance (red) = k(red) * illuminance(red) * albedo(red) The constant depends upon several factors including the speed of the lens, the sensitivity of the camera or sensor, the transmission characteristics of the color filter, the gain of the analog amplifier, the digital gain 5 applied by the software, and other factors. The constant k will usually be measured and corrected for as a correction constant or calibration of the camera corrects for these effects. The value of the constant can typically be determined during calibration, when the illumination from the LEDs is assumed to be fixed, and the albedo is calculated based on that assumption. 10 Reflectance is not absolute, but is a measure of what comes out of the camera. The sensor is typically a camera without an amplifier, a digital converter, or the lens housing. In one embodiment, the sensor is a solid state MOS sensor with a lens and associated electronic equipment. The frexel data can be processed to determine a reflectance and an illuminance for each light source, and that 15 information can be used to determine reflectance and surface profile. In one example, the reflectance is the average or mean of all measurements. The illuminance can be determined from the known brightness of light sources such as LEDs. Illuminance is the known light times the cosine of the angle of incident light relative to the normal. 20 One problem with obtaining reflectance data is that glare may be present at some angles, and that an accurate reading cannot be obtained. In one example, glare or glossiness can be eliminated with the use of polarizing materials to provide a cross polarization of the LEDs. In other examples the sensor can deliberately be positioned at a relatively large angle such as 60 degrees in order to eliminate glare. 25 Determining Position Frexel Location Relative to Sensor or Coordinate System The term (xf, yf, ze f, r, T) may represent the distance of the frexel i from the sensor, or may be an absolute position and orientation of the frexel with respect to a reference coordinate system. In one example, the determination of the distance from the frexel to the scanner may be made in two steps. A first step can be an 30 approximate mechanically-based measurement such as a constant height of the sensor from the skin. The second step can be an optical first derivative measurement to provide a fine adjustment. In one example, the fine adjustment is calculated by measuring an angle from the surface. In another embodiment, a fine adjustment may be made by using two light sources to send out two reference points or grids for detection by a sensor. 35 Mechanical Gross Estimate In one embodiment, the sensor may be attached to a helmet or a fixed booth in a manner that the sensor position may be determined relative to the helmet or booth. 12 WO 2007/021972 PCT/US2006/031441 In another embodiment shown in FIG. 26, the sensor 278 may be equipped with a cup 280, so that the sensor 278 maintains an average height from the skin. In another embodiment, the sensor may start from a known position, and keep track of its movements in order to 5 estimate its location. The sensor may measure the angle relative to the probe itself to determine the shape of a surface feature relative to constantly changing plane of probe. A gimbal may be used to provide a reference in space. The tracking may be used to follow the position of a hand, or hand-held scanner in space. The gimbal arrangement can provide regular feedback in a manner that is 10 analogous to stereo-mapping or GPS mapping relative to satellites, such as for crop dusting. Optical Fine Adjustment For finer alignment, an optical means may be used. For example, the first derivative of the z component of the skin may be obtained from shading, through multiple light and shadings from probes. The first derivative can provide a measure of the angle of the surface. 15 In one example, three light sources send out different patterns. The color and the shading provide data for determining surface relief so that a shaded relief map may be obtained from the LEDs. Frexel Orientation By determining the tilt of the frexel relative to two orthogonal axes, the orientation of the frexel can be 20 determined. The orientation of a frexel and its neighbors is an indication of the actual local surface texture of the skin. One aspect of the current invention is the ability to measure and compensate for both local reflective properties and local surface texture. In this example, there are four light sources which are designated as North, South, East, and West. The sensor 25 obtains data when each light source is on, and the other sources are off. The sensor may also obtain data for ambient lighting, with none of the four light sources on. The tilt of the frexel can be determined by comparing the North and South measurements. The difference between these measurements is a related to the tilt of the frexel along the East-West axis. The difference 30 between the East and West measurements is a related to the tilt of the frexel along the North-South axis. It is generally necessary to make a gamma correction by converting the data to linear space. The gamma correction is approximated by taking the square root of the data output by typical gamma 2 cameras. Light Sources 35 FIGS. 40A-B show configurations for light sources that may be used with one embodiment of the present invention. In this embodiment, a set of four light sources is used- LEDN, LEDs, LEDE, and LEDw. The light sources are placed in a diamond configuration where the sensor is positioned at the center of the LED layout. This configuration simplifies the mathematical analysis for calculating surface profile. 13 WO 2007/021972 PCT/US2006/031441 Mean Illumination In one embodiment, it is useful to employ the concept of mean illumination. Mean illumination is the average angle and diffusion of light reaching a particular surface. This defines how surface irregularities are typically shaded. For example, mean illumination for the entire body is overhead, and a typical orientation for a head is 5 vertical; therefore, a bump on a cheek is typically shaded at the bottom. For a child on the beach, typically the bump would be less tanned on the bottom because the average light throughout the day, integrating both sun angle and body angle to give average or mean illumination, is from over "head." Occasionally light is reversed from average. An example is lighting a face from underneath. However, this often gives a bizarre, sometimes sinister look, and is the exception that proves the rule. By correcting a defect for mean illumination, the best 10 correction on average is performed. Mean illumination is the interaction of mean light direction relative to gravity and the mean orientation of a particular frexel of skin relative to gravity. One method to obtain the angle of the skin is to use multiple diffuse or orthogonal light sources in a configuration which may include mirrors. The lights may be flashed repeated, 15 as strobe lights, so that hundreds of images may be taken of a small area, and the data can be averaged. From the angle of the skin relative to "up," one can calculate how much light reaches the skin under mean illumination and the angle of the skin relative to "up." A reasonable approximation to mean illumination can be made by turning on all lights sources at the same time, 20 or by adding images made by individual light sources. In one example, mean illumination is diffuse because lights and probes are perpendicular to the skin. A refinement of this technique will compensate for gloss effects on the skin. For example, several images with four lights sources may be used and an average taken of the images from the light sources. For example, the 25 average might be a median. One advantage of the median is that if specular reflection is caught by a minority of light sources, it would be filtered by median. The median would also filter shadows observed from a minority of light source images. This is important because the human body represents complex surfaces, i.e. a nose may be shiny when illuminated. 30 One way to create diffuse light is to introduce light from many light sources at many angles. Another way to create diffuse light is to reflect it from the scanner housing. Another option is to polarize the light. Example ofFrexel Data Representation An example of the data representation for a frexel is shown below: [(x, ys, zS, Cs, PS, YS), 35 (xf, yf, zf, ar, , yr ), {(refl)A, (refl)N, (refl)s, (refl)E, (refl)w}] In this example, (x,,ys,zs,as, s,ys) and (xf, yf, zf, af, 1 3 r, yfr ) represent the position and angular orientation of the scanner sensor and the frexel relative to a coordinate system. 14 WO 2007/021972 PCT/US2006/031441 Compression In some embodiments, the efficiency of the data processing can be improved by various compression methods, such as JPEG. Frexel Location on the Skin 5 Through Feature Mapping Computer mapping for feature recognition, known to those skilled in the art in areas such as computer gaming, can be used for tracking the location of the probe on the area of skin 302 and for determining enhancements appropriate for specific features. 10 For example, such computer mapping enables the identification of features such as a cheekbone, a nose, and an ear, so that the probe can orient its location with regard to a particular frexel, potentially in multiple passes over an area of skin. Moreover, the identification of a feature such as a cheekbone enables determination of appropriate 15 enhancements. For example, a red reflectance modifying agent may be applied to the center of a cheekbone to add color to a face. Dark reflectance modifying agents may be applied underneath the cheekbone to make the cheekbone appear to project more prominently. Skeleton Model In one embodiment, a map is built around a skeleton model so that the skeletal joints become reference points. 20 In this example, the joints are located, a stick figure is constructed, and a 3-D mesh is built around the stick figure. The map is relative to a predetermined model of human skeletal structure in the memory of a computing environment. Manikin-Like Model In one embodiment, the map is relative to a predetermined model of a human body. 25 Dynamic Model In one embodiment, the map is relative to the movement of skin over a predetermined model, such as a skeleton model or manikin-like model. Through Chemical Markers In other embodiments, chemical markers may be applied to the area of skin during the scan to help create the 30 map and enable subsequent tracking of the map with the area of skin 302. For example, ultraviolet markers may be used, such as dots which are visible under ultraviolet light, but not visible under conventional lighting. Single Pass or Multiple Pass In various embodiments, the scanning and correction can be accomplished in a single or multiple passes. For instance, a first pass may be performed to become acquainted with the subject, and a second or subsequent pass 15 WO 2007/021972 PCT/US2006/031441 may be performed to get additional data. Multiple passes at different orientations over the same area provide an opportunity for compensating for the effects of skin hair by observing the skin at different angles. Single Pass In one embodiment of the current invention, an application device comprising a scanner and an inkjet printer 5 makes a single pass over an area of skin. It scans the skin, identifies unattractive characteristics, calculates enhancements to make the skin more attractive, and quickly applies RMAs onto the skin to achieve those enhancements. Multiple Pass In a further embodiment, the application device makes multiple passes over the skin, each time improving the 10 scanning and the application of RMAs for the desired enhancement or enhancements. Example of Tracking Process In one example of a tracking process, a rough position is first determined, and then a more precise location is established. In a first approach, a rough estimate of location can be maintained from a known starting point through the use of gimbals in proximity to the probe to compute distance and direction traveled. In another 15 approach, a rough location can be determined from mechanical probes or gauges. In another approach, a rough location can be estimated mathematically by using the first derivative of the shading data. Once the rough location is known, a more precise location can be determined from the analysis of frexel orientation from shading data. This is analogous to a pilot determining position by first knowing an 20 approximate location and then locating land features that provide a more precise location. Tracking Over Time One advantage to the generation of maps is that changes in reflectance or surface profile can be determined by comparing an image from a first time with an image from a second time. These changes may represent changes in the health of a person, or may represent areas that require a "touch-up" of RMAs. 25 Determining, from the Optical Attributes of the Frexels, at Least One Measured Skin Characteristic Affecting Visual Attractiveness Pattern recognition may be used to identify features of the area of skin 302 that has been scanned. 30 Feature Identification Reflectance and Topology Feature identification may be based on patterns determined in scanned data, and may have to do with both the reflectance patterns and the surface topology of the area of skin. Mathematical pattern analysis of this data 35 allows identification of specific unattractive features that could benefit from enhancement techniques. As explained below, such features may typically be characterized by age-related and damage-related patterns that 16 WO 2007/021972 PCT/US2006/031441 are irregular or asymmetrical compared to the more regular and symmetrical genetic-based patterns of younger skin. The eve's perception of depth 5 At small distances, the human eye perceives depth by stereoscopy. At a typical human interaction range of a few feet, however, the eye perceives depth of human skin based on the reflectance of the skin. A difference in shading between adjacent areas of skin is perceived as a surface texture representing elevation or depth from the surface of the skin. As an example of that perception, FIG. 45 shows the text letters "RICK" which were created in PhotoshopTM. From an original image of flat letters, the software created the apparent shadows. The 10 human eye interprets the differences in reflectance by assuming that a light source is located in the upper left, and that shadows are created because the text has a raised profile. This perception of depth from differences in reflectance is also important in the perception of human beauty. The eye interprets differences in shading of skin to be surface texture. That perception of texture can be altered 15 by changing the reflectance of the skin. In the letter example for instance, the perception of raised letters can be dramatically altered by reducing the shadowing around the letters. The eye perceives the color of the skin and translates that color information into a perception of depth. One aspect of the current invention is to selectively change the reflectance of a portion of the skin in order to alter 20 this perception of depth. This alteration may be made in relatively small areas such as a bump on the skin; or the alteration may be made over larger areas, such as with traditional cosmetics, such as deliberately darkening an area around the eyes or cheeks. Examples of Unattractive Features Some examples of unattractive features in skin that can be identified from scanned data are 25 * Acne, * Age spots/sun damage, * Bruises, * Bumps, * Cellulite, 30 * Light spots, * Pitting, * Scars, * Damaged freckles, and * Wrinkles. 35 Other unattractive features that also may be identified have to do with artificial patterns that have been added to the skin, such as body painting and tattoos that have faded over time or that have been distorted by changing patterns of the skin itself such as sagging or wrinkling. These features can be identified and then refreshed through the application of RMAs to refresh or enhance their appearance. 17 WO 2007/021972 PCT/US2006/031441 Techniques for Identifying Unattractive Features Pattern for Age-Related Freckles in a Single Spectral Band For example, natural freckles are about 2mm across and are sharp edged and have the pattern 442 shown in FIG. 22B. Age-related freckles, caused for example by sun damage, have the pattern 446, shown in FIG. 22A. 5 As explained above, an age-related, random freckle 440, for example from sun damage, on an older person can be identified by its characteristic pattern in a single spectral band, as illustrated in FIG. 22. When scanned data of the random freckle 440 is put into a spectral band, it shows a rough, irregular pattern. Patterns in Multiple Spectral Bands 10 By breaking the scanned image into multiple spectral bands, such as RGB bands, the patterns of unattractive features may be identified with even greater clarity. For example, FIG. 25 is a generalized graph of patterns of unattractive features in RGB bands for an area of young skin, showing the empirically observed general patterns of * A scar 460, 15 * A freckle 468 from sun damage, * A varicose vein 476, * A new, bluish bruise 484, and * An older, yellow bruise 492. 20 The set of RGB patterns for each of these unattractive features is quite distinct and thus detectable through feature recognition. For example, the scar 460 shows patterns in the higher frequency range in all three bands 462, 464, and 466, unlike the other features. The freckle 468 dips more deeply into low frequencies in the blue band 474, than the blue-band patterns for the varicose vein 482, the bluish bruise 490, and the yellow bruise 498. The bluish bruise 484 has larger dips in the red pattern 486 and green pattern 488 than the yellow bruise 25 red pattern 494 and green pattern 496. The yellow bruise blue pattern 498 dips more deeply than the bluish bruise blue pattern 490. Advanced Feature Identification Through Mapping Mapping based on feature identification can add greatly increased capabilities for enhancement to mapping 30 based on reflectance and surface topology. * Maintain register over entire skin surface. * Translate 3-D to lightness/darkness using mean illumination, and include with lightness/darkness attribute, both for printing against or for aesthetic augmentation. 18 WO 2007/021972 PCT/US2006/031441 Means of Compensating for Special Conditions Compensating for Body Hair In one embodiment, the presence of skin hair may be compensated for by taking images in multiple passes while attempting to orient the hair in various directions. The orientation may be accomplished by a comb device 5 associated with the scanner. In other embodiments, a static electric charge may be used to cause the hair to rise relative to the skin. Determining a Desired State of the Skin Characteristic Principles of Attractiveness 10 The present invention employs general principles of attractiveness 500, examples of which are shown in FIG. 11. These principles are based on observation of attributes that many people find attractive and thus represent tendencies in human behavior. Means of Determining a Desired State of the Skin Characteristic 15 Approaches for corrections include pure mathematical techniques and artificial intelligence techniques. By contrast, artistic approaches are more intuitive and less quantitative. * Mathematical * Artificial Intelligence * Artistic 20 Mathematical Means Mathematical techniques include filtering to remove a portion of middle frequencies, and to remove a portion of asymmetric low frequencies. Another example of a pure mathematical technique is printing in opposition to an image in order to make the skin appear more uniform. This treatment can vary by spatial frequency, and it is typically preferable to have uniformity in the mid-frequency. Low frequency corrections may be more AI or 25 artistic based for correction over larger areas of the skin. In an embodiment, a low-pass filter may be performed with a desired range of wavelengths. In one example, one half inch to one inch wavelengths are filtered to remove a portion of the middle frequencies. As shown in FIG. 28, weaker middle frequencies 390 show less pronounced swings between light and dark points than stronger 30 middle frequencies 392. In an embodiment, the weak middle frequency components are removed to smooth an image. Performing a derivation of the low-pass filter In an embodiment, a low-pass filter may be performed, such as where a color value for a frexel is replaced by the average color value of its neighbors. 35 19 WO 2007/021972 PCT/US2006/031441 Artificial Intelligence Means Artificial intelligence techniques include expert systems for detecting particular skin features, and selection of correction strategies. In one embodiment, the skin features are correlated to a registry or map, to identify feature locations. The registry allows for improving faded or distorted body painting and tattooing. 5 Features Library Another aspect of AI techniques is the use of a features library for feature identification, and for comparison of actual features with idealized features. Artistic Means Computer Controls 10 In an embodiment, a human observer may optionally use means, such as a computer screen, keyboard, and mouse, to make further modifications in the perceived depth of the scanned area in order to accomplish aesthetic enhancements. A makeup artist or the customer may interact with the computer screen through controls to experiment with enhancements before the application. 15 A "cosmetic markup language" to provide general instructions such as to darken the top surface of bumps to the left of the nose; or to lighten varicose veins may be employed. The cosmetic markup language simplifies the correction process. Working with Multispectral Bands In an embodiment, effective techniques may be employed to enhance the patterns identified in multispectral 20 bands, such as RGB bands. For example, as explained above FIG. 25 shows RGB patterns for a scar 460, an age-related freckle 468, a varicose vein 476, a bluish bruise 484, and a yellow bruise 492, all in young skin. The following techniques are effective when the skin as a whole is being darkened.in middle frequencies to smooth it, as explained above. Scar 25 To enhance the scar 460, RMAs of magenta and yellow but not much cyan may be applied to it. This adds red color to the pale-looking scar 460. Varicose Vein To enhance the varicose vein 476, less of the darkening RMAs may be added to the areas surrounding the varicose vein 476. 30 Age-Related Freckle To enhance an age-related freckle 468, less of the darkening RMAs may be added to the area of the freckle 468. Bluish Bruise 20 WO 2007/021972 PCT/US2006/031441 To enhance a bluish bruise, less cyan RMA can be added during the general darkening. Yellow Bruise To enhance a yellow bruise, less yellow RMA can be added during the general darkening. 5 Applying at Least One Reflectance Modifying Agent Types of Reflectance Modifying Agents (RMAs) The current invention may utilize a variety of Reflectance Modifying Agents (RMAs), including * Analine, 10 * Food coloring, * UV, * Transparent Dyes, * Transparent Inks, * Pigments, 15 * Oxidizers, * Tanning agents, * Bleaches, and * Chemically altering agent. For example, a dye does not reflect light, but changes the skin reflectance, acting primarily by absorbing light. 20 In an embodiment, the RMAs can have a time delay, so that their application does not have an immediate effect but takes effect later through a triggering agent. For example, the RMAs can comprise one or more photosensitive materials that can be selectively exposed by a modulated beam of ultraviolet or other light or other forms of light and later developed by a chemical agent applied uniformly over the surface. For example a 25 photographic emulsion of a light based material may be used, of which silver based halides are a good example. Multiple Passes In an embodiment, the RMAs may be applied to the skin by scanning and printing almost at the same time and making multiple passes over the skin. Several advantages result from using multiple pass application. Micro 30 registration problems may be reduced because multiple passes permit dithering or blurring the image, as is well known to those skilled in the art. For example, multiple pass applications are useful for smoothing out the effects of hairs on the skin. Also, multiple pass applications of RMAs allow time for the skin to assimilate the RMAs, which is especially 35 important because some types of skin will absorb more than others. The process for multiple pass applications is to make a partial application of the RMAs, then to scan again the area of skin that has received the partial application. A further application of RMAs can be made, and still further multiple pass scanning and applications can be made to approach an aesthetic goal. 21 WO 2007/021972 PCT/US2006/031441 drop control application techniques Substances may be applied with "flow control" devices. These flow control devices typically may be 5 characterized as "drop control techniques" where individual droplets of the substance are controlled; or "non drop control techniques". Spray devices and electrostatic spray devices are non-drop control techniques where droplets are produced and controlled only in aggregate. Often in a spray device, a lack of individual droplet control, or "randomness" is 10 desired in order to produce a smooth application over a relatively large area. By contrast, in the current invention, it is desirable to provide very specific control of the amount and placement of medicinal agents. Examples of drop control include "fine flow control" where the flow of the substance is precisely controlled to deliver droplets as desired; and "inkjet technologies". An older inkjet technology includes supplying a 15 continuous flow of charged droplets past electrostatic deflector plates which are alternately charged so that the plates either permit a droplet to pass or deflect to a gutter. This technique was the original design basis for inkjet printers. Other inkjet technologies include "drop on demand" such as thermal devices provided by Hewlett Packard, and 20 piezoelectric devices such as provided by Epson and other printer manufacturers. In one embodiment of the current invention, the drop on demand technology is combined with charging the droplets. Another embodiment of the current invention is the use of the older inkjet technology in a manner that delivers charged droplets in a scanned direction beam. Modern inkjet printers have been optimized for printing on flat 25 surfaces over limited distances. The current invention prints on skin which is a dimensioned surface, and often requires a greater throw distance for the droplets. This greater throw distance can be facilitated with the better droplet aiming that is possible with a charged droplet. In another embodiment of the current invention, a non-inkjet drop control technique is used, such as fine flow 30 control techniques. As mentioned above, in this patent specification, the phrase "inkjet printer" is used for brevity represent any form of inkjet technology. 35 A particularly useful technique for the present invention is to employ "drop on demand" technology, a subset of drop control technology, which charges the droplets electrostatically. One of the advantages of applying charged droplets are that the applicator may be placed farther from the skin than is otherwise possible, whiled maintaining accuracy. Another advantage is that because hair is not grounded, the hair may be charged to the same level as the charged droplets so that the hair does not interfere with applications to the skin. For example, 22 WO 2007/021972 PCT/US2006/031441 drop on technology may be used to apply a single droplet of white pigment to spot in the face with pixel-level precision. In an embodiment, an inkjet printer may be used to apply the RMAs or medicinal agents onto the surface of 5 skin, printing at a resolution of 300 dpi (11.8 dpmm). In an embodiment, the inkjet printer may have multiple printer heads to speed the application. It may also traverse the body by robotics. 10 It is desirable to control the application of RMAs or medicinal agents to a desired spray range. In one example, an inkjet printer has a desired spray distance of about 1/8 inch (3.2 mm). Various techniques may be used to guide the printer element over the surface of the skin in order to maintain that desired spray distance, such as a cup, as shown in FIG. 26. 15 In an embodiment, the head of the inkjet printer has a comb to keep hairs on the skin even and in fixed pattern, to smooth the hairs. DETAILED DESCRIPTION OF EMBODIMENT - MAPPING-BASED MONITORING 20 Example - Generating a Map of the Sldkin This example demonstrates one method for generating a map of the skin, analyzing the map to generate a corrective plan, and executing the corrective plan. In this example, a corrective plan typically refers to one or more application of at least one medicinal agent. In other examples, the plan may include a closer monitoring of 25 a region of skin, without the application of a medicinal agent. Step 1- scan skin and generate a map of the skin In this example, the map of the skin is generated from data collected by scanning the skin at a first time. In this example, the general process of creating a map of the skin involves obtaining data by scanning the 30 frexels, and then processing that data to create the map. In this example, the processing includes determining the location of the scanning device and the frexel with respect to a reference coordinate system, determining the reflective properties of the frexel in multiple wavelengths, and determining the tilt or orientation of the frexel with respect to the coordinate system. Information about the frexel and its neighbors is then processed to make fine adjustments to the location of the frexel with respect to a portion of the body such as a face, so that a map 35 can be generated. This fine adjustment includes referencing the frexel to the face, such as by referencing the frexel relative to recognized facial features. a. Data representation An example of the data representation for a frexel is shown below: 23 WO 2007/021972 PCT/US2006/031441 [(x Y, Z , ., 3 ., ), (xf, yf, Zf,, Pf, yf ), {(refl)A, (refl)N, (refl)s, (refl)E, (refl)w}] 5 In this example, (x,,y,,z,,a,P,y,) and (xr, ye, zf, a, Pr, yfr) represent the position and angular orientation of the scanner sensor and the frexel relative to a coordinate system. b. Frexel location relative to sensor or coordinate system The data elements (x y, yZ f, f, 7r) may represent the distance of the frexel from the sensor, or may be an absolute position and orientation of the frexel with respect to a reference coordinate system. In one example, the 10 determination of the distance from the frexel to the scanner may be made in two steps. A first step can be an approximate mechanically-based measurement such as a constant height of the sensor from the skin. The second step can be an optical first derivative measurement to provide a fine adjustment. In one example, the fine adjustment is calculated by measuring an angle from the surface. In another embodiment, a fine adjustment may be made by using two light sources to send out two reference points or grids for detection by a sensor. 15 c. Reflectance and illumination data and calculations The data elements {(refl)A, (refl)N, (refl)s, (refl)E, (refl)w} represent reflective data for the frexel under ambient lighting conditions, and for each of four light sources, such as LEDs, which are arbitrarily designated as north south-east-west for ease of discussion. Other numbers of light sources, such as three sources, can be used, but the mathematics is simplified with four light sources. The (refl) represents one or more data point for the reflectance 20 measurement. The frexel data can be processed to determine a reflectance and an illuminance for each light source, and that information can be used to determine reflectance and surface profile. 25 In one example, the reflectance is the average or mean of all measurements. The illuminance can be determined from the known brightness of light sources such as LEDs. d Frexel orientation By determining the tilt of the frexel relative to two orthogonal axes, the orientation of the frexel can be determined. The orientation of a frexel and its neighbors is an indication of the actual local surface texture of the 30 skin. One aspect of the current invention is the ability to measure and compensate for both local reflective properties and local surface texture. In this example, there are four light sources which are designated as North, South, East, and West. The sensor obtains data when each light source is on, and the other sources are off. The sensor may also obtain data for 35 ambient lighting, with none of the four light sources on. The tilt of the frexel can be determined by comparing the North and South measurements. The difference between these measurements is a related to the tilt of the frexel along the North-South axis. The difference between the East and West measurements is a related to the tilt of the frexel along the East-West axis. 24 WO 2007/021972 PCT/US2006/031441 e. Data representation for derived values An idealized data representation for data from a frexel is shown below. Various compression methods can be used to reduce the data storage requirements. In this example, each data element is shown as a complete set in order to demonstrate methods of registering the data and creating a map. 5 frexel data [(x,y,z) NS tilt, EW tilt, (R, G, B visual color albedo), 10 time of acquisition The (x,y,z) represents the location of a frexel with respect to a coordinate system. The NS tilt represents the tilt of the frexel relative to the EW axis. The EW tilt represents the tilt of the frexel relative to the NS axis. 15 The (R,G, B visual color albedo) represents the measured reflectance of the frexel in the red, green, and blue spectrum. One aspect of the current invention is that data may be obtained for multiple wavelengths, and that different wavelength data is useful in identifying skin features. The human eye sees both reflectance and topology. In one embodiment of the present invention, data is 20 obtained for both reflectance and topology. Step 2- register the groups of frexels. The second step is to make some sense out of the data from a plurality of frexels. This portion of the example is analogous to the problem of mapping the earth's surface from satellite or aerial photographs. In the case of aerial photos, a large number of photographs are slightly scaled, rotated, and/or 25 translated in order to permit the images to be properly overlapped to reflect the actual earth surface. The map can then be generated from the properly overlapped images. In the present example, one source of complexity is that data is captured at slightly different acquisition times, and it is necessary to compensate for movements of the skin and slight errors in calculated position. 30 This motion aspect is analogous to modeling in a gaming application. In gaming, a model of the body may include a model of the skeleton so that the body may be related to the skeleton. Movement may first be applied to the skeleton, and then the position of the body can be calculated from knowing the position of the skeleton and knowing the relationship between the skeleton and the body. In the current invention, the problem is the 35 reverse, in that the shape of the body has been determined, and it is desirable to correct for motion during the measurement, a. Mapping afrexel to a map 25 WO 2007/021972 PCT/US2006/031441 In this example, it is desirable to associate a frexel, or a group of frexels, with a position on a map. For instance, the frexel may be a portion of a face, and the map is an idealized map of a face. In the case of a face, a model could be a rigid and upright face in an expressionless pose. 5 In one embodiment the determination of the desired amount of each of a plurality of dyes to be applied is made by * generating a map of the skin at a first time; and * analyzing the map to generate a corrective plan. 10 The corrective plan is then executed at a second time by making at least one pass with a device which includes a scanner and a dye applicator. The scanner provides data that is used to determine the location of the applicator and to determine how much additional medicinal agent is required for that location at each pass. 15 DETAILED DESCRIPTION OF EMBODIMENT - EXAMPLES OF METHOD To illustrate embodiments of the present invention, examples are given below for enhancement or monitoring processes for the following areas of human skin: * A face, * A leg, and 20 * A breast. Enhancing a Face Undesirable and Desirable Characteristics in a Face FIG. 13 represents a human face 235 with certain characteristics: 25 * A light spot 408, * A freckle 410, and * A non-uniformity 412 such as a scar. FIG. 14 shows a representation of a 2-D surface map 232 of the face shown in FIG. 13, resulting from the 30 scanning process used by the present invention and described above. This 2-D surface map 232 in FIG. 14 retains the characteristics listed above, which may be identified by pattern recognition: * A light spot 408, * A freckle 410, and * A non-uniformity 412 such as a scar. 35 Note that the 2-D surface map 232 typically includes a representation of depth in order to capture the shape of the face. To enhance such a face 235, shown in FIG. 13, according to the principles of attractiveness given above, it may 40 be desirable to reduce or delete from view the undesirable characteristics, such as the light spot 408 and the non 26 WO 2007/021972 PCT/US2006/031441 uniformity 412. At the same time, it may also be desirable to retain or even augment the appearance of a characteristic such as a freckle 410, which can be a characteristic of youthful-looking skin. Unlike prior cosmetic techniques, which tend to cover over both undesirable and desirable features with makeup, the present invention can distinguish between the two and treat them appropriately. 5 Putting the Scanned Image into Spatial Frequency Bands As shown in Step 606 of FIG. 31 and described below, the application algorithm 230 puts the scanned image into spatial frequency bands, in an embodiment, to permit identification of characteristics. 10 FIGS. 16A-E represent the patterns of the 2-D face 232, shown in FIG. 14, after the data has been put into spatial frequency bands. Albedo The top band in FIG. 16 represents the actual "albedo" of the 2-D surface map 232. A rise in the actual albedo graph identifies the light spot 408. A deep, sharp drop in the graph identifies a non-uniformity 412 such as a 15 scar. And an irregular section identifies a freckle 410. Illuminance (Shading) The spatial frequency bands also graph the actual illuminance (shading) of the 2-D surface map 232. Feature Recognition FIG. 15 shows that pattern recognition can also identify features in the scanned 2-D surface map 232, such as 20 * Hair 422, * An eyebrow 424, * An eye 426, * A cheekbone 428, * The nose 430, 25 * The mouth 432, and * The chin 434. By identifying such features, the application algorithm 230 can determine whether to make enhancement or medicinal applicaiton to those features. For example, it is normally undesirable to print on an eye 426. 30 Therefore, the application algorithm 230 can remove the area that represents the eye 426 from consideration. Tracking The application algorithm 230 may also use pattern recognition for tracking the location of the application device 246, for example the one shown in FIG. 3, on the area of skin 302. 35 As mentioned above, chemical markers may be alternately applied to the area of skin during the scan to help create the map and enable subsequent tracking of the map with the area of skin 302. For example, ultraviolet markers may be used. 27 WO 2007/021972 PCT/US2006/031441 Comparing Features with Idealized Features The application algorithm 230 may compare the mapped physical features with the idealized features in a features library 274, shown in FIG. 2, and use the comparison to modify features. In the current invention, this comparison may be a scanned skin feature compared to images from a medical database to assist in diagnosis. 5 Thus the application algorithm 230 may apply to scanned features global guidelines established in the features library 274, shown in FIG. 2. Determining the Actual Depth 10 Scanning the area of skin 302 provides the actual depth. Determining the Aim Depth In an embodiment, the aim depth can be the low spatial frequencies only of the actual depth. However, aesthetics may dictate additional sculpting, through further mathematical or manual input. The aim depth 15 encompasses the effect of illuminance on perceived depth or texture, and is related to the amount and angle of incident light. Carrying Out a Low-Pass Filter In an embodiment, a low-pass filter may be performed with one half inch to one inch (12.7 to 25.4 mm) wavelengths to determine the aim depth to accomplish smoothing. 20 Determining the Actual Illuminance Both actual and aim depths are translated into surface angle, as the first derivative, or slope, of depth. The surface angle is then translated into illuminance of the surface, as is well understood in 3-D modeling in applications such as gaming or animation graphics. Typically the assumed illumination angle and diffusion is 25 mean light reaching the human skin. Determining the Aim Illuminance An aim reflectance may be derived algorithmically again simply as the low-pass version of the actual reflectance. However, additional aesthetic attributes may be added through mathematic or manual input. 30 Determining the Actual Albedo The actual albedo is determined by the sensor of the application device, as described above. Determining the Aim Albedo 35 The aim albedo is determined by the principles of correction explained above. In this example, a generalized smoothing is performed, and specific feature correction is performed. For example, the light spot would be darkened, the freckle would be retained and possibly sharpened, and the scar 28 WO 2007/021972 PCT/US2006/031441 would be at least partially camouflaged by a general darkening of the skin and a specific darkening of the light area on top of the scar. The aim albedo is the desired perceived reflectance after calculating the smoothing and feature correction. 5 In other examples, the aim albedo may also include artistic strategies such as darkening one portion of a face relative to another. Applving Aesthetic Objectives 10 In an embodiment, a human observer or medical technician may optionally use means, such as a computer screen, keyboard, and mouse, to make further modifications in the actual depth of the scanned area in order to accomplish aesthetic enhancements. The Enhanced Appearance of the Face 15 FIG. 17 shows an example of the enhancement through the application of RMAs of the appearance of the face 235 portrayed in FIG. 13. The light spot 408 and non-uniformity 412 shown in FIG. 13 have been removed in FIG. 17. However, the freckle 410 has been retained in FIG. 17 as an attractive pattern of variety. Single-Pass or Multiple-Pass Systems 20 Single-Pass With sufficient computing power, the application device 246 will only need to make only one pass across the area of skin 302 to both scan the data and apply the RMAs 264. Multiple-Pass In an embodiment, the user moves the application device 246 over the area of skin 302 many times. The 25 application system then scans continually, creates a new 2-D surface map 233 after each pass, uses the 2-D surface maps 233 continually to identify the landscape of the area of skin 302 and calculate new printable enhancement images 234 with each pass, and applies only a portion of the RMAs 264 or medicinal agent, for example 10%-20% of the RMAs 264, on each pass. The use of multiple passes thus lets the application system 200 make a partial application of the RMAs 264, view the results of that application, and then make another 30 partial application for further improvements. The continuation of these passes can ensure increased accuracy in the desired result. Application of the RMAs 264 in multiple passes also reduces the possibility of streaking and allows the RMAs 264 to dry between applications for greater effectiveness. FIG. 21 illustrates how multiple passes may be used to apply a printable enhancement image 234 (exact aim) to 35 an unprinted surface 366. Overlap areas 29 WO 2007/021972 PCT/US2006/031441 In some examples of the current invention, it is desirable to make multiple passes of the applicator over an area. In the general case, as the applicator crosses over an area in a subsequent pass, some frexels will be seen for the first time, other frexels will have had a previous first pass, and still other frexels will have had two previous passes, etc. It is desirable to keep track of how many times each frexel has been passed over, so that this 5 information can be included in the control algorithm for applying a desired amount of RMA. It may be desired to correct, by example, 50% of the aim depositions of RMAs or medicinal agent on a first pass. In the observation phase of the second pass, it may be noted that the application has produced more or less than 50% of the desired correction. Suppose this was seen to be 60%; so, only 40% remains uncorrected, and in 10 addition it is now known that this part of the skin is responding with 6/5 stronger response to the RMA. So, by calculation only 5/6 x 4/5 = 2/3 of the RMA would be needed on the second pass to attain the desired effect. Suppose instead the algorithm chooses to deposit less than this on the second pass, then on a third pass makes a final observation and final calculation of efficiency and final deposition, to precisely titrate to the desired effect by feedback. 15 It is possible that the multiple passed could be in sequential scan order; so a top side of the probe always sees fresh skin, a middle processes an intermediate pass, and a bottom processes a strip of skin for the final pass. A more practical system allows random movement similar to the motion of an electric shaver, in which case software tracks the number of times a frexel of skin has been operated on. A sonic or tactile feedback could 20 indicate the completion for each frexel, like an electric shave changes sound depending on completion of effect under each pass. Since it is generally impractical to exactly meet an edge from a previous application pass, it is also generally desirable that the extreme portions of the applicator make a weaker application of RMA than in the middle of 25 the pass. For instance, if the applicator were moved left to right on this page, then a lesser amount of RMA than calculated would be applied by the top and bottom portions of the print head so that there was an opportunity on a subsequent pass to print additional RMA in those areas to provide a better overlap of passes. It is also desirable to make each pass in a different orientation relative to the skin to randomize measurement or deposition fluctuations due to hairs, skin texture, or pulling distortions of skin, and not to repeat the same paths. 30 For instance, if a first pass were made left to right, a second pass might be tilted slightly clockwise, and a third pass tilted slightly counterclockwise. Summary of Enhancement Process 35 FIG. 46 shows the general process of one embodiment of the present invention to visually enhance objects such as an area of skin comprising a human face, in an embodiment. "What the eye sees 380" represent the scanned data about the area of skin 302. In terms of optics, this data comprises * the albedo G1 - which is the degree of reflectance from the surface of the area of skin 302; * the illuminance G3 - which is the degree of illumination G3 of the area of skin 302; and 30 WO 2007/021972 PCT/US2006/031441 * the depth G5 - which is the distance from the scanner or other reference point to the portion of skin being measured * the "tilt"or orientation of the portion of skin being measured. This orientation, when combined with information from adjoining skin areas, describes a surface profile of the skin. 5 "What you want to see 382" represents an enhancement that would make more attractive "what the eye sees 380." This enhancement, which may be calculated mathematically and optionally through manual visual corrections, comprises * an aim albedo G2 - which is a more attractive degree of reflectance from the surface of the area of skin 10 302; * an aim illuminance G4 - which is a more attractive degree of illuminance G3 of the area of skin 302; and * an aim depth G6 - which is the desired perceived distance from the scanner or other reference point to the portion of skin being measured Note: In one embodiment, the correction to be applied is a mixture 15 of transparent dyes, such that the mix and the amount of the dye is determined in response to the perceived reflectance of the local area of the skin-which is related both to the actual reflectance and to the skin surface profile. Thus the correction applies a desired RMA to compensate for actual reflectance, and applies a shading to hide or enhance surface features. 20 In an embodiment, the mathematical calculations to create the aim albedo G2, aim illuminance G4, and aim depth G6 may be performed with particular effectiveness through mid frequency filtering. By calculating "what you want to see 382" according to the principles of attractiveness given above, a printable enhancement image 234 may be created for printing on the area of skin 302 to make that area of skin 302 more 25 attractive. Steps in the Enhancement Process FIG. 33 shows steps in a process for accomplishing the present invention's enhancement techniques in an embodiment: * Step 7100 of FIG. 33 - Using pattern recognition to map the physical features of the scanned area; 30 * Step 7200 of FIG. 33 - Determining the actual depth of the scanned area. * Step 7300 of FIG. 33 - Determining the aim depth of the scanned area. * Step 7400 of FIG. 33 - Determining the actual illuminance of the scanned area. * Step 7500 of FIG. 33 - Determining the aim illuminance of the scanned area. * Step 7600 of FIG. 33 - Determining the actual albedo of the scanned area. 35 * Step 7700 of FIG. 33 - Determining the aim albedo of the scanned area. FIG. 34 shows steps in a process for accomplishing step 7300 of FIG. 33. * Step 7310 of FIG. 34 - Carry out a low-pass filter. * Step 7320 of FIG. 34 - Compare the mapped physical features with the idealized features in the features library 274. 31 WO 2007/021972 PCT/US2006/031441 * Step 7330 of FIG. 34 - Use pattern recognition to retain desired characterisitics. * Step 7340 of FIG. 34 - Apply aesthetic objectives. FIG. 35 shows steps in a process for accomplishing step 7500 of FIG. 33. * Step 7510 of FIG. 35 - Perform a derivation of the low-pass filter. 5 * Step 7520 of FIG. 35 - Apply aesthetic objectives. Enhancing a Leg Undesirable and Desirable Characteristics in a 2-D Leg Fig 18 shows an illustration of a human leg 237 with the following undesirable and desirable characteristics: 10 * Cellulite 414, * Natural color differences 416, * Varicose veins 418, and * Age spots 420. 15 The spectral bands for these characteristics are also shown, including one for a printable enhancement image 234 that may be used to print enhancements onto the leg 237. To simplify the illustration, a 2-D skin map is portrayed as a 1-D graph following the dotted line across the surface of the skin. The actual depth along this line is graphed. In addition, one obtains an aim depth. The aim depth can be the 20 low spatial frequencies only of the actual depth. However, aesthetics often dictate additional sculpting, as is known in cosmetology. Both actual and aim depths are translated into surface angle, as the first derivative, or slope, of depth. The surface angle is then translated into illuminance of the surface, as is well understood in 3-D modeling in 25 applications such as gaming or animation graphics. Typically the assumed illumination angle and diffusion is mean light reaching the human skin. Printing on the skin has negligible effect on surface depth. However, the visual illusion of depth is obtained by printing the shadowing. Cellulite is not actually perceived stereoptically at more distance than approximately 30 six inches. The human eye perceives cellulite primarily by shadowing. Note how tanning produces pigmentation in opposition to mean illumination reaching the skin, and thus is in opposition to mean shading, thus making a sun-tanned human body appear smoother and more attractive. Note that rub-on tanning solutions do not have this characteristic of being sensitive to skin angle relative to light, and 35 thus fail to provide the same attractiveness. The leg example also illustrates pigmentations and varicose veins. An aim reflectance may be derived algorithmically again simply as the low-pass version of the actual reflectance; however, aesthetic attributes may be added, such as freckles, which may align with existing pigmentations, while excluding age spots. It may also 40 include other selected features, such as knee cap darkening. 32 WO 2007/021972 PCT/US2006/031441 It should be understood that the aim and actual reflectance curves can represent each color separately. For example, varicose veins may be blue or red, while pigmentation may be orange. Thus each color is independently corrected using colored inks, such as the process colors cyan, magenta, and yellow. 5 The perceived light visualized from the leg by a human observer is the illuminance * reflectance (albedo).. It is actually actual illuminance * actual reflectance, but is desired to be aim illuminance * aim reflectance. Thus to go from actual to aim, a multiplying (or dye) image should be deposited on the skin, that is 10 translated aim angle * aim reflectance translated actual angle actual reflectance where "translated aim angle" is the aim angle translated to a standard illumination assuming mean illumination; and "translated actual angle" is the actual angle translated to a standard illumination assuming mean 15 illumination. This provides the aim correctance, shown as the printable enhancement image 234. A separate aim correctance can be derived for each color, typically red, green, and blue to print, in order, cyan, magenta, and yellow. A problem arises that with dyes it is only generally practical to darken the skin. (In other embodiments, it is 20 possible to use limited amounts of whitening dyes or bleaching agents to selectively lighten areas.) Thus, as an expedient the aim paint is shifted (dotted line) so that more of the skin is correctable. This is equivalent to choosing a lower aim reflectance, for a more tanned appearance. Some details, such as blue varicose veins on a leg, may still be outside the correction range even with the 25 reasonable offset. These details can be corrected by depositing small areas of light pigment, than printing over with dyes to provide the right color. Alternately the extreme points can be left uncorrected. The relative error of uncorrected points is still much less noticeable if the adjoining skin is darkened somewhat. FIG. 19 shows an illustration of a human leg 238 after being enhanced through the present invention. The 30 following undesirable characteristics, which were shown in FIG. 18, have been reduced from view: * Cellulite 414, * Varicose veins 418, and * Age spots 420. However, the desirable natural color differences 416, which serve to make the 3-D quality of the knee cap 35 visible, have been retained. Enhancing a Breast FIG. 20 shows an example for changing the perception of a breast 239 from an actual 3D surface 342 under mean illumination 340 to an aesthetic aim 344 by determining the difference 346. Applying RMAs to 40 approximate this difference will alter the perceived appearance of the breast. 33 WO 2007/021972 PCT/US2006/031441 Single Pass Smoothing Example FIG. 37 represents a simple smoothing example for skin. An area of skin, such as one on the arm, is broken into a plurality of frexels at step 900. At step 910, at least one optical attribute of the frexels is determined. The 5 optical attribute is represented as R. There is a look up table which provides a quantity of a reflectance modifying agent to apply for each range of visual characteristic. At step 930, the quantity of RMA to be applied is determined from this look up table. The desired quantity of RMA is applied at step 940, thereby changing the appearance of the area of skin. This single pass example does not require a mapping of the skin. 10 Multiple Pass Smoothing Example FIG. 38 represents a multiple pass smoothing of skin. In this figure, the desired reflectance Rd is approached with a series of applications of a reflective modifying agent. The actual initial reflectance is determined at step 900 as Ra, and that value provides a first quantity of RMA to be applied in a first pass which is Qi. The application of that first amount of RMA, Qi, changes the reflectance from Ra to Rj. At the second pass Ri is used 15 in the look up table to determine the second amount of RMA (Q2) to be applied. When that second amount is applied, the reflectance is changed to R 2 . On the third pass, R2 is used to determine a third amount of RMA. (Q3) The resulting reflectance R 3 approaches the desired reflectance. The number of passes is not limited to three but may be more or less than that number. 20 Facial Map Example FIG. 39 represents a facial map example. In this example, the skin on a face is allocated into a plurality of frexels as before. The optical attribute is measured in step 910 as before, except that the frexel location is determined and specified and recorded so that there is location data for individual frexels. The data for individual frexels includes sensor location, the location of the frexel and one or more optical attributes. The 25 optical attributes maybe used to determine the reflectance, position, and orientation of the frexel at step 920. Each frexel then has an initial characteristic such as an actual reflectance. The frexel also has a desired final characteristic, such as the desired reflectance, and an amount of RMA to be applied in one or more passes. The amount of RMA is determined at step 940. The desired reflectance is determined from an enhancement strategy such as smoothing of the skin, filtering to remove middle frequency characteristics, feature recognition and 30 feature enhancement, and general artistic schemes. The desired quantity of RMA is determined from the difference between the desired reflectance and the actual reflectance. LED arrangement FIG. 40A is a schematic for sensor and LED arrangement. In this example a sensor is located along the axes of 35 four LEDs which are designated as north, south, east, and west. FIG. 40B is a cross section showing that the LEDs are typically directed to a point on the skin below the sensor. Typically the LEDs and sensors are provided in a housing, and the housing may have reflective properties to provide more diffuse or indirect light to the frexel in some applications. In other applications it is desirable to 40 orientate the light directly at the frexel in order to determine the tilt of the frexel. 34 WO 2007/021972 PCT/US2006/031441 Depth maybe determined by shadow parallax grid projected by LEDs from different angles. In another embodiment, two cameras maybe used in a stereoscopy approach. 5 Feature recognition FIG. 41 shows a simple feature recognition approach. A frexel map for a particular frexel "min" and its neighboring frexels is represented. Data for each frexel typically includes the time, position, reflectance and orientation of the frexel. Information can be represented graphically as demonstrated in the reflectance feature portion of the diagram. At step 910, the skin is scanned to measure an optical attribute. At step 920, a visual 10 characteristic such as reflectance is determined from the scan. At step 921, a facial map is generated to provide the actual visual characteristics as perceived by a viewer. At step 922, the frexel data is reviewed to identify local features and the parameters for the particular subject. An example of parameters is the range of readings in that subject, which can be used in normalization or other data manipulation. At step 924, enhancement strategies are applied. At step 925, an enhancement map is provided. The enhancement map includes the 15 amount of RMA to apply to a particular frexel in order to change its visual characteristics. FIG. 42 illustrates a frexel map for a portion of a face. This figure shows characteristics such as a pimple, frexel, light spot and a scar. Each of those characteristics is shown in an enlarged position with multiple frexels in the diagram. These areas can be represented and detected mathematically from the known properties of the 20 various skin features, so that feature recognition can be preformed automatically with mathematical analyses. Artistic strategy FIG. 43 represents an example of simple artistic strategy. When a frexel map of a face is generated, various shading strategies or overall global strategies for appearance can be provided. 25 DETAILED DESCRIPTION OF EMBODIMENT- SYSTEMS Operating Environment for Cosmetics FIG. 1 shows an embodiment of the present invention used to apply RMAs 264 to an area of skin 302. A party sets up an application system 200 comprising the following elements, which are explained in more detail below: 30 * a computing environment 100 - for example a personal computer, server, or portable computing device; * a scanner 220 - which electronically scans data about attributes of an area of skin 302; and * a means of application 240 - for example a printer - which can be used to apply RMAs 264, such as ink. 35 The computing environment 200 further comprises * an application algorithm 230; * storage 250 - which may be may be non-volatile data storage; * an application map 232 - which is created by application algorithm 230 to provide instructions for 40 applications onto an area of skin 302; 35 WO 2007/021972 PCT/US2006/031441 Sa printable enhancement image 234 - which is the set of instructions for applications onto an area of skin 302. Loosely Coupled Systems 5 In embodiments, the elements of application system 200 may comprise discrete units and be connected through links 142 and 144, which may comprise internal connections. For example, FIG. 2 shows an embodiment of loosely connected elements for applications onto an area of skin 302. A scanner 220, printer 241, and computing environment 100 communicate over a network 130 and links 142, 144, and 146. The network 130 may comprise the Internet, a private LAN (Local Area Network), a wireless network, a TCP/IP (Transmission 10 Control Protocol/Internet Protocol) network, or other communications system, and may comprise multiple elements such as gateways, routers, and switches. The links 142, 144, and 146 are compatible with the technology used for network 130. A features library 274 may be used to store the characteristics of human features, for example the eye, nose, and 15 mouth, for use by pattern recognition. The features library 274 may also be used to store idealized pattern for human features that may be used to make actual features appear more attractive. For example, an idealized pattern for human lips may be used to make actual lips appear fuller as well as redder. For the application map 232 shown in FIG. 1, a 2-D surface map 233, shown in FIG. 2, is used The 2-D surface map typically includes a representation of depth in order to capture the shape of the face. 20 In addition, registration means 270, mechanical or electronic, are used for tracking the location of the scanner 220 and printer 241 relative to the area of skin 302 Combined Scanner and Printer Connected with Computer 25 FIG. 3 shows an embodiment where an application device 246 comprises a scanner 220 and an inkjet printer 242 to apply RMAs 264 from a reservoir 262 to the area of skin 302. The application device 246 also communicates over a network 130. Reflectance Modifying Agents FIG. 4 shows that in an embodiment, the RMAs 264 may comprise magenta 265, yellow 266 and cyan 267 30 RMAs. In other embodiments, the RMAs 264 may additionally include black or brown and white. Application Device The application device 246 comprises the portable scanner 220 and a portable inkjet printer 242, shown in FIG. 3. In this example, the device has a height-determination means such as a tip or cup to hold the device at uniform a height of 1/8 to inch (3.2 to 6.4 mm) from skin. The elevation of the probe only has to be accurate 35 within a few millimeters. The device uses mirrors or two cameras. It typically makes ten passes to cover the 150 square inches (1000 square cm) of a face, and the time required to complete the process is comparable to that required for electric shaving. The device is under 2 inches (50 mm) in length. 36 WO 2007/021972 PCT/US2006/031441 Portable Scanner In an embodiment, the portable scanner 220 comprises an area array that lightly touches the surface of the area of skin 302 to be scanned. In another embodiment the portable scanner is moved without touching skin in the vicinity of the skin being scanned. During scanning, a white LED light source in the sensor flashes to apply 5 normal light, defined as light from above, to the area of skin 302. Measurements are taken when the LED is on and off, and the difference between the two measurements is subtracted to determine the contribution of the light source. Inkjet Printer In an embodiment, the inkjet printer 242 comprises an inkjet printer with .001 inch resolution and a reservoir 10 262 capable of holding RMAs 264. In an embodiment the RMAs 264 comprise transparent dyes, while in other embodiments they comprise inks or other useful chemicals. In one embodiment, FDA-approved RMAs are employed. As shown in FIG. 4, the RMAs 264 may comprise agents for the following colors: magenta 265, yellow 266, and cyan 267. They may comprise additional colors, such as black, brown, and white, as well. These colors can enable the inkjet printer 242 to create any color on the area of human skin. 15 Registration Means As mentioned above, registration means 270, mechanical or electronic, are used for tracking the location of the scanner 220 and printer 241 relative to the area of skin 302. In an embodiment, the registration means 270 may comprise accelerometers, which measure acceleration and tilt, and gimbals, which measure the rotation of an object in three dimensions and control that rotation, may also be included in the application device 246. These 20 devices help control movement and positioning and maintain the correct reflective angle for the application device 246. In another embodiment, registration means may comprise a global positioning-like service (GPS) used locally through high frequency radiation. 25 In still another embodiment, registration means may comprise a set of small flat-ended pins that are pressed lightly against the surface of the skin to make an impression. For example the pins may be pressed against a face to make a mask of the face. The movement of the pins in a frame may be tracked mechanically to provide the 3-D coordinates. 30 Portable Application Device As shown in FIG. 5, another embodiment of the present invention is a portable application device 260 comprising multiple elements for applying material onto skin, which does not require an external network. An embodiment of the portable application device 260 uses an inkjet printer 242 to apply ink 248 to the area of skin 35 302. Portable Application Device with Curved Surface One aspect of the current invention is to acquire and manipulate image data of human skin. In one embodiment, a first step is used to generate a map of a portion of the body, and that map is used to generate a specific plan of 37 WO 2007/021972 PCT/US2006/031441 selectively applying dyes at a later time. One embodiment of the current invention is to use a portable scanning device to acquire data for generating the map; and to use the portable scanning device in combination with a portable printing device to selectively apply dyes to a region of skin. FIG. 30 shows an embodiment of the present invention that may be employed for applying material onto skin, through communications over a 5 network and a portable application device with a curved surface. Mask or Helmhnet The curved surface may comprise, for example, a mask or helmet into which a human face may be inserted and an application device (scanner/printer) that circles the face. Use of such a curve surface requires feature recognition through artificial intelligence and mapping, so that the application device can calculate its location 10 on the face and its distance from the skin. One advantage of the curved surface device is that is requires no user action or training. Another is that the application device remains above the skin and so does not touch the wet RMAs. Booth 15 Another embodiment of the current invention is to use a booth or work station to scan a region of skin, such as a face or an entire body. FIG. 27 shows an embodiment of an application device 246 comprising a booth. In this case, as shown in FIG. 36, the area of skin 302 comprises an entire person who steps into the application device 246 through a door 20 282. The person might undress, step into the booth, as is typically done with tanning booths, and lie or stand for the application of RMAs. A scanner/applicator 284, comprising a scanner 220, inkjet printer 242 with a reservoir 262 and RMAs 264, and registration means 270, would move across the person's body to collect data, analyze the data, and make enhancements by applying RMAs. 25 In another embodiment, the booth may comprise a two-part cylinder that closes over all a person or over part of a person such as a face. Blotter FIG. 29 shows an embodiment with an application device 246 comprising a blotter. The blotter comprises a cup 280 to maintain an approximate appropriated distance from the area of skin 302, as explained above. 30 Instead of moving the blotter application device 246 in a single pass or multiple passes over the enter area of skin 302, the user places the blotter application device 246 over a small area of skin and holds it there briefly, to accomplish scanning, analysis, an application of RMAs in that small area, and then moves the blotter application device 246 to the next small area. 35 For the blotter application device 246, mechanical means would move the printer 242 over the area of skin 302 for the application of the RMAs 264. For example, FIG. 44 shows an inverted view of an application device 246 comprising a blotter. In an embodiment, the blotter application device 246 comprises four LEDs 290, two cameras 292, and a rotating inkjet printer 294 that moves about a central axis on the application device 246 like 38 WO 2007/021972 PCT/US2006/031441 the hand of a clock. The rotating inkjet printer 294 prints RMAs throughout the area of the blotter application device 246 except for the area of the central axis, which can be printed on by moving the blotter to an overlapping area for a second printing. 5 Light Sources FIGS. 40A-B are sample layouts for LEDs and a sensors for acquiring reflectance and skin orientation data. In one embodiment, a set of four light sources is used, such that the light sources are placed at the corners of a diamond, where the sensor is positioned at the center of the diamond layout. This configuration simplifies the 10 mathematical analysis for calculating surface profile. In an embodiment, it is useful to employ mean illumination. For this, multiple diffuse or orthogonal light sources may be used, in a configuration which may include mirrors. The lights may be flashed repeated, as strobe lights, so that hundreds of images may be taken of a small area can averaged for effectiveness. 15 Process for Employing an Application System for Cosmetics FIG. 6 shows a process for employing an application system 200, in an embodiment. This process comprises the following high-level steps, which will be explained in detail below: * Step 1000 in FIG. 6 -Setting up an application system 200 based on scanning an area of skin 302 to 20 determine attributes and applying RMAs 264 to that area of skin 302 in registers in agreement with or in opposition to the determined attributes; * Step 2000 in FIG. 6 - Scanning an area of skin 302; * Step 3000 in FIG. 6 - Analyzing the scanned data with an application algorithm 230; * Step 4000 in FIG. 6 - Creating a printable enhancement image 234 with the application algorithm 230; 25 * Step 5000 in FIG. 6 - Using the printable enhancement image 234 to apply RMAs 264 on the area of skin 302; and * Step 5002 in FIG. 6 - Optionally repeating steps 2000 through 5000. Setting Up an Application System 30 FIG. 7 shows a process for Step 1000 - setting up an application system 200, shown in FIG. 6, in an embodiment. The process comprises the following steps, which will be explained below: * Step 1010 in FIG. 7- Providing an application algorithm 230; * Step 1020 in FIG. 7 - Providing the application algorithm 230 on a computing environment 100; * Step 1030 in FIG. 7 - Providing storage 250 on the computing environment; 35 * Step 1040 in FIG. 7 - Integrating a means of scanning 220 an area of skin 302; and * Step 1050 in FIG. 7 - Integrating a means of application 240 of RMAs 264. Providing an Application Algorithm One or more programmers create an application algorithm 230 that, in an embodiment, controls the elements 40 and processes of the present invention outlined in FIG. 6 and explained above. After the application algorithm 39 WO 2007/021972 PCT/US2006/031441 230 has been created, it can be used on at least one computing environment 100, as shown in FIG. 1, and may be integrated with other elements of application system 200. For example, in an embodiment application algorithm 230 may be loaded on a computing environment 100 comprising a server. The computing environment 100 may be equipped with non-volatile storage 250 capable of storing data such as scanned data 5 from scanner 220. In various embodiments, the application algorithm can include default strategies which may be based on feature recognition, a feature-based lookup scheme, or general artistic objectives. 10 As shown in FIG. 8, in an embodiment the general functions to be accomplished by the application algorithm 230 are * Coordinating pixel-level mapping of skin by scanning; * providing feature recognition, or accepting manual selection of image enhancement strategies * Creating a printable enhancement image 234; and 15 * Coordinating the pixel-level application of substances to achieve the determined enhancements. Coordinating Pixel-Level Mapping of Skin by Scanning A primary function of the application algorithm 230 is to analyze scanned data about an area of a first instance of material 300 and create a 3-D application map 232 of the attributes of that area 300 for which application of a second instance of material 300 would be useful. A key part of this function is that the application algorithm 20 230 determines at each scanned point whether the application of the second instance of material should be in a register in agreement with the attributes of that area of the first instance of material or in a register in opposition to those attributes. This decision is based on instructions in the algorithm for what would be useful and advantageous for the area of the first instance of material 300. 25 FIG. 31 shows the steps involved in coordinating scanning: * Step 602 in FIG. 31 - Initiating scanning by scanner 220. When the application device 246 is turned on and moved over an area of skin 302, the scanner 220 begins scanning. * Step 604 in FIG. 31 - Sending scanned data to computing environment 100. 30 The application device 246 transmits its scanned data over link 144, network 130, and link 142 to computing environment 100. * Step 606 in FIG. 31 - Storing scanned data in storage 250. * Step 608 in FIG. 31 - Putting the scanned data into spatial frequency bands. Creating a Printable Enhancement Image 35 The goal of a cosmetics embodiment of the present invention is to understand and make use of the characteristics of the human visual system to make the observer perceive a person as younger than that person is. This may be considered a form of camouflage performed at the pixel level. It is important to note that the techniques of the present invention for accomplishing this goal do not wipe out all the detail in the area of skin 40 WO 2007/021972 PCT/US2006/031441 affected, but retain significant, desirable details that make the area of skin look real. To accomplish this goal, the present invention uses sophisticated techniques, explained below, to create a printable enhancement image 234 for making appropriate applications of the RMAs 264. 5 FIG. 9 shows a process for creating a printable enhancement image 234, in an embodiment. Step 6051 in FIG. 9 - Converting the 3-D scan to a 2-D surface map 233 in a computer model. FIG. 10 shows an example of how a 3-D human face 235 may be mapped to a 2-D surface map of that face 233, through well known techniques employs in computer modeling and gaming. For this 2-D mapping, in the small 10 (limit) all surfaces are flat, creating a razor model for the "base." Step 6052 in FIG. 9 - Using pattern recognition to identify specified features 310 of the 2-D surface map 233. For example, pattern recognition may be used to identify the eyes. 15 Step 6054 in FIG. 9 - Using the identified features 310 to further identify portions of the skin not to be enhanced 320. For example, it may be desirable to specify that the eyes not be enhanced with potentially irritating RMAs. Step 6056 in FIG. 9 - Eliminate from calculations the portions that are not to be enhanced. 20 For example, the eyes may be eliminated from calculations. Step 6058 in FIG. 9 - Employing enhancement techniques 600 on correctable features 330 to achieve an aesthetic goal. The enhancement techniques employed by the present invention are explained in detail below. 25 Coordinating the Pixel-level Application ofReflectance Modifying Agents to Achieve the Determined Enhancements As shown in FIG. 32, coordinating the pixel-level application of RMAs 264 to achieve the determined enhancements may be achieved through the following steps: * Step 6060 in FIG. 32 - Sending the printable enhancement image 234 to the application device 246. 30 * Step 6070 in FIG. 32 - Using the inkjet printer 242 to apply the RMAs 264 to print the printable enhancement image 234. Operation of an Embodiment The operation of the present invention can be illustrated with reference to the application device 246 and 35 computing environment 100 shown in FIG. 3. 41 WO 2007/021972 PCT/US2006/031441 Scanning The user moves the application device 246 across the area of skin 302 so that the scanner 220 can record data. For example, the area of skin 302 might be the user's face. The scanner 220 sends the scanned data over the network 130 to the computing environment 100 where the data is stored in storage 250. 5 In an embodiment, the user may be asked to employ a tapping or blotting motion of the probe, rather and making smooth passes as in moving an electric shaver over the face. This motion reduces smudging in the application of RMAs. 10 In an embodiment the user may be asked assume a neutral, motionless position, to present a neutral model. For example, for use with the face, a user may be asked to hold still, close the eyes, and have an expressionless face. For use with the entire body, the user may be asked to stand still in a specified position in a booth. Analyzing the Scanned Data 15 The application algorithm 230 puts the stored data into spatial frequency bands and uses pattern recognition to analyze them to determine the landscape of the area of skin 302 and the dimensions that require application of the RMAs 264. The application algorithm 230 uses its analysis to create in software a 2-D surface map 233 of the area of skin 20 302, which is stored in storage 250, for potential future use. Creating a Printable Enhancement Image The application algorithm 230 also creates a printable enhancement image 234 based on a 2-D surface map 233. 25 Note that alternately the printable enhancement image 234 can be made manually by an operator who displays the map on a computer screen and uses controls to make desired adjustments. Printing the Enhancement The application algorithm 230 sends the printable enhancement image 234 over the network 130 to the 30 application device 246 that triggers the inkjet printer 242 to apply the RMAs 264 from the reservoir 262 to area of skin 302. The inkjet printer 242 applies different quantities and mixes of the RMAs 264 to create desired results in different portions of the area of skin 302, at the pixel level, making the application very precise. In various embodiments, the scanning and printing components can be provided in hand-held, fixtured, or booth 35 systems. Example of Hand-Held Operation In a hand-held system, the device may be the size of an electric shaver or powder puff so that it may be blotted or moved across the skin. The device may be used in a single pass mode to provide a general smoothing of skin 40 appearance, or in a multiple pass mode where several passes over each area of the skin are used in order to 42 WO 2007/021972 PCT/US2006/031441 provide a relatively small correction on each pass. The system may include a feedback means such as a tone to indicate that the operation is complete. Example of Hand-held Scanner that Touches Skin 5 FIG. 26 shows a handheld scanner. In this example, the scanner housing touches the skin so that the application device, such as a printer head, is maintained at an approximate known distance and orientation with respect to the skin. Example of Helmet Guide for Facial Modeling and Printing 10 A helmet mode is an example of a fixtured system where the scanning and application device has designated limited travel paths. The fixtured system may include coordinate reference points, guide strips, and a movable probe. Application Example- Facial Makeup 15 For example, a user could move the application device 246 over his or her face and have RMAs 264 applied as a form of makeup to enhance the attractiveness of the face. These RMAs may comprise transparent dyes, or inks, or pigments that would even up the skin tone while retaining desirable details like beauty moles, add reddish color to cheeks, and hide flaws and scars in the skin, greatly enhancing the attractiveness of the skin to the human eye. Typically, in an embodiment the user would close his or her eyes and mouth to prevent exposure of 20 them to the RMAs 264. In another embodiment the system would use feature identification to recognize sensitive areas such as eyes and restrict itself from applying RMAs 264 to those sensitive areas. Touchups Once a 2-D surface map 233 and a printable enhancement image 234 for that face has been stored, they can be 25 used repeatedly to quickly apply the RMAs 264 to the face with the application device 246, for example for quick daily cosmetic touchups. Note that the printable enhancement image 234 may be both in a register in agreement with the attributes of areas of the face or in a register in opposition to those attributes. For example, a light area of skin may be left 30 relatively light or may be darkened, depending on the desired effect calculated by the application algorithm 230. Camouflaging a Bump through Simulated Tanning The present invention can identify a very small area with surface texture variations 400, shown in FIG. 12, 35 representing a tiny bump, for example. It can apply an ink or dye to the apparently lighter portion 404 of the bump 400, apparently lighter because it is receiving more illumination by virtue of the surface angle relative to the light source, and not darken the shaded, apparently darker portion 406 underneath the bump. This reduces the light and dark contrast associated with the dimensionality of the bump, making the skin look smoother. 40 43 WO 2007/021972 PCT/US2006/031441 DETAILED DESCRIPTION OF EMBODIMENT - Handheld Mark Applicator FIG. 47A is a side view of one embodiment of a handheld device for skin marks such as age spots, small scars, 5 and varicose vein. FIG. 47B is a front view of the device of FIG. 47A, and FIG. 47C is a top cross sectional view along section AA' of FIG. 47B. The mark applicator device 550 includes a housing 553 which provides an upper handle portion and a lower skin application portion. In this example, the device is about 1 2 by 2 inches (38 - 50 mm) wide and about 4-5 10 inches (100-127 mm) tall. In this example, an opening in bottom 554 of the housing is about V to inch square (12.7 -19.2 mm square). At least one light source is used. In this example, four light sources 551 are positioned in proximity to the 4 corners of a square tube. The light sources are typically white light LEDs, or combinations of LEDs such as red, 15 green, and blue to produce a white light, but the sources may also be of varying wavelengths to provide additional data for mark recognition. In some cases, a single light source may be used. The advantages of using separate wavelength light sources include greater sensitivity, better color accuracy, and higher resolution. In the booth and movable handheld embodiments described above, however, these advantages may not overcome the practical difficulties and time required to sequence four different lighting conditions for each set of frexels. 20 Most cameras are able to provide good color images from a white light source. In the current embodiment, however, the camera is not moved, and it is more practical to obtain an image from each of several colors of light sources, and from the white light produced when all of the light sources are on. Thus some examples of this applicator include light sources of different wavelengths, thus providing a better 25 white light and additional image data at a plurality of wavelengths in order to support more sophisticated feature recognition. In general, the light source or sources in this and other embodiments may be of a variety of wavelengths including visible light, infrared, and ultraviolet. The infrared wavelengths provide a better penetration of the 30 skin to support feature recognition. The lower portion of the tube preferably has a reflective surface such as a shiny or brushed aluminum or steel so that the light sources reflect from the housing walls and provide a uniform lighting to the exposed skin area. These reflective surfaces are analogous to an optical fiber. A camera 552 captures images of the exposed area 35 as described below. A print head 560 is moved across the opening in order to print a desired correction to the area, and to the mark in particular. Other components in the housing include a circuit board 562 and electronics; at least one RMA cartridge 564 and a battery 566. The term RMA is used here in the general sense and the cartridge or cartridges may contain pigments or other agents. 44 WO 2007/021972 PCT/US2006/031441 In operation, at step 7900 the device is placed over an area of skin which has a mark which the user desires to camouflage. The device is held in place for a predetermined period of time, or until the unit signals completion, such as with a status light or audible tone. The user then presses a switch on the housing (not shown) and the unit performs the following typical operations: 5 In response to the user pressing a switch on the housing at step 7910, the unit completes the following steps. At step 7920, the camera captures a first image at ambient light with the camera of the area of skin 10 exposed by the bottom opening. Even when the unit is pressed against the skin, some light travels through the skin and partially illuminates the area. At step 7930, the light sources are turned on. At step 7940, the camera captures a second image with the camera while the light sources are on. 15 At step 7950, the unit analyzes the images, which may include the following steps. Subtracting the first image from the second image at step 7952; identifying the mark at step 7954; and determining a desired modified reflectance for the mark and adjacent skin at step 7956. 20 At step 7960, determining a desired amount of RMA to print on the mark to achieve the desired modification. A generally opaque and white RMA would typically be used to camouflage the small marks of this embodiment. The substance would be similar to a classical makeup base, but typically lighter or more white that the base. In one example, the RMA is a pure white, or is white in one wavelength, such as a light pink. The RMA is preferably lighter than the skin so that small amounts 25 may be used over a mark in a manner than matches surrounding skin. At step 7970, printing the correction in one or more print head passes. One example print method includes printing a portion of the desired correction in a first pass at step 7972; taking an image of the area of skin after printing the first portion at step 7974; analyzing the image at step 7976; adjusting the 30 amount to be printed in the second pass according to the analysis of the image at step 7978; and printing at least a portion of the remaining correction amounts in a second pass at step 7979. Additional passes may be performed if desired. A "pass" in this example refers to the print head being moved over the skin area. All other components and the housing remain stationary. The second pass provides an opportunity to compare the predicted correction to the actual correction, and to compensate 35 for the difference. For instance if less correction is printed than desired, the unit may print more than the remaining calculated amount in a second pass; and if more correction is printed than desired, then the unit may print less than the remaining calculated amount in a second pass. DETAILED DESCRIPTION OF EMBODIMENT - specialized skin region applicator 45 WO 2007/021972 PCT/US2006/031441 In this embodiment, a unit is provided to print a specialized area of the skin such as lips, or around the eyes. The unit may be provided as a booth-type fixture, but is preferably portable, such as a handheld device. The device may include a portable support such as a chinrest to provide stability and alignment. 5 In an example embodiment for lips and surrounding skin areas, a device similar to the handheld mark applicator of the embodiment described above may be used. The unit typically has several differences to the mark applicator. In this example, the unit is typically larger than the mark applicator, and the opening, may be of a shape such as an ellipse which more closely matches the skin region. Since the skin region may have substantial curvature, the print head typically has a z-axis capability to be moved closer to the skin or further from the skin 10 as the head is moved over the region. The multiple light sources as described in the above embodiment are effective for providing a "shading" analysis of frexel orientation over small areas. Since a region like the lips has larger shape features as well as local features, it is desirable to supplement the shading analysis with stereoscopy methods. For instance the use 15 of two camera permits a comparison of the images to develop a stereoscopic analysis of the region, as well as a local shading analysis. The two approaches are thus complimentary. In this example, the device is placed over the lips; or in the case of a booth device, the lips are placed in the booth. Images are taken by a pair of cameras with multiple lights sources under various lighting conditions. The 20 image data from one or both cameras can be used to determine frexel orientation as described above. The image data from both cameras can also be used to develop a stereoscopic analysis. The analysis is used to develop a correction plan. The correction plan is executed by moving the print head over the region to apply one or more RMA- preferably in multiple passes. In this example, the print head has a z-axis 25 control so that the head may be brought closer to the lips or further from the lips as necessary. 30 Alternate Embodiments Other Hardware and Software It will also be apparent to those skilled in the art that different embodiments of the present invention may employ a wide range of possible hardware and of software techniques. For example the communication 35 between a Web service provider and client business computers could take place through any number of links, including wired, wireless, infrared, or radio ones, and through other communication networks beside those cited, including any not yet in existence. Also, the term computer is used here in its broadest sense to include personal computers, laptops, telephones with computer capabilities, personal data assistants (PDAs) and servers, and it should be recognized 40 that it could include multiple servers, with storage and software functions divided among the servers. A wide 46 WO 2007/021972 PCT/US2006/031441 array of operating systems, compatible e-mail services, Web browsers and other communications systems can be used to transmit messages among client applications and Web services. 5 DETAILED DESCRIPTION OF EMBODIMENT - DETERMINING REFLECTANCE AND TOPOLOGY OF SKIN AREAS FOR MEDICAL MONITORING In this embodiment, the devices and techniques for image acquisition and analysis which were described above for the cosmetic examples are provided to provide one or more image for medical analysis. 10 These techniques, include providing multiple light sources to permit fine shading analysis to evaluate fine detail of skin morphology; providing two or more camera to permit stereoscopic analysis of a region of skin, providing both shading and stereoscopic analysis; developing a map of a region of skin, comparing a map or feature to a library to assist in the analysis; and comparing maps acquired at different times to determine and record changes over time. 15 These techniques may be provided in various housings including, but not limited to, a full body booth; a partial booth for a region of skin such as a face, arm, let, torso, or back; a handheld scanning device which is moved over the skin; or a smaller handheld device which is placed in a stationary manner over specific skin marks such as moles. 20 DETAILED DESCRIPTION OF EMBODIMENT - SELECTIVELY APPLYING MEDICINAL AGENTS In this embodiment, the devices and techniques for image acquisition and analysis further include at least one print head or other drop control device to apply one or more medicinal agent in response to the analysis. 25 These techniques may be provided in the full range of housings as described above. DETAILED DESCRIPTION OF EMBODIMENT - ADAPTING COSMETIC SYSTEM TO MEDICAL 30 MONITORING OR TO SELECTIVELY APPLY MEDICINAL AGENTS In one embodiment, the monitoring of a region of skin is provided in the normal course of routine usage of a cosmetic device as described above. 35 In another embodiment, the monitoring of a region of skin is provided supplemental to the normal course of routine usage of a cosmetic device as described above. 47 WO 2007/021972 PCT/US2006/031441 In another embodiment, cosmetic device as described above is adapted to deliver one or more medicinal agent in one or more passes over a skin region, and in one or more treatment sessions. One or more additional application heads and reservoirs may be provided for the medicinal agents. 5 This patent application employs for medical monitoring and the optional application of medicines and other compounds beneficial to health the system and method explained by the patent application cross referenced above. In accordance with the present invention, the data gathered by the cosmetics system and method of the cross 10 referenced application is repurposed for medical purposes. This process can provide frequent monitoring of all or portions of the external surface of the human body for the early detection and treatment of medical problems. For example, daily monitoring through the present invention to track all skin lesions in a precancerous range, so that such lesions can be treated early, might make untreatable skin cancers a thing of the past. To give another example, such daily monitoring may provide teenagers with early diagnoses of results of various behaviors, such 15 as their getting acne from eating unhealthy foods and from improper skin care, in a manner analogous to biofeedback In addition, the system of the cross-referenced application may be used for the application of medicinal agents. 20 Process In an embodiment, the present invention uses the following process, shown in FIG. 48. * Step 8000 in FIG. 48 - Collecting scanned data through a scanning device; * Step 8100 in FIG. 48 - Defining a base line; * Step 8200 in FIG. 48 - Identifying medical problems; 25 * Step 8300 in FIG. 48 - Reporting on the medical problems; and * Step 8400 in FIG. 48 - Optionally applying medications when appropriate. Step 8000 in FIG. 48 - Collecting Scanned Data through a Scanning Device Any scanning device, known to those skilled in the art, may be used. In an embodiment, the scanning device 30 may comprise the computer-controlled cosmetics system explained by the patent application cross referenced above, which may be a whole-body system. Whenever a person uses the cosmetics system, scanned data about all or part of the external surface of that person's body is collected and quantified. The cosmetics system puts the data into one or more spectral bands, where reflectance and surface topography data may be analyzed to identify medically relevant patterns and changes to previously identified characteristics. 35 Many people are highly motivated to use cosmetics daily or even multiple times a day. Because cosmetic use is much more frequent that the typical use of medical examinations, much more data can be collected about the person more frequently and more consistently than has typically been possible through prior methods. 48 WO 2007/021972 PCT/US2006/031441 FIG. 50 is a bargraph showing the number of samples collected for medical monitoring of a person in a year by different data collection methods. Method A represents a single yearly exam made by a general practitioner. 5 Method B shows exams made every six months by a medical specialist. Method C shows exams made every three months by a specialist. 10 Method D represents data collected during weekly applications of makeup. Method E shows data collected when makeup is applied twice a week. Method F shows the much larger amount of data that can be collected when makeup is applied daily. 15 As FIG. 50 shows, a woman who applies makeup daily through the present invention would collect 365 times more medical information, in a much more statistically meaningful sample, than she would through a yearly medical exam. She would also collect medical data 365 times more frequently, which would greatly aid the early detection of medical problems. 20 More frequent monitoring may have many medical advantages. It can lead to earlier detection of medical problems. It can be used to create a better base line about various physical conditions in human beings than infrequent monitoring can, creating a history of these conditions over each person's whole lifetime. Many changes in physical conditions are normal. For example, muscles swell after exercise, and mosquito bites cause 25 bumps on a person's skin. But a bump that keeps recurring in the same place over time may be of more concern medically. Step 8100 in FIG. 48 - Defining a Base Line The system and method of the cross-referenced application cited above creates a 3-D model of the scanned 30 exterior surface of all or a portion of a human body. The present invention uses feature recognition software to analyze the scanned data represented in the 3-D model and establish a medical base line of a person's physical characteristics as shown in reflectance data and surface topology data. For example, reflectance data can show medically relevant colors of areas of skin and of other features. Light 35 areas of skin may represent relatively healthy areas. Darker areas may represent moles and freckles. Very light areas may represent scars. Redness in cheeks may represent healthy coloring. Surface topology data can show the relative depth of areas. For example, smooth areas of skin can be identified, as well as projections such as bumps and scars. Depressions such as pores and wrinkles can also be identified. 40 49 WO 2007/021972 PCT/US2006/031441 As a result, a base line of the scanned areas can be created and can be represented in a 3-D map showing medically relevant conditions. Step 8200 in FIG. 48 - Identifying Medical Problems 5 The feature recognition software can identify medical problems in several ways, as shown in FIG. 2. Step 8210 in FIG. 49 - Analyzing the Base Line 3-D Image The feature recognition software can potentially identify existing medical problems in the base line 3-D Image, such as acne, bruises, tumors, and melanoma, by their characteristic patterns of reflectance and surface 10 topology. When medical problems are identified, the present invention reports on them, as explained below, so that the problems can be treated. For example, changes showing that an area of skin has become recessed over time may indicate structural deterioration as the result of malignant cells. If such changes are identified, they are reported. Bulges or depressions in large areas of skin may indicate cancer, atrophy, dropsy, and other more deeply positioned pathologies. 15 Step 8220 in FIG. 49 - Identifying Changes to the Base Line Each time additional scanned data about a person is collected through the cosmetics system, the present invention further uses feature recognition software to track changes to the medical base line, representing changes in physical characteristics. This can provide frequent phasic data about medically relevant conditions, 20 and data as to what changes are cyclical for a particular individual. For example, reflectance data can show medically relevant changes in the colors of areas of skin and other features. Blue areas that suddenly appear may indicate bruising. Sudden red areas may indicate acne or sunburns. Changes in the color and patterns of moles may indicate cancer. It may be medically relevant to 25 determine whether such changes have occurred in a person previously. With regard to surface topology data, swelling in formerly smooth areas of skin may indicate tumors, bumps may indicate bruises and new scars, and depressions in skin may indicate internal disease, for example. 30 Step 8230 in FIG. 49 - Analyzing the Changes to Identify Medical Problems Identified changes are further analyzed, for example through Bayesien probability equations, in comparison with a data in a software library showing the characteristics of medical problems as expressed in patterns of reflectance and of surface topology. 35 Examples of Unhealthy Characteristics The following list presents a few examples of unhealthy characteristics in the external surface of the human body that the present invention identifies. 50 WO 2007/021972 PCT/US2006/031441 Acne Analysis of reflectance data: Pigment changes can identify the reddish eruptions of acne, typically caused by viruses. Blackheads may be present. White pus may be discharged. 5 Analysis of surface topology data: Changes in morphology can identify the swelling of acne sores. Breast Cancer Analysis of reflectance data: Surface changes can identify rapid changes in vascularization characteristic of cancer. Thicker veins or changing vascular patterns seen through the skin are suspicious because in their early 10 stages cancerous cells pull in blood to feed themselves. Sudden redness of the breast, especially when the redness is mottled, may be a further indication of pathological conditions, as conditioned by measurements elsewhere on the body indicating sun exposure. Analysis of surface topology data: Thicker veins can produce swollen areas on the surface of the skin. Later, 15 cancer can destroy surrounding tissues, so that an affected area shrinks or appears to sink. In general, lumps or thickening and changes in shape in the breast may indicate cancer Jaundice Analysis of reflectance data: A quick change in color to yellow may indicate a liver malfunction. 20 Melanoma Analysis of reflectance data: The well-known ABCD patterns are indicators of melanoma: Asymmetry - The two halves of a spot on the skin are unlike each other. Border - The border of a spot is irregular, scalloped, or poorly defined. 25 Color - The color of a spot varies from one area to another. The shades of color are tan, brown, black, and sometimes white, red, or blue. Diameter - The diameter of a spot is larger that 6 mm. Analysis of surface topology data: Lumpiness, irregular growth changes. 30 Varicose Veins Analysis of reflectance data: Reflectance changes to bluish can characterize varicose veins. Analysis of surface topology data: Swelling and enlargement can also characterize varicose veins. 35 Step 8300 in FIG. 48 - Reporting on the Medical Problems The present invention identifies patterns potentially indicating medical problems, either during the creation of an initial 3-D map of medically relevant patterns or during subsequent comparison of changes to a base line 3-D map. After identifying medical problems, the present invention reports about them. The means of reporting can 51 WO 2007/021972 PCT/US2006/031441 take different forms, known and not yet known, in different embodiments. For example, in different embodiments the means of reporting may be a printer, a computer display, or an e-mail. FIG. 51 shows an embodiment where the present invention comprises a computing environment 700 loosely 5 connected through a network 730 to a portable scanner and inkjet printer 842 that represent an element of the cross-reference application mentioned above for scanning an area of skin 902. The computing environment 700 is directly connected to a printer 710 and a display 720 and is also connected to a medical office 910 through the network 730. 10 The computing environment 700 can communicate with portable scanner and inkjet printer 842 and the medical office 910 over network 730 and links 742, 744, and 746. The network 730 may comprise the Internet, a private LAN (Local Area Network), a wireless network, a TCP/IP (Transmission Control Protocol/Internet Protocol) network, or other communications system, and may 15 comprise multiple elements such as gateways, routers, and switches. The links 742, 744, and 746 are compatible with the technology used for network 730. After identifying medical problems, the present invention employs software 740 on the computing environment 700 to report on those problems according to parameters in the software 730 specified by the user. For example, 20 the user may specify that medical problems be displayed on display 720, printed on printer 710, or sent to a medical office 910 by e-mail for review. Step 8400 in FIG. 48 - Optionally Applying Medications when Appropriate 25 Applications of Medications In an embodiment, the present invention may apply appropriate compounds to identified medical problems when appropriate. This can even be done without requiring the patient to take overt action. For example, an embodiment may automatically apply sun block to an area of skin that is frequently sunburned. Other embodiments may automatically apply acne medicine to acne at an early stage of the acne's development and 30 antiseptic ointment to scratches. Still another embodiment may apply living skin cells or other replacement materials to burned areas of a patient's skin with great accuracy. For some conditions, such as acne, medicine can be applied very precisely at the pixel level to only the areas that are affected with a disease or medical problem. This may allow stronger, more effective dosages of the 35 medicine to be used than is typically possible through standard broadcast topical applications., as well as applications at an earlier stage of the medical condition, because of early detection. Standard acne medicines may be harmful if used in concentrated dosages over wide areas of skin. But with the present invention acne medicine may be applied specificlly to the pores of acne spots in very small but concentrated dosages that are administered frequently, such as during daily makeup sessions. To cite another example, the present invention 52 WO 2007/021972 PCT/US2006/031441 can apply a medical compound right to the pore into which a mosquito has bitten, to soothe the itching and reduce the swelling and redness. This capability for precise application of medications allows varied application of them. For example, sun block 5 may be applied selectively, so that an upturned surface which would receive more light or areas more often exposed to the sun and therefore age faster, such as the back of hands, can be protected more strongly. When used cosmetically, areas of skin that will be tanned can be corrected at the frexel level by the application of sun screen. For example, applying more sun screen to a pigmented area, such as damaged skin, would cause that area to be lightened relative to surrounding areas that received less sun screen. Regions burned by the sun can 10 be avoided for aesthetic purposes similar and supplemental to the application of transparent cosmetics. In addition, the medical compounds can themselves be used for cosmetic purposes. For example, a white colored sun block ointment can be used to make an unattractive dark spot on skin light and thus more attractive. 15 System To apply medicines and medically beneficial compounds, the present invention may use the system provided by the cross-referenced application mentioned above, in an embodiment. However, modifications may be made to that system for the medical applications of the present invention. 20 As shown in FIG. 52, for example, a second reservoir 870 and inkjet printer head 922 for medical compounds may be added to the reservoir 862 and printer head 920 of the cross-referenced application. Moreover, the second reservoir 870 can store multiple compounds, for example sun block 872 and acne medication 874. The use of different reservoirs, one 862 for RMAs 864 and one 870 for medical compounds 872 and 874, can 25 help prevent the accidental application or mixing of medical compounds or ink. The use of a second printer head 922 accomplishes the same purposes, and further allows the use of a second printer head 922 with a configuration appropriate for the application of medicinal compounds. 53

Claims (20)

1. A method for monitoring a region human skin, the method comprising 5 allocating the region of skin into a plurality of fi'exels; measuring the reflective properties of each of the plurality of frexels; determining, from the reflective properties of each of the plurality of frexels, the morphology of the skin; and analyzing the skin, according to the reflective properties and the morphology. 10
2. The method of claim 1 further comprising using a cosmetic enhancement system to perform the steps of allocating the region of skin into a plurality of frexels, measuring the reflective properties of each of the plurality of frexels, and 15 determining, from the reflective properties of each of the plurality of frexels, the morphology of the skin.
3. The method of claim 1 further comprising applying preventative or curative medical compounds to specific frexels to treat medical problems. 20
4. The method of claim 1 wherein analyzing the skin according to the reflective properties and the morphology further comprises using feature recognition software by defining a base line of medically relevant attributes from the scanned data, and 25 identifying attributes that warrant preventative or curative medical attention or treatment; and reporting about attributes that warrant preventative or curative medical attention or treatment.
5. The method of claim 1, wherein using feature recognition software to identify attributes that warrant 30 preventative or curative medical attention or treatment comprises analyzing the base line; identifying changes to the base line that appear in subsequent scanned data; and analyzing the changes to identify medical problems. 35
6. The method of claim 1 wherein measuring the reflective properties of each of the plurality of frexels further comprises providing a plurality of light sources which may be controlled to be on or off; creating a plurality of lighting conditions, each lighting condition comprising selecting an on or off state for each of the light sources; 54 WO 2007/021972 PCT/US2006/031441 scanning the region of skin to obtain image data while repeating the plurality of lighting conditions; and determining from the image data, the reflectance of the plurality of frexels. 5
7. The method of claim 6 further comprising determining the position of a frexel being scanned.
8. The method of claim 6 further comprising determining from the image data, the tilt of the frexel with respect to a reference plane. 10
9. The method of claim 8 further comprising determining the local skin morphology from the tilt of the frexel and the tilt of other frexels in proximity to the frexel. 15
10. The method of claim 1 wherein analyzing the skin according to the reflective properties and the morphology further comprises building a first map of the region of skin, such that the map comprises the location and reflectance of each of the plurality of frexels at a first time. 20
11. The method of claim 10 further comprising building a second map of the region of skin, such that the map comprises the location and reflectance of each of the plurality of frexels at a second time; and comparing the first map and the second map. 25
12. A apparatus for monitoring a region human skin, the apparatus comprising a housing; at least one light source, such that the light source may be turned on and off; at least one image receiving element; and 30 a computation device, such that the computation device may analyze data from the image receiving element in order to determine the reflectance and tilt of specific frexels.
13. The apparatus of claim 12 further comprising a first preventative or curative medical compound reservoir; and 35 a first preventative or curative medical compound application device.
14. The apparatus of claim 12 wherein the first preventative or curative medical compound application device further comprises a drop control device. 40 55 WO 2007/021972 PCT/US2006/031441
15. The apparatus of claim 12 wherein the first preventative or curative medical compound application device further comprises at least one inkjet print head. 5
16. The apparatus of claim 12 wherein the light source is a plurality of light emitting diodes, such a first light emitting diode has a first wavelength of light emission, and a second light emitting diode has a second wavelength of light emission. 10
17. The apparatus of claim 13 comprising a second reflectance modifying agent reservoir.
18. The apparatus of claim 12 wherein the housing is a booth. 15
19. The apparatus of claim 12 wherein the housing is a handheld device, such that the device may be manually moved over a region of skin.
20 20. The apparatus of claim 12 wherein the housing is a handheld device, such that the device may be manually placed over a skin mark. 56
AU2006279800A 2005-08-12 2006-08-14 System and method for medical monitoring and treatment through cosmetic monitoring and treatment Ceased AU2006279800B2 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US70811805P 2005-08-12 2005-08-12
US60/708,118 2005-08-12
PCT/US2006/031441 WO2007021972A2 (en) 2005-08-12 2006-08-14 System and method for medical monitoring and treatment through cosmetic monitoring and treatment

Publications (2)

Publication Number Publication Date
AU2006279800A1 true AU2006279800A1 (en) 2007-02-22
AU2006279800B2 AU2006279800B2 (en) 2012-03-29

Family

ID=37459663

Family Applications (2)

Application Number Title Priority Date Filing Date
AU2006279800A Ceased AU2006279800B2 (en) 2005-08-12 2006-08-14 System and method for medical monitoring and treatment through cosmetic monitoring and treatment
AU2006279652A Ceased AU2006279652B2 (en) 2005-08-12 2006-08-14 System and method for applying a reflectance modifying agent to improve the visual attractiveness of human skin

Family Applications After (1)

Application Number Title Priority Date Filing Date
AU2006279652A Ceased AU2006279652B2 (en) 2005-08-12 2006-08-14 System and method for applying a reflectance modifying agent to improve the visual attractiveness of human skin

Country Status (12)

Country Link
US (6) US8007062B2 (en)
EP (3) EP1912796B1 (en)
JP (3) JP5254790B2 (en)
KR (2) KR20080058337A (en)
CN (2) CN101326064B (en)
AT (1) ATE546294T1 (en)
AU (2) AU2006279800B2 (en)
CA (3) CA2985955C (en)
ES (1) ES2790594T3 (en)
HK (1) HK1125339A1 (en)
RU (2) RU2367577C1 (en)
WO (2) WO2007021972A2 (en)

Families Citing this family (168)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090163809A1 (en) * 2004-06-03 2009-06-25 Kane Scott D Medical method and associated apparatus utilizable in accessing internal organs through skin surface
US7324688B2 (en) * 2005-02-14 2008-01-29 Mitsubishi Electric Research Laboratories, Inc. Face relighting for normalization of directional lighting
WO2007021972A2 (en) 2005-08-12 2007-02-22 Yeager Rick B System and method for medical monitoring and treatment through cosmetic monitoring and treatment
US7890152B2 (en) * 2007-02-11 2011-02-15 Tcms Transparent Beauty Llc Handheld apparatus and method for the automated application of cosmetics and other substances
US8184901B2 (en) 2007-02-12 2012-05-22 Tcms Transparent Beauty Llc System and method for applying a reflectance modifying agent to change a person's appearance based on a digital image
US8942775B2 (en) 2006-08-14 2015-01-27 Tcms Transparent Beauty Llc Handheld apparatus and method for the automated application of cosmetics and other substances
WO2008098235A2 (en) * 2007-02-11 2008-08-14 Yeager Rick B System and method for providing simulated images through cosmetic monitoring
US8027505B2 (en) 2007-02-11 2011-09-27 Tcms Transparent Beauty Llc System and method for providing simulated images through cosmetic monitoring
WO2008100878A1 (en) * 2007-02-12 2008-08-21 Yeager Rick B System and method for applying a reflectance modifying agent to change a person's appearance based on a digital image
KR20100015318A (en) * 2007-02-12 2010-02-12 릭 비. 예거 System and method for applying agent electrostatically to human skin
US10092082B2 (en) 2007-05-29 2018-10-09 Tcms Transparent Beauty Llc Apparatus and method for the precision application of cosmetics
RU2501505C2 (en) * 2007-05-29 2013-12-20 ТиСиЭмЭс ТРАНСПЭРЕНТ БЬЮТИ, ЭлЭлСи Device and method for accurate application of cosmetics
EP2180832A2 (en) * 2007-08-22 2010-05-05 Koninklijke Philips Electronics N.V. Method and apparatus for the optical characterization of surfaces
DE102008041972A1 (en) * 2007-09-13 2010-03-11 Beiersdorf Ag Procedure for wrinkle reduction
DE102008041971A1 (en) * 2007-09-13 2010-03-11 Beiersdorf Ag Method for the reduction of age spots
DE102007043698A1 (en) * 2007-09-13 2009-03-19 Beiersdorf Ag Selective product applicator
US7995816B2 (en) * 2007-09-24 2011-08-09 Baxter International Inc. Detecting access disconnect by pattern recognition
NO2255173T3 (en) * 2008-03-18 2018-06-23
US8812357B2 (en) * 2008-04-29 2014-08-19 Tracy Hughes Automated tanning system
US8425477B2 (en) * 2008-09-16 2013-04-23 Elc Management Llc Method and system for providing targeted and individualized delivery of cosmetic actives
US8358348B2 (en) 2008-05-09 2013-01-22 Elc Management Llc Method and system for automatic or manual evaluation to provide targeted and individualized delivery of cosmetic actives in a mask or patch form
US8597667B2 (en) 2008-05-09 2013-12-03 Elc Management Llc Targeted and individualized cosmetic delivery
FR2933582B1 (en) * 2008-07-10 2011-10-07 Oreal DEVICE FOR TREATING HUMAN KERATINIC MATERIALS
FR2933584B1 (en) * 2008-07-10 2011-12-23 Oreal METHODS OF TREATING KERATINIC MATERIALS AND APPARATUS FOR CARRYING OUT SAID METHODS
FR2933583B1 (en) * 2008-07-10 2011-10-07 Oreal DEVICE FOR APPLYING A COMPOSITION ON HUMAN KERATINIC MATERIALS
FR2933581B1 (en) * 2008-07-10 2011-12-02 Oreal MAKE-UP METHOD AND DEVICE FOR IMPLEMENTING SUCH A METHOD
FR2933611B1 (en) * 2008-07-10 2012-12-14 Oreal MAKE-UP PROCESS AND DEVICE FOR IMPLEMENTING SUCH A METHOD
FR2933585B1 (en) * 2008-07-10 2011-10-14 Oreal METHOD FOR MAKING THE SKIN FROM A COLOR MEASUREMENT THEREOF
FR2933586B1 (en) * 2008-07-10 2011-10-14 Oreal DEVICE FOR APPLYING A PRODUCT TO HUMAN KERATIN FIBERS
CN101650775B (en) * 2008-08-15 2011-09-14 深圳市飞瑞斯科技有限公司 Uniform light source-based human face recognition device
US8491926B2 (en) * 2008-09-16 2013-07-23 Elc Management Llc Method and system for automatic or manual evaluation to provide targeted and individualized delivery of cosmetic actives in a mask or patch form
JP5521172B2 (en) * 2008-09-22 2014-06-11 公益財団法人ヒューマンサイエンス振興財団 Image printing device
US20100139682A1 (en) * 2008-11-13 2010-06-10 Albert Durr Edgar Method and apparatus for precise deposition of hair care agents
US20100224205A1 (en) 2008-12-09 2010-09-09 Shekhar Mitra Device and methods for modifying keratinous surfaces
WO2010083405A2 (en) * 2009-01-16 2010-07-22 The Procter & Gamble Company Apparatus and methods for modifying keratinous surfaces
US20100208048A1 (en) * 2009-02-19 2010-08-19 Jennifer Cha Digital image storage system and human body data comparison method for medical and aesthetic applications
JP5071430B2 (en) 2009-04-23 2012-11-14 トヨタ自動車株式会社 Disc brake device
JP2011041036A (en) * 2009-08-12 2011-02-24 Sony Corp Image processing apparatus, image processing method, and electronic equipment
US8855751B2 (en) * 2010-02-26 2014-10-07 Empire Technology Development Llc Multidirectional scan and algorithmic skin health analysis
US8591413B2 (en) * 2010-02-26 2013-11-26 Empire Technology Development Llc Echogram detection of skin conditions
KR101619342B1 (en) * 2010-02-26 2016-05-11 (주)아모레퍼시픽 System and method for detecting skin color
US9202310B2 (en) * 2010-04-13 2015-12-01 Disney Enterprises, Inc. Physical reproduction of reflectance fields
EP2433805B1 (en) * 2010-09-28 2014-04-23 BlackBerry Limited Portable electronic device and method of controlling the electronic device to output information
US8590543B2 (en) 2010-11-17 2013-11-26 Mattel, Inc. Hair extension kit
EP2713872A4 (en) * 2011-05-26 2014-10-22 3Derm Systems Llc Stereoscopic plug-and-play dermatoscope and web interface
CA2861110A1 (en) * 2012-01-18 2013-07-25 Covidien Lp Wireless relay module for monitoring network status
US10660561B2 (en) * 2012-02-25 2020-05-26 Massachusetts Institute Of Technology Personal skin scanner system
US9668653B2 (en) * 2012-03-15 2017-06-06 Access Business Group International Llc, A Michigan Limited Liability Company Quantification of under-eye skin color
CA2892698C (en) * 2012-11-30 2022-08-30 Koninklijke Philips N.V. Method and apparatus for estimating the fall risk of a user
CN105101869B (en) * 2012-11-30 2018-06-12 皇家飞利浦有限公司 For identifying the method and apparatus of the transformation between sitting posture and stance
JP5372276B1 (en) 2013-03-22 2013-12-18 パナソニック株式会社 Makeup support device, makeup support method, and makeup support program
US20140378810A1 (en) * 2013-04-18 2014-12-25 Digimarc Corporation Physiologic data acquisition and analysis
AT514407B1 (en) * 2013-06-06 2015-06-15 Andreas Wampl Process for the preparation of a tattoo dye
US9458990B2 (en) * 2013-08-01 2016-10-04 3Gen, Inc. Dermoscopy illumination device with selective polarization and orange light for enhanced viewing of pigmented tissue
US10152529B2 (en) 2013-08-23 2018-12-11 Elwha Llc Systems and methods for generating a treatment map
US10010704B2 (en) * 2013-08-23 2018-07-03 Elwha Llc Systems, methods, and devices for delivering treatment to a skin surface
US9805171B2 (en) 2013-08-23 2017-10-31 Elwha Llc Modifying a cosmetic product based on a microbe profile
US9557331B2 (en) 2013-08-23 2017-01-31 Elwha Llc Systems, methods, and devices for assessing microbiota of skin
JP6268827B2 (en) * 2013-09-05 2018-01-31 カシオ計算機株式会社 Nail printing apparatus and printing method for nail printing apparatus
JP6273640B2 (en) * 2013-10-23 2018-02-07 パナソニックIpマネジメント株式会社 Captured image display device
FR3015887B1 (en) * 2013-12-27 2017-03-24 Oreal DEVICE AND METHOD FOR MAKE-UP BY TRANSFERRING KERATINIC MATERIALS
FR3015889B1 (en) 2013-12-27 2016-02-05 Oreal DEVICE FOR MAKE-UP BY TRANSFERRING KERATINIC MATERIALS
FR3015888B1 (en) 2013-12-27 2017-03-31 Oreal DEVICE FOR MAKE-UP BY TRANSFERRING KERATINIC MATERIALS
FR3015927A1 (en) 2013-12-27 2015-07-03 Oreal TRANSFER MAKEUP METHOD AND ASSOCIATED DEVICE.
FR3015872B1 (en) * 2013-12-27 2017-03-24 Oreal MAKE-UP DEVICE COMPRISING A PLURALITY OF COSMETIC INKS
FR3015870B1 (en) 2013-12-27 2016-02-05 Oreal DEVICE FOR MAKE-UP BY TRANSFERRING KERATINIC MATERIALS.
TWI518583B (en) * 2014-04-18 2016-01-21 三緯國際立體列印科技股份有限公司 Three dimensional printing apparatus and method for detecting printing anomaly
USD791933S1 (en) 2014-06-09 2017-07-11 The Procter & Gamble Company Cosmetic device
USD750772S1 (en) 2014-06-09 2016-03-01 The Procter & Gamble Company Applicator tip for cosmetic device
USD750225S1 (en) 2014-06-09 2016-02-23 The Procter & Gamble Company Cosmetic device in charging station
US10156036B2 (en) 2014-06-13 2018-12-18 The Procter & Gamble Company Device and methods for applying compositions to fabric surfaces
CN106456005B (en) 2014-06-13 2020-12-08 宝洁公司 Apparatus and method for modifying keratinous surfaces
US10518291B2 (en) 2014-06-13 2019-12-31 The Procter & Gamble Company Device and methods for modifying surfaces
US20160114349A1 (en) 2014-06-13 2016-04-28 The Procter & Gamble Company Device and methods for depositing materials on hard surfaces
US9925362B2 (en) 2014-06-13 2018-03-27 The Procter & Gamble Company Apparatus and methods for modifying keratinous surfaces
EP3154417A2 (en) 2014-06-13 2017-04-19 The Procter & Gamble Company Apparatus and methods for modifying keratinous surfaces
KR101904709B1 (en) 2014-06-13 2018-10-08 더 프록터 앤드 갬블 캄파니 Apparatus and methods for modifying keratinous surfaces
WO2015191831A2 (en) 2014-06-13 2015-12-17 The Procter & Gamble Company Cartridges for the deposition of treatment compositions on keratinous surfaces
US20150359714A1 (en) * 2014-06-13 2015-12-17 The Procter & Gamble Company Treatment compositions, apparatus and methods for modifying keratinous surfaces
US20160022009A1 (en) * 2014-07-25 2016-01-28 The Procter & Gamble Company Handheld treatment apparatus for modifying keratinous surfaces
US9955769B2 (en) 2014-07-25 2018-05-01 The Procter & Gamble Company Applicator heads for handheld treatment apparatus for modifying keratinous surfaces
US11083880B2 (en) 2014-07-25 2021-08-10 The Procter & Gamble Company Angled cartridge assembly for a dispensing device
US10314378B2 (en) 2014-07-25 2019-06-11 The Procter & Gamble Company Cartridge assembly for a dispensing device
US9949552B2 (en) 2014-07-25 2018-04-24 The Procter & Gamble Company Handheld treatment apparatus for modifying keratinous surfaces
US10188193B2 (en) 2014-07-25 2019-01-29 The Procter & Gamble Company Applicator heads for handheld treatment apparatus for modifying keratinous surfaces
GB201415611D0 (en) 2014-09-03 2014-10-15 Univ Strathclyde Apparatus for topical application of material
WO2016054164A1 (en) * 2014-09-30 2016-04-07 Tcms Transparent Beauty, Llc Precise application of cosmetic looks from over a network environment
US10219737B2 (en) * 2014-12-11 2019-03-05 Skin Depth Inc. Topical product dispensing tool
WO2016191217A1 (en) * 2015-05-22 2016-12-01 L'oreal Applicators for treating skin conditions
US10076646B2 (en) 2015-05-22 2018-09-18 L'oreal Rolling applicator for treating skin conditions
US10004885B2 (en) 2015-05-22 2018-06-26 L'oreal Imaging applicator for treating skin conditions
US10071233B2 (en) * 2015-05-22 2018-09-11 L'oreal Point applicator for treating skin conditions
US9949547B2 (en) 2015-06-11 2018-04-24 The Procter & Gamble Company Cartridges for use in an apparatus for modifying keratinous surfaces
US20160361240A1 (en) 2015-06-11 2016-12-15 The Procter & Gamble Company Apparatus and methods for modifying keratinous surfaces
JP6878314B2 (en) 2015-06-11 2021-05-26 ザ プロクター アンド ギャンブル カンパニーThe Procter & Gamble Company Equipment and methods for applying the composition to the surface
US11116302B2 (en) * 2015-06-11 2021-09-14 The Procter & Gamble Company Apparatus and methods for modifying keratinous surfaces
US9962532B2 (en) 2015-06-11 2018-05-08 The Procter & Gamble Company Cartridges for use in an apparatus for modifying keratinous surfaces
USRE49230E1 (en) 2015-06-11 2022-10-04 The Procter & Gamble Company Cartridges for use in an apparatus for modifying keratinous surfaces
US11234594B2 (en) 2015-06-15 2022-02-01 E.S.I Novel Ltd. Systems and methods for adaptive skin treatment
US9789683B2 (en) * 2015-07-09 2017-10-17 Funai Electric Co., Ltd. Ejection device
EP3341889A1 (en) 2015-08-24 2018-07-04 Koninklijke Philips N.V. Server-client architecture in digital pathology
US9607347B1 (en) 2015-09-04 2017-03-28 Qiang Li Systems and methods of 3D scanning and robotic application of cosmetics to human
US9811717B2 (en) 2015-09-04 2017-11-07 Qiang Li Systems and methods of robotic application of cosmetics
US9616668B1 (en) 2015-12-07 2017-04-11 The Procter & Gamble Company Servicing cassettes for handheld fluid jet apparatuses for use in modifying surfaces
US10166799B2 (en) 2015-12-07 2019-01-01 The Procter & Gamble Company Service stations for handheld fluid jet apparatuses
US9616692B1 (en) 2015-12-07 2017-04-11 The Procter & Gamble Company Systems and methods for providing a service interface mechanism
US9782971B2 (en) 2015-12-07 2017-10-10 The Procter & Gamble Company Cartridge servicing cases for fluid jet cartridge
US11590782B2 (en) 2015-12-07 2023-02-28 The Procter & Gamble Company Systems and methods for providing a service station routine
US10391042B2 (en) 2015-12-07 2019-08-27 The Procter & Gamble Company Treatment compositions, apparatus and methods for modifying keratinous surfaces
US11077689B2 (en) 2015-12-07 2021-08-03 The Procter & Gamble Company Systems and methods for providing a service station routine
KR20180109080A (en) 2016-02-08 2018-10-05 이퀄리티 코스메틱스, 인크. Apparatus and method for the formation and distribution of visually customized cosmetics
KR101783314B1 (en) 2016-03-09 2017-10-23 스케치온 주식회사 Solution system, device and method for image printing in human skin
CN109152628A (en) * 2016-04-12 2019-01-04 Tcms剔透美丽有限责任公司 The targeting of dental care compound deposits
KR102509934B1 (en) * 2016-08-01 2023-03-15 엘지전자 주식회사 Mobile terminal and operating method thereof
USD836244S1 (en) 2016-10-14 2018-12-18 The Procter & Gamble Company Hand held cosmetic device
USD838411S1 (en) 2016-10-14 2019-01-15 The Procter & Gamble Company Service station for a cosmetic device
US10511777B2 (en) 2016-11-08 2019-12-17 Thomas Nichols Personal care device with camera
US11122206B2 (en) 2016-11-08 2021-09-14 Preh Holding, Llc Personal care device with camera
KR102528866B1 (en) * 2016-12-20 2023-05-04 가부시키가이샤 시세이도 Dispensing control device, dispensing control method, program and recording medium
EP3589187B1 (en) * 2017-03-01 2024-04-17 Canfield Scientific, Incorporated Three dimensional image capture
CN106991379B (en) * 2017-03-09 2020-07-10 Oppo广东移动通信有限公司 Human skin recognition method and device combined with depth information and electronic device
US10671971B1 (en) * 2017-03-29 2020-06-02 Stealhnetwork Communications, Inc. RFID inventory and mapping system
EP3638195A1 (en) 2017-06-16 2020-04-22 The Procter and Gamble Company Computer asissted colorimetric make up method for camouflaging skin color deviations
CN110650674B (en) * 2017-06-16 2022-06-14 宝洁公司 Personal care device with auditory feedback
US11090238B2 (en) 2017-06-16 2021-08-17 The Procter & Gamble Company Array of cosmetic compositions for camouflaging tonal imperfections
KR200494840Y1 (en) 2017-07-25 2022-01-10 블리. 엘엘씨 Systems and devices for cosmetic customization
US20190029396A1 (en) * 2017-07-25 2019-01-31 Laurel Corrinne KAGAY Tanning methods and articles therefor
US20190073448A1 (en) * 2017-09-05 2019-03-07 International Business Machines Corporation Monitoring and determining the state of health of a user
CN109508587A (en) * 2017-09-15 2019-03-22 丽宝大数据股份有限公司 Biological information analytical equipment and its bottom adornment analysis method
KR101856909B1 (en) * 2017-10-26 2018-05-10 박창식 Apparatus for Measuring Skin Condition with Multiple Lights
US10646022B2 (en) 2017-12-21 2020-05-12 Samsung Electronics Co. Ltd. System and method for object modification using mixed reality
JP2019115653A (en) * 2017-12-26 2019-07-18 パナソニックIpマネジメント株式会社 Body appearance correction support method and device, and computer program
US10441379B2 (en) 2017-12-28 2019-10-15 3Gen, Inc. Multipurpose medical illuminator with magnification
US10433630B2 (en) * 2018-01-05 2019-10-08 L'oreal Cosmetic applicator system including trackable cosmetic device, and client device to assist users in makeup application
US10813857B2 (en) 2018-02-01 2020-10-27 The Procter & Gamble Company Heterogenous cosmetic ink composition for inkjet printing applications
US10610471B2 (en) 2018-02-01 2020-04-07 The Procter & Gamble Company Cosmetic ink composition comprising a surface tension modifier
JP7104180B2 (en) 2018-02-01 2022-07-20 ザ プロクター アンド ギャンブル カンパニー Systems and methods for distributing materials
US10849843B2 (en) 2018-02-01 2020-12-01 The Procter & Gamble Company Stable cosmetic ink composition
EP3530179A1 (en) * 2018-02-27 2019-08-28 Koninklijke Philips N.V. Obtaining images for use in determining one or more properties of skin of a subject
US10271629B1 (en) 2018-05-29 2019-04-30 Equality Cosmetics, Inc. Cosmetics portioning machine
US10575623B2 (en) 2018-06-29 2020-03-03 Sephora USA, Inc. Color capture system and device
US11798162B2 (en) 2018-07-16 2023-10-24 Honor Device Co., Ltd. Skin detection method and electronic device
WO2020061514A1 (en) * 2018-09-21 2020-03-26 L'oreal A system for sensing and associating a condition on a body part of the user with a three-dimensional environment when using a cosmetic device
CN111053356A (en) * 2018-10-17 2020-04-24 丽宝大数据股份有限公司 Electronic cosmetic mirror device and display method thereof
DE102018126568A1 (en) 2018-10-24 2020-04-30 J. Wagner Gmbh Method of applying a cosmetic substance to a person's skin
JP2020085856A (en) * 2018-11-30 2020-06-04 ポーラ化成工業株式会社 Estimation device, estimation method and estimation program
CN109591059B (en) * 2018-12-07 2020-12-11 安徽省华腾农业科技有限公司经开区分公司 Extended rotary shaver
FR3090949B1 (en) 2018-12-20 2020-12-25 Thales Sa PROCESS FOR CHARACTERIZING PIGMENTARY ORDERS OF AN INDIVIDUAL
CN113226158A (en) * 2018-12-27 2021-08-06 强生消费者公司 Device and method for selectively administering topical compositions using dynamic threshold
JP2022516046A (en) * 2018-12-27 2022-02-24 ジョンソン・アンド・ジョンソン・コンシューマー・インコーポレイテッド Application device and application method for topical composition guided by projected reference
US11030798B2 (en) 2019-01-30 2021-06-08 Perfect Mobile Corp. Systems and methods for virtual application of makeup effects based on lighting conditions and surface properties of makeup effects
CN109646178B (en) * 2019-01-30 2024-01-12 深圳市启明医药科技有限公司 Control device, control method and storage medium based on micro-droplet ejection device
JP2021000761A (en) * 2019-06-21 2021-01-07 ロレアル Cosmetic ink application device and cosmetic ink application method
JP2023502830A (en) 2019-08-16 2023-01-26 ジョンソン アンド ジョンソン コンシューマー インコーポレイテッド Device and method for applying cosmetic composition through grid end effector
US11633106B2 (en) 2019-10-24 2023-04-25 L'oreal System and method for changing a cosmetic formulation attribute
US11395714B2 (en) 2019-11-11 2022-07-26 Dermlite Llc Medical illuminator with variable polarization
US11657481B2 (en) * 2019-11-18 2023-05-23 Shinyfields Limited Systems and methods for selective enhancement of skin features in images
USD1000624S1 (en) 2019-12-27 2023-10-03 Thomas Nichols Personal care device with camera
US11615486B2 (en) * 2020-02-12 2023-03-28 Blue River Technology Inc. Upward facing light sensor for plant detection
KR102167185B1 (en) * 2020-03-18 2020-10-19 이승락 Skin test method and cosmetic composition manufacturing method using the same
CN113496459A (en) * 2020-04-01 2021-10-12 华为技术有限公司 Make-up assisting method, terminal device, storage medium, and program product
CA3188443A1 (en) * 2020-06-30 2022-01-06 Johnson & Johnson Consumer Inc. Device and method for controlling treatment for a skin condition using a tracer
JP2023536884A (en) 2020-08-03 2023-08-30 ジョンソン アンド ジョンソン コンシューマー インコーポレイテッド Systems and methods for selective application of cosmetic compositions to provide under-eye brightening
WO2022040392A1 (en) 2020-08-19 2022-02-24 Blee.LLC System and device for customization of cosmetics
CA3205941A1 (en) * 2020-12-21 2022-06-30 Johnson & Johnson Consumer Inc. Device and method for continuous even application of composition to skin
US11900556B2 (en) * 2021-06-30 2024-02-13 L'oreal Generation of digital 3D models of body surfaces with automatic feature identification
WO2024077379A1 (en) * 2022-10-11 2024-04-18 Fitskin Inc. Systems and methods for improved skin tone rendering in digital images
CN116603716B (en) * 2023-04-28 2024-03-29 宁波市融嘉轻合金科技有限公司 Die casting surface treatment method, system, intelligent terminal and storage medium

Family Cites Families (169)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4190056A (en) 1977-10-31 1980-02-26 General Medical Company Method and means for recording sweat gland activity
US4401122A (en) 1979-08-02 1983-08-30 Children's Hospital Medical Center Cutaneous methods of measuring body substances
JPS59171280A (en) 1983-03-17 1984-09-27 Sumitomo Electric Ind Ltd Picture display device for medical use
JPS59171280U (en) 1983-05-02 1984-11-15 三谷セキサン株式会社 perforated tube
US4628356A (en) 1984-10-15 1986-12-09 Imagex, Inc. Digital X-ray scanner
JPS61215306A (en) 1985-03-19 1986-09-25 Nippon Kanko Shikiso Kenkyusho:Kk Novel skin-beautifying agent
US4807991A (en) * 1986-04-07 1989-02-28 Electro-Organic Company Method of inspecting and repairing a structural defect in the surface of an object
US4882492A (en) 1988-01-19 1989-11-21 Biotronics Associates, Inc. Non-invasive near infrared measurement of blood analyte concentrations
CA2005616C (en) 1988-12-16 2000-05-09 Shigeo Iizuka Liquid dispenser
US5016173A (en) 1989-04-13 1991-05-14 Vanguard Imaging Ltd. Apparatus and method for monitoring visually accessible surfaces of the body
US5241468A (en) 1989-04-13 1993-08-31 Vanguard Imaging Ltd. Apparatus and method for spectral enhancement of body-surface images to improve sensitivity of detecting subtle color features
US5027817A (en) 1989-06-22 1991-07-02 New York University Statistical based display for positron emission tomography scans
GB9115275D0 (en) 1991-07-15 1991-08-28 Unilever Plc Colour cosmetic spray system
US7549424B2 (en) 1991-10-18 2009-06-23 Pro Surg, Inc. Method and apparatus for tissue treatment with laser and electromagnetic radiation
JPH05281041A (en) * 1992-03-31 1993-10-29 Omron Corp Spectroscope
US7970620B2 (en) 1992-11-17 2011-06-28 Health Hero Network, Inc. Multi-user remote health monitoring system with biometrics support
JPH06201468A (en) * 1992-12-25 1994-07-19 Hiroshi Maeda Led light emitting spectroscope
US5431911A (en) 1993-06-21 1995-07-11 Reynolds; Diane S. Skin treatment composition
JP3236731B2 (en) * 1994-02-23 2001-12-10 カネボウ株式会社 Skin surface analysis system and skin surface analysis method
US5714007A (en) 1995-06-06 1998-02-03 David Sarnoff Research Center, Inc. Apparatus for electrostatically depositing a medicament powder upon predefined regions of a substrate
GB2304187B (en) * 1995-08-07 1998-12-30 Dia Stron Ltd Translucency measurement
US6385487B1 (en) 1996-05-08 2002-05-07 Biophoretic Therapeutic Systems, Llc Methods for electrokinetic delivery of medicaments
US6151031A (en) 1996-09-09 2000-11-21 Hewlett-Packard Company Map builder system and method for enabling generic interfacing of an application with a display map generation process in a management system
CA2192036A1 (en) 1996-12-04 1998-06-04 Harvey Lui Fluorescence scope system for dermatologic diagnosis
US6122042A (en) * 1997-02-07 2000-09-19 Wunderman; Irwin Devices and methods for optically identifying characteristics of material objects
AU740638B2 (en) 1997-02-28 2001-11-08 Electro-Optical Sciences, Inc. Systems and methods for the multispectral imaging and characterization of skin tissue
JP3912834B2 (en) 1997-03-06 2007-05-09 有限会社開発顧問室 Face image correction method, makeup simulation method, makeup method, makeup support apparatus, and foundation transfer film
JPH1119050A (en) 1997-07-01 1999-01-26 Nippon Colin Co Ltd Image transmitting device for medical treatment
JPH1119051A (en) 1997-07-07 1999-01-26 Nippon Colin Co Ltd Image transmitting device for medical treatment
US6022124A (en) 1997-08-19 2000-02-08 Ppt Vision, Inc. Machine-vision ring-reflector illumination system and method
US6459919B1 (en) 1997-08-26 2002-10-01 Color Kinetics, Incorporated Precision illumination methods and systems
ATE328642T1 (en) * 1997-10-08 2006-06-15 Gen Hospital Corp PHOTOTHERAPEUTIC SYSTEMS
KR100253203B1 (en) 1997-10-29 2000-04-15 구자홍 The objection extraction method using motion picture
EP0931649A3 (en) 1998-01-27 2000-04-26 Eastman Kodak Company Apparatus and method for making a contoured surface having complex topology
US6578276B2 (en) 1998-01-27 2003-06-17 Eastman Kodak Company Apparatus and method for marking multiple colors on a contoured surface having a complex topography
US6487440B2 (en) 1998-07-08 2002-11-26 Lifespex, Inc. Optical probe having and methods for difuse and uniform light irradiation
US6292277B1 (en) 1998-10-16 2001-09-18 Lextron Systems, Inc. Method and apparatus for creating a white-light interference hologram from PC input photographic data
DE19850350C1 (en) * 1998-11-02 2000-09-28 Jena Optronik Gmbh Method and device for generating data for the diagnosis of the degree of damage to a patient's skin tissue
CA2315716C (en) 1998-11-03 2008-12-02 Shade Analyzing Technologies, Inc. Interactive dental restorative network
US6396599B1 (en) 1998-12-21 2002-05-28 Eastman Kodak Company Method and apparatus for modifying a portion of an image in accordance with colorimetric parameters
US5931166A (en) 1998-12-22 1999-08-03 Weber; Paul J. Fingernail decorating
US6286517B1 (en) 1998-12-22 2001-09-11 Pearl Technology Holdings, Llc Fingernail and toenail decoration using ink jets
US6341831B1 (en) 1999-03-09 2002-01-29 Paul J. Weber Skin decoration apparatus and method
US6477410B1 (en) 2000-05-31 2002-11-05 Biophoretic Therapeutic Systems, Llc Electrokinetic delivery of medicaments
JP2000331167A (en) 1999-05-25 2000-11-30 Medeikku Engineering:Kk Method and device for collating facial image
US6571003B1 (en) * 1999-06-14 2003-05-27 The Procter & Gamble Company Skin imaging and analysis systems and methods
WO2001001854A2 (en) * 1999-07-02 2001-01-11 Hypermed Imaging, Inc. Integrated imaging apparatus
US6293284B1 (en) 1999-07-07 2001-09-25 Division Of Conopco, Inc. Virtual makeover
FR2799022B1 (en) 1999-09-29 2002-02-01 Oreal MAKEUP ASSISTANCE DEVICE AND ASSEMBLY CONSISTING OF SUCH A DEVICE AND A DEVICE FOR DELIVERING A PRODUCT HAVING A PREDETERMINED BRDF, SELECTED BY THE MAKEUP ASSISTANCE DEVICE
IL132290A0 (en) 1999-10-08 2001-03-19 Ultra Cure Ltd A method and device for affecting an object by acoustic radiation
JP3729692B2 (en) 1999-10-19 2005-12-21 株式会社資生堂 Surface state analysis method
US6312124B1 (en) 1999-10-27 2001-11-06 Hewlett-Packard Company Solid and semi-flexible body inkjet printing system
US6993167B1 (en) 1999-11-12 2006-01-31 Polartechnics Limited System and method for examining, recording and analyzing dermatological conditions
US6250927B1 (en) 1999-11-29 2001-06-26 Jean Narlo Cosmetic application training system
US20030060810A1 (en) 2000-02-16 2003-03-27 Diego Syrowicz Method and apparatus for treating and/or removing an undesired presence on the skin of an individual
WO2001077976A2 (en) 2000-03-28 2001-10-18 Eyeweb, Inc. Image segmenting to enable electronic shopping for wearable goods and cosmetic services
US20020064302A1 (en) 2000-04-10 2002-05-30 Massengill R. Kemp Virtual cosmetic autosurgery via telemedicine
FR2810761B1 (en) * 2000-06-26 2003-09-05 Oreal COSMETIC TREATMENT PROCESS AND DEVICE, IN PARTICULAR FOR CARE, MAKE-UP OR COLORING
FR2810539B1 (en) 2000-06-26 2004-05-07 Oreal PROCESS AND DEVICE FOR TREATING, PARTICULARLY MAKEUP, COLORING OR COSMETIC CARE, OF PARTS OR THE WHOLE OF THE HUMAN OR ANIMAL BODY
JP2002011106A (en) 2000-06-28 2002-01-15 Nidek Co Ltd Laser therapeutic apparatus
JP4842424B2 (en) 2000-07-07 2011-12-21 株式会社ナリス化粧品 Method for measuring and evaluating skin texture, and use of a composition that exhibits effectiveness by the method
FR2812532B1 (en) 2000-08-01 2003-07-04 Oreal METHOD FOR ACQUIRING AN IMAGE OF A NON-DERMATHOGLYPHIC ZONE OF THE SKIN OR A HAIR ZONE BY MEANS OF AN ACQUISITION DEVICE INCLUDING A NON-OPTICAL SENSOR
KR100398362B1 (en) 2000-09-01 2003-09-19 스펙트론 테크 주식회사 Method and apparatus for measuring skin moisture by using near-infrared reflectance spectroscopy
US6959119B2 (en) 2000-11-03 2005-10-25 Unilever Home & Personal Care Usa Method of evaluating cosmetic products on a consumer with future predictive transformation
US8360973B2 (en) 2000-11-29 2013-01-29 L'oreal Process for acquiring scanned image data relating to an external body portion and/or a product applied thereto
US20020065452A1 (en) * 2000-11-29 2002-05-30 Roland Bazin Process for diagnosing conditions of external body portions and features of products applied thereto
FR2817745B1 (en) 2000-12-08 2003-03-28 Oreal MANUFACTURER AND METHOD OF MANUFACTURE
US20020081003A1 (en) 2000-12-27 2002-06-27 Sobol Robert E. System and method for automatically enhancing graphical images
US20040201694A1 (en) 2001-02-07 2004-10-14 Vladimir Gartstein Noninvasive methods and apparatus for monitoring at least one hair characteristic
ES2219585T3 (en) 2001-02-27 2004-12-01 Sms Demag Ag PROCEDURE FOR THE ROLLING OF PROFILE SHEET MATERIAL.
US6937749B2 (en) 2001-03-02 2005-08-30 L'oreal Process, system, and kit for evaluating the relief of the skin with a substrate
US6719467B2 (en) 2001-04-30 2004-04-13 Hewlett-Packard Development Company, L.P. Floor printer
US20020172419A1 (en) 2001-05-15 2002-11-21 Qian Lin Image enhancement using face detection
AUPR541801A0 (en) 2001-06-01 2001-06-28 Canon Kabushiki Kaisha Face detection in colour images with complex background
CH692772A5 (en) 2001-06-29 2002-10-31 Gecomwert Anstalt Simple and precise method of classification and cosmetic treatment of facial wrinkling
FR2826857B1 (en) 2001-07-09 2004-03-12 Oreal INSTRUMENT FOR OBSERVING SKIN OR HAIR
FR2826856B1 (en) 2001-07-09 2004-03-12 Oreal DEVICE FOR DETERMINING THE DEGREE OF A BODY TYPOLOGY CHARACTERISTIC
US20030208190A1 (en) 2001-07-20 2003-11-06 Cynthia Roberts Methods and instruments for refractive ophthalmic surgery
JP4485788B2 (en) 2001-07-27 2010-06-23 コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ Apparatus for skin treatment having a processor for determining the dose of a radiation pulse
JP2003057170A (en) 2001-08-13 2003-02-26 Shiseido Co Ltd Melanin evaluation apparatus
JP2003052642A (en) * 2001-08-13 2003-02-25 Shiseido Co Ltd Detector for epidermis-dermis interface of skin
JP2003057169A (en) * 2001-08-13 2003-02-26 Shiseido Co Ltd Three-dimensional image generation apparatus for skin
FR2829344B1 (en) * 2001-08-29 2004-10-01 Oreal DEVICE FOR ACQUIRING AT LEAST ONE IMAGE OF AT LEAST PART OF A PERSON'S FACE OR HAIR
US7027619B2 (en) 2001-09-13 2006-04-11 Honeywell International Inc. Near-infrared method and system for use in face detection
US20030063102A1 (en) 2001-10-01 2003-04-03 Gilles Rubinstenn Body image enhancement
US6622733B2 (en) 2001-10-22 2003-09-23 Hewlett-Packard Development Company, L.P. Cosmetic applicator
JP4394444B2 (en) 2001-10-24 2010-01-06 パワー ペーパー リミティド Devices and methods for controlled intradermal delivery of active agents
FR2831790B1 (en) 2001-11-06 2004-07-16 Oreal MEASUREMENT AND / OR ANALYSIS DEVICE OF AT LEAST ONE PARAMETER OF AN EXTERNAL PORTION OF THE HUMAN BODY
US7433102B2 (en) 2002-05-10 2008-10-07 Canon Kabushiki Kaisha Reproduction color prediction apparatus and method
US7082211B2 (en) 2002-05-31 2006-07-25 Eastman Kodak Company Method and system for enhancing portrait images
DE50302500D1 (en) 2002-06-26 2006-04-27 Webasto Ag Fahrzeugtechnik Adhesive tape for fastening a sealing element, and method for producing and fixing the sealing element
US7406184B2 (en) 2002-07-03 2008-07-29 Equinox Corporation Method and apparatus for using thermal infrared for face recognition
JP3530523B1 (en) 2002-07-09 2004-05-24 株式会社リック Digital zoom skin diagnostic device
US20050261641A1 (en) 2002-09-26 2005-11-24 Warchol Mark P Method for ophthalmic administration of medicament
US20040073186A1 (en) * 2002-10-15 2004-04-15 Cameron Mickey G. Apparatus for automatic application of compositions to the skin
EP1570684A4 (en) 2002-12-13 2010-06-30 Color Savvy Systems Ltd Method for using an electronic imaging device to measure color
FR2848822B1 (en) 2002-12-24 2006-01-13 Oreal COSMETIC COMPOSITION OF FOUNDATION TYPE FOR MAKE-UP OF DARK SKIN
US20040125996A1 (en) 2002-12-27 2004-07-01 Unilever Home & Personal Care Usa, Division Of Conopco, Inc. Skin diagnostic imaging method and apparatus
US7039222B2 (en) 2003-02-28 2006-05-02 Eastman Kodak Company Method and system for enhancing portrait images that are processed in a batch mode
US7006223B2 (en) 2003-03-07 2006-02-28 3Gen, Llc. Dermoscopy epiluminescence device employing cross and parallel polarization
ITMI20030541A1 (en) * 2003-03-20 2004-09-21 Ist Naz Stud Cura Dei Tumori APPARATUS FOR THE CHARACTERIZATION OF PIGMENTAL INJURY OF THE SKIN
US20040186373A1 (en) 2003-03-21 2004-09-23 Dunfield John Stephen Method and device for targeted epithelial delivery of medicinal and related agents
JP2004315416A (en) * 2003-04-15 2004-11-11 Naris Cosmetics Co Ltd Powder composition
JP4438310B2 (en) * 2003-04-16 2010-03-24 ロレアル Cosmetic additives
DE10317692A1 (en) * 2003-04-17 2004-11-11 Lts Lohmann Therapie-Systeme Ag Medical active substance patches with reduced optical conspicuity on the skin
ITRM20030184A1 (en) * 2003-04-22 2004-10-23 Provincia Italiana Della Congregazi One Dei Figli METHOD FOR AUTOMATED DETECTION AND SIGNALING
US7233693B2 (en) 2003-04-29 2007-06-19 Inforward, Inc. Methods and systems for computer analysis of skin image
US20050004475A1 (en) 2003-06-13 2005-01-06 L'oreal Method of comparing the appearance of a region of the human body at at least two different instants
JP3954533B2 (en) 2003-06-17 2007-08-08 株式会社モリテックス Skin observation device
US7208007B2 (en) 2003-08-07 2007-04-24 Cutera, Inc. System and method utilizing guided fluorescence for high intensity applications
JP2005078376A (en) 2003-08-29 2005-03-24 Sony Corp Object detection device, object detection method, and robot device
IL157786A (en) 2003-09-07 2010-11-30 Ahava Dead Sea Lab Ltd Personalized cosmetics
WO2005034747A1 (en) * 2003-09-15 2005-04-21 Beth Israel Deaconess Medical Center Medical imaging systems
US7309335B2 (en) * 2003-12-31 2007-12-18 Palomar Medical Technologies, Inc. Dermatological treatment with visualization
US7382400B2 (en) 2004-02-19 2008-06-03 Robert Bosch Gmbh Image stabilization system and method for a video camera
DE202004003148U1 (en) 2004-03-01 2005-03-24 Merlaku Kastriot Makeup application device for automatically applying makeup to user, applies makeup liquid or powder in form of fine jet
FR2872405B1 (en) 2004-07-02 2006-11-10 Biomedical Electronics METHOD FOR PARAMETERSING A SKIN TREATMENT DEVICE USING LIGHT SOURCES
US8026942B2 (en) 2004-10-29 2011-09-27 Johnson & Johnson Consumer Companies, Inc. Skin imaging system with probe
US7602942B2 (en) 2004-11-12 2009-10-13 Honeywell International Inc. Infrared and visible fusion face recognition system
US20060104507A1 (en) 2004-11-15 2006-05-18 Sozotek, Inc. Correction of image color levels
WO2006062895A2 (en) 2004-12-06 2006-06-15 Cambridge Research And Instrumentation, Inc. Systems and methods for in-vivo optical imaging and measurement
EP1677254A1 (en) 2004-12-30 2006-07-05 GSF-Forschungszentrum für Umwelt und Gesundheit GmbH Method and device for collimating an energy input beam
US20060159820A1 (en) 2005-01-14 2006-07-20 Susanne Rathjen Sweetener compositions with a sweetness and taste profile comparable to HFCS 55
JP2006271654A (en) 2005-03-29 2006-10-12 Canon Inc Cosmetic-application system and method using liquid delivery method
PL1878421T3 (en) 2005-04-28 2019-09-30 Sbi Pharmaceuticals Co., Ltd. External preparation for skin
US20070047761A1 (en) 2005-06-10 2007-03-01 Wasilunas Elizabeth A Methods Of Analyzing Human Facial Symmetry And Balance To Provide Beauty Advice
WO2007021972A2 (en) 2005-08-12 2007-02-22 Yeager Rick B System and method for medical monitoring and treatment through cosmetic monitoring and treatment
US20070134192A1 (en) 2005-12-08 2007-06-14 L'oreal Two-coat cosmetic product comprising an ester of dimerdilinoleic acid and of polyol(s)
JP2007231883A (en) 2006-03-02 2007-09-13 Toyota Motor Corp Air fuel ratio control device for internal combustion engine
US20070255589A1 (en) 2006-04-27 2007-11-01 Klinger Advanced Aesthetics, Inc. Systems and methods using a dynamic database to provide aesthetic improvement procedures
US8620038B2 (en) 2006-05-05 2013-12-31 Parham Aarabi Method, system and computer program product for automatic and semi-automatic modification of digital images of faces
US8184901B2 (en) 2007-02-12 2012-05-22 Tcms Transparent Beauty Llc System and method for applying a reflectance modifying agent to change a person's appearance based on a digital image
US7890152B2 (en) 2007-02-11 2011-02-15 Tcms Transparent Beauty Llc Handheld apparatus and method for the automated application of cosmetics and other substances
US8942775B2 (en) 2006-08-14 2015-01-27 Tcms Transparent Beauty Llc Handheld apparatus and method for the automated application of cosmetics and other substances
US8027505B2 (en) 2007-02-11 2011-09-27 Tcms Transparent Beauty Llc System and method for providing simulated images through cosmetic monitoring
KR20100015318A (en) 2007-02-12 2010-02-12 릭 비. 예거 System and method for applying agent electrostatically to human skin
RU2334866C1 (en) 2007-02-26 2008-09-27 Открытое акционерное общество "Татнефть" им. В.Д. Шашина Device for simultaneous-separate operation of multypay well
US10092082B2 (en) 2007-05-29 2018-10-09 Tcms Transparent Beauty Llc Apparatus and method for the precision application of cosmetics
US8337405B2 (en) 2007-06-08 2012-12-25 Raytheon Company System and method for automatic detection of anomalies in images
AT505355B1 (en) 2007-06-13 2009-04-15 Lechthaler Andreas DEVICE FOR IRRADIATING AN OBJECT, IN PARTICULAR THE HUMAN SKIN, WITH UV LIGHT
US20090231356A1 (en) 2008-03-17 2009-09-17 Photometria, Inc. Graphical user interface for selection of options from option groups and methods relating to same
WO2009115947A1 (en) 2008-03-18 2009-09-24 Koninklijke Philips Electronics N.V. Apparatus for skin imaging, system for skin analysis
FR2933581B1 (en) 2008-07-10 2011-12-02 Oreal MAKE-UP METHOD AND DEVICE FOR IMPLEMENTING SUCH A METHOD
KR101446975B1 (en) 2008-07-30 2014-10-06 디지털옵틱스 코포레이션 유럽 리미티드 Automatic face and skin beautification using face detection
US20100139682A1 (en) 2008-11-13 2010-06-10 Albert Durr Edgar Method and apparatus for precise deposition of hair care agents
US20100224205A1 (en) 2008-12-09 2010-09-09 Shekhar Mitra Device and methods for modifying keratinous surfaces
WO2010083405A2 (en) 2009-01-16 2010-07-22 The Procter & Gamble Company Apparatus and methods for modifying keratinous surfaces
US8182425B2 (en) 2010-05-18 2012-05-22 Johnson & Johnson Consumer Companies, Inc. Method for measuring skin hydration
EA023736B1 (en) 2010-08-24 2016-07-29 Юнилевер Нв Personal care composition
TW201212852A (en) 2010-09-21 2012-04-01 Zong Jing Investment Inc Facial cosmetic machine
WO2012103048A2 (en) 2011-01-24 2012-08-02 Tcms Transparent Beauty Llc Apparatus and method for rapid and precise application of cosmetics
US8908904B2 (en) 2011-12-28 2014-12-09 Samsung Electrônica da Amazônia Ltda. Method and system for make-up simulation on portable devices having digital cameras
TWI463955B (en) 2012-02-20 2014-12-11 Zong Jing Investment Inc Eye makeup device
JP5949030B2 (en) 2012-03-26 2016-07-06 カシオ計算機株式会社 Image generating apparatus, image generating method, and program
US9462872B2 (en) 2012-08-27 2016-10-11 Tcms Transparent Beauty, Llc Printer head with prism for sensing
JP5949331B2 (en) 2012-08-30 2016-07-06 カシオ計算機株式会社 Image generating apparatus, image generating method, and program
US8977389B2 (en) 2013-07-17 2015-03-10 ColorCulture Network, LLC Method, system and apparatus for dispensing products for a personal care service, instructing on providing a personal care treatment service, and selecting a personal care service
US9277799B2 (en) 2014-04-18 2016-03-08 United Arab Emirates University Henna applicator device and a method of operating the same
US20150359714A1 (en) 2014-06-13 2015-12-17 The Procter & Gamble Company Treatment compositions, apparatus and methods for modifying keratinous surfaces
KR101904709B1 (en) 2014-06-13 2018-10-08 더 프록터 앤드 갬블 캄파니 Apparatus and methods for modifying keratinous surfaces
WO2016054164A1 (en) 2014-09-30 2016-04-07 Tcms Transparent Beauty, Llc Precise application of cosmetic looks from over a network environment
TWI587242B (en) 2015-09-08 2017-06-11 宏達國際電子股份有限公司 Facial image adjustment method and facial image adjustment system
US9754351B2 (en) 2015-11-05 2017-09-05 Facebook, Inc. Systems and methods for processing content using convolutional neural networks
US9984282B2 (en) 2015-12-10 2018-05-29 Perfect Corp. Systems and methods for distinguishing facial features for cosmetic application
KR20180109080A (en) 2016-02-08 2018-10-05 이퀄리티 코스메틱스, 인크. Apparatus and method for the formation and distribution of visually customized cosmetics
CN109791702B (en) 2016-09-23 2023-09-29 苹果公司 Head portrait creation and editing
DK179948B1 (en) 2017-05-16 2019-10-22 Apple Inc. Recording and sending Emoji
US10375313B1 (en) 2018-05-07 2019-08-06 Apple Inc. Creative camera
DK179874B1 (en) 2018-05-07 2019-08-13 Apple Inc. USER INTERFACE FOR AVATAR CREATION

Also Published As

Publication number Publication date
ATE546294T1 (en) 2012-03-15
CA2618706A1 (en) 2007-02-22
US11445802B2 (en) 2022-09-20
AU2006279652B2 (en) 2012-10-25
EP1912795A1 (en) 2008-04-23
JP5039700B2 (en) 2012-10-03
WO2007022095A1 (en) 2007-02-22
JP2009504678A (en) 2009-02-05
JP2009504272A (en) 2009-02-05
RU2367577C1 (en) 2009-09-20
US20070049832A1 (en) 2007-03-01
CN101287607B (en) 2010-09-01
RU2385669C2 (en) 2010-04-10
RU2008109234A (en) 2009-09-20
CA2618519C (en) 2016-09-27
EP1912796B1 (en) 2013-05-15
CA2618706C (en) 2018-01-02
WO2007021972A3 (en) 2008-02-07
ES2790594T3 (en) 2020-10-28
CA2618519A1 (en) 2007-02-22
KR20080058337A (en) 2008-06-25
US20180310693A1 (en) 2018-11-01
US10016046B2 (en) 2018-07-10
US20070035815A1 (en) 2007-02-15
HK1125339A1 (en) 2009-08-07
EP2335576A2 (en) 2011-06-22
JP2013019909A (en) 2013-01-31
EP2335576B8 (en) 2020-04-15
KR20080059157A (en) 2008-06-26
CN101287607A (en) 2008-10-15
AU2006279800B2 (en) 2012-03-29
EP1912796A2 (en) 2008-04-23
US20150196109A1 (en) 2015-07-16
WO2007021972A2 (en) 2007-02-22
EP1912795B1 (en) 2012-02-22
EP2335576A3 (en) 2017-05-03
JP5254790B2 (en) 2013-08-07
CA2985955C (en) 2020-07-07
CN101326064B (en) 2010-08-11
US20110270200A1 (en) 2011-11-03
US8007062B2 (en) 2011-08-30
US20210251370A1 (en) 2021-08-19
US8915562B2 (en) 2014-12-23
US11147357B2 (en) 2021-10-19
CA2985955A1 (en) 2007-02-22
CN101326064A (en) 2008-12-17
EP2335576B1 (en) 2020-03-04
US9247802B2 (en) 2016-02-02
AU2006279652A1 (en) 2007-02-22

Similar Documents

Publication Publication Date Title
AU2006279800B2 (en) System and method for medical monitoring and treatment through cosmetic monitoring and treatment
US10486174B2 (en) System and method for applying a reflectance modifying agent electrostatically to improve the visual attractiveness of human skin
AU2013200395B2 (en) System and method for applying a reflectance modifying agent to improve the visual attractiveness of human skin
US8942775B2 (en) Handheld apparatus and method for the automated application of cosmetics and other substances
ES2539035T3 (en) Makeup method and device to implement such method

Legal Events

Date Code Title Description
PC1 Assignment before grant (sect. 113)

Owner name: TCMS TRANSPARENT BEAUTY LLC

Free format text: FORMER APPLICANT(S): YEAGER, RICK

FGA Letters patent sealed or granted (standard patent)
MK14 Patent ceased section 143(a) (annual fees not paid) or expired