EP2488371A1 - Manufacturing security documents using 3d surface parameterization and halftone dithering - Google Patents
Manufacturing security documents using 3d surface parameterization and halftone ditheringInfo
- Publication number
- EP2488371A1 EP2488371A1 EP09744312A EP09744312A EP2488371A1 EP 2488371 A1 EP2488371 A1 EP 2488371A1 EP 09744312 A EP09744312 A EP 09744312A EP 09744312 A EP09744312 A EP 09744312A EP 2488371 A1 EP2488371 A1 EP 2488371A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- pixels
- printing
- image
- brightness
- texture
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Withdrawn
Links
Classifications
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B42—BOOKBINDING; ALBUMS; FILES; SPECIAL PRINTED MATTER
- B42D—BOOKS; BOOK COVERS; LOOSE LEAVES; PRINTED MATTER CHARACTERISED BY IDENTIFICATION OR SECURITY FEATURES; PRINTED MATTER OF SPECIAL FORMAT OR STYLE NOT OTHERWISE PROVIDED FOR; DEVICES FOR USE THEREWITH AND NOT OTHERWISE PROVIDED FOR; MOVABLE-STRIP WRITING OR READING APPARATUS
- B42D25/00—Information-bearing cards or sheet-like structures characterised by identification or security features; Manufacture thereof
- B42D25/40—Manufacture
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B42—BOOKBINDING; ALBUMS; FILES; SPECIAL PRINTED MATTER
- B42D—BOOKS; BOOK COVERS; LOOSE LEAVES; PRINTED MATTER CHARACTERISED BY IDENTIFICATION OR SECURITY FEATURES; PRINTED MATTER OF SPECIAL FORMAT OR STYLE NOT OTHERWISE PROVIDED FOR; DEVICES FOR USE THEREWITH AND NOT OTHERWISE PROVIDED FOR; MOVABLE-STRIP WRITING OR READING APPARATUS
- B42D25/00—Information-bearing cards or sheet-like structures characterised by identification or security features; Manufacture thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N1/00—Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
- H04N1/40—Picture signal circuits
- H04N1/405—Halftoning, i.e. converting the picture signal of a continuous-tone original into a corresponding signal showing only two levels
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N1/00—Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
- H04N1/40—Picture signal circuits
- H04N1/405—Halftoning, i.e. converting the picture signal of a continuous-tone original into a corresponding signal showing only two levels
- H04N1/4055—Halftoning, i.e. converting the picture signal of a continuous-tone original into a corresponding signal showing only two levels producing a clustered dots or a size modulated halftone pattern
-
- B42D2035/06—
-
- B42D2035/26—
Definitions
- the invention relates to a method for manufacturing a security document by applying an object thereon.
- it relates to a technique where three dimensional object is applied onto the security document using a three-dimensional projection.
- the inven tion also relates to a security document manufactured by using this technique.
- Another class of techniques includes the application of special rendering techniques, such as special halftone dithering mechanisms.
- special rendering techniques such as special halftone dithering mechanisms.
- An example of such a technique is described in US 6 198 545.
- Disclosure of the Invention is to provide another rendering technique for manufacturing security documents .
- the method comprises the following steps:
- Such a surface model can e.g. be a real object, such as a building or a person, or it may be a virtual object, such as a torus.
- the surface model is e.g. represented by a mesh of nodes in three-dimensional space.
- a brightness or color Attributing, to each point on said surface model, a brightness or color:
- the color or brightness of the surface model can e.g. correspond to the coloring of the object it represents, and/or it may be derived from a shading of the object, i.e. the play of shadow and light on the object.
- the brightness or color is typically non ⁇ uniform over the object.
- a bijective mapping function for mapping said surface model into a two-dimensional parameter space, thereby defining a brightness or color image in said parameter space, wherein the color or brightness of each point in said image is given by the brightness or color of the point on the surface model corresponding to the point of the image:
- the bijective mapping function maps each point of the surface model into the two- dimensional parameter space, thereby defining the brightness or color image therein.
- step d) Rendering, in said parameter space, said image as a halftone screen pattern, which halftone screen pattern comprises an array of texture pixels, with each texture pixel being attributed to a point in said parameter space:
- the image created in step c) is "drawn" into an array of pixels, the so-called texture pixels, using a halftone screen pattern.
- step d) Using an inverse of the mapping function for mapping the texture pixels onto the surface model:
- the texture pixels obtained in step d) are now mapped back onto the original surface model.
- each point on the surface model now has a color or brightness defined by its corresponding texture pixel.
- the texture pixels are applied as a texture to the surface model .
- step f) Rendering the texture pixels in said printing plane into a two-dimensional array of printing pixels:
- the texture pixels as obtained in step f) are now rendered into the printing pixels (i.e. the pixels that will finally be printed onto the document) .
- step g) Applying said printing pixels to said se- curity document:
- the printing pixels of step g) are finally applied by suitable techniques to the document.
- This procedure will create a unique textured appearance of the model in the ' applied image, with the texture being such that it depends and represents not only the shape, but also the color or brightness of the original model.
- the invention also relates to a security document manufactured by this method.
- a security document may e.g. be a banknote, check or identification document.
- the document need not necessarily be of paper, but may e.g. also be of plastics.
- the document may be a credit card or identification card.
- Fig. 1 shows a banknote
- Fig. 2 shows an example of a surface model with shading
- Fig. 3 shows the surface model mapped into the parameter space
- Fig. 4 shows the halftone screen pattern in parameter space
- Fig. 5 shows an enlarged section of Fig. 4
- Fig. 6 shows the final print result
- Fig. 7 shows an enlarged section of Fig. 6.
- Fig. 1 shows a banknote having a sheet-like carrier 1 of plastic and/or paper material and carrying various features 2 - 5 thereon, such as
- Step a providing a surface model:
- a model of a non-flat two- dimensional surface model in three-dimensional space is provided.
- This surface model typically (but not necessarily) represents a real-world object. In the example of Fig. 2, it represents a cup, but it may e.g. also represent other types of objects, such as buildings, coins, animals or people.
- the security document is a document of identification, such as a passport or identity card, it may also represent the face of the bearer as scanned by suitable machinery.
- the model may be the same on each document of a series of the identical document, as in the example of the banknote shown in Fig. 1. Alternatively, it may be individual to each document, such as in the example of the document of identification mentioned above. An individual model for each document has the advantage that a counterfeiter cannot copy the document by scanning with ⁇ out applying the manufacturing steps described here.
- the surface model is typically represented by a series of nodes (each of which is characterized by a three-dimensional coordinate) and by a series of faces spanned by the nodes.
- nodes each of which is characterized by a three-dimensional coordinate
- faces spanned by the nodes.
- a mesh of triangles or quadlaterals can be used, where each corner of each triangle or quadlateral is given by a node.
- Step b attributing color or brightness:
- Each (visible) point of the surface should have a color or at least a brightness value attributed to it.
- a color may e.g. be expressed as a tuple of coordinates in a color space (such as RGB or CMYK) , while a single coordinate (such as gray level) is typically sufficient for brightness.
- the brightness or color may e.g. correspond to the brightness or color in real live. It can be a calculated value, such as in the example of Fig. 2, where the gray level for each (visible) point on the surface was calculated using conventional shading techniques, or it may have been recorded while scanning a real-world ob- j ect .
- the model including its color and brightness values, is to be mapped to a two- dimensional parameter space, i.e. a space having two coordinates, e.g. called u, v.
- This process is typically called “parameterization” and is e.g. used for applying a texture to a three-dimensional object.
- mapping function must be provided to map each point of the surface model into parameter space.
- the mapping function must be bijective (for points that are visible on the surface, at least) , so that its inverse can be used for the step e below.
- mapping function should advantageously minimize distortions, although some distortions cannot be0 avoided for non-trivial surface models .
- mappings are conformal mappings, such as "least-squares conformal” mappings, or isometric mappings (length-invariable mappings) , or equiareal mappings (mappings that leave areas invariant) .
- mapping functions are known to the skilled person and e.g. described by:
- mapping functions are known to the skilled person.
- mapping into parameter space can be a fully automated or semi-automatic process.
- it may e.g. be advantageous for a human operator to select certain parameters of the mapping function, such as the location of clipping lines, or to identify regions that should be stretched or compressed during mapping, in order to minimize visual distortions or artifacts.
- mapping of the surface model with its color or brightness into parameter space defines a color or brightness image in parameter space, with the color or brightness of each point in said image given by the color or brightness of the point on the surface model corresponding to the point of the image.
- FIG. 3 Such a color or brightness image for the model of Fig. 2 is shown in Fig 3.
- the image comprises three distinct sections 10, 11, 12.
- a first section 10 corresponds to the inner and outer surfaces, and the upper rim of the cup's body.
- a second section 11 corresponds to the cup's handle, and a third section 12 to the cup's bottom side.
- Step d halftone screening in parameter space:
- the image obtained in step c is rendered as a halftone screen pattern into an array of texture pixels.
- the image can be divided into a plurality of halftone cells, with each halftone cell spanning a plurality of pixels.
- the pixels within each cell are switched on or off as a function of the color or brightness of the image in the given cell, in particular as a function of the average color or average brightness of the image in the given cell.
- the average gray scale factor of the original image at a given cell is calculated and then a number of pixels proportional to this factor is switched on in the cell, while the others remain switched off.
- the image is divided into a plurality of vector-based symbols, with each symbol having a parameter that varies the coverage of the symbol.
- the symbols can e.g. be stars drawn as outlines, and the parameter can e.g. be the line thickness of the star's outline.
- the parameter can e.g. be 0, in which case the star is not drawn at all, or 1, in which case the star is drawn with such a large outline thickness that it basically covers all its assigned area, or the parameter can be any value in between.
- the parameter is varied according to the average gray scale factor of the original image at the area assigned to each symbol, and then the symbols are rendered into the texture pixels.
- the stars at dark regions are rendered with a thick line
- the stars at light regions are rendered with a thin line, thereby blackening more pixels in dark regions of the image.
- the object may e.g. be a solid (filled) star whose size varies according to the average gray level of the image.
- the second embodiment basically divides the image into a plurality of halftone cells, with each cell corresponding to the area assigned to a symbol.
- the texture pixels of a cell are again switched on or off as a function of a brightness or color level of said brightness or color image in said cell
- the algorithm for switching the pixels of a cell on or off as a function of the color or brightness of the image at said cell is the same for all cells.
- all cells representing a given brightness look a the same, which allows to generate a texture of high regularity that emphasizes the curvature of the model surface in the final document.
- Figs. 4 and 5 show a representation of the texture pixels in parameter space after halftone screen- ing.
- the present example is based on substantially quadratic pixel cells, with each cell having a roughly star-shaped cluster of white pixels in the center and dark pixels at its edges.
- the number of dark pixels in a cell is a function of the average gray level of the image in the given cell - the darker the image is in the given cell, the higher the number of dark pixel is.
- step c the inverse of the mapping function of step c is used for mapping the texture pixel of step d back onto the surface model.
- each texture pixel is attributed to a point on the surface model, such that it can be projected as described in the following step.
- Step f projecting the surface model:
- the surface model with its texture is projected into a two-dimensional printing plane, i.e. in a space whose coordinates correspond to locations on the document to be manufactured.
- Projection takes place by means of a three- dimensional projection, which is a mapping of the three- dimensional surface model into two-dimensional space by means of orthographic or perspective projection.
- This type of mapping corresponds to the mapping performed by a camera and generates an object in the plane that, when viewed, is perceived by the viewer as having depth.
- Orthographic and perspective projections are well known to the skilled person and need not be ex ⁇ plained in detail herein.
- each point in the printing plane is attributed to one of the texture pixels .
- Step g rendering in the printing plane:
- step f the printing plane obtained in step f is rendered into a two-dimensional array of printing pixels.
- each printing pixel is e.g. represented by a binary value that in ⁇ dicates if the given pixel is on or off.
- a suitable algorithm can e.g. calculate the average value (weighted by overlap) of all texture pixels overlapping the given printing pixel, and then switch the printing pixel on if the average value is above a given threshold.
- an ink dot is applied to the given location of the document.
- each printing pixel is e.g. represented by an n-tuple of binary values indicating which of n differently colored inks is/are to be applied at the given location.
- Step h application to ducument :
- step g the printing pixels of step g are physically applied the document using suitable techniques, such as:
- a diffractive structure such as a hologram, or any other type of structure, such as a structured metal layer, that is structured according to the printing pixels, - manufacturing a micro-engraving according to the printing pixels,
- the above techniques can be used to apply the printing pixels "directly” or "indirectly” to the document.
- the printing pixels are applied directly to the document itself, e.g. by means of printing ink on the document.
- the printing pixels are applied to a carrier other than the document, such as a separate security element, which carrier is then attached to the document by suitable techniques, such as gluing.
- the printing pixels can be applied to the document in such a manner that the rendered image can be verified by eye, or, alternatively, in such a manner that further tools are required to view them, such as optical polarizers, microlens-arrays or devices for making special inks visible.
- Figs. 6 and 7. an apparently three-dimensional object is applied onto the document.
- the object carries a texture, namely the texture calculated in step d and mapped onto the object in step e.
- the texture varies in such a manner that its brightness changes. This brightness is, in its turn, chosen such that it imitates light and shadow playing on the surface of the object.
- the three-dimensional effect conveyed by the texture is strongest if the structure of the texture remains visible and crisp in the final result. This can be achieved if the resolution of the printing pixels (the printer resolution) is sufficiently fine. More precisely, the resolution of the printing pixels should be suffi- ciently large such that, for a majority of the halftone cells, each cell extends over a plurality of printing pixels. In particular, in order to resolve the individual cells in the final printed product optimally, for a ma- jority of the halftone cells, each cell should extend over a number of printing pixels being at least equal to (in particular being larger than) the number of texture pixels in the cell. Notes:
- the present method allows to create a two- dimensional rendering of a three-dimensional textured object where the texture not only serves to emphasize the curvature and orientation of the object's surface, but also serves to encode the color or brightness of the same.
- the brightness of an object is not encoded in its texture, but overlaid with said texture, and halftone printing techniques need to be applied to the object af- ter its three-dimensional projection into the printing plane, which affects the clarity and crispness of the result .
- the surface model can be common for all of them.
- all banknotes of a given denomination and series may use an identical object seen from the same position and using the same texture.
- the surface model may be unique for each security document in a series of documents.
- the surface model may be a scanned version of the face of the bearer of the document. This makes counterfeiting particularly difficult because a counterfeiter needs to be able to replicate the present mechanism.
- the printing pixels may be applied to a blank or uniform background, or they may be applied to or superimposed with a structure, e.g. a patterned background.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Computer Graphics (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Manufacturing & Machinery (AREA)
- Facsimile Image Signal Circuits (AREA)
Abstract
Description
Claims
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/CH2009/000330 WO2011044704A1 (en) | 2009-10-15 | 2009-10-15 | Manufacturing security documents using 3d surface parameterization and halftone dithering |
Publications (1)
Publication Number | Publication Date |
---|---|
EP2488371A1 true EP2488371A1 (en) | 2012-08-22 |
Family
ID=41698079
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP09744312A Withdrawn EP2488371A1 (en) | 2009-10-15 | 2009-10-15 | Manufacturing security documents using 3d surface parameterization and halftone dithering |
Country Status (2)
Country | Link |
---|---|
EP (1) | EP2488371A1 (en) |
WO (1) | WO2011044704A1 (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109334216A (en) * | 2014-10-29 | 2019-02-15 | 惠普发展公司,有限责任合伙企业 | Three-dimensional halftone |
CN109716344A (en) * | 2016-09-16 | 2019-05-03 | 惠普发展公司有限责任合伙企业 | Indicate the data set of the various aspects of 3D object |
Families Citing this family (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10890692B2 (en) | 2011-08-19 | 2021-01-12 | Visual Physics, Llc | Optionally transferable optical system with a reduced thickness |
KR101981833B1 (en) | 2012-04-25 | 2019-05-23 | 비쥬얼 피직스 엘엘씨 | Security device for projecting a collection of synthetic images |
KR102014576B1 (en) | 2012-08-17 | 2019-08-26 | 비쥬얼 피직스 엘엘씨 | A process for transferring microstructures to a final substrate |
BR112015004922A2 (en) * | 2012-09-05 | 2017-07-04 | Lumenco Llc | pixel mapping, arrangement, and image processing for circle and square based micro-lens arrays to achieve maximum 3d volume and multi-directional motion |
RU2673137C9 (en) | 2013-03-15 | 2019-04-04 | Визуал Физикс, Ллс | Optical security device |
US9873281B2 (en) | 2013-06-13 | 2018-01-23 | Visual Physics, Llc | Single layer image projection film |
US10766292B2 (en) | 2014-03-27 | 2020-09-08 | Crane & Co., Inc. | Optical device that provides flicker-like optical effects |
MX2016012305A (en) | 2014-03-27 | 2017-02-23 | Visual Physics Llc | An optical device that produces flicker-like optical effects. |
CA3230729A1 (en) | 2014-07-17 | 2016-01-21 | Visual Physics, Llc | An improved polymeric sheet material for use in making polymeric security documents such as banknotes |
DE102015100280A1 (en) | 2015-01-09 | 2016-07-14 | Ovd Kinegram Ag | Method for the production of security elements and security elements |
CA2976218C (en) | 2015-02-11 | 2023-02-14 | Crane & Co., Inc. | Method for the surface application of a security device to a substrate |
EP3243668B1 (en) * | 2016-05-10 | 2020-07-15 | Agfa Nv | Manufacturing of a security document |
MX2019009459A (en) | 2017-02-10 | 2019-12-16 | Crane & Co Inc | Machine-readable optical security device. |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6198545B1 (en) * | 1994-03-30 | 2001-03-06 | Victor Ostromoukhov | Method and apparatus for generating halftone images by evolutionary screen dot contours |
-
2009
- 2009-10-15 WO PCT/CH2009/000330 patent/WO2011044704A1/en active Application Filing
- 2009-10-15 EP EP09744312A patent/EP2488371A1/en not_active Withdrawn
Non-Patent Citations (1)
Title |
---|
See references of WO2011044704A1 * |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109334216A (en) * | 2014-10-29 | 2019-02-15 | 惠普发展公司,有限责任合伙企业 | Three-dimensional halftone |
CN109334216B (en) * | 2014-10-29 | 2020-01-14 | 惠普发展公司,有限责任合伙企业 | Additive manufacturing method for producing an object from a digital representation of the object |
CN109716344A (en) * | 2016-09-16 | 2019-05-03 | 惠普发展公司有限责任合伙企业 | Indicate the data set of the various aspects of 3D object |
Also Published As
Publication number | Publication date |
---|---|
WO2011044704A1 (en) | 2011-04-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2011044704A1 (en) | Manufacturing security documents using 3d surface parameterization and halftone dithering | |
CA2453456C (en) | Images incorporating microstructures | |
US5371627A (en) | Random dot stereogram and method for making the same | |
US9092872B2 (en) | System and method for creating an animation from a plurality of latent images encoded into a visible image | |
EP1690697A1 (en) | Method to apply an invisible mark on a media | |
US11037038B2 (en) | Artwork generated to convey digital messages, and methods/apparatuses for generating such artwork | |
US20030021437A1 (en) | Images and security documents protected by micro-structures | |
AU2002345270A1 (en) | Images incorporating microstructures | |
JP2005512846A (en) | Certificate of value | |
WO2010032718A1 (en) | Forgery preventive printed matter, method for producing same, and recording medium in which dot data creation software is stored | |
JP4844894B2 (en) | Three-dimensional moire formation | |
EP2953796A1 (en) | Multiple shade latent images | |
CN114746904A (en) | Three-dimensional face reconstruction | |
JP2018504634A (en) | Method for forming a security element and security document | |
KR101123648B1 (en) | Printing method for expressing solid texture of oil painting | |
JP6134927B1 (en) | Image generation method, image generation apparatus, engraving manufacturing method, program, engraving and printed matter | |
JPWO2020096009A1 (en) | Moire visualization pattern generation method, moire visualization pattern generation device, and moire visualization pattern generation system | |
CN110869555B (en) | Paper preparation method capable of controlling watermark brightness and paper product | |
JP7185876B2 (en) | latent image print | |
JP4229270B2 (en) | Method and apparatus for simulating printed matter using two-dimensional data | |
EP2862345B1 (en) | Simulated embossing and imprinting | |
US20190389243A1 (en) | Optical illusion device | |
EP4360901A1 (en) | A security element and a data carrier | |
WO2003042923A1 (en) | Logic arrangements, storage mediums, and methods for generating digital images using brush strokes | |
Maciejewski et al. | Aesthetics of Hand-Drawn vs. Computer-Generated Stippling. |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
17P | Request for examination filed |
Effective date: 20120414 |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO SE SI SK SM TR |
|
DAX | Request for extension of the european patent (deleted) | ||
17Q | First examination report despatched |
Effective date: 20150311 |
|
GRAP | Despatch of communication of intention to grant a patent |
Free format text: ORIGINAL CODE: EPIDOSNIGR1 |
|
RIC1 | Information provided on ipc code assigned before grant |
Ipc: H04N 1/405 20060101ALI20150723BHEP Ipc: G06T 15/00 20110101ALI20150723BHEP Ipc: B42D 25/00 20140101AFI20150723BHEP |
|
INTG | Intention to grant announced |
Effective date: 20150814 |
|
GRAP | Despatch of communication of intention to grant a patent |
Free format text: ORIGINAL CODE: EPIDOSNIGR1 |
|
INTG | Intention to grant announced |
Effective date: 20160628 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN |
|
18D | Application deemed to be withdrawn |
Effective date: 20161109 |