Automated system for image archiving

Abstract
A method for producing universal object tracking implementations. This invention provides a functional implementation, from which any object-producing device can construct automatically generated archival enumerations. This implementation uses an encoding schemata based on location numbers, object numbers, and parent numbers. Location numbers encode information about logical sequence in the archive, object numbers encode information about the physical attributes of an object, and parent numbers record the conception date and time of a given object's parent. Parent-child relations are algorithmically derivable from location and parent number relationships, thus providing fully recoverable, cumulative object lineage information. Encoding schemata are optimized for use with all current 1, 2, and 3 dimensional barcode symbologies to facilitate data transportation across disparate technologies (e.g., imaging devices, card readers/producers, printers, medical devices). The implemented encoding schemata of this invention supports all manner of object forming devices such as image forming devices, medical devices and computer generated objects.
Description


FIELD OF INVENTION

[0002] This invention relates generally to archive, documentation and location of objects and data. More particularly this invention is a universal object tracking system wherein generations of objects which may be physical, electronic, digital, data and images can be related one to another and to original objects that contributed to a object without significant user intervention.



BACKGROUND OF THE INVENTION

[0003] Increasingly, images of various types are being used in a wide variety of industrial, digital, medical, and consumer uses. In the medical field, telemedicine has made tremendous advances that now allow a digital image from some medical sensor to be transmitted to specialists who have the requisite expertise to diagnose injury and disease at locations remote from where the patient lies. However, it can be extremely important for a physician, or indeed any other person to understand how the image came to appear as it does. This involves a knowledge of how the image was processed in order to reach the rendition being examined. In certain scientific applications, it may be important to “back out” the effect of a particular type of processing in order to more precisely understand the appearance of the image when first made.


[0004] Varieties of mechanisms facilitate storage and retrieval of archival information relating to images. However, these archival numbering and documentation schemes suffer from certain limitations. For example, classificatory schemata are used to facilitate machine sorting of information about a subject (“subject information”) according to categories into which certain subjects fit. Additionally tracking information, that is, information concerning where the image has been or how the image was processed, is also used together with classificatory schemata.


[0005] However, relying on categorizing schemata is inefficient and ineffective. On the one hand, category schemata that are limited in size (i.e. number of categories) are convenient to use but insufficiently comprehensive for large-scale applications, such as libraries and national archives. Alternatively if the classificatory schemata is sufficiently comprehensive for large-scale applications, it may well be far too complicated, and therefore inappropriate for small scale applications, such as individual or corporate collections of image data.


[0006] It is also an approach to provide customizable enumeration strategies to narrow the complexity of large-scale systems and make them discipline specific. Various archiving schemes are developed to suit a particular niche or may be customizable for a niche. This is necessitated by the fact that no single solution universally applies to all disciplines, as noted above. However, the resulting customized archival implementation will differ from, for example, a medical image to a laboratory or botanical image archive. The resulting customized image archive strategy may be very easy to use for that application but will not easily translate to other application areas.


[0007] Thus, the utility provided by market niche image archiving software simultaneously makes the resulting applications not useful to a wide spectrum of applications. For example, tracking schemata that describes art history categories might not apply to high-tech advertising.


[0008] Another type of archival mechanism suffering from some of the difficulties noted above is equipment-specific archiving. In this implementation a particular type of image device, such as a still camera, a video camera, a digital scanner, or other form of imaging means has its own scheme for imprinting or recording archival information relating to the image that is recorded. This is compounded when an image is construed as a data object(s) such as a collection of records related to a given individual, distributed across machines and databases. In the case with distributed data tracking, records (represented as images) may necessarily exist on multiple machines of the same type as well as multiple types of machines.


[0009] Thus, using different image-producing devices in the image production chain can cause major problems. For example, mixing traditional photography (with its archive notation) with digital touch-up processing (with its own different archive notation). Further, equipment-specific archive schemes do not automate well, since multiple devices within the same archive may use incompatible enumeration schemata.


[0010] Certain classification approaches assume single device input. Thus, multiple devices must be tracked in separate archives, or are tracked as archive exceptions. This makes archiving maintenance more time consuming and inefficient. For example, disciplines that use multiple cameras concurrently, such as sports photography and photo-journalism, confront this limitation.


[0011] Yet other archive approaches support particular media formats, but not multiple media formats simultaneously occurring in the archive. For example, an archive scheme may support conventional silver halide negatives but not video or digital media within the same archive.


[0012] Thus, this approach fails when tracking the same image across different media formats, such as tracking negative, transparency, digital, and print representation of the same image.


[0013] Yet another archive approach may apply to a particular state of the image, as the initial or final format, but does not apply to the full life-cycle of all images. For example, some cameras time- and date-stamp negatives, while database software creates tracking information after processing. While possibly overlapping, the enumeration on the negatives differs from the enumeration created for archiving. In another example, one encoding may track images on negatives and another encoding may track images on prints. However, such a state-specific approach makes it difficult automatically to track image histories and lineages across all phases of an image's life-cycle, such as creation, processing, editing, production, and presentation. Similarly, when used to track data such as that occurring in distributed databases, such a system does not facilitate relating all associated records in a given individual's personal record archive.


[0014] Thus, tracking information that uses different encoding for different image states is not particularly effective since maintaining multiple enumeration strategies creates potential archival error, or at a minimum, will not translate well from one image form to another.


[0015] Some inventions that deal with recording information about images have been the subject of U.S. patents in the past. U.S. Pat. No. 5,579,067 to Wakabayashi describes a “Camera Capable of Recording Information.” This system provides a camera which records information into an information recording area provided on the film that is loaded in the camera. If information does not change from frame to frame, no information is recorded. However, this invention does not deal with recording information on subsequent processing.


[0016] U.S. Pat. No. 5,455,648 to Kazami was granted for a “° Film Holder or for Storing Processed Photographic Film.” This invention relates to a film holder which also includes an information holding section on the film holder itself. This information recording section holds electrical, magnetic, or optical representations of film information. However, once the information is recorded, it is to used for purposes other than to identify the original image.


[0017] U.S. Pat. No. 5,649,247 to Itoh was issued for an “Apparatus for Recording Information of Camera Capable of Optical Data Recording and Magnetic Data Recording.” This patent provides for both optical recording and magnetic recording onto film. This invention is an electrical circuit that is resident in a camera system which records such information as aperture value, shutter time, photo metric value, exposure information, and other related information when an image is first photographed. This patent does not relate to recording of subsequent operations relating to the image.


[0018] U.S. Pat. No. 5,319,401 to Hicks was granted for a “Control System for Photographic Equipment.” This invention deals with a method for controlling automated photographic equipment such as printers, color analyzers, film cutters. This patent allows for a variety of information to be recorded after the images are first made. It mainly teaches methods for production of pictures and for recording of information relating to that production. For example, if a photographer consistently creates a series of photographs which are off center, information can be recorded to offset the negative by a pre-determined amount during printing. Thus the information does not accompany the film being processed but it does relate to the film and is stored in a separate database. The information stored is therefore not helpful for another laboratory that must deal with the image that is created.


[0019] U.S. Pat. No. 5,193,185 to Lanter was issued for a “Method and Means for Lineage Tracing of a Spatial Information Processing and Database System.”. This Patent relates to geographic information systems. It provides for “parent” and “child” links that relate to the production of layers of information in a database system. Thus while the this patent relates to computer-generated data about maps, it does not deal with how best to transmit that information along a chain of image production.


[0020] U.S. Pat. No. 5,008,700 to Okamoto was granted for a “Color Image Recording Apparatus using Intermediate Image Sheet.” This patent describes a system, where a bar code is printed on the image production media which can then be read by an optical reader. This patent does not deal with subsequent processing of images which can take place or recording of information that relates to that subsequent processing.


[0021] U.S. Pat. No. 4,728,978 was granted to Inoue for a “Photographic Camera.” This patent describes a photographic camera which records information about exposure or development on an integrated circuit card which has a semiconductor memory. This card records a great deal of different types of information and records that information onto film. The information which is recorded includes color temperature information, exposure reference information, the date and time, shutter speed, aperture value, information concerning use of a flash, exposure information, type of camera, film type, filter type, and other similar information. The patent claims a camera that records such information with information being recorded on the integrated circuit court. There is no provision for changing the information or recording subsequent information about the processing of the image nor is there described a way to convey that information through many generations of images.


[0022] Thus a need exists to provide a uniform tracking mechanism for any type of image, using any type of image-producing device, which can describe the full life-cycle of an image and which can translate between one image state and another and between one image forming mechanism and another. Such a mechanism should apply to any type of object, relationship, or data that can be described as an image.



SUMMARY OF THE INVENTION

[0023] It is therefore an object of the present invention to create an archival tracking method that includes relations, descriptions, procedures, and implementations for universally tracking objects, entities, relationships, or data able to be described as images.


[0024] It is a further object of the present invention to create a tracking method capable of describing any object, entity, or relationship or data that can be described as an image of the object, entity, relationship or data.


[0025] It is a further object of the present invention to create an encoding schemata that can describe and catalogue any image produced on any media, by any image producing device, that can apply to all image producing disciplines and objects, entities, relationships, or data able to be described as images for corresponding disciplines.


[0026] It is a further object of the present invention to implement to archival scheme on automated data processing means that exist within image producing equipment.


[0027] It is a further object of the present invention to apply to all image-producing devices including devices producing objects, entities, relationships, or data that is able to be described as images.


[0028] It is a further object of the present invention to support simultaneous use of multiple types of producing devices including devices producing objects, entities, relationships, or data that is able to be described as images.


[0029] It is a further object of the present invention to support simultaneous use of multiple producing devices of the same type.


[0030] It is a further object of the present invention to provide automatic parent-child encoding.


[0031] It is a further object of the present invention to track image lineages and family trees.


[0032] It is a further object of the present invention to provide a serial and chronological sequencing scheme that uniquely identifies all objects, entities, relationships, or data that is able to be described as images in an archive.


[0033] It is a further object of present invention to provide an identification schemata that describes physical attributes of all objects, entities, relationships, or data that is able to be described as images in an archive.


[0034] It is a further object of the present invention to separate classificatory information from tracking information.


[0035] It is a further object of the present invention to provide an enumeration schemata applicable to an unlimited set of media formats used in producing objects, entities, relationships, or data that is able to be described as images.


[0036] It is a further object of the present invention to apply the archival scheme to all stages of life-cycle, from initial formation to final form of objects, entities, relationships, or data that is able to be described as images.


[0037] It is a further object of the present invention to create self-generating archives, through easy assimilation into any device including devices producing objects, entities, relationships, or data that is able to be described as images.


[0038] It is a further object of the present invention to create variable levels of tracking that are easily represented by current and arriving barcode symbologies, to automate data transmission across different technologies (e.g., negative to digital to print).


[0039] These and other objects of the present invention will become clear to those skilled in the art from the description that follows.



BRIEF DESCRIPTION OF THE INVENTION

[0040] The present invention is a universal object tracking method and apparatus for tracking and documenting objects, entities, relationships, or data that is able to be described as images through their complete life-cycle, regardless of the device, media, size, resolution, etc., used in producing them.


[0041] Specifically, the automated system for image archiving (“ASIA”) encodes, processes, and decodes numbers that characterize objects, entities, relationships, or data that is able to be described as images and related data. Encoding and decoding takes the form of a 3-number association: 1) location number (serial and chronological location), 2) image number (physical attributes), and 3) parent number (parent-child relations).



Flexibility of Encoding

[0042] A key aspect of the present inventions is that any implementation of the system and method of the present invention is interoperable with any other implementation. For example, a given implementation may use the encoding described herein as complete database record describing a multitude of images. Another implementation may use the encoding described herein for database keys describing medical records. Another implementation may use the encoding described herein for private identification. Still another may use ASIA encoding for automobile parts-tracking. Yet all such implementations will interoperate.


[0043] This design of the present invention permits a single encoding mechanism to be used in simple devices as easily as in industrial computer systems.



Parent-Child Relations

[0044] The system and method of the present invention includes built-in “parent-child” encoding that is capable of tracking parent-child relations across disparate producing mechanisms. This native support of parent-child relations is included in the encoding structure, and facilitates tracking diverse relations, such as record transactions, locations of use, image derivations, database identifiers, etc. Thus, it is a function of the present invention that parent-child relations are used to track identification of records from simple mechanisms, such as cameras and TVs through diverse computer systems.



Uniqueness

[0045] The system and method of the present invention possesses uniqueness that can be anchored to device production. Thus the encoding described herein can bypass the problems facing “root-registration” systems (such as facing DICOM in the medical X-ray field). Additionally, the encoding described herein can use a variety of ways to generate uniqueness. Thus it applies to small, individual devices (e.g. cameras), as well as to fully automated, global systems (e.g., universal medical records).


[0046] In global systems, uniqueness resembles Open Software Foundation's DCE UUIDs (and Microsoft GUIDs), except that the encoding described herein includes a larger logical space. Such design facilitates interoperability across wide domains of application: e.g., from encoding labels for automatically generated photographic archival systems, to secure identification, to keys for distributed global database systems. This list is not meant as a limitation and is only illustrative in nature. Other applications will be apparent to those skilled in the art from a review of the specification herein.



Media Traversability

[0047] The encoding described herein applies equally well to “film” and “filmless” systems, or other such distinctions. This permits the same encoding mechanism for collections of records produced on any device, not just digital devices. Tracking systems can thus track tagged objects as well as digital objects. Similarly, since the encoding mechanism is anchored to device-production, supporting a new technology is as simple as adding a new device. This in turn permits the construction of comprehensive, automatically generated tracking mechanism to be created and maintained, which require no human effort aside from the routine usage of the devices.


[0048] Using the encoding described herein, a simple implementation of the system requires less than 2K of code space. A more complex industrial implementation requires less than 200K of code space.



DEFINITIONS

[0049] The following definitions apply throughout this specification:


[0050] Field: A record in an ASIA number.


[0051] Component: A set of fields grouped according to general functionality.


[0052] Location component: A component that identifies logical location.


[0053] Parent component: A component that characterizes a relational status between an object and it's parent.


[0054] Image component: A component that identifies physical attributes.


[0055] Schema: A representation of which fields are present in length and encoding.


[0056] Length: A representation of the lengths of fields that occur in encoding.


[0057] Encoding: Data, described by length and encoding.



BRIEF DESCRIPTION OF THE FIGURES

[0058]
FIG. 1. illustrates an overview of the present invention


[0059]
FIG. 1A. illustrates the overarching structure organizing the production of system numbers


[0060]
FIG. 1B. illustrates the operation of the system on an already existing image


[0061]
FIG. 2 illustrates the formal relationship governing encoding and decoding


[0062]
FIG. 3 illustrates the encoding relationship of the present invention


[0063]
FIG. 4 illustrates the relationships that characterize the decoding of encoded information


[0064]
FIG. 5 illustrates the formal relations characterizing all implementations of the invention


[0065]
FIG. 6 illustrates the parent-child encoding of the present invention in example form


[0066]
FIG. 7 illustrates the processing flow of ASIA







DETAILED DESCRIPTION OF THE INVENTION

[0067] The present invention is a method and apparatus for formally specifying relations for constructing image tracking mechanisms, and providing an implementation that includes an encoding schemata for images regardless of form or the equipment on which the image is produced.


[0068] The numbers assigned by the system and method of the present invention are automatically generated unique identifiers designed to uniquely identify objects within collections thereby avoiding ambiguity. When system numbers are associated with objects they are referred to herein as “tags.” Thus, the expression “system tag encoding” refers to producing and associating system numbers with objects.


[0069]
FIG. 1A illustrates the overarching structure organizing the production of system numbers.


[0070] The present invention is organized by three “components”: “Location,” “Image,” and “Parent.” These heuristically group 18 variables are called “fields.” Fields are records providing data, and any field (or combination of fields) can be encoded into an system number. The arrow ( ) labeled ‘Abstraction’ points in the direction of increasing abstraction. Thus, the Data Structure stratum provides the most concrete representation of the system and method of the present invention, and the Heuristics stratum provides the most abstract representation of the system of present invention.


[0071] As noted above, the system and method of the present invention comprises and organizational structure includes fields, components, and relations between fields and components. The following conventions apply and/or govern fields and components:


[0072] Base 16 schema. Base 16 (hex) numbers are used, except that leading ‘0’ characters are excluded in encoding implementations. Encoding implementations MUST strip leading ‘0’ characters in base 16 numbers.


[0073] Decoding implementations MUST accept leading ‘0’ characters in base 16 numbers.


[0074] UTF-8 Character Set —The definition of “character” in this specification complies with RFC 2279. When ‘character’ is used, such as in the expression “uses any character”, it means “uses any RFC 2279 compliant character”.


[0075] Software Sets —ASIA implementations inherit core functionality unless otherwise specified through the image component's set field. See Section 2.2.3.3 Image for details about set.


[0076] Representation Constraints —Certain constraints MAY be imposed on field representations, such as the use of integers for given fields in given global databases. This will be handled on a per domain basis, such as by the definition of a given global database. When needed, given applications MAY need to produce multiple ASIA numbers for the same item, to participate in multiple databases.



Fields

[0077] A field is a record in an ASIA number. Using any field (or collection of fields) MAY (1) distinguish one from another ASIA number, and (2) provide uniqueness for a given tag in a given collection of tags.


[0078] ASIA compliance requires the presence of any field, rather than any component. Components heuristically organize fields.



Components

[0079] A component is a set of fields grouped according to general functionality. Each component has one or more fields. ASIA has three primary components: location, image, and parent.
1TABLE 1ComponentsComponentDescriptionLocationLogical locationParentParent informationImagePhysical attributes


[0080] Referring to Table 1 (above) Components are illustrated. This table 1 lists components and their corresponding description. The following sections specifically describe components and their corresponding fields.


[0081] A tag's location (see Table 1) component simply locates an ASIA number within a given logical space, determined by a given application. The characteristics of the location component are illustrated below in Table 2.
2TABLE 2LocationFieldDescriptionRepresentationGenerationFamily relation depthUses any characterSequenceEnumeration of groupUses any characterTimeDate/time madeUses any characterAuthorProducing agentUses any characterDeviceDevice usedUses any characterUnitEnumeration in groupUses any characterRandomNontemporal uniquenessUses any characterCustomReserved for applicationsUses any character


[0082] Table 2 Location (above) lists location fields, descriptions, and representation specifications.


[0083] The following definitions apply to component fields.


[0084] generation —identifies depth in family relations, such as parent-child relations. For example, ‘1’ could represent “first generation”, ‘2’ could represent “second generation”, and so forth.


[0085] sequence —serially enumerates a group among groups. For example, sequence could be the number of a given roll of 35 mm film in a photographer's collection.


[0086] time —date-stamps a number. This is useful to distinguish objects of a given generation. For example, using second enumeration could (“horizontally”) distinguish siblings of a given generation.


[0087] Author —identifies the producing agent or agents. For example, a sales clerk, equipment operator, or manufacturer could be authors.


[0088] device —identifies a device within a group of devices. For example, cameras in a photographer's collection, or serial numbers in a manufacturer's equipment-line, could receive device assignments.


[0089] unit —serially enumerates an item in a group. For example, a given page in a photocopy job, or a frame number in a roll of film, could be units.


[0090] random —resolves uniqueness. For example, in systems using time for uniqueness, clock resetting can theoretically produce duplicate time-stamping. Using random can prevent such duplicate date-stamping.


[0091] custom —is dedicated for application-specific functionality not natively provided by ASIA, but needed by a given application.


[0092] The following notes apply to the system of the present invention.


[0093] Uniqueness —ASIA often generates uniqueness through time, nit, and random.


[0094] Inplementation needs will determine which field or combination of fields are used to guarantee uniqueness.


[0095] time —ASIA uses ISO 8601:1988 date-time marking, and optional fractional time. Date-stamping can distinguish tags within a generation. In such cases, time granularity MUST match or exceed device production speed or 2 tags can receive the same enumeration. For example, if a photocopy machine produces 10 photocopies per minute, time granularity MUST at least use 6 second time-units, rather than minute time-units. Otherwise, producing 2 or more photocopies could produce 2 or more of the same time-stamps, and therefore potentially also 2 or more of the same ASIA numbers.


[0096] Author —Multiple agents MUST be separated with “,” (comma). random —ASIA uses random as one of three commonly used mechanisms to generate uniqueness (see uniqueness, above). It is particularly useful for systems using time which may be vulnerable to clock resetting. Strong cryptographic randomness is not required for all applications.



The Parent

[0097] A tag's parent component characterizes an object's parent. This is an system number, subject to the restrictions of any system number as described herein. Commonly, this contains time, random, or unit. The following notes apply:


[0098] representation constraints


[0099] The representation constraints for the parent field are those of the database appropriate to it. Representation constraints MAY differ between the parent field and the system number of which the field is a part.



Image

[0100] A tag's image component describes the physical characteristics of an object. For example, an image component could describe the physical characteristics of a plastic negative, steel part, silicon wafer, etc. Table 3 Image lists and illustrates image component fields and their general descriptions.
3TABLE 3ImageFieldDescriptionRepresentationCategoryCharacterizing domainUses any characterSizeDimensionalityUses any characterBitDynamic range (“bit depth”)Uses any characterPushExposureUses any characterMediaMedia representationUses any characterSetSoftware packageUses any characterResolutionResolutionUses any characterStainChromatic representationUses any characterFormatObject embodimentUses any character


[0101] The following definitions describes component fields.


[0102] category: identifies characterizing domain. For example, in photography category could identify “single frame”, to distinguish single frame from motion picture photography.


[0103] size: describes object dimensionality. For example, size could describe an 8×11 inch page, or 100×100×60 micron chip, etc.


[0104] bit: dynamic range ( “bit depth”). For example, bit could describe a “24 bit” dynamic range for a digital image.


[0105] Push: records push or pull. For example, push could describe a “1.3 stop over-exposure” for a photographic image.


[0106] Media: describes the media used to represent an object. For example, media could be “steel” for a part, or “Kodachrome” for a photographic transparency.


[0107] Set: identifies a software rendering and/or version. For example, set could be assigned to “HP: 1.3”.


[0108] Resolution: describes resolution of an object. For example, resolution could represent dots-per-inch in laser printer output.


[0109] Stain: describes chromatic representation. For example, stain could represent “black and white” for a black and white negative.


[0110] Format: describes object embodiment. For example, format could indicate a photocopy, negative, video, satellite, etc. representation.


[0111] The following notes apply to the image data.


[0112] set —when nonempty, permits the remapping of any field value. When empty, core functionality defaults are active. To add revisions, the delimiter “:” (colon) is added to a given root. For example, to add revision number “1.3” to root “HP”, set would be “HP: 1.3”.


[0113] While the functionality of components and fields. Have been illustrated above, Table 4 (below) assembles these into the formal organization from which the the system data structure of the present invention is derived. This ordering provides the basis for the base 16 representation of the schema.
4TABLE 4FieldsComponentFieldLocation12345678GenerationsequencetimeauthordeviceunitrandomcustomParent9parentImage101112131415161718Categorysizebitpushmediasetresolutionstainformat


[0114] This Table 4 (above) assembles these data into the formal organization from which ASIA data structure is derived. This ordering provides the basis for the base 16 representation of the schema.


[0115] Table 5 (below) illustrates the data structure used to encode fields into an ASIA tag. An ASIA tag has five parts: schema, :, length, :, encoding.
5TABLE 5Data StructurePart1Part2Part3Part4Part5Schema:length:encoding


[0116] Each “part” has one or more “elements,” and “elements” have 1-to-1 correspondences across all “parts.” Consider the following definitions for Table 5.


[0117] Schema —An arbitrarily long base 16 integer, whose bits represent the presence of component fields represented in length and encoding, and listed in Table 4. If the least significant bit is set the field 1 is encoded.


[0118] : —The opening delimiter separating schema and length.


[0119] length —Comma separated lengths for the fields represented in encoding, whose presence is indicated in schema.


[0120] : —The closing delimiter separating length and encoding.


[0121] encoding —A concatenated list of fields instantiating the elements indicated by schema and determined by length.


[0122] It is helpful to illustrate an ASIA number. Consider example (1) (below):


[0123] 1C:15,2,1:19981211T112259GE1


[0124] In this ASIA number, schema is “1C”, a hex (base 16) integer (stripped of leading zeros) indicating the presence of three fields (see Table 4): time, author, device. In turn, length is ‘15,2,1’, indicating that the first field is 15 long, the second 2 long, and the third 1 long. The encoding is ‘1998121 1T112259GE1’, and includes the fields identified by schema and determined by length. The encoding has 3 fields: ‘19981211 T112259’ (time), ‘GE’ (author), and ‘1’ (device).


[0125] Table 6 Location, (below) illustrates the fields and the description of the filed used to specify “location.”
6TABLE 6LocationFieldDescriptionGenerationUses any integer.SequenceUses any integer.TimeSee “time” (above)AuthorUses any character.DeviceUses any character.UnitUses any integer.RandomUses any integer.CustomUses any character.


[0126] parent —uses the definition of the location component's time field (see Table 6 above.).


[0127] Table 7: Image, (below) illustrates the fields and description associated with “image.”
7TABLE 7ImageFieldDescriptioncategorySee Table 8 CategoriessizeSee Table 9 Size/res. SyntaxSee Table 10 MeasureSee Table 11 Size examplesbitSee Table 12 BitpushSee Table 13 PushmediaSee Table 14 Reserved media slotsSee Table 15 Color transparency filmSee Table 16 Color negative filmSee Table 17 Black & white filmSee Table 18 Duplicating & internegative filmSee Table 19 FacsimileSee Table 20 PrintsSee Table 21 DigitalsetSee Table 22 Software SetsresolutionSee Table 9 Size/res. SyntaxSee Table 10 MeasureSee Table 23 Resolution examplesstainSee Table 24 StainformatSee Table 25 Format


[0128] The category field has 2 defaults as noted in Table 8 (below)..
8TABLE 8CategoriesLiteralDescriptionSSingle frameMMotion picture


[0129] The size field has 2 syntax forms, indicated in Table 9 Size/res. Syntax (below).


[0130] Table 10 Measure (below) provides default measure values that are used in Table 9.


[0131] Table 11 Size examples provides illustrations of the legal use of size. Consider the following definitions.


[0132] dimension —is a set of units using measure.


[0133] measure —is a measurement format.


[0134] n{+}—represents a regular expression, using 1 or more numbers (0-9).


[0135] Ic{*}—represents a regular expression beginning with any single letter (a-z; A-Z), and continuing with any number of any characters.


[0136] X-dimension —is the X-dimension in an X-Y coordinate system, subject to measure.


[0137] Y-dimension —is the Y-dimension in an X-Y coordinate system, subject to measure.


[0138] X —is a constant indicating an X-Y relationship.
9TABLE 9Size/res. syntaxCategoryIllustrationNamesDimensionmeasureType 1n{+}lc{*}NamesX-dimensionXY-dimensionmeasureType 2n{+}Xn{+}lc{*}


[0139] Table 10 illustrates default values for measure. It does not preclude application-specific extensions.
10TABLE 10MeasureCategoryLiteralDescriptionSharedDIDots per inch(dpi)DEDots per foot(dpe)DYDots per yard(dpy)DQDots per mile(dpq)DCDots per centimeter(dpc)DMDots per millimeter(dpm)DTDots per meter(dpt)DKDots per kilometer(dpk)DPDots per pixel(dpp)NMicron(s)MMillimeter(s)CCentimeter(s)TMeter(s)KKilometer(s)IInch(s)EFoot/FeetYYard(s)QMile(s)PPixel(s)LLine(s)RRow(s)OColumn(s)BColumn(s) & row(s). . .etc.Size UniqueFFormatSSheet. . .etc.Res. UniqueSISO. . .etc.


[0140] Table 11 entitled “Size Examples” (below) illustrates the Syntax, literal, description and measures associated with various sizes of images. This listing is not meant as a limitation but is illustrative only. As other sizes of images are created, these too will be able to be specified by the system of the present invention.
11TABLE 11Size examplesSyntaxLiteralDescriptionMeasureType 1135F35 mmformat120FMediumformat220FFullformat4 × 5F4 × 5format. . .. . .etc.Type 29 × 14C9 × 14centimeter3 × 5I3 × 5inch4 × 6I4 × 6inch5 × 7I5 × 7inch8 × 10I8 × 10inch11 × 14I11 × 14inch16 × 20I16 × 20inch20 × 24I20 × 24inch24 × 32I24 × 32inch24 × 36I24 × 36inch32 × 40I32 × 40inch40 × 50I40 × 50inch50 × 50I50 × 50inch40 × 50P40 × 50pixels100 × 238P100 × 238pixels1024 × 1280P1024 × 1280pixelsA1S59.4 × 84.0 cmsheetA2S42.0 × 59.4 cmsheetA3S29.7 × 42.0 cmsheetA4S21.0 × 29.7 cmsheetA5S14.85 × 21.0 cmsheetA6S10.5 × 14.85 cmsheetA7S7.42 × 10.5 cmsheetA1RS84.0 × 59.4 cmsheetA2RS59.4 × 42.0 cmsheetA3RS42.0 × 29.7 cmsheetA4RS29.7 × 21.0 cmsheetA5RS21.0 × 14.85 cmsheetA6RS14.85 × 10.5 cmsheetA7RS10.5 × 7.42 cmsheetB1S70.6 × 100.0 cmsheetB2S50.0 × 70.6 cmsheetB3S35.3 × 50.0 cmsheetB4S25.0 × 35.3 cmsheetB5S17.6 × 25.0 cmsheetB6S13.5 × 17.6 cmsheetB7S8.8 × 13.5 cmsheetB1RS100.0 × 70.6 cmsheetB2RS70.6 × 50.0 cmsheetB3RS50.0 × 35.3 cmsheetB4RS35.3 × 25.0 cmsheetB5RS25.0 × 17.6 cmsheetB6RS17.6 × 13.5 cmsheetB7RS13.5 × 8.8 cmsheetC1S64.8 × 91.6 cmsheetC2S45.8 × 64.8 cmsheetC3S32.4 × 45.8 cmsheetC4S22.9 × 32.4 cmsheetC5S16.2 × 22.9 cmsheetC6S11.46 × 16.2 cmsheetC7S8.1 × 11.46 cmsheetC1RS91.6 × 64.8 cmsheetC2RS64.8 × 45.8 cmsheetC3RS45.8 × 32.4 cmsheetC4RS32.4 × 22.9 cmsheetC5RS22.9 × 16.2 cmsheetC6RS16.2 × 11.46 cmsheetC7RS11.46 × 8.1 cmsheetJISS18.2 × 25.7 cmsheetUSS8.5 × 11 insheetUSRS8.5 × 11 insheetLEGALS8.5 × 14 insheetEXECUTIVES7.25 × 10.5 insheetFOOLSCAPS13.5 × 17.0 insheet. . .. . .etc.


[0141] Table 12 (below) lists legal values for the bit field.
12TABLE 12BitLiteralDescription 88 bit dynamic range2424 bit dynamic range. . .etc.


[0142] Table 13 (below) lists legal values for the push field.
13TABLE 13PushLiteralDescription+1  Pushed +1 stops −.3Pulled −.3 stops. . .etc.


[0143] Values of media (illustrated in Table 14 below) are keyed to the format field, thus permitting reuse of given abbreviations for multiple contexts. The mapping of dependencies is illustrated in Table 25..
14TABLE 14Reserved media slotsReserved ForLiteralDescriptionUnknownXXXXUnknown filmSpecificationUR0For future useUR1UR2UR3UR4UR5UR6UR7UR8UR9UserUX0CustomizationUX1UX2UX3UX4UX5UX6UX7UX8UX9


[0144] Table 15 “Color Transparency Filem” (below) illustrates the company. Literal and description field available for existing transparency films. As new transparency flms emerge, these too can be accommodated by the present invention.
15TABLE 15Color transparency filmCompanyLiteralDescriptionAgfaAASCAgfa Agfapan Scala Reversal (B&W)ACRSAgfa Agfachrome RSACTXAgfa Agfachrome CTXARSXAgfa Agfacolor Professional RSX ReversalFujiFCRTPFuji Fujichrome RTPFCSEFuji Fujichrome SensiaFRAPFuji Fujichrome AstiaFRDPFuji Fujichrome Provia Professional 100FRPHFuji Fujichrome Provia Professional 400FRSPFuji Fujichrome Provia Professional 1600FRTPFuji Fujichrome Professional TungstenFRVPFuji Fujichrome Velvia ProfessionalIlfordIICCIlford IlfochromeIICDIlford Ilfochrome DisplayIICMIlford Ilfochrome MicrographicKonicaCAPSKonica APS JXCCSPKonica Color Super SR ProfessionalKodakK5302Kodak Eastman Fine Grain Release PositiveFilm 5302K7302Kodak Fine Grain Positive Film 7302KA2443Kodak Aerochrome II Infrared Film 2443KA2448Kodak Aerochrome II MS Film 2448KE100SWKodak Ektachrome Professional E100SW FilmKE100SKodak Ektachrome Professional E100S FilmKE200Kodak Ektachrome Professional E200 FilmKEEEKodak Ektachrome EliteKEEO100Kodak Ektachrome Electronic Output Film 100KEEO200Kodak Ektachrome Electronic Output Film 200KEEO64TKodak Ektachrome Electronic Output Film 64TKEEPKodak Ektachrome E ProfessionalKEESKodak Ektachrome ESKEEWKodak Ektachrome EWKEIRKodak Ektachrome Professional Infrared EIRFilmKEKKodak EktachromeKELLKodak Ektachrome Lumiere ProfessionalKELXKodak Ektachrome Lumiere X ProfessionalKEPDKodak Ektachrome 200 Professional FilmKEPFKodak Ektachrome ProfessionalKEPHKodak Ektachrome Professional P1600 FilmKEPJKodak Ektachrome 320T Professional Film,TungstenKEPL400Kodak Ektachrome Professional 400X FilmKEPLKodak Ektachrome 200 Professional FilmKEPLKodak Ektachrome Plus ProfessionalKEPNKodak Ektachrome 100 Professional FilmKEPOKodak Ektachrome P ProfessionalKEPRKodak Ektachrome 64 ProfessionalKEPTKodak Ektachrome 160T Professional Film,TungstenKEPYKodak Ektachrome 64T Professional Film,TungstenKETPKodak Ektachrome T ProfessionalKETTKodak Ektachrome TKEXPKodak Ektachrome X ProfessionalKCCRKodak KodachromeKPKAKodak Kodachrome Professional 64 FilmKPKLKodak Kodachrome Professional 200 FilmKPKMKodak Kodachrome Professional 25 FilmKVSSO279Kodak Vericolor Slide Film SO-279KVSKodak Vericolor Slide FilmPolaroidPPCPKodak Professional High Contrast PolychromeReservedSee Table 14SeattleSFWSSeattle Film WorksFW3MTSCS3M ScotchColor SlideTSCT3M ScotchColor T slide


[0145] Table 16 (below) illustrates the types of color negative film that can be accommodated by the present invention. Again this list is not meant as a limitation but is illustrative only.
16TABLE 16Color negative filmCompanyLiteralDescriptionAgfaACOPAgfa Agfacolor OptimaAHDCAgfa Agfacolor HDCAPOTAgfa Agfacolor Triade Optima ProfessionalAPOAgfa Agfacolor Professional OptimaAPPAgfa Agfacolor Professional PortraitAPUAgfa Agfacolor Professional UltraAPXPSAgfa Agfacolor Professional Portrait XPSATPTAgfa Agfacolor Triade Portrait ProfessionalATUTAgfa Agfacolor Triade Ultra ProfessionalFujiFHGPFuji Fujicolor HG ProfessionalFHGFuji Fujicolor HGFNHGFuji Fujicolor NHG ProfessionalFNPHFuji Fujicolor NPH ProfessionalFNPLFuji Fujicolor NPL ProfessionalFNPSFuji Fujicolor NPS ProfessionalFPIFuji Fujicolor PrintFPLFuji Fujicolor Professional, Type LFPOFuji Fujicolor PositiveFRGFuji Fujicolor Reala GFRFuji Fujicolor RealaFSGPFuji Fujicolor Super G PlusFSGFuji Fujicolor Super GFSHGFuji Fujicolor Super HG 1600FSFuji Fujicolor SuperKodakK5079Kodak Motion Picture 5079K5090Kodak CF1000 5090K5093Kodak Motion Picture 5093K5094Kodak Motion Picture 5094KA2445Kodak Aerocolor II Negative Film 2445KAPBKodak Advantix Professional FilmKCPTKodak Kodacolor PrintKEKAKodak Ektar AmateurKEPGKodak Ektapress GoldKEPPRKodak Ektapress Plus ProfessionalKGOPKodak Gold PlusKGOKodak GoldKGPXKodak Ektacolor Professional GPXKGTXKodak Ektacolor Professional GTXKPCNKodak Professional 400 PCN FilmKPHRKodak Ektar Professional FilmKPJAMKodak Ektapress MultispeedKPJAKodak Ektapress 100KPJCKodak Ektapress Plus 1600 ProfessionalKPMCKodak Pro 400 MC FilmKPMZKodak Pro 1000 FilmKPPFKodak Pro 400 FilmKPRMCKodak Pro MCKPRNKodak ProKPRTKodak Pro TKRGDKodak Royal GoldKVPS2LKodak Vericolor II Professional Type LKVPS3SKodak Vericolor III Professional Type SKVPKodak Vericolor Print FilmKonicaCCIPKonica Color Impresa ProfessionalCCSRKonica SRGPolaroidPOCPPolaroid OneFilm Color PrintReservedSee Table 14


[0146] Table 17 (below) illustrates a list of black and white film that can be accommodated by the present invention. Again this list is illustrative only and is not meant as a limitation.
17TABLE 17Black & white filmCompanyLiteralDescriptionAgfaAAORAgfa Agfapan OrthoAAPXAgfa Agfapan APXAPANAgfa AgfapanIlfordIDELIlford Delta ProfessionalIFP4Ilford FP4 PlusIHP5Ilford HP5 PlusIPFPIlford PanF PlusIPSFIlford SFX750 InfraredIUNIIlford UniversalIXPPIlford XP2 PlusFujiFNPNFuji NeopanKodakK2147TKodak PLUS-X Pan Professional 2147, ESTARThick BaseK2147Kodak PLUS-X Pan Professional 2147, ESTARBaseK4154Kodak Contrast Process Ortho Film 4154,ESTAR Thick BaseK4570Kodak Pan Masking Film 4570, ESTAR ThickBaseK5063Kodak TRI-X 5063KA2405Kodak Double-X Aerographic Film 2405KAI2424Kodak Infrared Aerographic Film 2424KAP2402Kodak PLUS-X Aerographic II Film 2402,ESTAR BaseKAP2412Kodak Panatomic-X Aerographic II Film 2412,ESTAR BaseKEHCKodak Ektagraphic HCKEKPKodak EktapanKH13101Kodak High Speed Holographic Plate, Type131-01KH13102Kodak High Speed Holographic Plate, Type131-02KHSIETKodak High Speed Infrared, ESTAR Thick BaseKHSIEKodak High Speed Infrared, ESTAR BaseKHSIKodak High Speed InfraredKHSO253Kodak High Speed Holographic Film,ESTAR Base SO-253KLPD4Kodak Professional Precision Line Film LPD4KO2556Kodak Professional Kodalith Ortho Film 2556KO6556Kodak Professional Kodalith Ortho Film 6556,Type 3KPMF3Kodak Professional Personal Monitoring Film,Type 3KPNMFAKodak Professional Personal Neutron MonitorFilm, Type AKPXEKodak PLUS-X Pan Professional, RetouchingSurface, Emulsion & BaseKPXPKodak PLUS-X Pan Professional, RetouchingSurface, EmulsionKPXTKodak PLUS-X Pan Professional, RetouchingSurface, Emulsion & BaseKPXXKodak Plus-XKPXKodak PLUS-X Pan FilmKRECKodak Recording 2475KSAF1Kodak Spectrum Analysis Film, No. 1KSAP1Kodak Spectrum Analysis Plate, No. 1KSAP3Kodak Spectrum Analysis Plate, No. 3KSWRPKodak Short Wave Radiation PlateKTMXCNKodak Professional T-MAX Black andWhite Film CNKTMYKodak Professional T-MAXKTMZKodak Professional T-MAX P3200 FilmKTP2415Kodak Technical Pan Film 2415,ESTAR-AH BaseKTPKodak Technical Pan FilmKTRPKodak TRI-Pan ProfessionalKTRXPTKodak TRI-X Pan Professional 4164,ESTAR Thick BaseKTRXPKodak TRI-Pan ProfessionalKTXPKodak TRI-X Professional, Interior TungstenKTXTKodak TRI-X Professional, Interior TungstenKTXKodak TRI-X ProfessionalKVCPKodak Verichrome PanKonicaCIFRKodak Infrared 750PolaroidPPGHKonica Polagraph HCPPLBPolaroid Polablue BNPPPNPolaroid Polapan CTReservedSee Table 14


[0147] Referring to Table 18 (below) Duplicating and Internegative Films are illustrated. Again this list is for illustrative purposes only and is not meant as a limitation.
18TABLE 18Duplicating & internegative filmCompanyLiteralDescriptionAgfaACRDAgfa Agfachrome Duplication Film CRDFujiFCDUFuji Fujichrome CDU DuplicatingFCDU1Fuji Fujichrome CDU Duplicating, Type IFCDU2Fuji Fujichrome CDU Duplicating, TypeIIFITNFuji Fujicolor Internegative IT-NKodakK1571Kodak 1571 InternegativeK2475REKodak Recording Film 2475K4111Kodak 4111KC4125Kodak Professional Professional CopyFilm 4125K6121Kodak 6121KA2405Kodak Double-X Aerographic Film 2405KA2422Kodak Aerographic Direct DuplicatingFilm 2422KA2447Kodak Aerochrome II Duplicating Film2447KARA2425Kodak Aerographic RA Duplicating Film2425, ESTAR BaseKARA4425Kodak Aerographic RA Duplicating Film4425, ESTAR Thick BaseKARAKodak Aerographic RA Duplicating FilmKCINKodak Commercial Internegative FilmKE5071Kodak Ektachrome Slide DuplicatingFilm 5071KE5072Kodak Ektachrome Slide DuplicatingFilm 5072KE6121Kodak Ektachrome Slide DuplicatingFilm 6121KE7121KKodak Ektachrome Duplicating Film7121, Type KKESO366Kodak Ektachrome SE Duplicating FilmSO-366KS0279Kodak S0279KS0366Kodak S0366KSO132Kodak Professional B/W DuplicatingFilm SO-132KV4325Kodak Vericolor Internegative 4325KVINKodak Vericolor Internegative FilmReservedSee Table 14


[0148] Referring to Table 19, Facsimile types and formats are illustrated. This listing is not meant as a limitation and is for illustrative purposes only.
19TABLE 19FacsimileCategoryLiteralDescriptionDigitalSee Table 21FacsimileDFAXHDigiBoard, DigiFAX Format, Hi-ResDFAXLDigiBoard, DigiFAX Format, Normal-ResG1Group 1 FacsimileG2Group 2 FacsimileG3Group 3 FacsimileG32DGroup 3 Facsimile, 2DG4Group 4 FacsimileG42DGroup 4 Facsimile, 2DG5Group 4 FacsimileG52DGroup 4 Facsimile, 2DTIFFG3TIFF Group 3 FacsimileTIFFG3CTIFF Group 3 Facsimile, CCITT RLE 1DTIFFG32DTIFF Group 3 Facsimile, 2DTIFFG4TIFF Group 4 FacsimileTIFFG42DTIFF Group 4 Facsimile, 2DTIFFG5TIFF Group 5 FacsimileTIFFG52 DTIFF Group 5 Facsimile, 2DReservedSee Table 14


[0149] Referring to Table 20 (below) examples of the types of print paper that can be classified by the present invention are illustrated. Again, this list is for illustrative purposes only and is not meant as a limitation.
20TABLE 20PrintsCompanyLiteralDescriptionAgfaACRAgfacolor RCABFAgfa Brovira, fiber, B&WABSRCAgfa Brovira-speed RC, B&WAPFAgfa Portriga, fiber, B&WAPSRCAgfa Portriga-speed RC, B&WARRFAgfa Record-rapid, fiber, B&WACHDAgfacolor HDCAMCC111FBAgfacolor Multicontrast Classic MC C 111 FB,double weight, glossy surfaceAMCC118FBAgfacolor Multicontrast Classic MC C 118 FB,double weight, fine grained matte surfaceAMCC1FBAgfacolor Multicontrast Classic MC C 1 FB, singleweight, glossy surfaceAMCP310RCAgfacolor Multicontrast Premium RC 310, glossysurfaceAMCP312RCAgfacolor Multicontrast Premium RC 312, semi-mattesurfaceAPORGAgfacolor Professional Portrait Paper, glossy surfaceCN310APORLAgfacolor Professional Portrait Paper, semi-mattesurface CN312APORMAgfacolor Professional Portrait Paper, lustre surfaceCN319ASIGGAgfacolor Professional Signum Paper, glossy surfaceCN310ASIGMAgfacolor Professional Signum Paper, matte surfaceCN312KonicaCCOLKonica ColorFujiFCHPFujicolor HG ProfessionalFCPIFujicolor PrintFCSPFujicolor Super G Plus PrintFCT35GFujichrome paper, Type 35, glossy surfaceFCT35HGFujichrome reversal copy paper, Type 35, glossysurfaceFCT35HLFujichrome reversal copy paper, Type 35, lustresurfaceFCT35HMFujichrome reversal copy paper, Type 35, mattesurfaceFCT35LFujichrome paper, Type 35, lustre surfaceFCT35MFujichrome paper, Type 35, matte surfaceFCT35PGFujichrome Type 35, polyester, super glossly surfaceFSFA5GFujicolor paper super FA, Type 5, glossy SFA5surfaceFSFA5LFujicolor paper super FA, Type 5, lustre SFA5 surfaceFSFA5MFujicolor paper super FA, Type 5, matte SFA5 surfaceFSFA5SGFujicolor paper super FA5, Type C, glossy surfaceFSFA5SLFujicolor paper super FA5, Type C, lustre surfaceFSFA5SMFujicolor paper super FA5, Type C, matte surfaceFSFA5SPGFujicolor paper super FA, Type 5P, glossy SFA PsurfaceFSFA5SPLFujicolor paper super FA, Type 5P, lustre SFA PsurfaceFSFA5SPMFujicolor paper super FA, Type 5P, matte SFA PsurfaceFSFAGFujicolor paper super FA, Type 5, glossy surfaceFSFALFujicolor paper super FA, Type 5, lustre surfaceFSFAMFujicolor paper super FA, Type 5, matte surfaceFSFAS5PGFujicolor paper super FA, Type P, glossy SFA 5PsurfaceFSFAS5PLFujicolor paper super FA, Type P, lustre SFA 5PsurfaceFSFAS5PMFujicolor paper super FA, Type P, matte SFA 5PsurfaceFSFASCGFujicolor paper super FA, Type C, glossy surfaceFSFASCLFujicolor paper super FA, Type C, lustre surfaceFSFASCMFujicolor paper super FA, Type C, matte surfaceFTRSFAFujitrans super FAFXSFAFujiflex super FA polyester (super gloss), Fujiflex SFAsurfaceIlfordICF1KIlfochrome Classic Deluxe Glossy Low ContrastICLM1KIlfochrome Classic Deluxe Glossy Medium ContrastICPM1MIlfochrome Classic RC GlossyICPM44MIlfochrome Classic RC PearlICPS1KIlfochrome Classic Deluxe GlossyIGFBIlford Galerie FBIILRA1KIlfocolor DeluxeIIPRAMIlfocolor RCIMG1FDWIlford Multigrade Fiber, Double WeightIMG1FWIlford Multigrade Fiber WarmtoneIMG1RCDLXIlford Multigrade RC DLXIMG1RCPDWIlford Multigrade RC Portfolio, Double WeightIMG1RCRIlford Multigrade RC RapidIMG2FDWIlford Multigrade II Fiber, Double WeightIMG2FWIlford Multigrade II Fiber WarmtoneIMG2RCDLXIlford Multigrade II RC DLXIMG2RCPDWIlford Multigrade II RC Portfolio, Double WeightIMG2RCRIlford Multigrade II RC RapidIMG3FDWIlford Multigrade III Fiber, Double WeightIMG3FWIlford Multigrade III Fiber WarmtoneIMG3RCDLXIlford Multigrade III RC DLXIMG3RCPDWIlford Multigrade III RC Portfolio, Double WeightIMG3RCRIlford Multigrade III RC RapidIMG4FDWIlford Multigrade IV Fiber, Double WeightIMG4FWIlford Multigrade IV Fiber WarmtoneIMG4RCDLXIlford Multigrade IV RC DLXIMG4RCPDWIlford Multigrade IV RC Portfolio, Double WeightIMGFSWGIlford Multigrade Fiber, Single Weight, glossyIPFPIlford PanF PlusISRCDIlford Ilfospeed RC, DeluxeKodakB&W Selective Contrast PapersKPC1RCEKodak Polycontrast RC, medium weight, fine-grained,lustreKPC1RCFKodak Polycontrast RC, medium weight, smooth,glossyKPC1RCNKodak Polycontrast RC, medium weight, smooth,semi-matteKPC2RCEKodak Polycontrast II RC, medium weight, fine-grained, lustreKPC2RCFKodak Polycontrast II RC, medium weight, smooth,glossyKPC2RCNKodak Polycontrast II RC, medium weight, smooth,semi-matteKPC3RCEKodak Polycontrast III RC, medium weight, fine-grained, lustreKPC3RCFKodak Polycontrast III RC, medium weight, smooth,glossyKPC3RCNKodak Polycontrast III RC, medium weight, smooth,semi-matteKPMFFKodak Polymax Fiber, single weight, smooth, glossyKPMFNKodak Polymax Fiber, single weight, smooth, semi-matteKPMFEKodak Polymax Fiber, single weight, fine-grained,lustreKPM1RCFKodak Polymax RC, single weight, smooth, glossyKPM1RCEKodak Polymax RC, single weight, fine-grained, lustreKPM1RCNKodak Polymax RC, single weight, smooth, semi-matteKPM2RCFKodak Polymax II RC, single weight, smooth, glossyKPM2RCEKodak Polymax II RC, single weight, fine-grained,lustreKPM2RCNKodak Polymax II RC, single weight, smooth, semi-matteKPMFAFKodak Polymax Fine-Art, double weight, smooth,glossyKPMFANKodak Polymax Fine-Art, double weight, smooth,semi-matteKPPFMKodak Polyprint RC, medium weight, smooth, glossyKPPNMKodak Polyprint RC, medium weight, smooth, semi-matteKPPEMKodak Polyprint RC, medium weight, fine-grained,lustreKPFFSKodak Polyfiber, single weight, smooth, glossyKPFNDKodak Polyfiber, double weight, smooth, semi-matteKPFGLKodak Polyfiber, light weight, smooth, lustreKPFNSKodak Polyfiber, smooth, single weight, semi-matteKPFNDKodak Polyfiber, double weight, smooth, semi-matteKPFGDKodak Polyfiber, double weight, fine-grained, lustreB&W Continuous Tone PapersKAZOFKodak AZO, fine-grained, lustreKB1RCFKodak Kodabrome RC Paper, smooth, glossyKB1RCG1Kodak Kodabrome RC, premium weight (extra heavy)1, fine-grained, lustreKB1RCNKodak Kodabrome RC Paper, smooth, semi-matteKB2RCFKodak Kodabrome II RC Paper, smooth, glossyKB2RCG1Kodak Kodabrome II RC, premium weight (extraheavy) 1, fine-grained, lustreKB2RCNKodak Kodabrome II RC Paper, smooth, semi-matteKBRKodak Kodabromide, single weight, smooth, glossyKEKLGKodak Ektalure, double weight, fine-grained, lustreKEKMSCFKodak Ektamatic SC single weight, smooth, glossyKEKMSCNKodak Ektamatic SC, single weight, smooth, semi-matteKEKMXRALFKodak Ektamax RA Professional L, smooth, glossyKEKMXRALNKodak Ektamax RA Professional L, smooth, semi-matteKEKMXRAMFKodak Ektamax RA Professional M, smooth, glossyKEKMXRAMNKodak Ektamax RA Professional M, smooth, semi-matteKELFA1Kodak Elite Fine-Art, premium weight (extra heavy) 1,ultra-smooth, high-lustreKELFA2Kodak Elite Fine-Art, premium weight (extra heavy) 2,ultra-smooth, high-lustreKELFA3Kodak Elite Fine-Art, premium weight (extra heavy) 3,ultra-smooth, high-lustreKELFA4Kodak Elite Fine-Art, premium weight (extra heavy) 4,ultra-smooth, high-lustreKK1RCG1Kodak Kodabrome RC, premium weight (extra heavy)1, fine-grained, lustreKK1RCG2Kodak Kodabrome RC, premium weight (extra heavy)2, fine-grained, lustreKK1RCG3Kodak Kodabrome RC, premium weight (extra heavy)3, fine-grained, lustreKK1RCG4Kodak Kodabrome RC, premium weight (extra heavy)4, fine-grained, lustreKK1RCG5Kodak Kodabrome RC, premium weight (extra heavy)5, fine-grained, lustreKK2RCG1Kodak Kodabrome II RC, premium weight (extraheavy) 1, fine-grained, lustreKK2RCG2Kodak Kodabrome II RC, premium weight (extraheavy) 2, fine-grained, lustreKK2RCG3Kodak Kodabrome II RC, premium weight (extraheavy) 3, fine-grained, lustreKK2RCG4Kodak Kodabrome II RC, premium weight (extraheavy) 4, fine-grained, lustreKK2RCG5Kodak Kodabrome II RC, premium weight (extraheavy) 5, fine-grained, lustreKPMARCW1Kodak P-Max Art RC, double weight 1, suede double-matteKPMARCW2Kodak P-Max Art RC, double weight 2, suede double-matteKPMARCW3Kodak P-Max Art RC, double weight 3, suede double-matteB&W Panchromatic PapersKPSRCHKodak Panalure Select RC, H grade, medium weight,smooth, glossyKPSRCLKodak Panalure Select RC, L grade, medium weight,smooth, glossyKPSRCMKodak Panalure Select RC, M grade, medium weight,smooth, glossyColor Reversal PapersKER1FKodak Ektachrome Radiance Paper, smooth, glossyKER1NKodak Ektachrome Radiance Paper, smooth, semi-matteKER1SFKodak Ektachrome Radiance Select Material, smooth,glossyKER2FKodak Ektachrome Radiance II Paper, smooth, glossyKER2NKodak Ektachrome Radiance II Paper, smooth, semi-matteKER2SFKodak Ektachrome Radiance II Select Material,smooth, glossyKER3FKodak Ektachrome Radiance III Paper, smooth,glossyKER3NKodak Ektachrome Radiance III Paper, smooth, semi-matteKER3SFKodak Ektachrome Radiance III Select Material,smooth, glossyKERCFKodak Ektachrome Radiance Copy Paper, smooth,glossyKERCHCFKodak Ektachrome Radiance HC Copy Paper,smooth, glossyKERCHCNKodak Ektachrome Radiance HC Copy Paper,smooth, semi-matteKERCNKodak Ektachrome Radiance Copy Paper, smooth,semi-matteKERCTFKodak Ektachrome Radiance Thin Copy Paper,smooth, glossyKERCTNKodak Ektachrome Radiance Thin Copy Paper,smooth, semi-matteKEROMKodak Ektachrome Radiance Overhead Material,transparent ESTAR Thick BaseColor Negative Papers & Transparency MaterialsKD2976EKodak Digital Paper, Type 2976, fine-grained, lustreKD2976FKodak Digital Paper, Type 2976, smooth, glossyKD2976NKodak Digital Paper, Type 2976, smooth, semi-matteKDCRAKodak Duraclear RA Display Material, clearKDFRAFKodak Duraflex RA Print Material, smooth, glossyKDT2Kodak Duratrans Display Material, translucentKDTRAKodak Duratrans RA Display Material, translucentKECCKodak Ektacolor, Type CKECEKodak Ektacolor Professional Paper, fine-graned,lustreKECFKodak Ektacolor Professional Paper, smooth, glossyKECNKodak Ektacolor Professional Paper, smooth, semi-matteKECKodak EktacolorKEP2EKodak Ektacolor Portra II Paper, Type 2839, fine-grained, lustreKEP2FKodak Ektacolor Portra II Paper, Type 2839, smooth,glossyKEP2NKodak Ektacolor Portra II Paper, Type 2839, smooth,semi-matteKEP3EKodak Ektacolor Portra III Paper, fine-grained, lustreKEP3FKodak Ektacolor Portra III Paper, smooth, glossyKEP3NKodak Ektacolor Portra III Paper, smooth, semi-matteKES2EKodak Ektacolor Supra II Paper, fine-grained, lustreKES2FKodak Ektacolor Supra II Paper, smooth, glossyKES2NKodak Ektacolor Supra II Paper, smooth, semi-matteKES3EKodak Ektacolor Supra III Paper, fine-grained, lustreKES3FKodak Ektacolor Supra III Paper, smooth, glossyKES3NKodak Ektacolor Supra III Paper, smooth, semi-matteKESEKodak Ektacolor Supra Paper, fine-grained, lustreKESFKodak Ektacolor Supra Paper, smooth, glossyKESNKodak Ektacolor Supra Paper, smooth, semi-matteKET1Kodak Ektatrans RA Display Material, smooth, semi-matteKEU2EKodak Ektacolor Ultra II Paper, fine-grained, lustreKEU2FKodak Ektacolor Ultra II Paper, smooth, glossyKEU2NKodak Ektacolor Ultra II Paper, smooth, semi-matteKEU3EKodak Ektacolor Ultra III Paper, fine-grained, lustreKEU3FKodak Ektacolor Ultra III Paper, smooth, glossyKEU3NKodak Ektacolor Ultra III Paper, smooth, semi-matteKEUEKodak Ektacolor Ultra Paper, fine-grained, lustreKEUFKodak Ektacolor Ultra Paper, smooth, glossyKEUNKodak Ektacolor Ultra Paper, smooth, semi-matteInkjet Papers & FilmsKEJFC50HGKodak Ektajet 50 Clear Film LW4, Polyester Base,clearKEJFEFSGKodak Ektajet Film, Type EF, semi-glossKEJFLFSGKodak Ektajet Film, Type LF, semi-glossKEJFW50HGKodak Ektajet 50 White Film, Polyester Base, highglossKEJP50SGKodak Ektajet 50 Paper, RC Base, semi-glossKEJPCKodak Ektajet Coated PaperKEJPCHWKodak Ektajet Heavy Weight Coated PaperKEJPEFSGKodak Ektajet Paper, Type EF, semi-glossKEJPLFSGKodak Ektajet Paper, Type LF, semi-glossPolaroidPOCPOneFilm Color PrintPPCPProfessional High Contrast PolychromePPGHPolagraph HCPPLBPolablue BNPPPNPolapan CTReservedSee Table 14


[0150] Referring to Table 21, the types of digital file formats that may be accommodated by the present invention are illustrated. This list is for illustrative purposes only and is not meant as a limitation.
21TABLE 21DigitalCategoryLiteralDescriptionDigitalACADAutoCAD database or slideASCIASCII graphicsATKAndrew Toolkit raster objectAVIMicrosoft videoAVSAVS X imageBIOBiorad confocal fileBMPMicrosoft Windows bitmap imageBMPMMicrosoft Windows bitmap image, monochromeBPGMBentleyized Portable Graymap FormatBRUSDoodle brush fileCGMCGM colorCDRCorel DrawCIFCIF file format for VLSICGOGCompressed GraphOn graphicsCMUWCMU window manager bitmapCMXCorel VectorCMYKRaw cyan, magenta, yellow, and black bytesCQTCinepak QuicktimeDVITypesetter DeVice Independent formatEPSAdobe Encapsulated PostScriptEPSFAdobe Encapsulated PostScript file formatEPSIAdobe Encapsulated PostScript Interchange formatFIGXfig image formatFITFlexible Image Transport SystemFLCFLC movie fileFLIFLI movie fileFSTUsenix FaceSaver(tm) fileG10XGemini 10X printer graphicsGEMGEM image fileGIFCompuServe Graphics imageGIF8CompuServe Graphics image (version 87a)GOULGould scanner fileGRARaw gray bytesHDFHierarchical Data FormatHIPSHIPS fileHISImage HistogramHPLJHewlett Packard LaserJet formatHPPJHewlett Packard PaintJetHTMHypertext Markup LanguageHTM2Hypertext Markup Language, version 2HTM3Hypertext Markup Language, version 3HTM4Hypertext Markup Language, version 4ICONSun iconICRNCSA Telnet Interactive Color Rastergraphic formatIFFElectronic ArtsILBMAmiga ILBM fileIMGImg-whatnot fileJBGJoint Bi-level image experts Group fileinterchange formatJPGJoint Photographic experts Group fileinterchange formatLISPLisp Machine bitmap fileMACPApple MacPaint fileMAPColormap intensities and indicesMATRaw matte bytesMCIMCI formatMGRMGR bitmapMIDMID formatMIFImageMagick formatMITSMitsubishi S340-10 Color sublimationMMMMMM move fileMOVMovie formatMP2Motion Picture Experts Group (MPEG)interchange format, level 2MP3Motion Picture Experts Group (MPEG)interchange format, level 3MPGMotion Picture Experts Group (MPEG)interchange format, level 1MSPMicrosoft PaintMTVMTV ray tracer imageNKNNikon formatNULNULL imagePBMPortable BitMapPCDKodak Photo-CDPCXZSoft IBM PC PaintbrushPDFPortable Document FormatPGMPortable GrayMap formatPGNPortable GrayMap formatPI1Atari Degas .pi1 FormatPI3Atari Degas .pi3 FormatPICApple Macintosh QuickDraw/PICTPLOTUnix Plot(5) formatPNGPortable Network GraphicsPNMPortable anymapPPMPortable pixmapPPTPowerpointPRTPRT ray tracer imagePS1Adobe PostScript, level 1PS2Adobe PostScript, level 2PSDAdobe PhotoshopQRTQRT ray tracerRADRadiance imageRASCMU raster image formatRGBRaw red, green, and blue bytesRGBARaw red, green, blue, and matte bytesRLEUtah Run length encoded imageSGIIrix RGB imageSIRSolitaire file formatSIXLDEC sixel color formatSLDAutoCAD slide fileSPCAtari compressed Spectrum fileSPOTSPOT satelite imagesSUNSUN RasterfileTGATarga True VisionTIFTagged Image FormatTILTile image with a textureTXTRaw textUILMotif UIL icon fileUPCUniversal Product Code bitmapUYVYYUV bit/pixel interleaved (AccomWSD)VICVideo Image Communication And Retrieval(VICAR)VIDVisual Image DirectoryVIFKhoros Visualization imageWRLVirtual reality modeling languageX1BMX10 bitmapXBMX11 bitmapXCCConstant image of X server colorXIMXIM fileXPMX11 pixmapXWDX Window system window DumpXXXImage from X server screenYBMBennet Yee “face″ fileYUVAbekas YUY fileYUV3Abekas Y- and U- and Y-file, 3ZEISZeiss confocal fileZINCZinc bitmapFacsimileSee Table 19ReservedSee Table 14


[0151] Table 22 provides default software set root values. Implementations MAY add to, or extend values in Table 22.
22TABLE 22Software Sets3CDescription3M3MADAdobeAGAGFAAIMAIMS LabsALSAlesisAPPApolloAPLAppleARMArt MediaARLArtelAVMAverMedia TechnologiesATTAT&TBRBronicaBORBorlandCNCanonCASCasioCOContaxCRCorelDNDenebaDLDeLormeDIDiamondDGDigitalDIGDigitechEPEpsonFOSFostexFUFujiHASHasselbladHPHPHTIHitachiILIilfordIDXIDXIYIiyamaJVCJVCKDSKDSKKKodakIBMIBMINGIntergraphLEILeicaLEXLexmarkLUCLucentLOTLotusMAMMamiyaMACMackieMAGMAG InnovisionMATMatrox GraphicsMETMetaCreationsMSMicrosoftMTMicrotechMKMicrotekMINMinoltaMTSMitsubishiMCXMicrografxNECNECNTSNetscapeNTKNewTekNKNNikonCNCanonPXPentaxOPCOpcodePNCPinnaclePOPolaroidROLRolandRORolleiNSNixdorf-SiemensOLYOlympusORO'ReillyPANPanasonicPRCPrinceton GraphicsQTQuicktimeRICRicohSAMSamsungSANSANYOSHASharpSHIShin HoSKSoftkeySNSonySUNSUNTASTascamTETEACTKXTektronixTOSToshibaULSUlead systemsUMXUMAXVWSViewSonicVIDVideonicsWGWangXXUnknownXEXeroxYAYashicaYAMYamahaXUnknown. . .etc.


[0152] Table 23 (below) illustrates values for the “resolution ” filed. The resolution field behaves the way that the size field behaves. Table 23 provides specific examples of resolution by way of illustration only. This table is not meant as a limitation.
23TABLE 23Resolution examplesLiteralDimensionMeasureType 150S50ISO200S200ISO300DC600dpc1200DI1200dpi. . .. . .etc.Type 2640 × 768P640 × 768pixels1024 × 1280P1024 × 1280pixels1280 × 1600P1024 × 1280pixels. . .. . .etc.


[0153] Table 24 lists legal values for the stain field as might be used in chemical testing.
24TABLE 24StainLiteralDescription0Black & White1Gray scale2Color3RGB (Red, Green, Blue)4YIQ (RGB TV variant)5CYMK (Cyan, Yellow, Magenta, Black)6HSB (Hue, Sat, Bright)7CIE (Commission de l'Eclairage)8LAB. . .etc.


[0154] Table 25 (below) lists legal values for the format field. Table 25 also identifies media dependences. For example, when format is ‘F’ the value of field media will be determined by Table 19.
25TABLE 25FormatLiteralDescriptionMediaAAudio-visualunspecifiedTTransparencyTable 15NNegativeTables 16-18FFacsimileTable 19PPrintTable 20CPhotocopyTable 20DDigitalTable 21VVideoSee Negative. . .etc.etc.


[0155] Referring now to FIG. 1 an overview of the present invention is illustrated. This figure provides the highest-level characterization of the invention. FIG. 1 itself represents all components and relations of the ASIA.


[0156] Reference conventions. Since FIG. 1 organizes all high-level discussion of the invention, this document introduces the following conventions of reference.


[0157] Whenever the text refers to “the invention” or to the “Automated System for Image Archiving”, it refers to the aggregate components and relations identified in FIG. 1.


[0158] Parenthesized numbers to the left of the image in FIG. 1 Invention represents layers of the invention. For example, ‘Formal specification’ represents the “first layer” of the invention.


[0159] In FIG. 1 Invention, each box is a hierarchically derived sub-component of the box above it. ‘ASIA’ is a sub-component of ‘Formal objects’, which is a sub-component of ‘Formal specification’. By implication, thus, ASIA is also hierarchically dependent upon ‘Formal specification.’ The following descriptions apply.


[0160] Formal specification 1. This represents (a) the formal specification governing the creation of systems of automatic image enumeration, and (b) all derived components and relations of the invention's implementation.


[0161] Formal objects 2. This represents implied or stated implementations of the invention.


[0162] ASIA 3. This is the invention's implementation software offering.


[0163] It is useful to discuss an overview of the present invention as a framework for the more detailed aspects of the invention that follow. Referring first to FIG. 1A an overview of the original image input process according to the present invention is shown. The user first inputs information to the system to provide information on location, author, and other record information. Alternatively, it is considered to be within the scope of the present invention for the equipment that the user is using to input the required information. In this manner, data is entered with minimum user interaction. This information will typically be in the format of the equipment doing the imaging. The system of the present invention simply converts the data via a configuration algorithm, to the form needed by the system for further processing. The encoding/decoding engine 12 receives the user input information, processes into, and determines the appropriate classification and archive information to be in coded 14. The system next creates the appropriate representation 16 of the input information and attaches the information to the image in question 18. Thereafter the final image is output 20, and comprises both the image data as well as the appropriate representation of the classification or archive information. Such archive information could be in electronic form seamlessly embedded in a digital image or such information could be in the form of a barcode or other graphical code that is printed together with the image on some form of hard copy medium.


[0164] Referring to FIG. 1B the operation of the system on an already existing image is described. The system first receives the image and reads the existing archival barcode information 30. This information is input to the encoding/decoding engine 32. New input information is provided 36 in order to update the classification and archival information concerning the image in question. This information will be provided in most cases without additional user intervention. Thereafter the encoding/decoding engine determines the contents of the original barcoded information and arrives at the appropriate encoded data and lineage information 34.


[0165] This data and lineage information is then used by the encoding/decoding engine to determine the new information that is to accompany the image 38 that is to be presented together with the image in question. Thereafter the system attaches the new information to the image 40 and outputs the new image together with the new image related information 42. In this fashion, the new image contains new image related information concerning new input data as well as lineage information of the image in question. Again, such archive information could be in electronic form as would be the case for a digital image or such information could be in the form of a barcode or other graphical code that is printed together with the image on some form of hard copy medium.


[0166] Referring to FIG. 2 the formal relations governing encoding 4, decoding 5, and implementation of the relations 6 are shown. Encoding and decoding are the operations needed to create and interpret the information on which the present invention relies. These operations in conjunction with the implementation of the generation of the lineage information give rise to the present invention. These elements are more fully explained below.



Encoding

[0167] Introduction. This section specifies the formal relations characterizing all encoding of the invention, as identified in FIG. 2 Formal specification.


[0168] Rather than using a “decision tree” model (e.g., a flow chart), FIG. 3 uses an analog circuit diagram. Such a diagram implies the traversal of all paths, rather than discrete paths, which best describes the invention's, encoding relations.


[0169] Component descriptions. Descriptions of each component in FIG. 3 Encoding follow.


[0170] Apparatus input 301 generates raw, unprocessed image data, such as from devices or software. Apparatus input could be derived from image data, for example, the digital image from a scanner or the negative from a camera system.


[0171] Configuration input 303 specifies finite bounds that determine encoding processes, such as length definitions or syntax specifications.


[0172] The resolver 305 produces characterizations of images. It processes apparatus and configuration input, and produces values for variables required by the invention.


[0173] Using configuration input, the timer 307 produces time stamps. Time-stamping occurs in 2 parts:


[0174] The clock 309 generates time units from a mechanism. The filter 311 processes clock output according to specifications from the configuration input. Thus the filter creates the output of the clock in a particular format that can be used later in an automated fashion. Thus the output from the clock is passed through the filter to produce a time-stamp.


[0175] User data processing 313 processes user specified information such as author or device definitions, any other information that the user deems essential for identifying the image produced, or a set of features generally governing the production of images.


[0176] Output processing 315 is the aggregate processing that takes all of the information from the resolver, timer and user data and produces the final encoding that represents the image of interest.



Decoding

[0177] Referring to FIG. 4 the relationships that characterize all decoding of encoded information of the present invention are shown. The decoding scheme shown in FIG. 4 specifies the highest level abstraction of the formal grammar characterizing encoding. The set of possible numbers (the “language”) is specified to provide the greatest freedom for expressing characteristics of the image in question, ease of decoding, and compactness of representation. This set of numbers is a regular language (i.e., recognizable by a finite state machine) for maximal ease of implementations and computational speed. This language maximizes the invention's applicability for a variety of image forming, manipulation and production environments and hence its robustness.


[0178] Decoding has three parts: location, image, and parent. The “location” number expresses an identity for an image through use of the following variables.
26generationGeneration depth in tree structures.sequenceSerial sequencing of collections or lots of images.time-stampDate and time recording for chronological sequencing.authorCreating agent.deviceDevice differentiation, to name, identify, and distinguishcurrently used devices within logical space.locationResReserved storage for indeterminate future encoding.locationCusReserved storage for indeterminate user customization.


[0179] The “image” number expresses certain physical attributes of an image through the following variables.
27categoryThe manner of embodying or “fixing” a representation,e.g., “still” or “motion”.sizeRepresentation dimensionality.bit-or-pushBit depth (digital dynamic range) or push status ofrepresentation.setOrganization corresponding to a collection of tabularspecifiers, e.g. a “Hewlett Packard package ofmedia tables”.mediaPhysical media on which representation occurs.resolutionResolution of embodiment on media.stainCategory of fixation-type onto media, e.g. “color”.formatPhysical form of image, e.g. facsimile, video, digital, etc.imageResReserved storage for indeterminate future encoding.imageCusReserved storage for user customization.The “parent” number expresses predecessor imageidentity through the following variables.time-stampDate, and time recording for chronological sequencing.parentResReserved storage, for indeterminate future encoding.parentCusReserved storage, for indeterminate user customization.


[0180] Any person creating an image using “location,” “image,” and “parent” numbers automatically constructs a representational space in which any image-object is uniquely identified, related to, and distinguished from, any other image-object in the constructed representational space.



Implementation

[0181] Referring to FIG. 5, the formal relations characterizing all implementations of the invention are shown. Three components and two primary relations characterize any implementation of the encoding and decoding components of the present invention. Several definitions of terms are apply.


[0182] “schemata” 51 are encoding rules and notations.


[0183] “engine ” 53 refers to the procedure or procedures for processing data specified in a schemata.


[0184] “interface” 55 refers to the structured mechanism for interacting with an engine.


[0185] The engine and interface have interdependent relations, and combined are hierarchically subordinate to schemata. The engine and interface are hierarchically dependent upon schemata.



Formal Objects

[0186] The present invention supports the representation of (1) parent-child relations, (2) barcoding, and (3) encoding schemata. While these specific representations are supported, the description is not limited to these representations but may also be used broadly in other schemes of classification and means of graphically representing the classification data.



Parent-child Implementation

[0187] Parent-child relations implement the ‘schemata’ and ‘engine’ components noted above. The following terms are used in conjunction with the parent child implementation of the present invention:


[0188] “conception date” means the creation date/time of image.


[0189] “originating image” means an image having no preceding conception date.


[0190] “tree” refers to all of the parent-child relations descending from an originating image.


[0191] “node” refers to any item in a tree.


[0192] “parent” means any predecessor node, for a given node.


[0193] “parent identifier” means an abbreviation identifying the conception date of an image's parent.


[0194] “child” means a descendent node, from a given node.


[0195] “lineage” means all of the relationships ascending from a given node, through parents, back to the originating image.


[0196] “family relations” means any set of lineage relations, or any set of nodal relations.


[0197] A conventional tree structure describes image relations.



Encoding

[0198] Database software can trace parent-child information, but does not provide convenient, universal transmission of these relationships across all devices, media, and technologies that might be used to produce images that rely on such information. ASIA provides for transmission of parent-child information both (1) inside of electronic media, directly; and (2) across discrete media and devices, through barcoding.


[0199] This flexibility implies important implementational decisions involving time granularity and device production speed.



Time Granularity & Number Collision

[0200] This invention identifies serial order of children (and thus parents) through date- and time-stamping. Since device production speeds for various image forming devices vary across applications, e.g. from seconds to microseconds, time granularity that is to be recorded must at least match device production speed. For example, a process that takes merely tenths of a second must be time stamped in at least tenths of a second.


[0201] In the present invention any component of an image forming system may read and use the time stamp of any other component. However, applications implementing time-stamping granularities that are slower than device production speeds may create output collisions, that is, two devices may produce identical numbers for different images. Consider an example in which multiple devices would process and reprocess a given image during a given month. If all devices used year-month stamping, they could reproduce the same numbers over and over again.


[0202] The present invention solves this problem by deferring decisions of time granularity to the implementation.


[0203] Implementation must use time granularity capable of capturing device output speed. Doing this eliminates all possible instances of the same number being generated to identify the image in question. In the present invention, it is recommended to use time intervals beginning at second granularity, however this is not meant to be a limitation but merely a starting point to assure definiteness to the encoding scheme. In certain operations, tenths of a second (or yet smaller units) may be more appropriate in order to match device production speed.



Specification

[0204] All images have parents, except for the originating image which has a null ( ‘O’) parent. Parent information is recorded through (1) a generation depth identifier derivable from the generation field of the location number, and (2) a parent conception date, stored in the parent number. Two equations describe parent processing. The first equation generates a parent identifier for a given image and is shown below.


[0205] Equation 1: Parent identifiers. A given image's parent identifier is calculated by decrementing the location number's generation value (i.e. the generation value of the given image), and concatenating that value with the parent number's parent value. Equation 1 summarizes this:


parent identifier=prev(generation)•parent  (1)


[0206] To illustrate parent-child encoding, consider an image identified in a given archive by the following key:


B0106-19960713T195913=JSA@1-19 S135F-OFCP@100S:2T-0123 19960613T121133


[0207] In this example the letter “B” refers to a second generation. The letter “C” would mean a third generation and so forth. The numbers “19960713” refers to the day and year of creation, in this case Jul. 13, 1996. The numbers following the “T” refers to the time of creation to a granularity of seconds, in this case 19:59:13 (using a 24 hour clock). The date and time for the production of the parent image on which the example image relies is 19960613T121133, or Jun. 13, 1996 at 12:11:33.


[0208] Equation 1 constructs the parent identifier:


parent identifier=prev(generation)•parent


[0209] or,


parent identifier=prev(B)•(19960613T121133)


=A•19960613T121133


=A19960613T121133


[0210] The location number identifies a B (or “2nd”) generation image. Decrementing this value identifies the parent to be from the A (or “1st”) generation. The parent number identifies the parent conception date and time, (19960613T121133). Combining these, yields the parent identifier A19960613T121133, which uniquely identifies the parent to be generation A, created on 13 Jun. 1996 at 12:11:13PM (T121133).


[0211] Equation 2 evaluates the number of characters needed to describe a given image lineage.


[0212] Equation 2: Lineage lengths. Equation 2 calculates the number of characters required to represent any given generation depth and is shown below:
1lineagelength=len(key)+(generation-1)(depth)*len(parent)(identifier)


[0213] Example: 26 generations, 1079 family relations. Providing a 26 generation depth requires a 1 character long definition for generation (i.e. A-Z). Providing 1000 possible transformations for each image requires millisecond time encoding, which in turn requires a 16 character long parent definition (i.e. gen. 1-digit, year-4 digit, month 2-digit, day 2-digit, hour 2-digit, min. 2-digit, milliseconds 3-digit). A 1 character long generation and 16 character long parent yield a 17 character long parent identifier.


[0214] Referring to FIG. 6, the parent child encoding of the present invention is shown in an example form. The figure describes each node in the tree, illustrating the present invention's parent-child support.


[0215]

601
is a 1st generation original color transparency.


[0216]

603
is a 2nd generation 3×5 inch color print, made from parent 601.


[0217]

605
is a 2nd generation 4×6 inch color print, made from parent 601.


[0218]

607
is a 2nd generation 8×10 inch color internegative, made from parent 601.


[0219]

609
is a 3rd generation 16×20 inch color print, made from parent 607.


[0220]

611
is a 3rd generation 16×20 inch color print, 1 second after 609, made from parent 607.


[0221]

613
is a 3rd generation 8×10 inch color negative, made from parent 607.


[0222]

615
is a 4th generation computer 32×32 pixel RGB “thumbnail” (digital), made from parent 611.


[0223]

617
is a 4th generation computer 1280×1280 pixel RGB screen dump (digital), 1 millisecond after 615, made from parent 611.


[0224]

619
is a 4th generation 8.5×11 inch CYMK print, from parent 611.


[0225] This tree (FIG. 6) shows how date- and time-stamping of different granularities (e.g., nodes 601,615, and 617) distinguish images and mark parents. Thus, computer screen-dumps could use millisecond accuracy (e.g., 615,617), while a hand-held automatic camera might use second granularity (e.g., 601). Such variable date,- and time-stamping guarantees (a) unique enumeration and (b) seamless operation of multiple devices within the same archive.



Processing Flow

[0226] Referring to FIG. 7 the processing flow of ASIA is shown.


[0227] Command 701 is a function call that accesses the processing to be performed by ASIA Input format 703 is the data format arriving to ASIA. For example, data formats from Nikon, Hewlett Packard, Xerox, Kodak, etc., are input formats.


[0228] ILF (705,707, and 709) are the Input Language Filter libraries that process input formats into ASIA-specific format, for further processing. For example, an ILF might convert a Nikon file format into an ASIA processing format. ASIA supports an unlimited number of ILFs.


[0229] Configuration 711 applies configuration to ILF results. Configuration represents specifications for an application, such as length parameters, syntax specifications, names of component tables, etc.


[0230] CPF (713,715, and 717) are Configuration Processing Filters which are libraries that specify finite bounds for processing, such pre-processing instructions applicable to implementations of specific devices. ASIA supports an unlimited number of CPFs.


[0231] Processing 719 computes output, such as data converted into numbers.


[0232] Output format 721 is a structured output used to return processing results.


[0233] OLF (723, 725, 727) are Output Language Filters which are libraries that produce formatted output, such as barcode symbols, DBF, Excel, HTML, LATEX, tab delimited text, WordPerfect, etc. ASIA supports an unlimited number of OLFs.


[0234] Output format driver 729 produces and/or delivers data to an Output Format Filter.


[0235] OFF (731, 733, 735) are Output Format Filters which are libraries that organize content and presentation of output, such as outputting camera shooting data, database key numbers, data and database key numbers, data dumps, device supported options, decoded number values, etc. ASIA supports an unlimited number of OLFs.



Applications

[0236] The design of parent-child encoding encompasses several specific applications. For example, such encoding can provide full lineage disclosure, and partial data disclosure.





[0237] Application 1: Full Lineage Disclosure, Partial Data Disclosure


[0238] Parent-child encoding compacts lineage information into parent identifiers. Parent identifiers disclose parent-child tracking data, but do not disclose other location or image data. In the following example a given lineage is described by (1) a fully specified key (location, image, and parent association), and (2) parent identifiers for all previous parents of the given key. Examples illustrates this design feature.



Example 1: 26 Generations, 1079 Family Relations

[0239] Providing a 26 generation depth requires a 1 character long definition for generation. Providing 1000 possible transformations for each image requires millisecond time encoding, which in turn requires a 16 character long parent definition. A 1 character long generation and 16 character long parent yield a 17 character long parent identifier (equation 1).


[0240] Documenting all possible family relations requires calculating the sum of all possible nodes. This is a geometric sum increasing by a factor of 1000 over 26 generations. The geometric sum is calculated by the following equation:
2sum=factor(generation-1)-1factor-1or,sum=1000(26-1)-11000-1=1081-1999=1.00·1079(3)


[0241] For 26 generations, having 1000 transformations per image, the geometric sum yields 1079 possible family relations. To evaluate the number of characters needed to represent a maximum lineage encoded at millisecond accuracy across 26 generations, the following equation is used (noted earlier):
3lineagelength=len(key)+(generation)-1(depth)·lens(parent)(2)(identifier)orlineagelength=(100)+(26-1)·(17)=525


[0242] Thus, the present invention uses 525 characters to encode the maximum lineage in an archive having 26 generations and 1000 possible transformations for each image, in a possible total of 1079 family relations.


[0243] Example 2: 216 generations, 10649 family relations. The upper bound for current 2D symbologies (e.g., PDF417, Data Matrix, etc.) is approximately 4000 alphanumeric characters per symbol. The numbers used in this example illustrate, the density of information that can be encoded onto an internally sized 2D symbol.


[0244] Providing a 216 generation depth requires a 2 character long definition for generation. Providing 1000 possible transformations for each image requires millisecond time encoding, which in turn requires a 16 character long parent definition. A 2 character long generation and 16 character long parent yield an 18 character long parent identifier.


[0245] To evaluate the number of characters needed to represent a maximal lineage encoded at millisecond accuracy across 216 generations, we recall equation 2:
4lineagelength=len(key)+(generation)-1(depth)·lens(parent)(2)(identifier)orlineagelength=(100)+(216-1)·(18)=3970


[0246] In an archive having 216 generations and 1000 possible modifications for each image, a maximal lineage encoding requires 3970 characters.


[0247] Documenting all possible family relations requires calculating the sum of all possible nodes. This is a geometric sum increasing by a factor of 1000 over 216 generations. To calculate the geometric sum, we recall equation 3:
5sum=factor(generation-1)-1factor-1or,sum=1000(216-1)-11000-1=10651-1999=1.00·10649


[0248] For 216 generations, having 1000 transformations per image, the geometric sum yields 10641 possible family relations. Thus, this invention uses 3970 characters to encode a maximal lineage, in an archive having 216 generations and 1000 possible transformations for each image, in a possible total of 10649 family relations.


[0249] Conclusion. The encoding design illustrated in Application 1:


[0250] Full lineage disclosure, partial data disclosure permits exact lineage tracking. Such tracking discloses full data for a given image, and parent identifier data for a given image's ascendent family. Such design protects proprietary information while providing full data recovery for any lineage by the proprietor.


[0251] A 216 generation depth is a practical maximum for 4000 character barcode symbols, and supports numbers large enough for most conceivable applications. Generation depth beyond 216 requires compression and/or additional barcodes or the use of multidimensional barcodes. Furthermore, site restrictions may be extended independently of the invention's apparati. Simple compression techniques, such as representing numbers with 128 characters rather than with 41 characters as currently done, will support 282 generation depth and 10850 possible relations.



Application 2: Full Lineage Disclosure, Full Data Disclosure

[0252] In direct electronic data transmission, the encoding permits full transmission of all image information without restriction, of any archive size and generation depth. Using 2D+barcode symbologies, the encoding design permits full lineage tracking to a 40 generation depth in a single symbol, based on a 100 character key and a theoretical upper bound of 4000 alphanumeric characters per 2D symbol. Additional barcode symbols can be used when additional generation depth is needed.



Application 3: Non-tree-structured Disclosure

[0253] The encoding scheme of the present invention has extensibility to support non-tree-structured, arbitrary descent relations. Such relations include images using multiple sources already present in the database, such as occurring in image overlays.



Conclusion

[0254] Degrees of data disclosure. The invention's design supports degrees of data disclosure determined by the application requirements. In practicable measures the encoding supports:


[0255] 1. Full and partial disclosure of image data;


[0256] 2. Lineage tracking to any generation depth, using direct electronic data transmission;


[0257] 3. Lineage tracking to restricted generation depth, using barcode symbologies, limited only by symbology size restrictions.


[0258] Further, ASIA supports parent-child tracking through time-stamped parent-child encoding. No encoding restrictions exist for electronic space. Physical boundaries within 2D symbology space promote theoretical encoding guidelines, although the numbers are sufficiently large so as to have little bearing on application of the invention. In all cases, the invention provides customizable degrees of data disclosure appropriate for application in commercial, industrial, scientific, medical, etc., domains.



Barcoding Implementation

[0259] Introduction. The invention's encoding system supports archival and classifications schemes for all image-producing devices, some of which do not include direct electronic data transmission. Thus, this invention's design is optimized to support 1D-3D+barcode symbologies for data transmission across disparate media and technologies.



1D Symbology

[0260] Consumer applications may desire tracking and retrieval based on 1 dimensional (1D) linear symbologies, such as Code 39. Table 5 shows a configuration example which illustrates a plausible encoding configuration suitable for consumer applications.


[0261] The configuration characterized in Table 5 yields a maximal archive size of 989,901 images (or 19,798 images a year for 50 years), using a 4 digit sequence and 2 digit unit. This encoding creates 13 character keys and 15 character long, Code 39 compliant labels. A database holds full location, image, and parent number associations, and prints convenient location number labels, for which database queries can be made.
28<generation> = 1 character<sequence> = 4 digits<date> = 6 digits<unit> = 2 digitsconstants = 2 charactersTotal =15 characters



Table 5: Configuration Example

[0262] With such a configuration, a conventional 10 mil, Code 39 font, yields a 1.5 inch label. Such a label conveniently fits onto a 2×2 inch slide, 3×5 inch prints, etc. Note, that this encoding configuration supports records and parent-child relations through a conventional “database key” mechanism, not through barcode processing.


[0263] A system and method for automated coding of objects of different types has been disclosed. It will be appreciated by those skilled in the art that the present invention can find use in a wide variety of applications. The fact that various tables have been disclosed relating to images and image forming mechanisms should not be read as a limitation, but is presented by way of example only. Other objects such as software, databases, and other types of aggregations of information can equally take advantage of the present invention.


Claims
  • 1. A system for universal object tracking comprising: an object forming apparatus; a CPU integral to the image forming apparatus; user input means connected to the CPU for receiving user input; logic stored in the CPU for receiving user input and creating archive data based upon the user input; and a graphic code producer responsive to the CPU for producing graphic codes representative of the archive data.
  • 2. The system for universal object tracking of claim 1 wherein the object forming apparatus is taken from the group consisting of imaging forming apparatus, digital data forming apparatus, electronic data forming apparatus.
  • 3. The system for universal object tracking of claim 1 wherein the object forming apparatus is a digital camera.
  • 4. The system for universal object tracking of claim 1 wherein the object forming apparatus is a video camera.
  • 5. The system for universal object tracking of claim 1 wherein the object forming apparatus is a digital image processor.
  • 6. The system for universal object tracking of claim 1 wherein the object forming apparatus is a medical image sensor.
  • 7. The system for universal object tracking of claim 6 wherein the medical image sensor is a magnetic resonance imager.
  • 8. The system for universal object tracking of claim 6 wherein the medical image sensor is an X-ray imager.
  • 9. The system for universal object tracking of claim 6 wherein the medical image sensor is a CAT scan imager.
  • 10. The system for universal object tracking of claim 1 wherein the user input means is a push button input.
  • 11. The system for universal object tracking of claim 1 wherein the user input means is a keyboard.
  • 12. The system for universal object tracking of claim 1 wherein the user input means is voice recognition equipment.
  • 13. The system for universal object tracking of claim 1 wherein the graphic codes are one-dimensional.
  • 14. The system for universal object tracking of claim 1 wherein the graphic codes are two-dimensional.
  • 15. The system for universal object tracking of claim 1 wherein the graphic codes are three-dimensional.
  • 16. The system for universal object tracking of claim 1 wherein the logic comprises configuration input processing for determining bounds for the archive data generation based on configuration input; a resolver for determining the correct value of archive data representing the image forming apparatus and the configuration input; and a timer for creating date/time stamps.
  • 17. The system for universal object tracking of claim 16 wherein the timer further comprises a filter for processing the time stamp according to configuration input rules.
  • 18. The system for universal object tracking of claim 16 wherein the configuration input comprises at least generation, sequence, data, unit, and constants information.
  • 19. The system for universal object tracking of claim 1 further comprising a graphic code reader connected to the CPU for reading a graphic code on an image representing archive information; and A decoder for decoding the archive information represented by the graphic code.
  • 20. The system for universal object tracking of claim 19 wherein the logic further comprises: logic for receiving a second user input and creating lineage archive information relating to the image based upon the archive information and the second user input; and logic for producing graphic code representative of the lineage archive data.
  • 21. The system for universal object tracking of claim 1 wherein the archive data comprises location attributes of an image.
  • 22. The system for universal object tracking of claim 1 wherein the archive data comprises physical attribute of an image.
  • 23. The system for universal object tracking of claim 1 wherein each image in an image archive has unique archive data associated with each image.
  • 24. The system for universal object tracking of claim 21 wherein the location data comprises at least: image generation depth; serial sequence of lot within an archive; serial sequence of unit within a lot; date location of a lot within an archive; date location of an image within an archive; author of the image; and device producing the image.
  • 25. The system for universal object tracking of claim 16 wherein the timer tracks year in the range of from 0000 to 9999.
  • 26. The system for universal object tracking of claim 16 wherein the timer tracks all 12 months of the year.
  • 27. The system for universal object tracking of claim 16 wherein the timer tracks time in at least hours and minutes.
  • 28. The system for universal object tracking of claim 16 wherein the timer tracks time in fractions of a second.
  • 29. The system for universal object tracking of claim 16 wherein the system is ISO 8601:1988 compliant.
  • 30. The system for universal object tracking of claim 22 wherein the physical attributes comprise at least: image category; image size; push status; digital dynamic range; image medium; image resolution; image stain; and image format.
  • 31. The system for universal object tracking of claim 20 wherein the lineage archive information comprises a parent number.
  • 32. The system for universal object tracking of claim 31 wherein the parent number comprises at least: a parent conception date; and a parent conception time.
  • 33. A method for universally tracking objects comprising: inputting raw object data to an object forming apparatus; inputting object-related data; creating first archive data based upon the object-related data; and translating the first archive data into a form that can be attached to the raw object data.
  • 34. The method for universally tracking objects of claim 33 wherein the raw object data is from a film based camera.
  • 35. The method for universally tracking objects of claim 33 wherein the raw object data is from a digital camera.
  • 36. The method for universally tracking objects of claim 33 wherein the raw object data is from a video camera.
  • 37. The method for universally tracking objects of claim 33 wherein the raw object data is from a digital image processor.
  • 38. The method for universally tracking objects of claim 33 wherein the raw object data is from a medical image sensor.
  • 39. The method for universally tracking objects of claim 38 wherein the medical object sensor is a magnetic resonance imager.
  • 40. The method for universally tracking objects of claim 38 wherein the raw object data is from an X-ray imager.
  • 41. The method for universally tracking objects of claim 38 wherein the raw object data is from a CAT scan imager.
  • 42. The method for universally tracking objects of claim 33 wherein the inputting object related data occurs without user intervention.
  • 43. The method for universally tracking objects of claim 33 wherein the inputting of object related data occurs via push button input.
  • 44. The method for universally tracking objects of claim 33 wherein the inputting of object related data occurs via voice recognition equipment.
  • 45. The method for universally tracking objects of claim 33 wherein the inputting of object related data occurs via a keyboard.
  • 46. The method for universally tracking objects of claim 33 wherein the form of the translated archive data is an electronic file.
  • 47. The method for universally tracking objects of claim 33 wherein the form of the translated data is a graphic code.
  • 48. The method for universally tracking objects of claim 47 wherein the graphic code is one dimensional.
  • 49. The method for universally tracking objects of claim 47 wherein the graphic code is two dimensional.
  • 50. The method for universally tracking objects of claim 47 wherein the graphic code is three dimensional.
  • 51. The method for universally tracking objects of claim 33 wherein the object data comprises image data and second archive data.
  • 52. The method for universally tracking objects of claim 33 further comprising reading4 the second archive data; and creating lineage archive information relating to the object based upon the first archive information and second archive information.
  • 53. The method for universally tracking objects of claim 33 wherein the inputting of object related data comprises configuration input processing for determining bounds for the archive data generation based upon configured input; determining the correct value of archive data representing the object forming apparatus and configuration input; and date/time s tamping the object related data.
  • 54. The method for universally tracking objects of claim 53 wherein date/time stamping is filtered according to configuration input rules.
  • 55. The method for universally tracking objects of claim 33 wherein the configuration input comprises at least generation, sequence, data, unit, and constants information.
  • 56. The method for universally tracking objects of claim 33 wherein the first archive data comprises location attributes of an object.
  • 57. The method for universally tracking objects of claim 33 wherein the first archive data comprises physical attributes of an object.
  • 58. The method for universally tracking objects of claim 56 wherein the location attributes comprise at least: object generation depth; serial sequence of lot within an archive; serial sequence of unit within a lot; date location of a lot within an archive; date location of an object within an archive; author of the object; and device producing the object.
  • 59. The method for universally tracking objects of claim 57 wherein the physical attributes of an object comprise at least: object category; image size; push status; digital dynamic range; image medium; software set; image resolution; image stain; and image format.
  • 60. The method for universally tracking objects of claim 52 wherein the lineage archive information comprises a parent number.
  • 61. The method for universally tracking objects of claim 52 wherein the parent number comprises at least: a parent conception date; and a parent conception time.
  • 62. The system for universal object tracking of claim 1 wherein the input means comprises a magnetic card reader.
  • 63. The system for universal object tracking of claim 1 wherein the input means comprises a laser scanner.
  • 64. The system for universal object tracking of claim 31 wherein the physical attributes further comprise; imageRes; and imageCus.
  • 65. The method for universally tracking objects of claim 33 wherein the inputting object related data is via a magnetic card reader.
  • 66. The method for universally tracking objects of claim 33 wherein the inputting of object related data is via a laser scanner.
  • 67. The method of universally tracking objects of claim 33 wherein the inputting of object related data is via an optical reader.
REFERENCE TO RELATED APPLICATION

[0001] This application is a continuation in part of co-pending application Ser. No.09/111,896 filed Jul. 8, 1998 entitled “System and Method for Establishing and Retrieving Data Based on Global Indices” and application No. 60/153,709 filed Sep. 13, 1999 entitled “Simple Data Transport Protocol Method and Apparatus” from which priority is claimed.

Provisional Applications (1)
Number Date Country
60153709 Sep 1999 US
Continuations (1)
Number Date Country
Parent 09503441 Feb 2000 US
Child 10118588 Apr 2002 US
Continuation in Parts (1)
Number Date Country
Parent 09111896 Jul 1998 US
Child 09503441 Feb 2000 US