This application is a National Stage of International Application No. PCT/EP2010/055014, filed Apr. 16, 2010, and published in German as WO 2010/121957 A1 on Oct. 28, 2010. This application claims the benefit and priority of German application 10 2009 018 318.3, filed Apr. 22, 2009. The entire disclosures of the above applications are incorporated herein by reference.
This section provides background information related to the present disclosure which is not necessarily prior art.
1. Technical Field
The invention relates to an automated teller machine comprising at least one camera that produces image data. In particular, the invention relates to an automated teller machine that is configured as a cash dispenser.
2. Discussion
In the area of self-service automats, in particular cash dispensers, criminal activities in the form of manipulation are frequently undertaken with the goal of spying out sensitive data, in particular PINs (personal identity numbers) and/or card numbers of users of the automated teller machine. Manipulation attempts are known specifically in which so-called skimming devices, such as keypad overlays and similar, are installed illegally in the operating area or on the control panel. Such keypad overlays often have their own power supply, as well as a processor, a memory and an operating program so that an unsuspecting user is spied on when entering his PIN or inserting his bank card. The data mined in this way are then sent over a transmitter integrated into the keypad overlay to a remote receiver or stored in a memory in the overlay. Many of the skimming devices encountered today can be distinguished only with great difficulty by the human eye from original controls (keypad, card reader, etc.).
In order to frustrate such manipulation attempts, surveillance systems are often used that have one or more cameras installed close to the site of the automated teller machine and capture images of the entire control panel and often the area occupied by the user as well. One such solution is described in DE 201 02 477 U1. Images of both the control panel and the user area immediately in front of said panel can be captured using camera surveillance. One additional sensor is provided in order to distinguish whether a person is in the user area.
An object of the present invention is to propose a solution for camera surveillance that allows reliable detection of manipulation attempts even without the use of an additional sensor system. As part of this a high-quality data base is to be created and provided for the detection of manipulation attempts.
Accordingly, an automated teller machine is proposed in which at least one camera is provided generating image data for surveillance of the automated teller machine, wherein to detect manipulation attempts at the automated teller machine the at least one camera captures images of one or more of the elements provided in the control panel and generates image data from several individual images, and wherein the camera is connected to a data processing unit that preprocesses the image data generated into a resulting image that helps with manipulation detection. Preferably the at least one camera generates the image data from the individual images as a function of predefined criteria, specifically at predefined time intervals and/or under different lighting conditions, or ambient brightness. Predefined camera settings, particularly exposure times and/or image rates can be taken into account. The data processing unit combines these image data (individual image data) using image data preprocessing, specifically creating an average, creating a median and/or what is termed exposure blending into the resulting image, or total image, that is then available for manipulation detection. Resulting or total images (resulting image sequence) can be computed continuously at intervals to be available for a comparison to detect manipulation attempts.
At least one additional camera can be provided that is similarly mounted at or in the automated teller machine in close proximity to the control panel and captures images of at least one of the control elements, such as the keypad, card entry slot, money dispensing compartment. The image data, or individual recordings, generated by this additional camera can, in conjunction with image data from the other camera, be combined into a sequence of resulting images.
The resulting images obtained from the individual images exhibit a substantially higher image data quality than the respective individual images. A high-quality data base in the form of preprocessed image data is prepared for manipulation detection.
In so doing, it may be advantageous if the multiple individual image recordings are generated depending on at least one predefined function that sets different exposure times for the individual image recordings. This ensures that no individual image recordings are made with the same exposure time, which in turn is advantageous for exposure blending. In this context, provision can be made for the at least one predefined function to match at least one ramp function that sets increasing and/or decreasing exposure times for a series of individual image recordings. In accordance with this, the first individual image recording starts with the shortest exposure time of 0.5 ms, as an example, and with the subsequent recordings the exposure time is successively increased until, with the final image, a maximum exposure time of 2000 ms, for example, has been reached. Alternatively, the ramp can run downward, i.e. the exposure times trend downward, i.e. become successively shorter. The total duration of all individual image recordings can also be predetermined and be, for example, 10 seconds. It is also of advantage if one of the predefined functions specifies the different exposure times in such a way that they lie within a specific valuation range, for example within a first lower valuation range that extends from 0.5 ms to 1000 ms. This valuation range is preferably applied to what is known as the day mode, i.e. for the event that a brightness and/or contrast value from at least one of the individual image recordings exceeds a predefined threshold value. In night mode, i.e. when a brightness and/or contrast value of at least one of the individual image recordings falls below a predefined threshold value, the different exposure times are grouped within a second upper valuation range that may extend from 1000 ms to 2000 ms. The functions can also be combined into a function sequence.
Consequently it is also of advantage if the at least one camera generates image data for the individual image recordings dependent on events, particularly on events captured by this or by another camera. Such events may be, for example, sudden brightening or darkening of the image. Another example may be operating signals (actuation of the keypad or similar). In this respect it may be advantageous if individual image recordings are made not (only) while the event is taking place but also thereafter.
The data processing unit preferably combines the image data generated from the individual image recordings using one or more suitable image data processing methods, i.e. exposure blending. Image segmenting and/or edge detection can also be used. In this connection it is of advantage if the data processing unit segments the individual image recordings into several sub-regions assigned to the at least one captured element and processes the individual image data differently by segment. Provision can be made for the data processing unit to compile the resulting image from the sub-regions of different individual image recordings. Provision can also be made for the data processing unit to process the image data from the sub-regions using different image processing methods and/or using different variations of image data processing. The sub-regions preferably include at least a close-up or interior region and a surrounding or outer region of the captured elements, such as the slit area and the surrounding region of a card entry slot. Provision can also be made for one of the sub-regions to include a transitional region between the inner region and the outer region of the element.
The data processing unit is preferably designed in such a way that it performs both the image data preprocessing as well as the actual image data evaluation, i.e. that it computes from the individual image data the preprocessed image data for the resulting image and evaluates said data to detect manipulation attempts using image processing. To do this, the data processing unit has at its disposal a first stage receiving the preprocessed image data for the actual image processing or image data evaluation, where specifically shadow removal, edge detection, vectorizing and/or segmenting can be carried out. The data processing unit also has a second stage downstream from the first stage for feature extraction, specifically using blob analysis, edge position and/or color distribution. The data processing additionally has a third stage downstream from the second stage for classification.
The data processing unit is preferably integrated into the self-service terminal.
The elements provided in the control panel of the self-service terminal, images of which are captured by the at least one camera, include, for example, a cash dispensing drawer, a keypad, an installation panel, a card insert slot, and/or a monitor. Provision is also made for the data processing unit to trigger an alarm, disable the self-service terminal and/or trigger the additional camera when it detects a manipulation attempt at the captured elements by processing the preprocessed image data of the resulting image. This additional camera can be a portrait camera, i.e. a camera that captures an image of that area in which the user, or more specifically his head, is positioned while using the self-service terminal. In this way a portrait of the user can be taken if the need arises. It is also intended that the particular camera and/or the data processing unit is/are deactivated during operation and/or maintenance of the self-service terminal.
The invention and the advantages resulting therefrom are described hereinafter using embodiments and with reference to the accompanying schematic drawings.
The drawings described herein are for illustrative purposes only of selected embodiments and not all possible implementations, and are not intended to limit the scope of the present disclosure.
a-d show three individual image recordings as examples and a resulting image obtained therefrom;
a shows the installation location of the camera that is integrated into the card insert slot;
b reproduces the coverage area of this camera from
a)-c) show different functional sequences in the form of falling and/or rising ramps.
Corresponding reference numerals indicate corresponding parts throughout the several views of the drawings.
Example embodiments will now be described more fully with reference to the accompanying drawings.
Cameras CAMS and CAMD are located are located within a boundary, frame or similar and are mounted there. Each of these cameras CAMS or CAMD captures images from the outside in each case of at least one of the elements arranged in the control panel of the automated teller machine, for example the cash dispensing drawer 1 (shutter) and/or the keypad 2. The lateral camera CAMS preferably captures images of precisely these two elements 1 and 2; the top view camera CAMD captures images of still more elements in addition (see also
Besides the cameras positioned immediately at or in the control panel, the additional camera CAMO is located in the upper housing section of the automated teller machine ATM and is directed at the area in which the user stands when operating the automated teller machine. In particular this camera CAMO captures images of the head or face of the user and is therefore described here also as a portrait camera.
A different perspective, that of the top view camera CAMD, is clarified using
Using
a-c show as examples three individual images F1, F2 and F3 recorded at different times by the side camera CAMS (compare
As can be seen from
The resulting image R is compiled by combining individual image data, where by comparing the individual images with each other the effects of interference are detected and eliminated. For example, many sub-regions can be utilized from individual image F1, except for the area with the reflection, where individual image F1 reproduces the surface texture of the housing and of the operating elements particularly well. Likewise, many sub-regions, except for the area of the keypad and the surroundings in front of the automated teller machine, can be used from individual image F2, where the edges of the housing and of the controls in particular are reproduced clearly recognizable. Individual image F3 also has many usable sub-regions, with the keypad in particular being reproduced without any interference.
The resulting image F3 can then be computed from the different sub-regions and the numerous image components of the individual images F1 to F3. In contrast to the individual images, the resulting image does not reproduce any actual image recording but instead is the equivalent of an optimally computed image composition that shows the captured region, or the control elements, in a form free of interference. The result is to achieve a very high image quality that far surpasses the quality of the individual images. In this way an optimal foundation for the later actual image data evaluation is created.
Methods known from other fields, such as exposure blending for example, can intrinsically be used for pre-processing the image data. Individual images recorded with different exposure times are combined in such a manner that over- and underexposed areas are largely avoided and more details are preserved. The individual photographs from a series of exposures are combined, where the brightest spots in an image are replaced with the corresponding spots from the next darker image.
As is illustrated from
a)-c) illustrate different functional sequences, each with a specific ramp shape:
a) shows a first increasing ramp function MD that specifies exposure times in a lower range of values so that an exposure time T=0.5 ms is set for the first individual image “1”, and longer exposure times are set in each case for the subsequent individual image recordings. The lower range of values W1 that applies to the day mode goes to a maximum exposure time of 1000 ms for example.
b) illustrates a composite increasing ramp that initially specifies exposure times in the lower range of values in accordance with day mode MD. Then longer exposure times in the upper range of values in accordance with night mode are specified.
c) shows an increasing ramp in which the transition from day mode function MD to night mode function MN overlaps. Many other functional progressions are conceivable and can be adapted to the circumstances. In CCTV mode, for example, two to four images per second are made.
The individual image recordings can, for example, be made depending on lighting conditions. Exposure times can also be dependent on different parameters, such as the location of the automated teller machine (indoors, outdoors), type and/or installation location of the camera, lighting conditions, etc.
Edge detection can also be utilized, for example, as illustrated in
Three individual image recordings F1′ to F3′ that were taken by the side camera CAMS (see
a to 5c illustrate a further variation or additional measure for image data preprocessing of individual image recordings F1″, F2″, F3″, etc. Here the image data undergo median formation pixel by pixel.
Image data processing, which can also be carried out based on image data from several cameras, is performed in a data processing unit that also performs the actual image evaluation and is shown in
Reference is made first to
Images of predefined sub-regions are captured and measured optically to detect manipulations caused by outside intervention, changes and the like. Deviations from reference values (normal status regarding image structure, image content, weighting of pixel areas, etc.) can be detected quickly and positively. Different image processing methods (algorithms), or image processing steps (routines), are carried out within a data processing unit described more precisely later (see
b illustrates the coverage area of camera CAMK segmented into different sub-regions and shows clearly that said coverage area is essentially subdivided into three sub-regions I, II and III.
The first sub-region I principally captures images of the interior region of the card entry slot, the actual card slit, sub-region III covers the outer region of the card entry slot, sub-region II covers the transition region lying between the other two. In conjunction with
Different types of skimming modules, overlays or manipulations can be detected very precisely through the internal camera position in which camera CAMK is arranged to the side in the card entry slot 4 and captures images of sub-regions I to III. This method of installation makes it possible to segment images corresponding to sub-regions I to III and to measure said sub-regions individually. The difference in contrast between the sub-regions can be put to good use in segmenting the image recording.
The camera CAMK is oriented here in such a way that an image of a person (user or attacker) standing in front of the automated teller machine can be captured with sub-region III. These image data can be compared in particular with those from the portrait camera CAMO (see
The lighting L (see
Different methods are employed in image data processing, in particular a combination of segmenting and edge detection. The data processing unit (see
Data processing will be described in greater detail using
Camera CAMK is configured here as a color camera with a minimum resolution of 400×300 pixels. With saturated lighting, a color value distribution-based method to detect overlays and the like can be used. Camera CAMK has a wide-angle lens so that good images of the outer region (sub-region III in
In the example described here at least the cameras CAMS, CAMDA and CAMK mounted in proximity to the control panel are connected to the data processing unit 10 (see
installation of a keypad overlay,
installation of a complete overlay at the lower/bottom installation panel,
installation of an overlay at the cash dispensing drawer (shutter) and/or installation of objects to record security information, particularly PINs, such as mini-cameras, camera cell phones and similar spy cameras.
In order to detect the presence of overlays, an optical measurement of the imaged elements, such as the keypad 2, is performed inside the data processing unit 10 with the aid of the cameras CAMS and CAMD, in order to detect discrepancies clearly in the event of manipulation. Tests on the part of the applicant have shown that reference discrepancies in the millimeter range can be detected clearly. To detect foreign objects (spy camera), a combination of edge detection and segmenting can be used in order to detect clearly the contours of foreign objects in the control panel (e.g. mini-cameras). The requisite image data processing is performed principally in the data processing unit described hereinafter.
A pre-stage or control stage ST controls the individual image recordings from the cameras to generate individual image data D or D′ from which, using the method described above, preprocessed image data D* can be computed for the actual data evaluation.
For the actual image data processing and evaluation a first stage 11 for image processing of said data, a second stage 12 for feature extraction and a third stage 13 for classifying the processed data are provided. Stage 13 in turn is connected to an interface 14 over which the various alarm or surveillance devices can be activated or controlled. These devices include image falsification or manipulation detection (IFD). The first stage 11, used for image processing, is in turn connected to a second interface 15 over which a link to the CCTV unit 20 is established. Remote surveillance or remote diagnosis, for example, can be conducted with the aid of this CCTV unit.
Control stage ST is responsible for controlling the cameras CAMS and CAMK to generate the individual image data D or D′. The subsequent first stage 11 computes from said data the prepared image data D* (computed complete image data), where here in particular steps such as shadow removal, edge detection, vectorizing and segmenting are carried out. The downstream second stage 12 is used for feature extraction that can be carried out, as an example, using blob analysis, edge positioning and/or color distribution. Blob analysis, for example, is used for detecting cohesive regions in an image and for conducting measurements on the blobs. A blob (binary large object) is an area of contiguous pixels having the same logic status. All pixels in an image that are part of a blob are in the foreground. All other pixels are in the background. In a binary image pixels in the background have values that correspond to zero, while each pixel not equal to zero is part of a binary object.
Then, in stage 13 a classification is made that determines on the basis of the extracted features whether a hostile manipulation has occurred at the self-service terminal, or automated teller machine, or not.
The data processing unit 10 can, for example, be implemented by means of a personal computer that is linked to the automated teller machine ATM or is integrated into said ATM. Besides camera CAMS and CAMK that capture images of the sub-regions of the control panel CP already mentioned, the additional camera CAMO can be installed on the automated teller machine ATM (refer to
Store a photograph of the attacker, when both individual cameras CAMS and/or CAMK and the supplementary portrait camera CAMO can be activated,
Alarm the active automated teller machine applications and/or a central management server and/or a person, for example, by e-mail,
Introduce counter-measures that include disabling or shutting down the automated teller machine,
Transmit data, specifically images, of the manipulation detected, for example over the Internet or a central office.
The operator of the automated teller machine can configure the scope and the type of measures, or countermeasures, taken using the system described here.
As described above, several cameras can be provided, installed directly at the control panel, where cameras CAMS and CAMD capture images of the control panel from the outside, and camera CAMK captures images of the card entry slot from the inside. A supplementary portrait camera can be installed in addition (see CAMO in
All the cameras preferably have a resolution of at least 2 megapixels. The lenses used have an acquisition angle of about 140 degrees and greater. In addition, the exposure time of the cameras used can be freely adjusted over a broad range from 0.25 msec, for example, up to 8000 msec (8 secs.). In this way, it is possible to adjust to the widest possible range of lighting conditions. Tests by the applicant have shown that a camera resolution of about 10 pixels per degree can be obtained. Referred to a distance of one meter, it is possible to achieve an accuracy of 1.5 mm per pixel. This means, in turn, that a manipulation can be detected reliably using a reference deviation of 2 to 3 mm. The closer the camera lens is to the imaged element or observed object, the more precise the measurement. As a result, a precision of less than 1 mm can be achieved in closer regions.
Depending on where the automated teller machine will be used, for example outside or inside, as well as on the existing light conditions, it may be of advantage to install the camera CAMS in the lateral part of the housing of the automated teller machine ATM or in the upper part of the housing. Various possibilities for surveillance exist depending on the camera position. When monitoring the different elements, or sub-regions, the following in particular can be achieved:
Capturing images of the cash dispensing drawer (shutter) 1 permits checking for manipulation in the form of cash trappers, i.e. special overlays. Capturing images of the keypad area makes it possible to determine manipulation attempts using overlays or changes to security lighting. Capturing images of the installation panel makes it possible in particular to detect complete overlays. Capturing images of the card entry slot 4, particularly using an integral camera, makes it possible to detect manipulations in this area.
It has been shown that discrepancies of 2 mm can be clearly detected in particular at the keypad and the card slot. Discrepancies at the rear outer edge of the installation panel can be detected starting at 4 mm. Discrepancies at the lower edge of the shutter can be detected starting at 8 mm.
The data processing unit 10 (refer to
The image data from the different cameras CAMS, CAMD and/or CAMK are also compared with one another to determine, for example, whether individual cameras have been manipulated. If, as an example, camera CAMD was masked, there is a discrepancy with the image recordings from the other cameras. It can be established very quickly from the brightness of the images whether darkening occurs at only a single camera so that manipulation or masking can be assumed. The combination and evaluation of several camera signals or image data increases the robustness of manipulation surveillance and prevention of false alarms. Some of the uses for the image data or information are as follows:
Distinguishing between artificial and natural darkening: if a camera is masked, the image it has recorded is inconsistent with the images from the other cameras. If the natural light (daylight) or the artificial light (area lighting) fails, the effect is the same at all cameras or at least similar. Otherwise the system detects a manipulation attempt.
Detection of deception attacks on the camera array, for example with photographs pasted in front of them: if an individual camera shows a different image (brightness, movement, colors, particularly regarding the outer region), this indicates attempted deception.
Increasing robustness, particularly when the card entry slot is masked: If it is covered, the integral camera (see CAMK in
Furthermore, the surroundings can be inspected, for example, for emission of the lighting for the card entry slot 4. Connecting the system to the Internet over interface 23 makes it possible to drive the camera, or the different cameras, by remote access. The image data obtained can also be transmitted over the Internet connection to a video server. So the respective camera acts almost as a virtual IP camera. The CCTV unit 20 described above in particular serves the possibility of such video surveillance, where the interface 15 to the CCTV unit is laid out for the following functions:
Retrieving an image, adjusting the image rate, the color model, and image resolution, triggering an event in the CCTV service when preparing a new image and/or possible visual enhancement of detected manipulation in an image provided.
The system is designed such that in normal operation (e.g. withdrawing money, account status inquiry, etc.) no false alarms are created by hands and/or objects in the image. For this reason, manipulation detection is deactivated in the period of normal use of an ATM. Also, time periods of cleaning or other brief uses (filing bank statements, interaction before and after the start of a transaction) should not be used for manipulation detection. Essentially, only fixed and immobile manipulation attempts are preferably analyzed and detected. The system is designed such that surveillance operates even under a great variety of light conditions (day, night, rain, cloud, etc.). Similarly, briefly changing light conditions, such as light reflections, passing shadows and the like are compensated for or ignored in the image processing in order to prevent a false alarm. In addition, events of a technical nature, such as a lighting failure and the like, can be taken into consideration. These and other special cases are detected for classification and solved in particular by the third stage.
The method carried out by the system described for detecting manipulation exhibits in particular the following sequence (refer to
First, preprocessed total image data D* are computed from the original individual image data D or D′ that are used as the starting point for the actual data evaluation.
In a first step, an image is initially recorded, where the camera parameters are adjusted to generate suitable images. In so doing, a series of images, or corresponding image data D or D′, is recorded that serves as the basis, or reference, for preprocessing.
Then the image data are processed further, where said data are processed such that they are as suitable as possible for evaluation. For example, several images are combined into a target image and optimized using image enhancement algorithms. The following steps in particular are performed:
Shadow removal, deletion of moving objects, elimination of noise and/or combination of differently exposed recordings.
Some of the adjustments to the cameras are for different exposure times, to eliminate reflections and to compile well lighted areas. The images are preferably compiled over a predetermined period in order to obtain the best possible images for manipulation detection. Feature extraction is performed in a third step (stage 12) in which image analysis methods are applied to the preprocessed images or image data in order to inspect said images or image data for specific features, such as edge positions or color distributions. A number or a value is assigned to each feature that indicated how well the corresponding feature was found in the scanned image. The values are collected in what is known as a features vector.
In a further step, a classification is carried out (Stage 13), i.e. the features vector is passed on to a classification sequence to reach a decision whether manipulation exists or not. Types of classifiers are used that are able to indicate a confidence, i.e. a probability or certainty, with which the decision holds true. The classification mechanisms used may include, for example:
Learning classifier systems, Bayes classifiers, support vector machines (SVM) or decision trees (CART or C 4.5).
The system described here is preferably modular in construction, in order to make different configurations possible. The actual image processing and the CCTV connection are implemented in different modules (refer to
The system presented here is also suitable for documenting the manipulations detected, or archiving said manipulations digitally. In the event of a detected manipulation, the images recorded, along with corresponding meta-information, such as time stamp, type of manipulation, etc., are saved on a hard disc in the system or on a connected PC. Messages can also be forwarded to a platform for the purposes of reporting, such as error reports, status reports (deactivation, change of mode), statistics, suspected manipulation and/or alarm reports. In the event of an alarm, a suitable message containing the specific alarm level can be transmitted to the administration interface or interface. The following possibilities can additionally be implemented at said interface:
Retrieving camera data, such as the number of cameras, construction status, serial number, etc., master camera data, or adjustment of camera parameters and/or registration for alarms (notifications).
The invention presented here is specifically suitable for reliably detecting hostile manipulations at a self-service terminal, such as an automated teller machine. To this end, the control panel is continuously and automatically monitored by at least one camera. Using image data processing, the elements captured by the camera are measured optically to identify deviations from reference data. It has already been shown that discrepancies in the range of mere millimeters can be identified reliably. A combination of edge detection and segmenting is preferably used for detecting foreign objects so that contours of objects left behind can be clearly detected and identified. In the event of attempted manipulation, countermeasures or actions can be initiated.
The invention clearly increases the reliability with which manipulations can be detected through the combination proposed here of several cameras and intelligent image data processing.
In a preferred embodiment the invention has the following camera arrangement:
One camera at the card entry slot, one camera at the control panel and one camera in the upper area of the automated teller machine for recording portrait photos or videos. In addition, the cameras are connected to the data processing unit previously described. Inside the data processing unit the image data or information acquired by the cameras is used in the following and other ways:
Detection of or distinguishing between artificial and natural darkening: If one camera is masked, the image it recorded is inconsistent with the images from the other cameras. If natural or artificial light fails, the effect appears at all cameras equally. Detection of deception attacks on the camera system, e.g. using stuck on photographs: If one camera shows another image (different brightness, movement, colors, etc.), this indicates a deception attempt. Increasing robustness of masking detection at the card entry slot: If the card entry slot is masked, the integral camera there CAMK shows a different image of the outer region than the other cameras.
The preprocessing of the camera image data described here, in which low-distortion or distortion-free total images are computed from individual recordings results in an increase in the reliability of detection of manipulation attempts and also serves to prevent false alarms.
In summary, a self-service terminal is proposed that has at least one camera to detect manipulation attempts that captures images of one or several elements provided in the control panel, such as a keypad, cash dispensing drawer, card entry slot, and generates image data from several individual image recordings. The at least one camera is connected to a data processing unit that preprocesses the image data (individual image data) generated into a resulting image. The preprocessed image data of the resulting image can be computed from the individual image data, for example, using exposure blending and represent a very good data base for data evaluation for manipulation detection.
The present invention was described using the example of an automated teller machine but is not restricted thereto, rather it can be applied to any type of self-service terminal.
The foregoing description of the embodiments has been provided for purposes of illustration and description. It is not intended to be exhaustive or to limit the invention. Individual elements or features of a particular embodiment are generally not limited to that particular embodiment, but, where applicable, are interchangeable and can be used in a selected embodiment, even if not specifically shown or described. The same may also be varied in many ways. Such variations are not to be regarded as a departure from the invention, and all such modifications are intended to be included within the scope of the invention.
Number | Date | Country | Kind |
---|---|---|---|
10 2009 018 318 | Apr 2009 | DE | national |
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/EP2010/055014 | 4/16/2010 | WO | 00 | 10/12/2011 |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2010/121957 | 10/28/2010 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
7881497 | Ganguli et al. | Feb 2011 | B2 |
7948538 | Asoma | May 2011 | B2 |
20080266424 | Asoma | Oct 2008 | A1 |
20090201372 | O'Doherty et al. | Aug 2009 | A1 |
20100259626 | Savidge | Oct 2010 | A1 |
Number | Date | Country |
---|---|---|
20102477 | May 2001 | DE |
20318489 | Feb 2004 | DE |
2351585 | Jan 2001 | GB |
2351585 | Jan 2001 | GB |
WO-2007093977 | Aug 2007 | WO |
Entry |
---|
International Preliminary Report on Patentability (Chapter I of the Patent Cooperation Treaty) in German (with English translation) for PCT/EP2010/055014, issued Oct. 25, 2011. |
International Search Report (in German with English Translation) for PCT/EP2010/055014, mailed Sep. 16, 2010; ISA/EP. |
English translation of Chinese Office Action for Application No. 2010-80027721.1 dated Mar. 25, 2014 (4 pages). |
Number | Date | Country | |
---|---|---|---|
20120038775 A1 | Feb 2012 | US |