The present invention relates to methods of processing input data to generate corresponding processed output data. Moreover, the present invention also concerns further methods of processing the processed output data to regenerate a representation of the input data. Furthermore, the present invention also relates to apparatus operable to implement these methods, and also to systems including such apparatus. Additionally, the invention is susceptible to being implemented by hardware or, alternatively, software executable on computing hardware. The invention is pertinent to electronic devices, for example mobile telephones (cell phones), video recorders, computers, optical disc players and electronic cameras although not limited thereto.
In contemporary electronic apparatus and systems, it has been found that superior picture quality can be presented to viewers when such pictures are derived from digitized image data in comparison to analogue image signals. Such benefit pertains not only to broadcast image content, for example satellite TV, but also pre-recorded image content, for example as contemporarily provided from DVDs. On account of image sequences being capable when digitized of creating a relatively large amount of data, various schemes for compressing image data have been developed; some of these schemes have given rise to established international standards such as a series of MPEG standards. MPEG is an abbreviation for Moving Picture Expert Group.
In MPEG2 compression, it possible to compress digitized image data to generate MPEG compressed image data; such compression is capable of providing a data size reduction in a range of 40:1 to 60:1. An MPEG encoder is operable to classify a sequence of images into intra-(I) frames, predictive-(P) frames and bi-directional (B) frames. Use of the I-frames arises on account of group of pictures (GOP) structures being employed in the encoder. For example, a GOP structure can comprise a sequence of frames IPPBBBPPBBB which aims to achieve best quality for I-frames, less quality for P-frames, and wherein the B-frames are arranged to employ information from “past and future” frames, namely bi-directional information. GOP structures are determined prior to MPEG encoding and groupings employed are independent of video content information. Successive images within a GOP often change more gradually such that considerable data compression can be achieved by merely describing changes, for example in terms of flow vectors; such compression is achieved by use of the aforesaid P-frames and B-frames. During MPEG2 data compression, the images in the sequence are divided into macroblocks, wherein each macroblock conveniently comprises a two-dimension field of 16×16 pixels. Such macroblock generation involves dividing images into two fields in interlaced format. Each field includes half the number of lines of pixels of corresponding frames and the same number of columns of pixels of corresponding frames. Thus, a 16×16 frame macroblock becomes an 8×16 macroblock in a corresponding field. The aforesaid flow vectors are used to describe evolution of macroblocks from a given earlier image in the sequence to macroblocks of a subsequent image thereof.
In generating the MPEG compressed data, a transform is used to convert information of pixel brightness and color for selected macroblocks into corresponding parameters in the compressed data. According to the MPEG standards, a discrete cosine transformation (DCT) is beneficially employed to generate the parameters. The parameters are digital values representing a transform of digitized luminance and color information of corresponding macroblock pixels. Moreover, the parameters are conventionally quantized and clipped to be in a range of 1 to 31, namely represented by five binary bits in headers included in the MPEG compressed data. Moreover, a table look-up method is conveniently employed for quantizing DCT coefficients to generate the parameters.
In order to try to ensure that MPEG encoding of image data corresponding to a sequence of images yields manageable MPEG encoded output data rates, it is conventional practice to utilize a complexity calculator, for example as described in a published U.S. Pat. No. 6,463,100. The complexity calculator is operable to calculate spatial complexity of an image stored in memory. Moreover, the complexity calculator is coupled to a bit rate controller for controlling quantization rate for maintaining encoded output data rate within allowable limits, the bit rate controller being operable to control the quantization rate as a function of spatial complexity as computed by the complexity calculator. In particular, quantization employed in generating the output data is made coarser when high spatial complexity is identified by the complexity calculator and less coarse for lower spatial complexity. Thus, the spatial complexity is used to control the bit rate control for quantization. Also, a defined bit rate is allocated to a group of pictures (GOP) according to a transfer bit rate and bits are allocated to each image according to the complexity of each picture depending upon whether it is an I-frame, P-frame or B-frame.
Although data compression techniques described in U.S. Pat. No. 6,463,100 are capable of providing further data compression, it is found in practice that such compression can give rise to undesirable artifacts, especially when rapid changes of scene occur giving rise to momentarily potentially high data rates. In devising the present invention, the inventor has attempted to address this problem of undesirable artifacts when high degrees of data compression are used, thereby giving rise to more acceptable image quality after subsequent image data decompression.
An object of the present invention is to provide an improved method of processing a video input signal comprising a sequence of images in a data processor to generate corresponding processed output data representative of the sequence of images.
According to a first aspect of the invention, there is provided a method of processing a video input signal in a data processor to generate corresponding processed output data, said method including steps of:
(a) receiving the video input signal at the data processor, said video input signal including a sequence of images wherein said images are each represented by pixels;
(b) grouping the pixels to generate at least one group of pixels per image;
(c) transforming the at least one group to corresponding representative transform parameters;
(d) coding the transform parameters of the at least one group to generate corresponding quantized transform data;
(e) processing the quantized transform data to generate the processed output data representative of the video input signal,
characterized in that coding the transform parameters in step (d) is implemented using quantization step sizes which are dynamically variable as a function of spatio-temporal information conveyed in the sequence of images.
The invention is of advantage in that it is capable of generating processed output data which is a more acceptable representation of the video input signal for a given volume of data.
Optionally, in the method, the at least one group corresponds to at least one block of pixels. Use of pixel blocks renders the method applicable to improve conventional image processing methods which are based on block representations.
Optionally, in the method, the quantization step sizes employed for a given group are determined as a function of spatio-temporal information which is local thereto in the sequence of images. Use of both local spatial and local temporal information is of considerable benefit in that bits of data present in the processed output data can be allocated more effectively to more suitably represent the input video signal, whilst not requiring prohibitive computing resources in making such an allocation of bits.
Optionally, in the method, the quantization step sizes are determined as a function of statistical analysis of spatio-temporal information conveyed in the sequence of images. Such statistical analysis is susceptible to giving rise to statistical parameters which are more suitable indicators to determine parts of images in the input video signal which need to be processed to greater accuracy.
Optionally, in the method, the quantization step sizes are determined as a function of a normal flow arising within each group in said sequence of images, said normal flow being a local component of image velocity associated with the group. More optionally, in the method, the normal flow is computed locally for each group from at least one of image brightness data and image color data associated with the group. Use of the normal flow as a parameter for determining appropriate quantization steps is found in practice to provide better data compression results at subsequent decompression in comparison to other contemporary advanced image compression techniques.
Optionally, in the method, the statistic analysis of the normal flow involves computing a magnitude of a mean and a variance of the normal flow for each group. In practice, the variance of the normal flow is especially useful for determining where most efficiently to allocate bits when compression sequences of images.
Optionally, in the method, adjustment of the quantization step sizes for a given group is implemented in a linear manner substantially according to a relationship:
q
—
sc
—
m=((δ·q—sc)±(λ·Γ(x)))
wherein
Γ(x)=x·e−(x−1), namely a shifted Gamma or Erlang function giving rise to non-linear modulation;
x=normal flow magnitude variance;
λ=a multiplying coefficient;
δ=a multiplying coefficient; and
q_sc=a quantization scale.
Such a relationship is capable of yet further resulting in more efficient allocation of bits when compressing sequences of images.
Optionally, the method is adapted to employ a discrete cosine transform (DCT) in step (c) and to generate groups of pixels in accordance with MPEG standards. Adapting the method to contemporary MPEG standards is capable of rendering the method workable with existing systems and equipment with relatively little change thereto being required.
According to a second aspect of the invention, there is provided processed video data generated according to the method according to the first aspect of the invention, said data being processed using quantization step sizes which are dynamically variable as a function of spatio-temporal information present in a sequence of images represented by said processed video data.
Optionally, the processed video data is stored on a data carrier, for example a DVD.
According to a third aspect of the invention, there is provided a processor for receiving video input signals and generating corresponding processed output data, the processor being operable to apply the method according to the first aspect of the invention in generating the processed output data.
According to a fourth aspect of the invention, there is provided a method of decoding processed input data in a data processor to generate decoded video output data corresponding to a sequence of images, characterized in that said method includes steps of:
(a) receiving the processed input data at the data processor;
(b) processing the processed input data to generate corresponding quantized transform data;
(c) processing the quantized transform data to generate transform parameters of at least one group of pixels of the sequence of images, said processing of the transform data utilizing quantization having quantization step sizes;
(d) decoding the transform parameters into corresponding groups of pixels; and
(e) processing the groups of pixels to generate the corresponding sequence of images for inclusion in the decoded video output data,
wherein the data processor is operable in step (d) to decode using quantization steps sizes that are dynamically variable as a function of spatio-temporal information conveyed in the sequence of images.
Optionally, in the method, the at least one group of pixels correspond to at least one block of pixels.
Optionally, in the method, the quantization step sizes employed for a given group are made dependent on spatio-temporal information which is local to the given group in the sequence of images. More optionally, in the method, the quantization step sizes are determined as a function of statistical analysis of spatio-temporal information conveyed in the sequence of images.
Optionally, in the method, the quantization step sizes are determined as a function of a normal flow arising within each group in said sequence of images, said normal flow being a local component of image velocity associated with the group.
Optionally, in the method, said normal flow is computed locally for each group from at least one of image brightness data and image color data associated with the group.
Optionally, in the method, said statistic analysis of the normal flow involves computing a magnitude of a mean and a variance of the normal flow for each macroblock.
Optionally, in the method, adjustment of the quantization step sizes for a given group is implemented in a linear manner substantially according to:
q
—
sc
—
m=((δ·q—sc)±(λ·Γ(x)))
wherein
Γ(x)=x·e−(x−1), namely a shifted Gamma or Erlang function giving rise to non-linear modulation;
x=normal flow magnitude variance;
λ=a multiplying coefficient;
δ=a multiplying coefficient; and
q_sc=a quantization scale
Optionally, the method is adapted to employ a discrete cosine transform (DCT) in step (d) and to process groups of pixels in accordance with MPEG standards.
According to a fifth aspect of the invention, there is provided a processor for decoding processed input data therein to generate video output data corresponding to a sequence of images, said processor being operable to employ a method according to the fourth aspect of the invention for generating the video output data.
According to a sixth aspect of the invention, there is provided an apparatus for processing video data corresponding to a sequence of images, said apparatus including at least one of: a processor according to the third aspect of the invention, a processor according to the fifth aspect of the invention. Optionally, said apparatus is implemented as at least one of: a mobile telephone, a television receiver, a video recorder, a computer, a portable lap-top computer, a portable DVD player, a camera for taking pictures.
According to a seventh aspect of the invention, there is provided a system for distributing video data, said system including:
(a) a first processor according to the third aspect of the invention for receiving video input signals corresponding to a sequence of images and generating corresponding processed output data;
(b) a second processor according to the fifth aspect of the invention for decoding the processed output data therein to generate video data corresponding to the sequence of images; and
(c) a data conveying arrangement for conveying the encoded data from the first processor to the second processor.
Optionally, in the system, said data conveying arrangement includes at least one of: a data storage medium, a data distribution network. For example, the system can be implemented via the Internet or via a mobile telephone (cell-phone) network.
According to an eighth aspect of the invention, there is provided software for executing in computing hardware for implementing the method according to the first aspect of the invention.
According to a ninth aspect of the invention, there is provided software for executing in computing hardware for implementing the method according to the fourth aspect of the invention.
It will be appreciated that features of the invention are susceptible to being combined in any combination without departing from the scope of the invention.
Embodiments of the invention will now be described, by way of example only, with reference to the following diagrams wherein:
Referring to
(a) via a data communication network, for example the Internet;
(b) via a terrestrial wireless broadcast network, for example via a wireless local area network (WAN), via satellite transmission or via ultra-high frequency transmission; and
(c) via a data carrier such as a magnetic hard disc, an optical disc such as a DVD, a solid-state memory device such as a data memory card or module.
The first and second processors 20, 30 are susceptible to being implemented using custom hardware, for example application specific integrated circuits (ASICs), in computing hardware operable to execute suitable software, and in any mixture of such hardware and computing hardware with associated software. The present invention is especially concerned with data compression processes occurring in the first processor 20 as will be described in greater detail later.
Referring to
It is known, as elucidated in the foregoing, to vary the quantization step applied to the parameters p1 to pn on an image frame-by-frame basis. Moreover, it is also known to render the quantization step size to be a function of spatial information included within each of the frames, for example spatial complexity. The first processor 20 is distinguished from such known approaches in that the quantization step size is varied within frames or groups of macroblocks, each group including one or more macroblocks. Moreover, the quantization step size is both a function of spatial complexity around each group and also temporal activity around each group.
For example, in the processor 20, the macroblock 130 gives rise to the parameters 170 as depicted, these parameters 170 being subsequently quantized using a quantization step size represented by 180, wherein the step size 180 is a function of spatial complexity information derived from, amongst others, the spatially neighboring macroblocks 134, 136, as well as temporal information derived from the temporally neighboring macroblocks 132, 138.
By varying the quantization step size on a macroblock basis, it is possible to include detail in the output data 200 relating to image features that are most perceptible to viewers and thereby enhance image quality for a given volume of output data 200. Thus, the processor 20 is capable of using bits in the output data 200 more optimally than has hitherto been possible for enhancing regenerated image quality in the second processor 30.
In summary, the inventor has appreciated that normal flow arising within images in the sequence 100 is a useful parameter for controlling the aforesaid quantization step size. Normal flow takes into account information pertaining to object shape, object texture fine features and its apparent motion. Optionally, the inventor has found that a variance of the normal flow magnitude is an especially useful measure for determining most optimal quantization step size to employ when processing any given macroblock of group of macroblocks within an image frame. For example, the quantization scale, and hence quantization step size, q_sc_m is beneficially substantially a function of the variance of the normal flow magnitude as provided in Equation 1.1 (Eq. 1.1):
q
—
sc
—
m=((δ·q—sc)±(λ·Γ(x))) Eq. 1.1
wherein
Γ(x)=x·e−(x−1), namely a shifted Gamma or Erlang function giving rise to non-linear modulation;
x=normal flow magnitude variance;
λ=multiplying coefficient;
δ=multiplying coefficient; and
q_sc=quantization scale.
Moreover, the inventor has found from experiments that the variance v varies considerably such that it is not ideal as a parameter from which to directly derive an appropriate value of quantization step for processing each macroblock or group of macroblocks. The inventor has appreciated, although such variance not appearing superficially ideal to use, that it is beneficial to take into account the probability distribution of the variances, for example a tail in a probability distribution, so that the variance v can be processed to generate an appropriate number from which the quantization step size can be derived.
The present invention is of benefit in that it is capable of improving image quality locally within an image, especially when the amount of spatial texture is high as well as when the local details also vary in time. If adaptive quantization according to the present invention is not used for more complex sequences of images, for example videos, visual artifacts will occur; such visual artifacts include, for example, blockiness. Conventionally, in contradistinction to the present invention, a uniform quantization scale used for all macroblocks in a given image will result in corresponding macroblocks potentially containing more spatial and temporal texture than necessary or details will not be provided with an appropriate number of bits to represent all the details adequately. Thus, an adaptive quantization scheme according to the present invention is capable of reducing the probability of noticeable blockiness being observed, such reduction being achieved by a more appropriate distribution of bits per frame, namely frame macroblocks, based on spatial texture, temporal texture and image motion.
An embodiment of the invention of the invention will now be described in more detail.
The aforesaid normal flow is defined as a normal component, namely parallel to a spatial image gradient, of a local image velocity or optical flow. The normal image velocity can be decomposed at each pixel in the sequence of images 100 into normal and tangential components as depicted in
As illustrated in
The image brightness is denoted by I(x, y) for a point P. This brightness is, for derivation purposes, constant as the point P moves from a first position (x, y) at a time t to a second position (x′, y′) at a time t′=t+Δt. Spatial co-ordinates of the point P are therefore expressible pursuant to Equation 1.2 (Eq. 1.2):
(x′, y′)=(x, y)+{right arrow over (V)}·Δt Eq. 1.2
wherein {right arrow over (V)} is a velocity vector pertaining to the movement from the first to the second position, this vector including corresponding vector components vx and vy as illustrated in
To an approximation when ΔT is relatively small, Equations 1.3 (Eqs. 1.3) pertain:
x′=x+(vx·Δt)
y′=y+(vy·Δt)
t′=t+Δt Eq. 1.3
A Taylor expansion can then be applied to approximately equate brightness at the first and second positions, namely I(x′, y′, t′)≈I(x, y, t) in Equation 1.4 (Eq. 1.4) wherein a Taylor expansion of I(x′, y′, t′) is shown up to first order in Δt, where higher order expansion terms are ignored:
Since I(x′, y′, t′)≈I(x, y, t), it is possible to derive from Equation 1.4 a corresponding Equation 1.5 (Eq. 1.5):
wherein
{right arrow over (a)}·{right arrow over (b)} denotes in Equation 1.5 the scalar product of vectors {right arrow over (a)} and {right arrow over (b)}; and
From inspection of
from which a magnitude of the normal flow vector {right arrow over (v)}n can be computed according to Equation 1.9 (Eq. 1.9):
and a unit vector direction of the normal flow vector {right arrow over (v)}n can be computed according to Equation 1.10 (Eq. 1.10):
The normal flow as provided in Equations 1.9 and 1.10 in distinction to image velocity, also serves as a measure of local image brightness gradient orientation. Variability in direction of the normal flow vector as provided by Equation 1.10 is also an implicit measure of an amount of image spatial texture per unit area of image, this measure being useable to determine suitable quantization step sizes to use when implementing the present invention.
In the processor 20, Equations 1.9 and 1.10 are computed in a discrete manner by approximating I(x, y, t) by I[i][j][k] wherein i, j and k are indices. By adopting such a discrete approach, it is then feasible to compute approximations of spatial and temporal derivatives using an image brightness cube representation indicated generally by 250 in
Given two successive image frames I1 and I2 present in the sequence of images 120 as illustrated in
The average computed in Step 3 is conveniently denoted by μB. Similarly, the variance computed in Step 2 is conveniently denoted by σB. Values for μB and σB for a group of N×N pels, namely an image block of size N×N pels, are computable in the processor 20 using Equations 2.1 and 2.2 (Eq. 2.1 and 2.2):
Optionally, when performing image processing in the processor 20, the groups of pels are selected to be blocks of pels, for example blocks of 8×8 pels or 16×16 pels. Use of such blocks results in images being tessellated into square blocks; any remainder of the picture remains untessellated. Generation of the blocks of peels is handled by the encoder 20; however, the input video beneficially has appropriate image dimensions so that interrelated peels do not occur. More optionally, in order to reduce residual untessellated image regions, a rectangular tessellation can be used and the variance of the normal flow employed; however, such an approach of employing rectangular groupings can potentially cause alignment problems with regard to standards such as MPEG 8×8 (DCT) pr MPEG 16×16 (MC).
In executing processing in the processor 20, computation of feature values within each group, for example block, is realized either:
(a) at each pels, namely pixel, for which |∇I(x, y, t)| is larger than a predetermined threshold T; or
(b) at feature points for which |∇I(x, y, t)| is larger than a pre-determined threshold TGr.
Beneficially, the thresholds T and TGr are set such that T<TGr.
The embodiment of the invention described in the foregoing is susceptible to including further refinements. A first optional feature is image registration. Moreover, a second optional feature is smoothing as a post-processing of normal flow magnitude variance.
Inclusion of image registration in processing functions executed by the processor 20 is capable of taking into account effects arising due to fast camera motion, for example panning and zooming operations. This feature is added to the steps outlined in Table 1 in the form of a velocity compensation per group of pels, for example per macroblock. A reason for needing to include such compensation arises on account of Equations 1.9 and 1.10 (Eq. 1.9 and 1.10) being approximations, namely a first order Taylor expansion of Δt which is only reasonably accurate for small to medium image velocity values. By registering consecutive images with respect to their global image velocity, it is possible to compute the aforesaid normal flow for a given image and its register pair image instead of consecutive images. Such motion compensation then renders the aforesaid approximation appropriate to use; once the images have been registered, for example to compensate for camera motion, the residual motion for which the normal flow is computed is sufficiently small to satisfy the constraints of the approximation employing a Taylor expansion. Conveniently, a 3DRS method of velocity estimation is employed per macroblock when implementing the motion compensation; the 3DRS method was developed by Philips BV and exploits a characteristics that any per macroblock block-based motion estimation is suitable for registration.
Inclusion of smoothing as a post-processing of normal flow magnitude variance is preferably implemented in the processor 20 by using first order neighborhood information as depicted in
When performing image processing as described in the foregoing in the processor 20, it is convenient to employ groups of pels implemented as 8×8 pixels which align with a standard MPEG image grid. These groups correspond to I-frame DCT/IDCT computation and describe spatial detail information. Alternatively, when performing image processing as elucidated above in the processor 20, it is also convenient to employ groups of pels implemented as 16×16 pixels which align with a MPEG image grid when processing P-frame and B-frame macroblocks for performing motion compensation (MC) in block-based motion estimation compliant with MPEG/H.26x video standards. Such an implementation allows for spatio-temporal information to be described.
In the foregoing, it is described that the quantization step size is varied as a function of normal flow, optionally the variance of the normal flow magnitude or statistics thereof, such as mean and variance. The quantization step size is in turn determined by the quantization scale denoted by q_sc which is adaptively modified as a function of the normal flow variance. From experiments, it has been appreciated by the inventor that the normal flow magnitude variance σv
q
—
m_low=((δlow·q)+(λlow·Γ(x))) Eq. 3.1
q
—
m_mid=((δmid·q)−(λmid·Γ(x))) Eq. 3.2
q
—
m_high=((δhigh·q)−(λhigh·Γ(x))) Eq. 3.3
wherein q-m and q are parameters describing the modulated and un-modulated quantization scales respectively. Moreover, an expression Γ(x)=x·exp(−(x−1)) is included to represent a Gamma function. Parameters δ and λ are adjustable parameters. Moreover, the addition “+” in Equation 3.1 is included for modeling image areas corresponding to a low magnitude of normal flow magnitude variance. Furthermore, the subtractions “−” in Equations 3.2 and 3.3 are included for coping best with textured regions in images. Terms “low”, “mid” and “high” are included to denote low, medium and high quantization scale factors respectively.
Use of multi-partitioning is of advantage in obtaining more favorable data compression in the output data 200 as a continuous range of potential quantization scale factors, and hence quantization step sizes, does not need to be supported by the processor 20. For example, modulated quantization scale factor selected per group of pels for tri-partitioning can be represented with two data bits in the output data 200 even despite the scale factors adopted for the partitioning being of greater resolution, for example pursuant to a 5-bit scale. Optionally, the number of multi-partitions is at least 5 times less than the actual resolution possible for the scale factors.
The present invention is capable of improving the visual quality of DVD+RW recordings when employed in DVD+RW devices. Moreover, the invention is also relevant to high-performance televisions for which appropriate de-interlacing and presented image sharpness improvement is a contemporary technological problem, especially in view of the increased use of digital display devices wherein new types of digital display artifacts are encountered. Furthermore, the invention is also relevant to mobile telephones (cell phones) personal data assistants (PDAs), electronic games and similar personal electronic devices capable of presenting images to users; such devices are contemporarily often provided with electronic pixel-array cameras whose output signals are subject to data compression prior to being stored, for example on a miniature hard disc drive, optical disc drive or in solid-state memory of such devices. The present invention also pertains to image data communicated, for example by wireless, to such devices.
In the system 10, the second processor 30 is designed to accept the compressed data 40 and decompress it, applying where required variable quantization steps size within each image frame represented in the data 40 for generating the data 60 for presentation on the display 80 to the user 90. When regenerating groups of pels, for example macroblocks, the processor 30 applies variable quantization steps size in regenerating parameters which are subject to an inverse transform, for example an inverse discrete cosine transform (IDCT), to regenerate groups of pels, for example macroblocks, for reassembling a representation of the sequence of images 100; the inverse discrete cosine transform (IDCT) is conveniently implemented by way of a look-up table. The processor 30 is thus designed to recognize the inclusion of additional parameters in the data 40 indicative of quantization step size to employ; optionally, these parameters can be indicative of particular a multi-partitioning pre-declared quantization scale factors in a manner as outlined with reference to Equations 3.1 to 3.3 in the foregoing.
Processing operations performed in the processor 30 are schematically illustrated in
Processing operations performed in the processor 20, for example to implement Steps 1 to 5 as described in Table 1, are schematically illustrated in
As described earlier, the processors 20, 30 are conveniently implemented by way of computing hardware operable to execute suitable software. However, other implementations are possible, for example dedicated custom digital hardware.
It will be appreciated that embodiments of the invention described in the foregoing are susceptible to being modified without departing from the scope of the invention as defined by the accompanying claims.
In the accompanying claims, numerals and other symbols included within brackets are included to assist understanding of the claims and are not intended to limit the scope of the claims in any way.
Expressions such as “comprise”, “include”, “incorporate”, “contain”, “is” and “have” are to be construed in a non-exclusive manner when interpreting the description and its associated claims, namely construed to allow for other items or components which are not explicitly defined also to be present. Reference to the singular is also to be construed to be a reference to the plural and vice versa.
Operable to employ a method means that there are means (e.g. one for each step) arranged or arrangeable to perform the method steps, e.g. as software running on a processor or hardware like an ASIC.
Number | Date | Country | Kind |
---|---|---|---|
05100068.5 | Jan 2005 | EP | regional |
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/IB2006/050004 | 1/2/2006 | WO | 00 | 6/27/2007 |