The present invention relates to a device for providing image data of adjustable anatomical planes, to a system for intervention planning, to a method for providing image data of adjustable anatomical planes as well as to a computer program and a computer readable medium for providing image data of adjustable anatomical planes.
As an example, during an intervention on the mitral valve, physicians may use 3D ultrasound to image the structure. In order to understand the anatomy, or in order to estimate relevant anatomical parameters, or in order to plan/guide the intervention, they need to assess the target structure at certain locations. In an example, pre-set view planes are created within the 3D volume in which the relevant information is visible. The manual setting can be a tedious process. In another example, view planes are encoded in a model, and after segmentation, the pre-encoded planes can be retrieved. However, it may be challenging to encode all potentially relevant planes in the model.
There may thus be a need to provide a more user friendly way of images in selected viewing planes.
The object of the present invention is solved by the subject-matter of the independent claims: further embodiments are incorporated in the dependent claims. It should be noted that the following described aspects of the invention apply also for the device for providing image data of adjustable anatomical planes, for the system for intervention planning and for the method for providing image data of adjustable anatomical planes.
According to the present invention, a device for providing image data of adjustable anatomical viewing planes is provided. The device comprises a data input, a data processor, a user interface and an output interface. The data input is configured to receive 3D image data of a region of interest of a subject. The data processor is configured to establish anatomical context for identifying anatomical reference locations within the 3D image data. The data processor is also configured to generate a reference coordinate system based on the identified anatomical reference locations. The data processor is further configured to define an anchor point. The data processor is still further configured to compute a viewing plane as a selected anatomical imaging plane based on the anchor point and at least one plane-related parameter. The data processor is configured to set the anchor point using the established anatomical knowledge. The user interface is configured for entering the at least one plane-related parameter for determining the viewing plane as the selected anatomical imaging plane by the user. The output interface is configured to provide a representation of a view in the selected anatomical imaging plane.
As an effect, the user is provided with a facilitated and less tedious way of achieving different viewing planes that can be customized for particular needs.
In another claimed option, a device for providing image data of adjustable anatomical viewing planes is provided. The device comprises a data input, a data processor, a user interface and an output interface. The data input is configured to receive 3D image data of a region of interest of a subject. The data processor is configured to establish anatomical context for identifying anatomical reference locations within the 3D image data. The data processor is also configured to generate a reference coordinate system based on the identified anatomical reference locations. The data processor is further configured to define an anchor point. The data processor is still further configured to compute a viewing plane as a selected anatomical imaging plane based on the anchor point and at least one plane-related parameter. The user interface is configured for entering the at least one plane-related parameter for determining the viewing plane as the selected anatomical imaging plane by the user. The output interface is configured to provide a representation of a view in the selected anatomical imaging plane.
According to an example, the output interface comprises a display configured to display the selected plane. As an additional or alternative option, the user interface comprises an adjustment control element configured to adjust the location of the displayed plane. The data processor is configured to provide the adjustment in real-time.
According to an example, the display is configured to display a control image indicating the spatial position of the selected plane within the anatomical context.
According to an example, the anchor point defines the point that will be visible in all possible planes when the user rotates the plane, and, for example, around which the view plane will be rotated.
As an option, provided alternatively or in addition, the at least one plane-related parameter comprises at least one of the group of an angle, a viewing direction, one or more image lines and one or more image points.
According to an aspect, the anchor point is pre-defined, such as the mid-point between antero-lateral and postero-medial commissures. The anchor point is provided for establishing the point around which the (viewing) plane is rotated. In an example, the user shifts the anchor point from the default location along the perpendicular to the AL-PM direction that can be defined by the user. After the anchor point is defined, the user defines an angle. e.g. either by a certain value or with the use of a slider, and this angle is the one that determines the actual plane as the viewing plane. The angle may be provided as a deviation from a perpendicular direction to the AL-PM direction. As an example, an angle which when 0, is the actual perpendicular direction.
According to an example, for establishing anatomical context, the data processor is configured to segment the 3D image data to provide a segmentation of at least a part of the 3D image data. The data processor is also configured to establish anatomical context for identifying the anatomical reference locations within the 3D image data based on the segmentation.
In an option, for segmenting the 3D image data, the data input is configured to receive an anatomical model. The data processor is configured to adapt the model to the 3D image data; and to base the segmenting on the adapted model.
In an additional or alternative option, for establishing anatomical context, the data processor is configured to provide machine learning approaches comprising at least one of the group of machine learning, deep-learning and convolutional neural networks to classify voxels or detect landmarks.
According to an example, the data processor is configured to define the reference system by a first direction along the mitral valve plane normal, a second direction along the antero-lateral and postero-medial direction, and a third direction that is varied accordingly.
In an example, the data processor is configured to apply a user-selected viewing plane for multiple interventions. The multiple interventions can be provided as pre-selected interventions. The interventions can also be referred to as dedicated interventions.
According to the present invention, also a system for intervention planning is provided. The system comprises a device for providing image data of adjustable anatomical viewing planes according to one of the preceding claims. The system also comprises an imaging device for providing images of a region of interest of an object. The imaging device is configured to generate image data of the object und to supply the image data to the device for providing image data of adjustable anatomical planes.
According to an example, the imaging device is an ultrasound imaging device configured to provide 3D ultrasound data.
In an option, the ultrasound imaging device is configured to acquire a 2D ultrasound image according to the selected anatomical imaging plane.
According to the present invention, also a method for providing image data of adjustable anatomical viewing planes is provided. The method comprises the following steps:
In another claimed option, a method for providing image data of adjustable anatomical viewing planes is provided. The method comprises the following steps: In a first step. 3D image data of a region of interest of a subject is received. In a second step, anatomical context for identifying anatomical reference locations within the 3D image data is established. In a third step, a reference coordinate system is generated based on the identified anatomical reference locations. In a fourth step, an anchor point of a viewing plane is established. In a fifth step, at least one plane-related parameter for determining the viewing plane as a selected anatomical imaging plane is entered by the user. In a sixth step, a viewing plane as the selected anatomical imaging plane is computed based on the anchor point and the at least one plane-related parameter. In a seventh step, a representation of the selected anatomical imaging plane is provided.
According to an example, a pre-set/already defined anchor point is provided. i.e. is available, such as the middle point between the AL and PM, which can be then adjusted/changed along the normal to the AL-PM direction. The pre-set/already defined anchor point may be provided by the system itself of may be adjusted by user entry.
As an example, the anchor point is set by the system. As an option, a location is between the commissures, or at the mitral valve center. In an option, the system sets the anchor point using the anatomical knowledge from the system and knowledge about the intended procedure.
As another example, further adjustments to the anchor point are possible, such as by the system or the user.
After the anchor point has been set, the definition of potential viewing planes is left with one degree of freedom, like the viewing angle. This provides the potential planes, as shown in
According to an aspect, imaging with adjustable anatomical planes is provided by establishing a reference system based on anatomical features. By providing one degree of freedom for imaging plane, the user can adapt the viewing plane in an intuitive way and adjust the representation to the current intervention and also to user's needs.
These and other aspects of the present invention will become apparent from and be elucidated with reference to the embodiments described hereinafter.
Exemplary embodiments of the invention will be described in the following with reference to the following drawings:
Certain embodiments will now be described in greater details with reference to the accompanying drawings. In the following description, like drawing reference numerals are used for like elements, even in different drawings. The matters defined in the description, such as detailed construction and elements, are provided to assist in a comprehensive understanding of the exemplary embodiments. Also, well-known functions or constructions are not described in detail since they would obscure the embodiments with unnecessary detail. Moreover, expressions such as “at least one of”, when preceding a list of elements, modify the entire list of elements and do not modify the individual elements of the list.
A first arrow 24 indicates the input of the 3D image data of a region of interest of a subject. e.g. from a database or from an imaging device. A second arrow 26 indicates the output of the representation of a view in the selected anatomical imaging plane.
A frame 28 indicates that the data input 12, the data processor 14, the user interface 16 and the output interface 18 can be arranged within a common housing or structure. In another option, the data input 12, the data processor 14, the user interface 16 and the output interface 18 are provided as separate components, data-connected with each other.
In an option, the imaging device 102 is an ultrasound imaging device configured to provide 3D ultrasound data. A mobile ultrasound transducer 104 for imaging an object 106 is indicated. Optionally, the ultrasound imaging device is configured to acquire a 2D ultrasound image according to the selected anatomical imaging plane.
The device 10 for providing image data may be provided in the context of an operation console or control arrangement, shown in the right foreground. A display 116 is shown, together with a keyboard 118, a mouse 120, a control panel 122 and graphic tablet 124. The imaging device 102 is data-connected with the device 10 for providing image data of adjustable anatomical planes, as indicated with connecting line 124.
The system 100 for intervention planning can also be referred to as system for intervention guidance.
In an option, not shown, the imaging device is an X-ray imaging device configured to provide 3D X-ray imaging data, such as CT data.
In an option, the acquired 2D view is an additional step once the adapted view plane is provided. The acquired 2D view is so-to-speak in addition to the 2D view in form of the adapted image plane, but the acquired 2D view in better quality.
In another option, the acquired 2D view is provided, once the adapted view plane is defined, but not actually provided. The acquired 2D view is so-to-speak provided as the adapted view plane instead of the 2D view in form of the adapted image plane, but the acquired 2D view is having a better quality as a direct presentation of the adapted image plane would have.
As an aspect. 3D image data. e.g. ultrasound image data, is acquired, and based on the geometrical information, a reformatting is provided within the 3D volume to select and display a selected plane, without any further ultrasound image data acquisition.
In an example, an ultrasound imaging system may provide higher ultrasound image quality in 2D than the image data in 3D. Providing the option to acquire 2D image along the selected and determined viewing plane allows to provide the user, such as physicians, with 2D images with an increased degree of fine details at a specific location (for which the viewing plane has been adapted). Besides the reformatting of the 3D volume, by achieving the defined 2D viewing plane, it is provided to control an ultrasound imaging device to acquire the specific 2D view depending on the current user selection.
In an example, it is provided that the device is configured for the following steps:
From that, cither: a slide through the acquired 3D volume at that location is shown: or a smaller echo volume is re-acquired covering the selected plane with higher frame rate and/or image quality (assuming the ultrasound probe is still in the same location and the reference frame still matches the anatomy). As an option for the re-acquiring, it is provided the steps of: acquiring the exact plane in 2D mode, if geometrically possible. i.e. if it intersects with the transducer position; acquiring a plane in 2D that is as close as possible to the selected plane at the location of the mitral annulus, for example, a plane defined by the intersection of the selected planed and the mitral annulus plane (defines a line)+the probe position (in combination defines a plane); and acquiring a smaller 3D volume. e.g. a thin slab fully including the selected plane. Because the scanned volume is smaller, the frame rate, for example, can be higher.
In an example, the device 10 for providing image data of adjustable anatomical planes is provided for cardiac interventions, such as valve treatment or replacement procedures.
The data input 12 relates to providing the image data to the processor 14. The image data is thus received from a data supply or data source and forwarded to the data processing. The image data may be provided directly from an imaging device or from a data storage. The data input 12 is also referred to as data image input or image data input unit or data receiving module.
The data processor 14 relates to an arrangement for processing data. i.e. for computing steps to process the data. The data processor 14 is also referred to as processing unit or computing unit or processing module.
The user interface 16 relates to a device for allowing user interaction, such as entering control commands for the data processing. An example of data entry is the entering of the at least one plane-related parameter. Another example of the data entry is the selection or change of the viewing plane or imaging plane. The user interface 16 is also referred to as user interaction unit or user interaction module.
The at least one plane-related parameter comprises at least one of the group of an angle, a viewing direction, one or more image lines and one or more image points.
The output interface 18 relates to providing the computed image plane data, for example for displaying purposes. The output interface 18 is also referred to as processed data supply or output unit or output module.
In a first option, the “representation of a view in the selected anatomical imaging plane” refers to the generation of a 2D image from the 3D image data in the 2D viewing plane.
In a second option, provided in addition or as an alternative, the “representation of a view in the selected anatomical imaging plane” refers to the targeted acquisition of a 2D ultrasound image in the 2D viewing plane.
In an option, the user interface 16 and the output interface 18 are provided in an integrated manner as bi-directional user interface (not shown in detail).
The image data of adjustable anatomical planes can also be referred to as representations of adjustable anatomical planes.
As a result, the present invention thus provides multi-purpose anatomical planes for intervention planning with a single degree-of-freedom. As an effect, during an intervention, the physician can focus on the procedure, since not being forced to deal with multiple measurement systems between interventions which is difficult and prone to errors. The user is thus provided with an option to determine whatever viewing plane is needed, which means a more user-friendly way of providing data than to refine e.g. pre-encoded planes.
To this end, information is extracted from the model for certain relevant structures and the remaining refinement for achieving the most appropriate view is chosen by the user. This can be done for each problem at hand.
In an example, the anchor point of a viewing plane is determined, respectively defined. i.e. the data processor 14 is further configured to determine or define the anchor point.
In an example, a common reference frame and a user interface for multiple mitral valve interventions is provided. From the segmentation, some anatomical constraints are defined.
For example the normal direction of the mitral valve and the anchor point around which planes are rotated are defined.
For example, the center of the mitral valve is defined, and then the anchor point can be defined relative to the center and normal direction of the mitral valve.
With a single control, such as a slider, the user can adjust the plane in a way that it covers a set of anatomically interesting locations. e.g. planes cutting primarily on a direct path from the anterior to the posterior leaflet. Solely by the choice of the anchor point and the plane-related parameter, such as the angle as an option, the method can be applied to interventions that directly target the mitral leaflets, or procedures that target the mitral annulus. In an example, the user interface and the relevant anatomical boundaries are the same for both procedures. For example, the initial position may be a reference view along the AP-direction in the center of the mitral valve, and a certain position of the slider may correspond to a plane that cuts through the commissure points. Furthermore, in an example, external information is included into the same reference frame. i.e. setting the plane to the location of a tracked device, or to the location in which high regurgitant flow has been detected. Such locations can be encoded within the reference coordinate system so that the user controls can be updated and the physician can further navigate from there.
In an example, for the segmentation, the data processor 14 comprises a segmentor, also referred to as segmentation unit. In another example, for the setting up the reference coordinate system, the data processor 14 comprises a geometrical post-processing definer, also referred to as geometrical post-processing unit. The geometrical post-processing definer or geometrical post-processing unit can be provided as separate processor or be integrated in a common processor circuit. In a further example, for the computing of the view plane, the data processor 14 comprises a view plane updater, also referred to as view plane update unit.
In an option, also shown in
In another option, also shown in
As an example, the data processor 14 is configured to provide the adjustment in real-time. The term real-time refers to an update within a time span that the user receives as neglectable. e.g. within less than five seconds, such as within less than two or less than one second.
In an example, the user interface 16 comprises a slider element that controls an angle between the central direction between and the displayed plane. The central direction can further be defined as the perpendicular plane to the commissures plane going through the mitral valve center.
In an example, the user interface 16 comprises a slider element that controls an angle of the displayed plane.
If the plane is adjusted by enforcing it to cut through a distinct anatomical point. e.g. a region of high regurgitant flow, this plane position corresponds to an angle in the defined coordinate system, and the slider can be adjusted to that angle position. Then, the user can further manipulate the plane from this position.
In an example, a variable view plane unit is represented by an interface that can take as inputs either an angle value or a point chosen by the user. This will be used to compute the new view plane, and in the case of the latter, the angle from the given point will also be computed and updated. In an example, an interactive component for modifying views is provided.
In an example, the anchor point defines the point that will be visible in all possible planes when the user rotates the plane and around which the rotation is taking place.
For establishing anatomical context, the data processor 14 may be configured to:
In the first approach, the model is used for the segmentation of the image data. In an example, the model comprises labels for indicating predefined anatomical features, such as anatomical landmarks. As a result, segmented image data of the patient is provided and, in addition, also geometrical landscape landmarks can be provided.
In the second approach, the knowledge for identifying the anatomical structures within the image data is provided by the machine learning algorithm.
In an example, for the generation of the reference coordinate system, the data processor 14 is configured to provide at least one of the group of:
In an example, the direction normal to the mitral valve plane is provided roughly normal to the mitral valve plane. As an example, the term roughly relates to a deviation of maximum up to +/−25°. e.g. up to +/−10° or up to +/−5°.
In an example, for determining the direction normal to the mitral valve plane, such as the direction roughly normal, the data processor 14 is configured to determine a regression plane fitting through at least a part of all vertices of the mitral annulus or by taking a cross product between vectors connecting the anterior-posterior end and connection the antero-lateral and postero-medial end of the mitral commissures.
In an example, the data processor 14 is configured to define the reference system by a first direction along the mitral valve plane normal, a second direction along the antero-lateral and postero-medial direction, and a third direction that is varied accordingly.
The antero-lateral and postero-medial direction refers to the direction from the antero-lateral commissure point to the postero-medial commissure point.
The data processor may 14 be configured to use information about locations of actual high regurgitant flow to set the plane location.
As an option, the display 20 is configured to display a control image indicating the spatial position of the selected plane within the anatomical context.
The term spatial position refers to location and orientation in space in relation to the anatomical structure.
The data processor 14 is configured to apply a user-selected viewing plane for multiple interventions.
In an example, the data processor 14 is configured to apply the user-selected viewing plane also for other imaging modalities.
The generating of the reference coordinate system is also referred to as setting up a reference coordinate system.
In an example, the anchor point of the viewing plane is determined, respectively defined.
In an example, the anchor point is provided. i.e. determined by the system, as a default option, such as the middle point between the antero-lateral and the postero-medial commissures.
In another example, the anchor point is shifted from the default option along the perpendicular direction to the direction from the antero-lateral commissure point to the postero-medial commissure point by the user for determining the viewing plane as a selected anatomical imaging plane is provided.
In an example, as the at least one plane-related parameter, the user enters a viewing angle. In an example, the user enters a viewing direction, or one or more image lines or one or more image points. For example, the user enters or adapts the viewing angle, or other plane-related parameter, via a slider or via entering a specific value.
In another example, the plane related parameter is an image point. An example of an image point is a point within the 3D image data, Another example of an image point is a point that is determined from the anatomical model, such as an A1 or A2 anatomical point. In an example, an image line is defined by two geometrical points in the image space. In another example, the image content is considered for the image line. e.g. defined by an image gradient.
The anchor point defines the point that will be visible in all possible planes when the user rotates the plane. As also explained below, in an example, the anchor point determines for which procedure the planes are provided. With the selected anchor point being fixed, the user then still has one degree of freedom. e.g. an angle, a clicked point or the location of high regurgitant flow to fully define the plane.
The anchor point can also be referred to as reference position.
In an example of the method, the anchor point, or reference position, for determining a viewing plane can be adjusted by the user.
In an example, the choice of the anchor point is used to determine for which procedure the planes are provided. e.g. cardioband procedures or mitral annulus procedures.
In another example, a shift of the anchor point towards the mitral valve center is allowed, while keeping the same vertical directions in the middle between the projected commissures.
In an example, the anchor point or reference position is varied by an angle or definition of another point instead of the mitral valve center and a new view is then defined.
As an example, the angle is provided. e.g. by a user slider that does not change the anchor point.
In an example, the image data is ultrasound image data.
In an example of the method, for establishing anatomical context, it is provided the step of segmenting the 3D image data providing a segmentation of at least a part of the 3D image data. The establishing anatomical context for identifying the anatomical reference locations within the 3D image data is based on the segmentation.
In an example of the method, for segmenting the 3D image data, an anatomical model is provided. In an option, the model is adapted to the 3D image data. In another option, the segmenting is based on the adapted model.
In an example, after adaptation of the model to the image, anatomical reference locations are extracted from the model. e.g. by looking at which mesh triangles or vertices carry along a certain set of anatomical labels. For example, this can be the location of the mitral annulus, the leaflets, the commissures or the center of the mitral valve.
In an example of the method, for establishing anatomical context, machine learning approaches are provided comprising at least one of the group of deep-learning and convolutional neural networks to classify voxels or detect landmarks.
In an example of the method, for the generation of the reference coordinate system it is provided at least one of the group of:
In an example of the method, for determining the direction normal to the mitral valve plane, a regression plane is determined fitting through at least a part of all vertices of the mitral annulus or by taking a cross product between vectors connecting the anterior-posterior end and connection the antero-lateral and postero-medial end of the mitral commissures.
In an example, the point representing the center of the mitral valve is provided as the mean of at least a part of all mitral annulus vertices.
In an example of the method, the reference system is defined by a first direction along the mitral valve plane normal, a second direction along the antero-lateral and postero-medial direction, and a third direction that is varied accordingly.
In an example, the third direction is the direction perpendicular to the antero-lateral and postero-medial direction in the plane defined by the mitral valve normal. This third direction is also referred to as the reference direction or the 0°) plane.
In an example of the method, information about locations of actual high regurgitant flow are used to set the plane location.
For example, results from previous evaluation of flow in another tool is used to set the plane location.
In another example, an anatomical point from the segmentation is used to set the plane location. e.g. the center of a scallop of the posterior mitral leaflet, such as “P1”.
As an example, the adapting of viewing planes is provided for products and systems targeting at intervention planning. e.g. for cardiac interventions, like ultrasound applications for cardiac interventions. The adapting of viewing planes may also be implemented within cloud-based services or applications.
In an example, information about the strength of the regurgitant jet at different plane angles is visible inside or in the vicinity of the control slider. For example, the complete slider could have a color-coded shading, and the color could represent a flow quantity (like red for high flow and blue for low flow). In an example, the underlying value is derived as the maximum flow velocity in the plane, or as an integral over the complete regurgitant jet visible in the plane.
In an option, it is provided the step of saving a particular manually selected plane, so that the same view can be investigated in relation with other subjects for consistency and without having to repeat the plane selection step. In an example, the saving of a particular manually selected plane is provided by saving the computed angle and the anchor point.
In a further option, it is provided the step of applying a particular manually selected plane also for other imaging modalities.
In the top row, a mitral valve center 308 and normal direction 310 to the mitral valve plane in the left part together with a direction 312 from an antero-lateral (AL) mitral commissure 314 to a postero-medial (PM) mitral commissure 316 are indicated.
In the bottom row, a resulting reference system 318 is indicated in the left and right part around which a plane can be defined. The reference system 318 is shown with three different axis. One of the axis, indicated with reference numeral 320, is the reference position used for defining the viewing plane.
It is noted that
The term “subject” may also be referred to as individual. The “subject” may further also be referred to as patient, although it is noted that this term does not indicate whether any illness or disease is actually present with the subject.
In another exemplary embodiment of the present invention, a computer program or a computer program element is provided that is characterized by being adapted to execute the method steps of the method according to one of the preceding embodiments, on an appropriate system.
The computer program element might therefore be stored on a computer unit or be distributed over more than one computer units, which might also be part of an embodiment of the present invention. This computing unit may be adapted to perform or induce a performing of the steps 10) of the method described above. Moreover, it may be adapted to operate the components of the above described apparatus. The computing unit can be adapted to operate automatically and/or to execute the orders of a user. A computer program may be loaded into a working memory of a data processor. The data processor may thus be equipped to carry out the method of the invention.
Aspects of the invention may be implemented in a computer program product, which may be a collection of computer program instructions stored on a computer readable storage device which may be executed by a computer. The instructions of the present invention may be in any interpretable or executable code mechanism, including but not limited to scripts, interpretable programs, dynamic link libraries (DLLs) or Java classes. The instructions can be provided as complete executable programs, partial executable programs, as modifications to existing programs 20) (e.g. updates) or extensions for existing programs (e.g. plugins). Moreover, parts of the processing of the present invention may be distributed over multiple computers or processors.
As discussed above, the processing unit, for instance a controller implements the control method. The controller can be implemented in numerous ways, with software and/or hardware, to perform the various functions required. A processor is one example of a controller which employs one or more microprocessors that may be programmed using software (e.g., microcode) to perform the required functions. A controller may however be implemented with or without employing a processor, and also may be implemented as a combination of dedicated hardware to perform some functions and a processor (e.g., one or more programmed microprocessors and associated circuitry) to perform other functions.
Examples of controller components that may be employed in various embodiments of the present disclosure include, but are not limited to, conventional microprocessors, application specific integrated circuits (ASICs), and field-programmable gate arrays (FPGAs).
This exemplary embodiment of the invention covers both, a computer program that right from the beginning uses the invention and a computer program that by means of an update turns an existing program into a program that uses the invention.
Further on, the computer program element might be able to provide all necessary steps to fulfil the procedure of an exemplary embodiment of the method as described above.
According to a further exemplary embodiment of the present invention, a computer readable medium, such as a CD-ROM, is presented wherein the computer readable medium has a computer program element stored on it which computer program element is described by the preceding section. A computer program may be stored and/or distributed on a suitable medium, such as an optical storage medium or a solid-state medium supplied together with or as part of other hardware, but may also be distributed in other forms, such as via the internet or other wired or wireless telecommunication systems.
However, the computer program may also be presented over a network like the World Wide Web and can be downloaded into the working memory of a data processor from such a network. According to a further exemplary embodiment of the present invention, a medium for making a computer program element available for downloading is provided, which computer program element is arranged to perform a method according to one of the previously described embodiments of the invention.
It has to be noted that embodiments of the invention are described with reference to different subject matters. In particular, some embodiments are described with reference to method type claims whereas other embodiments are described with reference to the device type claims. However, a person skilled in the art will gather from the above and the following description that, unless otherwise notified, in addition to any combination of features belonging to one type of subject matter also any combination between features relating to different subject matters is considered to be disclosed with this application. However, all features can be combined providing synergetic effects that are more than the simple summation of the features.
While the invention has been illustrated, and described in detail in the drawings and foregoing description, such illustration and description are to be considered illustrative or exemplary and not restrictive. The invention is not limited to the disclosed embodiments. Other variations to the disclosed embodiments can be understood and effected by those skilled in the art in practicing a claimed invention, from a study of the drawings, the disclosure, and the dependent claims.
In the claims, the word “comprising” does not exclude other elements or steps, and the indefinite article “a” or “an” does not exclude a plurality. A single processor or other unit may fulfil the functions of several items re-cited in the claims. The mere fact that certain measures are re-cited in mutually different dependent claims does not indicate that a combination of these measures cannot be used to advantage. Any reference signs in the claims should not be construed as limiting the scope.
Number | Date | Country | Kind |
---|---|---|---|
21178479.8 | Jun 2021 | EP | regional |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/EP2022/064686 | 5/31/2022 | WO |