SEGMENTING AN IMAGE

Information

  • Patent Application
  • 20200334820
  • Publication Number
    20200334820
  • Date Filed
    October 18, 2018
    5 years ago
  • Date Published
    October 22, 2020
    3 years ago
Abstract
There are provided systems and methods for segmenting an image comprising one or more anatomical structures into segments. A system (100) comprises a memory (106) comprising instruction data representing a set of instructions and a processor (102) configured to communicate with the memory (106) and to execute the set of instructions. The set of instructions, when executed by the processor (102), cause the processor (102) to receive a first user input indicating a line in the image, define two regions in the image, one on either side of the indicated line, each region abutting the indicated line on its respective side and, for each of the two regions, determine a segment corresponding to an anatomical structure comprised in the region, using the region as a seed for determining the segment.
Description
TECHNICAL FIELD OF THE INVENTION

This disclosure relates to the field of image processing and, in particular, to a method and system for segmenting an image comprising one or more anatomical structures into segments.


BACKGROUND TO THE INVENTION

Segmentation is often used in image processing to divide (or segment) an image into segments. Each segment corresponds to a region in the image and this can help with image analysis. Segmentation is often used in medical imaging whereby computer-implemented methods may be used to automatically (or semi-automatically) extract and analyse anatomical features in medical images.


There are many different methods of segmentation, including graph partitioning methods and region-growing methods. These methods take a seed or starting position in the image as an indication of a portion of a structure of interest. The seeds are then used to split the image into segments corresponding to each structure in the image (one for each seed group). Successful segmentation therefore requires correctly designated regions (e.g. correctly indicated seed points) that the original volume of interest (VOI) should be segmented into. In some methods, a user indicates areas that belong to the different segments that the VOI will be divided into. The indication is often given by marking regions (or points) in the image that belong to the different structures. These regions are translated to groups of seed points in the image and are then fed into the segmentation algorithm.


There are many methods to define seed point groups. Commonly in these methods, the user marks up areas on each separate structure with a brush-like tool on the image, and the marked areas are converted into groups of seed points. This may be inefficient for the user however and error prone. Furthermore, the accuracy of the results can depend on the locations of the areas marked by the user (e.g. the locations of the provided seed points).


As such, existing methods of placing seed points may be unintuitive and cumbersome and may not result in optimally placed seed points.


There is thus a need for methods and systems that overcome these problems.


SUMMARY OF THE INVENTION

As noted above, there are various challenges associated with segmenting an image comprising one or more anatomical structures into segments.


Therefore according to a first aspect, there is provided a system configured for segmenting an image comprising one or more anatomical structures into segments. The system comprises a memory comprising instruction data representing a set of instructions and a processor configured to communicate with the memory and to execute the set of instructions. The set of instructions, when executed by the processor, cause the processor to receive a first user input indicating a line in the image and define two regions in the image, one on either side of the indicated line, each region abutting the indicated line on its respective side. For each of the two regions, the set of instructions, when executed by the processor, further cause the processor to determine a segment corresponding to an anatomical structure comprised in the region, using the region as a seed for determining the segment.


The above system is based on the realisation that, the closer the seed point groups are to the lines along which the segments are supposed to be cut, the better the segmentation process works. Thus it is realized that, the most effective placement of the seed points is on either side of the line along which the image is to be cut, e.g. as close to line as possible. By enabling a user to indicate a line in the image and defining regions either side of and abutting up to the line in this way, therefore, the methods and systems herein can provide more accurate results. Furthermore, the methods and systems herein provide a more intuitive approach for the user to placing seed points, which is similar to an actual cutting operation. In this way, the segmentation process is made more intuitive for the user and overall more accurate.


In some embodiments, causing the processor to determine a segment corresponding to an anatomical structure may comprise causing the processor to determine whether the region overlaps a portion of an anatomical structure in the image and use the overlapped portion of the anatomical structure as the seed for determining the segment.


In some embodiments, the image may comprise a plurality of image components and causing the processor to determine a segment corresponding to an anatomical structure may comprise causing the processor to use each image component in the overlapped portion of the anatomical structure as a seed point for determining the segment.


In some embodiments, the image may comprise a plurality of image components and causing the processor to determine a segment corresponding to an anatomical structure may comprise causing the processor to use each image component in the region having a value within a predefined range as a seed point for determining the segment.


In some embodiments, causing the processor to determine a segment corresponding to the anatomical structure may comprise causing the processor to determine the segment using a graph cut segmentation process.


In some embodiments, each region may comprise a band lying parallel to the line, wherein the parallel bands are symmetrical.


In some embodiments, the set of instructions, when executed by the processor, may further cause the processor to receive a second user input indicating a change in width of at least one of the regions and adjust the width of the at least one region according to the received second user input.


In some embodiments, the second user input may comprise a movement of an actuator or a movement on a touch screen and the width of the at least one region is changed in proportion to the magnitude of the movement of the actuator or the movement on the touchscreen.


In some embodiments, the first user input may comprise the locations of two or more points on the image and the set of instructions, when executed by the processor, may cause the processor to determine a spline through the two or more points. Causing the processor to define two regions in the image may comprise causing the processor to define two regions in the image, one on either side of the determined spline, each region abutting the indicated spline on its respective side.


In some embodiments, the spline may comprise a Bezier spline.


In some embodiments, the line may comprise a plurality of adjustable control points and the set of instructions, when executed by the processor, may further cause the processor to receive a third user input to adjust one or more of the adjustable control points and update the shape of the line based on the adjustment indicated in the received third user input.


In some embodiments, the set of instructions, when executed by the processor may further cause the processor to receive a fourth user input indicating another line in the image, define two further regions in the image, one on either side of the other line, each further region abutting the other line on its respective side, link each of the two regions with a further region and, for each of the two regions, determine a segment corresponding to an anatomical structure comprised in both the region and the linked further region, using both the region and the linked further region as seed points for determining the segment.


In some embodiments, the set of instructions, when executed by the processor may further cause the processor to control a user interface to display the image, the line and the two regions, change the perspective of the image and change the perspective of the line and the two regions, in accordance with the change of perspective of the image.


In some embodiments, the set of instructions, when executed by the processor, may further cause the processor to repeatedly receive a first user input, define two regions in the image and, for each of the two regions, determine a segment corresponding to an anatomical structure, on an image of a determined segment, to divide the determined segment into further segments.


According to a second aspect, there is a computer-implemented method of segmenting an image comprising one or more anatomical structures into segments, the method comprising receiving a first user input indicating a line in the image, defining two regions in the image, one on either side of the indicated line, each region abutting the indicated line on its respective side and, for each of the two regions, determining a segment corresponding to an anatomical structure comprised in the region, using the region as a seed for determining the segment.


According to a third aspect, there is provided a computer program product comprising a computer readable medium, the computer readable medium having computer readable code embodied therein, the computer readable code being configured such that, on execution by a suitable computer or processor, the computer or processor is caused to perform the method as described above.


There is thus provided an improved method and system for segmenting an image comprising one or more anatomical structures into segments.





BRIEF DESCRIPTION OF THE DRAWINGS

For a better understanding of the embodiments, and to show more clearly how they may be carried into effect, reference will now be made, by way of example only, to the accompanying drawings, in which:



FIG. 1 is a block diagram of a system for segmenting an image comprising one or more anatomical structures into segments according to an embodiment;



FIG. 2 illustrates an example image according to an embodiment;



FIG. 3 illustrates a schematic example of an image, an indicated line and two defined regions according to an embodiment;



FIG. 4 illustrates an example image whereby image components within a predefined range are highlighted to a user according to an embodiment;



FIGS. 5a, b and c illustrate an example embodiment whereby a line is indicated by a user in a first plane of a three-dimensional image and another line is indicated in a second plane of the three-dimensional image according to an embodiment;



FIGS. 6a, b and c indicate how the width of the regions can be adjusted according to an embodiment;



FIG. 7 indicates output segments, each segment corresponding to a different anatomical structure in the image, according to an embodiment;



FIG. 8 illustrates a computer-implemented method for segmenting an image comprising one or more anatomical structures into segments according to an embodiment; and



FIG. 9 illustrates a computer-implemented method for segmenting an image comprising one or more anatomical structures into segments according to a further example embodiment.





DETAILED DESCRIPTION OF THE EMBODIMENTS

As noted above, there is provided an improved method and system for segmenting an image comprising one or more anatomical structures into segments, which overcomes the existing problems.



FIG. 1 shows a block diagram of a system 100 according to an embodiment that can be used for segmenting an image (e.g. such as a medical image) comprising one or more anatomical structures into segments. With reference to FIG. 1, the system 100 comprises a processor 102 that controls the operation of the system 100 and that can implement the method described herein.


The system 100 further comprises a memory 106 comprising instruction data representing a set of instructions. The memory 106 may be configured to store the instruction data in the form of program code that can be executed by the processor 102 to perform the method described herein. In some implementations, the instruction data can comprise a plurality of software and/or hardware modules that are each configured to perform, or are for performing, individual or multiple steps of the method described herein. In some embodiments, the memory 106 may be part of a device that also comprises one or more other components of the system 100 (for example, the processor 102 and/or one or more other components of the system 100). In alternative embodiments, the memory 106 may be part of a separate device to the other components of the system 100.


In some embodiments, the memory 106 may comprise a plurality of sub-memories, each sub-memory being capable of storing a piece of instruction data. In some embodiments where the memory 106 comprises a plurality of sub-memories, instruction data representing the set of instructions may be stored at a single sub-memory. In other embodiments where the memory 106 comprises a plurality of sub-memories, instruction data representing the set of instructions may be stored at multiple sub-memories. For example, at least one sub-memory may store instruction data representing at least one instruction of the set of instructions, while at least one other sub-memory may store instruction data representing at least one other instruction of the set of instructions.


Thus, according to some embodiments, the instruction data representing different instructions may be stored at one or more different locations in the system 100. In some embodiments, the memory 106 may be used to store images, information, data, signals and measurements acquired or made by the processor 102 of the system 100 or from any other components of the system 100.


The processor 102 of the system 100 can be configured to communicate with the memory 106 to execute the set of instructions. The set of instructions, when executed by the processor may cause the processor to perform the method described herein. The processor 102 can comprise one or more processors, processing units, multi-core processors and/or modules that are configured or programmed to control the system 100 in the manner described herein. In some implementations, for example, the processor 102 may comprise a plurality of (for example, interoperated) processors, processing units, multi-core processors and/or modules configured for distributed processing. It will be appreciated by a person skilled in the art that such processors, processing units, multi-core processors and/or modules may be located in different locations and may perform different steps and/or different parts of a single step of the method described herein.


Returning again to FIG. 1, in some embodiments, the system 100 may comprise at least one user interface 104. In some embodiments, the user interface 104 may be part of a device that also comprises one or more other components of the system 100 (for example, the processor 102, the memory 106 and/or one or more other components of the system 100). In alternative embodiments, the user interface 104 may be part of a separate device to the other components of the system 100.


A user interface 104 may be for use in providing a user of the system 100 (for example, a medical personnel, a healthcare provider, a healthcare specialist, a care giver, a subject, or any other user) with information resulting from the method according to embodiments herein. The set of instructions, when executed by the processor 102 may cause processor 102 to control one or more user interfaces 104 to provide information resulting from the method according to embodiments herein. Alternatively or in addition, a user interface 104 may be configured to receive a user input. In other words, a user interface 104 may allow a user of the system 100 to manually enter instructions, data, or information. The set of instructions, when executed by the processor 102 may cause processor 102 to acquire the user input from one or more user interfaces 104.


A user interface 104 may be any user interface that enables rendering (or output or display) of information, data or signals to a user of the system 100. Alternatively or in addition, a user interface 104 may be any user interface that enables a user of the system 100 to provide a user input, interact with and/or control the system 100. For example, the user interface 104 may comprise one or more switches, one or more buttons, a keypad, a keyboard, a mouse, a mouse wheel, a touch screen or an application (for example, on a tablet or smartphone), a display screen, a graphical user interface (GUI) or other visual rendering component, one or more speakers, one or more microphones or any other audio component, one or more lights, a component for providing tactile feedback (e.g. a vibration function), or any other user interface, or combination of user interfaces.


In some embodiments, as illustrated in FIG. 1, the system 100 may also comprise a communications interface (or circuitry) 108 for enabling the system 100 to communicate with interfaces, memories and/or devices that are part of the system 100. The communications interface 108 may communicate with any interfaces, memories and devices wirelessly or via a wired connection.


It will be appreciated that FIG. 1 only shows the components required to illustrate this aspect of the disclosure and, in a practical implementation, the system 100 may comprise additional components to those shown. For example, the system 100 may comprise a battery or other power supply for powering the system 100 or means for connecting the system 100 to a mains power supply.


In more detail, as noted above, the memory 106 comprises instruction data representing a set of instructions. Briefly, the set of instructions, when executed by the processor 102 of the system 100 cause the processor 102 to receive a first user input indicating a line in the image and define two regions in the image, one on either side of the indicated line, each region abutting the indicated line on its respective side. For each of the two regions, the set of instructions, when executed by the processor 102 of the system 100 further cause the processor to determine a segment corresponding to an anatomical structure comprised in the region, using the region as a seed for determining the segment.


As noted briefly above, the system herein is based on the realisation that the closer the seed point groups are to the lines along which the segments are to be cut, the better the segmentation process works. Thus, it is realised that the most effective placement of the seed points is on either side of the line along which the image is to be cut, e.g. as close to said line as possible. Therefore, by enabling a user to indicate a line in the image and defining regions either side of, and abutting up to, the indicated line in this way, the methods and systems described herein provide a more accurate approach to placing seed points, which improves the segmentation results. Furthermore, the process is made more intuitive to the user as it is similar to an actual cutting operation. Also, less user input may also be required compared to, for example, methods whereby the user shades in portions of each anatomical structure to be segmented.


In any of the embodiments described herein, the image can be a two-dimensional image, a three-dimensional image, or any other dimensional image. In some embodiments, the image may comprise a multi-planar reconstructed (MPR) image. The image may comprise a plurality of image components. In embodiments where the image is a two-dimensional image, the image components are pixels. In embodiments where the image is a three-dimensional image, the image components are voxels.


The image can, for example, be a medical image, or any other type of image comprising one or more anatomical structures. The image may be acquired using any imaging modality. Examples of medical image modalities include, but are not limited to, a computed tomography (CT) image or CT volume (for example, from a CT scan) such as a C-arm CT image, a spectral CT image or a phase contrast CT Image, an x-ray image (for example, from an x-ray scan), a magnetic resonance (MR) image (for example, from an MR scan), an ultrasound (US) image (for example, from an ultrasound scan), fluoroscopy images, nuclear medicine images, or any other image comprising one or more anatomical structures. Although examples have been provided for the type of image, a person skilled in the art will appreciate that the teachings provided herein may equally be applied to any other type of image comprising one or more anatomical structures.


As mentioned earlier, the image comprises one or more anatomical structures. For example, the image may comprise one or more organs (such as a heart, one or more lungs, or any other organ, or any combination of organs), bones (such as one or more rib bones, one or more vertebrae or any other bones, or any combination of bones), and/or any other type of anatomical structure. Herein, “anatomical structure” is used generally and may refer to a group of anatomical structures (e.g. a group of organs), one or more individual anatomical structures (e.g. an individual organ) and/or one or more portions of an anatomical structure (e.g. one or more portions of an organ). It will therefore be appreciated that the image may not necessarily comprise an entire anatomical structure and may, for example, only comprise part of an anatomical structure and/or, if other anatomical structures are also present in the segmented image, part(s) of the other anatomical structures.



FIG. 2 illustrates an example image comprising one or more anatomical structures. In this embodiment, the example image comprises a multi-planar reconstructed (MPR) image and the one or more anatomical structures comprise structures in a mid-section of a subject. In particular, the image shows vertebrae 202, the kidneys 204 and the aorta 206.


As mentioned earlier, the system described herein can be used for segmenting an image. The person skilled in the art will be familiar with segmentation and different types of segmentation processes. Briefly, segmentation is where an image is segmented (or divided, converted or partitioned) into a plurality of different segments (or portions, blocks, or shapes). For example, each segment may correspond to a group of image components that are related in some way (e.g. all image components that lie within a boundary, predefined area or region), or share a common characteristic (e.g. connected image components having values within a predefined range). Segmentation may be used to delineate or highlight parts of an anatomical structure in the image, shapes in the image and/or other objects that have identifiable outlines (e.g. boundaries or edges) in the image. The segmentation process may be a computer-implemented segmentation process (e.g. an “auto-segmentation” algorithm that automatically segments images for the user). Examples of computer-implemented segmentation processes include, for example, graph partitioning methods (examples of graph partitioning methods include normalized cuts, random walker, minimum cut, isoperimetric partitioning, and segmentation-based object categorization) and region-growing methods, and the person skilled in the art will be familiar with such segmentation processes, amongst others.


As described above, certain segmentation processes require seed points to be provided by the user. The seed points provide an indication of the location of a portion of (each) structure (e.g. anatomical structure) in the image for which a segment is to be determined. In segmentation processes such as graph based processes, the accuracy of the outputted segments depends in part on the location of the provided seed points. It has been recognised herein, that when two anatomical structures are close together in an image, the closer the seed points are to the line (e.g. dividing line) between the structures, along which the segments are to be cut, the better the algorithm works. It is realized that the most effective placement of seed points where two structures are closely placed is directly either side of the line along which the structures are to be cut, e.g. as close to the dividing line as possible.


As such, herein, the set of instructions, when executed by the processor 102, cause the processor 102 to receive a first user input indicating a line in the image. The user may provide the first user input, for example, using a user interface 104. In some embodiments, the first user input comprises the locations of two or more points on the image through which a line is to be placed. The user may indicate the two or more points on the image, for example, using a user interface 104 (such as a mouse and by clicking on the image using the mouse to indicate the locations of each of the points). In some embodiments, the set of instructions, when executed by the processor 102, may cause the processor 102 to determine (e.g. fit) a line through the two or more points. The skilled person will be familiar with different methods that can be used to fit a line between points in an image. For example, in some embodiments, the processor 102 may be caused to determine (e.g. fit) a spline through the two or more points. In some embodiments, the spline may be a Bezier spline. Although examples have been given for the way in which a line can be fitted between points, the person skilled in the art will appreciate that these examples are non-exhaustive and that other line fitting methods may also be used (for example, a least squares fitting method and/or one or more other curve fitting methods).


In general, the first user input indicates a line in the image that divides two anatomical structures in the image, which are to be segmented into segments (e.g. two segments, one for each anatomical structure on either side of the line).


In some embodiments, the line (or, for example, where a spline is determined, the determined spline) may comprise a plurality of adjustable control points. The adjustable control points may be adjustable by the user (e.g. by dragging and dropping) in order to reposition (for example, to better place) the line over the optimal cut point (e.g. the line along which the structures are to be cut) in the image. In some embodiments, the set of instructions, when executed by the processor 102 may further cause the processor 102 to receive a third user input to adjust one or more of the adjustable control points and update the shape of the line based on the adjustment indicated in the received third user input. Adjusting a control point may comprise, for example, adding a control point, removing a control point and/or moving a control point so as to change the shape of the line (or spline). In embodiments where a spline is determined, the spline may comprise a plurality of adjustable control points and the third user input to adjust a control point may cause the processor 102 to re-determine a new spline that fits to the adjusted control points.


As noted above, the set of instructions, when executed by the processor 102, further cause the processor 102 to define two regions in the image, one on either side of the indicated line, each region abutting the indicated line on its respective side. In embodiments where the line comprises a spline, being caused to define two regions, one on either side of the line may comprise being caused to define two regions in the image, one on either side of the determined spline, each region abutting the indicated spline on its respective side.


In some embodiments, each region may comprise a band that lies parallel to (e.g. along the length of) the line (or spline). In some embodiments, the parallel bands may be symmetrical. In some embodiments, therefore, the regions may be shaped as “ribbons” lying either side of the indicated line (or spline).


Therefore in some embodiments, the two regions may appear as a pair of stripes, one stripe on either side of the indicated line (or spline).



FIG. 3 shows a schematic representation of an embodiment in which this is illustrated. In this embodiment, an image comprises anatomical structures, namely vertebrae 302 and a kidney 304. In the embodiment of FIG. 3, the image is to be segmented into a first segment comprising the vertebrae 302 and a second segment comprising the kidney 304. To this end, the user provides a first user input that is received by the processor 102 in the form of three points 306. The set of instructions when executed by the processor 102, may cause the processor 102 to determine a line 308 (such as a spline) through the three points and defines two regions 310a and 310b in the image, one on either side of the determined line 308. As can be seen in FIG. 3, the two regions in this embodiment lie either side of the determined spline. Each region comprises a band that lies parallel to, and abuts up to, the determined line 308.


Turning back now to FIG. 1, in some embodiments, the set of instructions, when executed by the processor 102, may further cause the processor 102 to receive a second user input (for example from a user interface 104) indicating a change in width of at least one of the regions and to adjust the width of the at least one region according to the received second user input. For example, in some embodiments, a second user input may be received from an actuator (e.g. a mechanism or component that a user can push or pull, such as a wheel on a computer mouse or a button on a joystick). More specifically, the second user input may comprise a movement of an actuator. For example, in some embodiments, the user may hover over the line (or spline) and move the actuator to change the width of the defined regions. In alternative embodiments, a second user input may comprise (or be received in the form of) a movement on a touch screen (e.g. a swipe on a touch screen). The set of instructions when executed by the processor 102, may cause the processor 102 to adjust the width of the at least one region in proportion to the magnitude of the movement of the actuator or the movement on the touchscreen (e.g. based on the magnitude of the received second user input). Adjusting a width (e.g. expanding or contracting the width) of a region based on an input, e.g. from an actuator such as a mouse wheel or a touch screen movement, is more efficient than if the user manually draws a new region in the image. In this way, the process described herein is more user efficient and user friendly.


As mentioned earlier, the set of instructions, when executed by the processor 102, may cause the processor 102 to determine, for each of the two regions, a segment corresponding to an anatomical structure comprised in the region, using the region as a seed for determining the segment.


In some embodiments, each region may be used as a seed for a segment in the segmentation. For example, each image component (e.g. each pixel or voxel) may be used as a seed point for determining the segment. In some embodiments, the co-ordinates of each image component in a region may be used as a seed point for determining the segment. Using the full region as a seed point may be appropriate, for example, where the edge of the anatomical structure lies along (or very close to) the line (or spline).


In some embodiments, a subset of the image components in a region may be used as seed points (or a seed point group) with which to determine the segment. For example, in some embodiments, the set of instructions, when executed by the processor 102 may cause the processor 102 to determine whether the region overlaps a portion of an anatomical structure in the image and use the overlapped portion of the anatomical structure as the seed for determining the segment. For example, in the example in FIG. 3, the portion of the region 310a overlapping the kidney 304 may be used as a seed region with which to segment the kidney 304 and the portions of the region 310b that overlap the vertebrae 302 may be used as seed regions to segment the vertebrae 302. In some embodiments, the set of instructions, when executed by the processor 102, may cause the processor 102 to use each image component in the overlapped portion of the anatomical structure as a seed point for determining the segment. In this way, only image components in the region that form part of an anatomical structure of interest are used as seed points with which to determine a segment for the anatomical structure.


In some embodiments, the processor 102 may be caused to use each image component in the region having a value within a predefined range as a seed point for determining the segment. The predefined range may be chosen, for example, as (e.g. to coincide with) the range of image component values observed for a specific anatomical structure in the image. For example, if in FIG. 3, image components corresponding to the kidney had numerical values between x1 and x2, the predetermined range may be set to encompass the values x1 and x2 so as to pick out the image components in the region 310a corresponding to the kidney 304. In this way, the predetermined range can be used to target specific anatomical structures within the region(s) to create appropriate seeds for the segmentation.


In some embodiments, image components with values lying within the predefined range may be determined (e.g. located) in a pre-processing stage. For example, in some embodiments, the processor 102 may cause the image to be displayed to the user, for example, by controlling a user interface 104 to display the image. Image components with values within the predetermined range may be highlighted to the user, for example in a different colour to other image components in the image.



FIG. 4, shows an embodiment in which this is illustrated. In this embodiment, the image comprises an image of the torso comprising anatomical structures, namely kidneys 402 and vertebrae 404. Image components within a predetermined range corresponding to the values of the kidneys 402 and vertebrae 404 are highlighted (e.g. appear brighter) to the user in FIGS. 4a, 4b and 4c. The user can thus more easily view the anatomical structures in the image to provide a user input to the processor 102, to provide an indication of a line in the image along which the highlighted anatomical structures should be split (or cut).


The image may be segmented using any appropriate segmentation process that takes seed points as input. In some embodiments, causing the processor 102 to determine a segment corresponding to the anatomical structure may comprise causing the processor 102 to determine the segment using a graph cut segmentation process. In alternative embodiments, causing the processor 102 to determine a segment corresponding to the anatomical structure may comprise causing the processor 102 to determine the segment using a region growing segmentation process. Although examples of suitable segmentation processes have been provided, it will be understood that any other segmentation processes that take seed points as starting points from which to determine the locations of segments may also be used to determine the segments.


In some embodiments, the set of instructions, when executed by the processor 102, may further cause the processor 102 to control a user interface (such as a user interface 104) to display the image, the line and the two regions. The processor 102 may further be caused to change the perspective of the image (for example, to rotate or translate the image, flip the image, or zoom-in or zoom out from the image). The processor 102 may be further caused to change the perspective of the displayed line and the two regions, in accordance with the change of perspective of the image. For example, if the image is rotated, the displayed line and the displayed regions may be similarly rotated so that they stay fixed in the image with respect to the displayed anatomical structures. In this way, the user can view the image from different angles and perspectives. This can, for example, be useful for the user to check the accuracy of the placement of the line and the regions in the image, from different angles and perspectives before the image is segmented.


In some embodiments, the set of instructions, when executed by the processor 102, may further cause the processor 102 to repeatedly: i) receive a first user input indicating a line in the image; ii) define two regions in the image, one on either side of the indicated line, each region abutting the indicated line on its respective side; and iii) determine, for each of the two regions, a segment corresponding to an anatomical structure comprised in the region, using the region as a seed for determining the segment; on an image of a determined segment, to divide the determined segment into further segments. For example, an image may be segmented into two segments, each segment corresponding to a different anatomical structure in the image. The user may concentrate on one of the two segments and segment this segment further (e.g. into further segments or sub-segment) using any of the processes described with respect to the processor 102 above.


In this way, the user can sequentially break an image down by cutting (or dividing) the image up into progressively smaller segments, each iteration dividing the chosen segment into two.


In some embodiments, the set of instructions, when executed by the processor 102, may further cause the processor 102 to receive a fourth user input indicating another (e.g. second) line in the image. The other line may be indicated by the fourth user input in the same way as the indicated line was indicated by the first user input (which may be thought of as e.g. the “first” line), as described in any of the embodiments above. The set of instructions, when executed by the processor 102, may further cause the processor 102 to define two further regions in the image, one on either side of the other indicated line, each further region abutting the other (e.g. second) indicated line on its respective side. Each region either side of the indicated line (e.g. first indicated line) may be linked (or associated) with one of the two further regions defined either side of the other indicated line. For each linked pair of regions (e.g. one associated with the indicated line and one associated with the other line), the processor 102 may be caused to determine a segment corresponding to an anatomical structure comprised in both the region and the linked further region, using both the region and the linked further region as seed points for determining the segment. In this way, one of the regions determined for the indicated line and one of the further regions determined for the other line can both be used as seed points for segmenting the same anatomical structure. As such, the user may indicate more than one line (e.g. cut) in the image, to more clearly define how two anatomical structures are separated in the image.


In some embodiments, where the image is a three dimensional image, the user may indicate the line in a first slice through a three dimensional image and the other line in a second, different slice through the three dimensional image. The first and second slices may be in different dimensions of the three dimensional image, for example, the indicated line may lie in a plane that is perpendicular to the plane in which the other indicated line lies. This is merely exemplary however and the skilled person will appreciate that the line and the other line may be drawn in any planes or combination of planes in a three dimensional image.


As noted above, each region determined either side of the indicated line may be linked (or associated) with a further region determined with respect of the other indicated line and both the region defined with respect to the line and the linked further region defined with respect to the other line may be used as seed points for determining an anatomical structure comprised in the region and the linked further region. In some embodiments, the regions determined either side of the indicated line and the further regions determined either side of the other line may be colour coded to indicate to the user which region is linked with which further region (and thus which regions and further regions will be used as seed points for an anatomical structure in the image). In some embodiments, the user may be able to change (e.g. flip) the links (or associations) to ensure that each indicated region and linked further indicated region overlap with the same anatomical structure in the image. Although examples have been given for the way in which more than one line may be indicated in the image and used to segment the image, the skilled person will appreciate that other processes are also possible. For example, the user may be permitted to indicate any number of lines (e.g. three or more) in the image from which linked regions may be defined to use as seed points for determining segments in the image. In this way, the user is able to easily provide seed points for segmenting anatomical structures in more than one dimension in the image.



FIGS. 5, 6 and 7 illustrate an embodiment in which this is illustrated, whereby a three dimensional image of a midsection is segmented into two segments, one corresponding to the kidneys and the vertebrae and a second corresponding to the aorta. FIGS. 5a, b and c show three planes through the three dimensional image. In this embodiment, image components having values within a predetermined range of values that correspond to a range of values for the kidneys 502, vertebrae 504 and aorta 506 are highlighted to the user (this is indicated by the brighter portions of FIG. 5). The user indicates a line 508 in the first plane shown in FIG. 5a separating the aorta 506 and vertebrae 504 in that plane and indicates another line 510 in the second plane shown in FIG. 5b separating the aorta 506 and vertebrae 504 in the second plane.


Two regions 512a (indicated by the spotted region) and 512b (indicated by the dashed region) are defined either side of the indicated line 508 and a further two regions 514a (indicated by the spotted region in FIG. 5b) and 514b (indicated by the dashed region in FIG. 5b) are defined either side of the other indicated line 510. The indicated region 512a in FIG. 5a overlapping the vertebrae and the further indicated region 514a overlapping the vertebrae 504 in FIG. 5b are linked such that both regions overlapping the vertebrae 504 are used as seed points to segment the vertebrae 504. Similarly, the indicated region 512b overlapping the aorta and the further indicated region 514b overlapping the aorta 506 are linked such that both regions overlapping the aorta 506 are used as seed points to segment the aorta 506. In this embodiment, the linked regions are colour coded (e.g. regions overlapping the aorta 506 are colour coded a first colour and the regions overlapping the vertebrae 504 are coloured a second colour, where the first colour and the second colour is different) to highlight to the user which regions are used as seed points for which anatomical structures. The user may change which region links up to which further region, if for example, a region overlapping the aorta 506 is inadvertently linked with a region overlapping the vertebrae 504, so as to ensure that linked regions overlap the same structures.



FIGS. 6a, b and c show the same image process as in FIG. 5, except in this embodiment, a second user input is received (e.g. a movement of a mouse wheel) to indicate that the widths of the regions 512a and 512b either side of the indicated line 508 and the other indicated line 510 should be reduced in size to better overlap the aorta 506 and the vertebrae 504 respectively. As such, the widths of the regions 512a and 512b are reduced accordingly (compared to FIG. 5).


The image is then segmented using graph partition methods according to this illustrated embodiment, taking the linked pairs of indicated regions and further regions (512a and 514a and 512b and 514b respectively) as seed points to segment the vertebrae 504 and aorta 506 in the image. FIG. 7 shows the output of the segmentation process, showing the two segments that have been produced, one corresponding to the vertebrae, kidneys and associated structures 702 and the other to the aorta 704. In this way, an image can easily and accurately be segmented into two segments, with minimum user input.



FIG. 8 illustrates a computer-implemented method 800 of segmenting an image comprising one or more anatomical structures into segments, according to an embodiment. The illustrated method 800 can generally be performed by or under the control of the processor 102 of the system 100. The method may be partially or fully automated according to some embodiments.


Briefly, with reference to FIG. 8, the method 800 comprises receiving a first user input indicating a line in the image (at block 802 of FIG. 8), and defining two regions in the image, one on either side of the indicated line, each region abutting the indicated line on its respective side (at block 804 of FIG. 8). The method 800 also comprises, for each of the two regions: determining a segment corresponding to an anatomical structure comprised in the region, using the region as a seed for determining the segment (at block 806 of FIG. 8).


As noted above, generally, the closer the seed point groups are to the lines along which the segments are to be cut, the better the segmentation process works. Thus, the most effective placement of the seed points is on either side of the line along which the image is to be cut, e.g. as close to line as possible. By enabling a user to indicate a line in the image and defining regions either side of and abutting up to the line in this way, the method of FIG. 8 provides a more intuitive and accurate approach to placing seed points. Less user input may be required compared to, for example, methods whereby the user shades in portions of each anatomical structure to be segmented. Furthermore, the method herein is accurate than such shading methods.



FIG. 9 shows a computer implemented method 900 according to another embodiment. The illustrated method 900 can generally be performed by or under the control of the processor 102 of the system 100. The method may be partially or fully automated according to some embodiments. In this illustrated embodiment, in a block 902, a scanned volume is received, such as an MPR image. In block 904 a volume of interest is specified. For example, the scanned volume may be reduced in size to focus the segmentation on a particular portion of the scanned volume.


In block 906, a first user input is received, indicating a line in the image. In this embodiment, the first user input indicates two or more points in the image and a line (e.g. a spline) is fit through the indicated points. Receiving a first user input in the form of two or more points and fitting a line through the points in this manner was described above with respect to system 100 and the details therein will be understood to equally apply to block 906. Block 906 further comprises defining two regions in the image, one either side of the indicated line, each region abutting the indicated line on its respective side. Block 906 may further comprise modifying the regions, for example, increasing the width of the region in response to a second user input. Block 906 may also further comprise representing the line with a series of control points, receiving a third user input indicating an adjustment to one of the control points, changing the control point and re-fitting (e.g. re-determining) a line passing through the adjusted control points.


Determining regions, changing the width of said regions in response to user input and adjusting the control points of a line were all described above with respect to system 100 and the details therein will be understood to equally apply to these aspects of block 906. Block 906 may further comprise receiving a fourth user input indicating another line in the image (for example, the fourth user input may comprise an indication of two or more other points in the image and another line (or spline) may be fit through the two or more other points, as described above with respect to the first user input.) Two further regions may be defined in the image, one either side of the other line, each further region abutting the other line on its respective side. Each of the two regions may then be linked with a further region (e.g. each region may be linked with a further region overlapping the same anatomical structure as that region). Receiving a fourth user input indicating another line, defining further regions either side of the other line and linking each region with a further region were all described above with respect to system 100 and the details therein will be understood to equally apply to these aspects of block 906.


In block 908, the user assesses whether the line (and/or the other line if a fourth user input is received in block 906) is correctly placed between the two structures that are to be segmented. In particular, the user assesses whether the line covers the intended separation region e.g. the line reflects the line along which the structures are to be cut (or separated) into two segments. The user then determines whether the defined regions (and/or the further defined regions) cover at least a portion of the anatomical structure of interest (e.g. that the regions can be used as seed points for the anatomical structures of interest). If the user determines that the line does not correctly align with the separation region (e.g. cut point between the two segments), or that one or both of the regions do not overlap with an anatomical structure of interest, then the method returns to block 906 whereby the user may provide further user inputs to adjust the positions of the line and/or the widths of the regions until the conditions of block 908 are satisfied.


If, at block 908, the user determines that the line correctly lies between the structures of interest (e.g. aligns with the separation region or cut point) and the regions overlap with the anatomical structure of interest (e.g. are appropriately placed), then the method moves on to block 910 which comprises, for each of the two regions, determining a segment corresponding to an anatomical structure comprised in the region, using the region as a seed for determining the segment. If a fourth input is received in block 906 indicating another line in the image, then at block 908, the method may (e.g. alternatively or additionally) comprise determining a segment corresponding to an anatomical structure comprised in both the region and the linked further region, using both the region and the linked further region as seed points for determining the segment. Determining a segment was described above with respect to system 100 and the details therein will be understood to equally apply to these aspects of block 910.


In block 912, the user assesses whether the segmentation (or separation) of the anatomical structures was successful. If the segmentation was unsuccessful, or the user is dissatisfied with the result, then the method returns to block 906 whereby, as described above, the user may adjust the position of the line and/or the width of the regions and re-run the segmentation.


If it is determined in block 912 that the segmentation was successful, then the method ends at block 914.


There is also provided a computer program product comprising a computer readable medium, the computer readable medium having computer readable code embodied therein, the computer readable code being configured such that, on execution by a suitable computer or processor, the computer or processor is caused to perform the method or methods described herein. Thus, it will be appreciated that the disclosure also applies to computer programs, particularly computer programs on or in a carrier, adapted to put embodiments into practice. The program may be in the form of a source code, an object code, a code intermediate source and an object code such as in a partially compiled form, or in any other form suitable for use in the implementation of the method according to the embodiments described herein.


It will also be appreciated that such a program may have many different architectural designs. For example, a program code implementing the functionality of the method or system may be sub-divided into one or more sub-routines. Many different ways of distributing the functionality among these sub-routines will be apparent to the skilled person. The sub-routines may be stored together in one executable file to form a self-contained program. Such an executable file may comprise computer-executable instructions, for example, processor instructions and/or interpreter instructions (e.g. Java interpreter instructions). Alternatively, one or more or all of the sub-routines may be stored in at least one external library file and linked with a main program either statically or dynamically, e.g. at run-time. The main program contains at least one call to at least one of the sub-routines. The sub-routines may also comprise function calls to each other.


An embodiment relating to a computer program product comprises computer-executable instructions corresponding to each processing stage of at least one of the methods set forth herein. These instructions may be sub-divided into sub-routines and/or stored in one or more files that may be linked statically or dynamically. Another embodiment relating to a computer program product comprises computer-executable instructions corresponding to each means of at least one of the systems and/or products set forth herein. These instructions may be sub-divided into sub-routines and/or stored in one or more files that may be linked statically or dynamically.


The carrier of a computer program may be any entity or device capable of carrying the program. For example, the carrier may include a data storage, such as a ROM, for example, a CD ROM or a semiconductor ROM, or a magnetic recording medium, for example, a hard disk. Furthermore, the carrier may be a transmissible carrier such as an electric or optical signal, which may be conveyed via electric or optical cable or by radio or other means. When the program is embodied in such a signal, the carrier may be constituted by such a cable or other device or means. Alternatively, the carrier may be an integrated circuit in which the program is embedded, the integrated circuit being adapted to perform, or used in the performance of, the relevant method.


Variations to the disclosed embodiments can be understood and effected by those skilled in the art, from a study of the drawings, the disclosure and the appended claims. In the claims, the word “comprising” does not exclude other elements or steps, and the indefinite article “a” or “an” does not exclude a plurality. A single processor or other unit may fulfil the functions of several items recited in the claims. The mere fact that certain measures are recited in mutually different dependent claims does not indicate that a combination of these measures cannot be used to advantage. A computer program may be stored/distributed on a suitable medium, such as an optical storage medium or a solid-state medium supplied together with or as part of other hardware, but may also be distributed in other forms, such as via the Internet or other wired or wireless telecommunication systems. Any reference signs in the claims should not be construed as limiting the scope.

Claims
  • 1. A system configured for segmenting an image comprising one or more anatomical structures into segments, the system comprising: a memory comprising instruction data representing a set of instructions;a processor configured to communicate with the memory and to execute the set of instructions, wherein the set of instructions, when executed by the processor, cause the processor to: receive a first user input indicating a line in the image;define two regions in the image, one on either side of the indicated line, each region abutting the indicated line on its respective side; andfor each of the two regions: determine a segment corresponding to an anatomical structure comprised in the region, using the region as a seed for determining the segment.
  • 2. The system as in claim 1 wherein causing the processor to determine a segment corresponding to an anatomical structure comprises causing the processor to: determine whether the region overlaps a portion of an anatomical structure in the image; anduse the determined overlapped portion of the anatomical structure as the seed for determining the segment.
  • 3. The system as in claim 2, wherein the image comprises a plurality of pixels or voxels and causing the processor to determine a segment corresponding to an anatomical structure comprises causing the processor to: use each pixels or voxels in the overlapped portion of the anatomical structure as a seed point for determining the segment.
  • 4. The system as in claim 1, wherein the image comprises a plurality of pixels or voxels and causing the processor to determine a segment corresponding to an anatomical structure comprises causing the processor to: use each pixels or voxels in the region having a value within a predefined range as a seed point for determining the segment.
  • 5. The system as in claim 1, wherein causing the processor to determine a segment corresponding to the anatomical structure comprises causing the processor to: determine the segment using a graph cut segmentation process.
  • 6. The system as in claim 1, wherein each region comprises a band lying parallel to the line, and wherein the parallel bands are symmetrical.
  • 7. The system as in claim 1, wherein the set of instructions, when executed by the processor, further cause the processor to: receive a second user input indicating a change in width of at least one of the regions; andadjust the width of the at least one region according to the received second user input.
  • 8. The system as in claim 7, wherein the second user input comprises a movement of an actuator or a movement on a touch screen and the width of the at least one region is changed in proportion to the magnitude of the movement of the actuator or the movement on the touchscreen.
  • 9. The system as in claim 1, wherein the first user input comprises the locations of two or more points on the image and the set of instructions, when executed by the processor, cause the processor to: determine a spline through the two or more points; andwherein causing the processor to define two regions in the image comprises causing the processor to: define two regions in the image, one on either side of the determined spline, each region abutting the determined spline on its respective side.
  • 10. The system as in claim 1, wherein the line comprises a plurality of adjustable control points; and wherein the set of instructions, when executed by the processor, further cause the processor to: receive a third user input to adjust one or more of the adjustable control points; andupdate the shape of the line based on the adjustment indicated in the received third user input.
  • 11. The system as in claim 1, wherein the set of instructions, when executed by the processor, further cause the processor to: receive a fourth user input indicating another line in the image;define two further regions in the image, one on either side of the other line, each further region abutting the other line on its respective side;associate each of the two regions with a further region; andfor each of the two regions: determine a segment corresponding to an anatomical structure comprised in both the region and the associated further region, using both the region and the associated further region as seed points for determining the segment.
  • 12. The system as in claim 1, wherein the set of instructions, when executed by the processor, further cause the processor to: control a user interface to display the image, the line and the two regions;change the perspective of the image; andchange the perspective of the line and the two regions, in accordance with the change of perspective of the image.
  • 13. A computer-implemented method of segmenting an image comprising one or more anatomical structures into segments, the method comprising: receiving a first user input indicating a line in the image;defining two regions in the image, one on either side of the indicated line, each region abutting the indicated line on its respective side; andfor each of the two regions: determining a segment corresponding to an anatomical structure comprised in the region, using the region as a seed for determining the segment.
  • 14. The computer program product comprising a computer readable medium, the computer readable medium having computer readable code embodied therein, the computer readable code being configured such that, on execution by a suitable computer or processor, the computer or processor is caused to perform the method as claimed in claim 13.
Priority Claims (1)
Number Date Country Kind
17197610.3 Oct 2017 EP regional
PCT Information
Filing Document Filing Date Country Kind
PCT/EP2018/078484 10/18/2018 WO 00