Embodiments of the invention are illustrated by way of example and not limitation in the figures of the accompanying drawings, in which like reference numerals indicate corresponding, analogous or similar elements, and in which:
It will be appreciated that for simplicity and clarity of illustration, elements shown in the figures have not necessarily been drawn to scale. For example, the dimensions of some of the elements may be exaggerated relative to other elements for clarity.
In the following detailed description, numerous specific details are set forth in order to provide a thorough understanding of embodiments of the invention. However it will be understood by those of ordinary skill in the art that the embodiments of the invention may be practiced without these specific details. In other instances, well-known methods, procedures, components and circuits have not been described in detail so as not to obscure the embodiments of the invention.
Unless specifically stated otherwise, as apparent from the following discussions, it is appreciated that throughout the specification discussions utilizing terms such as “processing,”“computing,” “calculating,” “determining,” or the like, refer to the action and/or processes of a workstation, or similar electronic computing device, that manipulates and/or transforms data represented as physical, such as electronic, quantities within the computing system's registers and/or memories into other data similarly represented as physical quantities within the computing system's memories, registers or other such information storage, transmission or display devices.
An in-vivo sensing device may transmit identification data, which may include data that relates to or identifies the device, for example, device type, such as a model or brand, device components, supporting mechanisms, supporting software, compatibility requirements or other identifying data. Identification data may indicate what type of sensory data the sensing device collects, for example, image data, pH data, etc. Identification data may include areas in a patient's body where the sensing device may be used, for example, the colon, esophagus, etc. Identification data may include geographical zones or areas, for example, nations or geographical regions, where the sensing device and/or supporting system components may properly function, may be allowed to function, or may be compatible with other applications and/or systems. Identification data may include data that uniquely identifies the sensing device, for example, a code, serial number or electronic signature. In one embodiment, no two sensing devices may have precisely the same identification data. In other embodiments, a group or type of sensing devices may have the same identification data or a common portion of identification data.
The sensing device may also transmit sensory data, for example, image data, that the sensing device captures or collects while traversing a body. The sensing device may include an image sensor or camera, or components for sensing physiological parameters of a body lumen such as, pH, temperature, pressure, electrical impedance, etc.
Devices according to embodiments of the present invention may be similar to embodiments described in U.S. Pat. No. 7,009,634 to Iddan et al., entitled “Device for In-Vivo Imaging”, and/or in U.S. Pat. No. 5,604,531 to Iddan et al., entitled “In-Vivo Video Camera System”, and/or in U.S. patent application Ser. No. 10/046,541, filed on Jan. 16, 2002, published on Aug. 15, 2002 as United States Patent Application Publication No. 2002/0109774, all of which are hereby incorporated by reference. An external reception system or receiver unit, a processor and a monitor, e.g., in a workstation, such as those described in the above publications, may be suitable for use with some embodiments of the present invention. Devices and systems as described herein may have other configurations and/or other sets of components. For example, some embodiments of the present invention may be practiced using an endoscope, needle, stent, catheter, etc. Some in-vivo devices may be capsule shaped, or may have other shapes, for example, a peanut shape or tubular, spherical, conical, or other suitable shapes.
Reference is made to
Sensing device 4 may include a control block 26, a transmitter 28, one or more memory units 33, a receiver 30, a processor 47, an antenna 32, a power source 34, and a sensing system 24. In one embodiment, sensing system 24 may include an imaging system that may include for example an optical window 36, at least one illumination source 38, such as, for example, a light emitting diode (LED), an imaging sensor 40, and an optical system 42. Sensing device 4 may include one or more registers or memory units 33, which may be included for example in processor 47, control block 26 or transmitter 28. In one embodiment, control block 26, processor 47 and transmitter 28, or all or part of their functionality may be combined in one unit. In one embodiment, components of sensing device 4 may be sealed within a device body, shell or container (the body, shell or container may include more than one piece).
According to one embodiment of the present invention, identification data 73 may be stored in the sensing device 4, for example, in memory unit 33, transmitter 28, processor 47, or any other storage area. In other embodiments identification data 73 may be stored using, for example, hard wired non-solid state devices, for example using one or more switches.
Transmitter 28 may transmit identification data 73. Identification data 73 may be transmitted automatically or in response to a system, program or administrator's request. Data, including for example sensory data and identification data 73, may be transmitted from the in-vivo sensing device 4 to receiver 6 via a wireless or hard-wired medium 11 while inside the patient's body. Receiver 6 may receive, record and/or store the data transmitted by transmitter 28. Receiver 6, which may be positioned close to or worn on a subject, may receive a stream of data transmitted by sensing device 4. Workstation 8 may download or access the stream of data from receiver 6 via, for example, a wireless or hard-wired medium 11, and may analyze and/or display the stream of data. In one embodiment, workstation 8 may download, store, use or display identification data 73 and sensory data, separately. In alternate embodiments workstation 8 may receive data transmitted directly from sensing device 4, rather than using receiver 6 as an intermediary.
In one embodiment, identification data 73 transmitted by sensing device 4, may be used to determine if sensing device 4 meets system 2 requirements, for example, identified by system 2 component's requirement data 75. Requirement data 75 may include, for example, data that specifies a system 2 component's requirement or standard, such that in order for the system 2 component to use sensory data transmitted by sensing devices 4, sensing devices 4 must transmit identification data 73 that substantially fulfills the requirement or standard. System 2 components, for example, receiver 6 and/or workstation 8, and applications and software thereof may include requirement data 75. In some embodiments, system 2 component requirement data 75 may be stored in the system 2 components themselves. For example, requirement data 75 may be stored in memory 17 of workstation 8 or memory 56 of receiver 6. Requirement data 75 may include, for example, read only data, electronic signatures or other types of data. Different system 2 components, as well as different hardware or software programs within a system 2 component, may have different identification requirements.
In one embodiment, a system 2 component may compare identification data 73 transmitted by the sensing device 4 with the requirement data 75. For example, sensing devices 4 must transmit identification data 73, which may be accepted by the system 2 component, for example, workstation 8, in order for system 2 components to work with sensing device 4. The system 2 component may read identification data 73. The system 2 component may read requirement data 75, which may be for example retrieved from memory. The system 2 component may compare analogous portions of identification data 73 and requirement data 75 to determine if the two data sets substantially match.
For example, workstation 8 may have requirement data 75 that specifies that workstation 8 may only use sensory data from sensing devices 4 that collect image data. Thus, if identification data 73 transmitted by sensing device 4 identifies sensing device 4 as an imaging device, workstations 8 may accept sensory data transmitted by sensing device 4.
In some embodiments, requirement data 75 may be entered at workstation 8, for example, by a user at a terminal. For example, a user may select a type of data or display program to be used by system 2, or configure workstation 8 or install software in workstation 8. For example, a user may configure workstation 8 by selecting a range of acceptable values, such that workstation 8 may only use sensory data from sensing devices 4 that transmit identification data 73 that falls within the range. In other embodiments, component requirement data 75 may include fixed or constant data, for example, pre-programmed, in hardware or software. In some embodiments, requirement data 75 or identification data 73 may be read-only data or may be protected or encrypted, such that the data may not be altered by a user.
In some embodiments, identification data 73 may include data indicating nations or geographical regions, in which sensing device 4 is intended to be used, function properly or comply with other system 2 components and applications. System 2 components may only accept or use sensory data from sensing device 4 if the regions in which sensing device 4 is intended to be used sufficiently matches region requirements of system 2 components. For example, a receiver 6 intended to be used in the United Kingdom may not receive, record and/or store sensory data transmitted by a sensing device 4 intended to be used in Australia.
In other embodiments, if identification data 73 includes data identifying the model, brand or type associated with sensing device 4, then system 2 components or applications may automatically access software such as programs, displays or modules that are compatible or preferably used with that model, brand or type of sensing device 4. For example, workstation 8 may accept identification data 73 including a model, version number, code or electronic signature, associated with sensing device 4, and may determine if identification data 73 matches requirement data 75 in the software. If identification data 73 sufficiently matches requirement data 75 in the software, workstation 8 may access or activate software or hardware that includes data that matches at least a portion of identification data 73. Thus, appropriate system 2 mechanisms may be accessed without instructions from a user. Receiver 6 and workstation 8 may accept identification data 73 and alter operations based on the identification data 73.
Identification data 73 may include data identifying the intended region in a patient's body from which sensing device 4 may collect sensory data, for example, the colon. Upon accepting such identification data 73, system 2 components or applications may access appropriate programs, displays, modules or software, for example, for viewing sensory data collected from that region. For example, system 2 may include localization tools or devices that may provide data on the location of sensing device 4 as it traverses the GI tract. Workstation 8 may access a preferred localization display application for the intended region in the patient's body from which sensing device 4 collects sensory data. For example, workstation 8 may access a generic localization display program and superimpose a diagram, map or schematic illustration of the region, for example, the GI tract, on a generic display.
In one embodiment, identification data 73 may include data that uniquely identifies sensing device 4, for example, a unique identifier such as a serial number, code or electronic signature. Multiple sensing devices 4 traversing one or more patients' bodies may transmit sensory data to receiver 6, for example, at overlapping times. Identification data 73 may be attached, grouped with or tagged onto the sensory data according to embodiments of the invention. Receiver 6 may separate the sensory data into separate image streams according to from which sensing device 4 the identification data 73 indicates the sensory data was transmitted. Thus, data collected from multiple sensing devices 4 at the same or overlapping time may be stored, used and displayed separately.
In some embodiments, additional identification data may be accepted at workstation 8, for example, that is entered or selected by a user. Such identification data may be used by system 2 components according to embodiments of the invention. In some embodiments, additional identification data may overwrite or replace transmitted identification data 73.
In some embodiments, identification data 73 transmitted by sensing device 4 may be stored, in a data structure or storage or memory location, with, or associated with sensory data transmitted by the same sensing device 4, for example, in receiver 6 or workstation 8, and be used for reference purposes. For example, identification data 73 may be used to identify sensing device 4 that collected the sensory data. In one embodiment, each frame of image data may include identification data. In another embodiment, each file of image data may include identification data. Other methods of associating identification data with sensory data may be used.
A non-exhaustive list of examples of processing system or workstation 8 includes a original equipment manufacturer (OEM) dedicated work station, a desktop personal computer, a server computer, a laptop computer, a notebook computer, a hand-held computer, and the like.
Receiver 6 may include a memory 56, for example, to store sensory and/or identification data transmitted from sensing device 4, a processor 16, an antenna 58, a receiver (RX), a transmitter 62, a program memory 64, a random access memory (RAM) 66, boot memory 68, a power source 82, and a communication controller, such as, for example, a universal serial bus (USB) controller 70. According to other embodiments of the invention, transmitter 62 may be a unit separate from receiver 6.
Processor 16 may control the operation of receiver 6, transmitter 62, and USB controller 70 through, for example, a bus 74. In addition, receiver 6, transmitter 62, processor 16 and USB controller 70 may exchange data, such as, for example, sensory data received from sensing device 4, or portions thereof, over bus 74. Other methods for control and data exchange are possible.
One or more antenna(s) 58 may be mounted inside or outside receiver 6 and both receiver 60 and transmitter 62 may be coupled to antenna 58. Transmitter 62 may transmit wireless messages to sensing device 4 through antenna 58. Receiver 6 may receive transmissions, for example, from sensing device 4 through antenna 58.
Receiver 6 may communicate with workstation 8 via connection or medium 12. For example, receiver 6 may transfer bits of wireless communication, for example, sensory data, identification data or other data stored in memory 56 to workstation 8, and may receive controls, and other digital content, from workstation 8. Although the invention is not limited in this respect, medium 12 may be, for example, a USB cable and may be coupled to USB controller 70 of receiver 6. Alternatively, medium 12 may be wireless, and receiver 6 and workstation 8 may communicate wirelessly.
A non-exhaustive list of examples of antennae 32 and 58 includes dipole antennae, monopole antennae, multilayer ceramic antennae, planar inverted-F antennae, loop antennae, shot antennae, dual antennae, omni-directional antennae, coil antennae or any other suitable antennas. Moreover, antenna 32 and antenna 58 may be of different types.
Sensing device 4 may be or may include an autonomous swallowable capsule, for example, an imaging capsule, but sensing device 4 may have other shapes and need not be swallowable or autonomous. Embodiments of sensing device 4 are typically autonomous, and are typically self-contained. For example, sensing device 4 may be a capsule or other unit where all the components including for example power components are substantially contained within a container or shell, and where sensing device 4 does not require any wires or cables to, for example, receive power or transmit information. Sensing device 4 may communicate with an external receiving and display system to provide display of data, control, or other functions. For example, in an autonomous system power may be provided by an internal battery or a wireless receiving system. Other embodiments may have other configurations and capabilities. For example, components may be distributed over multiple sites or units. Control information may be received from an external source.
A non-exhaustive list of examples of memory units 33 includes, for example, semiconductor devices such as registers, latches, electrically erasable programmable read only memory devices (EEPROM), flash memory devices, etc. At least one memory unit 33 may store identification data 73.
Power source 34 may include batteries, such as, for example, silver oxide batteries, lithium batteries, capacitors, or any other suitable power source. Power source 34 may receive power from an external power source, for example, by a magnetic field or electric field that transmits to the device.
Imaging sensor 40 may be for example a solid state imaging sensor or imager, a complementary metal oxide semiconductor (CMOS) imaging sensor, a charge coupled device (CCD) imaging sensor, a “camera on chip” imaging sensor, or any other suitable imaging sensor. A 256×256 or 320×320 pixel imager may be used. Pixel size may be, for example, between 5 and 6 micron. According to some embodiments pixels may be each fitted with a micro lens. Other numbers or dimensions may be used.
Control block 26 may control, at least in part, the operation of sensing device 4. For example, control block 26 may synchronize time periods, in which illumination source 38 produce light rays, time periods, in which imaging sensor 40 captures images, and time periods, in which transmitter 28 transmits the images. In addition, control block 26 may produce timing signals and other signals necessary for the operation of transmitter 28, receiver 30 and imaging sensor 40. Moreover, control block 26 may perform operations that are complimentary to the operations performed by other components of sensing device 4, such as, for example, image data buffering. Identification data 73 may be used to control the mode or setting for control block 26, processor 47 or image sensor 40. Control block 26 may include any combination of logic components, such as, for example, combinatorial logic, state machines, controllers, processors, memory elements, and the like.
Control block 26, transmitter 28, optional receiver 30 and imaging sensor 40 may be implemented on any suitable combination of semiconductor dies or chips. For example, and although the invention is not limited in this respect, control block 26, transmitter 28 and optional receiver 30 may be parts of a first semiconductor die or chip, and imaging sensor 40 may be a part of a second semiconductor die. Such a semiconductor die may be an application-specific integrated circuit (ASIC) or may be part of an application-specific standard product (ASSP). According to some embodiments semiconductor dies may be stacked. According to some embodiments some or all of the components may be on the same semiconductor die.
Illumination source 38 may produce light rays 44 that may penetrate through optical window 36 and may illuminate an inner portion 46 of a body lumen. A non-exhaustive list of examples of body lumens includes the gastrointestinal (GI) tract, a blood vessel, a reproductive tract, or any other suitable body lumen.
Reflections 50 of light rays 44 from inner portion 46 of a body lumen may penetrate optical window 36 back into sensing device 4 and may be focused by optical system 42 onto imaging sensor 40. Imaging sensor 40 may receive the focused reflections 50, and in response to an image capturing command from control block 26, imaging sensor 40 may capture image data or an image of inner portion 46 of a body lumen. Control block 26 may receive the image of inner portion 46 from imaging sensor 40 over wires 54, and may control transmitter 28 to transmit the image of inner portion 46 through antenna 32 into wireless medium 11. Optional processor 47 may modify control block 26 operations.
Sensing device 4 may passively or actively progress along a body lumen. Consequently, a stream of sensory data of inner portions of a body lumen may be transmitted from sensing device 4 into wireless medium 11.
Sensing device 4 may transmit captured images embedded in, for example, “wireless communication frames”. A payload portion of a wireless communication frame may include a captured image or other sensing data and may include additional data, such as, for example, identification data 73, telemetry information and/or cyclic redundancy code (CRC) and/or error correction code (ECC). In addition, a wireless communication frame may include an overhead portion that may contain, for example, framing bits, synchronization bits, preamble bits, and the like. Identification data 73 may be sent separately from image frames.
Receiver 30 may receive wireless messages via wireless medium 11 through antenna 32, and control block 26 may capture these messages. A non-exhaustive list of examples of such messages includes modifying the operations of sensing device 4, for example, activating or de-activating image capturing by sensing device 4 and/or activating or de-activating transmissions from sensing device 4, based on transmitted identification data 73.
Typically, the sensing device transmits data that are fixed in size. Typically, the sensing device collects data at a constant rate. For example, sensing device 4 may capture an image once every half second, and, after capturing such an image data may be transmitted the image to receiver 6 as an encoded image possibly over a series of imaging and transmission periods. A transmission or imaging period may be a period of time during which the sensing device may collect, generate and/or transmit a stream of sensory data. For example, in each of a series of transmission periods, a frame of image data may be captured and transmitted. Other constant and/or variable capture rates and/or transmission rates may be used. Typically, the image data recorded and transmitted is digital color image data, although in alternate embodiments other image formats (e.g., black and white image data) may be used. In one embodiment, each frame of image data may include, for example, 256 rows of 256 pixels or 320 rows of 320 pixels each, each pixel including data for color and brightness, according to known methods. Other data formats may be used.
In one embodiment, identification data 73 may be transmitted once at the start and/or once at the end of the collection and/or transmission of sensory data from sensing device 4. In such embodiments, identification data 73 may be used to indicate or command the start or end of data transmissions from sensing devices 4. For example, after receiver 6 receives identification data 73, indicating the completion of the transmission or reception of image data corresponding to an image frame. Upon receiving such indications, receiver 6 may de-activate receiving operations. In another embodiment, identification data 73 may be transmitted once at the start and/or once at the end of the movement of sensing device 4 across a region of a patient's body. Such markers may be used by receiver 6 and/or workstation 8 to sort or group sensory data (e.g., by image or frame).
The location of identification data 73 in transmitted data streams may be fixed or otherwise indicated, for example, by a data marker, pointer or an address, which may be easily accessible to a user or program applications. This may enable receiver 6, workstation 8 or a user to efficiently locate and access identification data 73.
In one embodiment, sensing device 4 may transmit identification data 73 separately from sensory data. For example, if sensory data corresponding to an image frame is not transmitted (e.g. due to functional error) identification data 73 corresponding to the image frame may still be transmitted.
In another embodiment, sensing device 4 may transmit identification data 73 together with sensory data, for example, in substantially the same data block, data stream or transmission or imaging period. Identification data 73 may be transmitted with sensory data, for example, with every or substantially every data transmission, image frame transmission or during substantially every transmission or imaging period. In one embodiment, receiving identification data may indicate the completion of the transmission of image data corresponding to an image frame. In some embodiments, relatively low data transmission rates may be used, for example, in accordance with regulations. Transmitting identification data 73 with substantially every image data transmission may enable receiver 6 and/or workstation 8 to access the identification data 73 without requesting it from sensing device 4, which may be temporally inefficient or may take time, where time constraints may be an issue. In another embodiment, identification data 73 may be transmitted less often than sensory data.
Reference is made to
In one embodiment sub-block 200 and sub-block 202 may package data in lines, sets, items or units of data that are typically a fixed size. For example, sub-block 202 may include a fixed number of bytes corresponding to the, for example, 256×262 pixels or 320×320 pixels of an image frame. In one embodiment, sensory data corresponding to each pixel in the image frame may have a fixed size, for example, 8 bits or 12 bits. Other block sizes or data formats may be used. Data block 204 may be any suitable length or size. While the length or size of data block 204 is typically fixed across transmission periods, the length may vary in some embodiments and/or transmissions.
Sub-block 200 may store multiple types of identification data 73. In one embodiment, specific types of identification data 73 may be grouped or transmitted in specific segments of sub-block 200, for example, in portions of in sub-block 200 that are fixed in size and position. Thus, system 2 components may automatically or efficiently access a desired specific type of identification data 73. For example, the unique identifier, geographical region data, body region data and model data may be transmitted in portions 250, 260, 270 and 280 of sub-block 200, respectively. Portions 250, 260, 270 and 280 of sub-block 200 may be arranged in any order in sub-block 200. Other data may be transmitted adjacent to or in between portions 250, 260, 270 and 280 of sub-block 200. Data block 204 may include a marker or address that identifies the location of identification data 73 in data block 204.
In operation 400, an in-vivo sensing device may collect sensory data. Sensory data may include, for example, image data collected or captured using an imaging system. For example, an autonomous in-vivo imaging device may capture image data. In other embodiments, sensory data may include, for example, data relating to pH, temperature, pressure, electrical impedance, or other sensed information.
In operation 410, identification data may be transmitted. The identification data may be transmitted alone or with the sensory data. Identification data may be attached to or grouped, packaged, transmitted or associated with sensory data, for example, in a data block or transmission period. In one embodiment, during one transmission period, data may be transmitted that includes image data and identification data.
In operation 420, a receiver may receive identification data, and may record or store the identification data. The receiver may send the identification data to a processing system such as a workstation via a wireless or hard-wired medium. The identification data may be sent alone or with the sensory data.
In operation 430, an in-vivo sensing system may use the identification data. For example, the workstation and/or receiver may store, process, display or use the sensory data in a suitable manner, for example, as allowed by the identification data. For example, identification data may be used to verify component compatibility or permissions, to allow access, or to select compatible system software or components, preferred operation settings, programs or software. System operation may be modified according to for the identification data. Identification data may have other meaning or functionality.
In one embodiment, a system component may compare the identification data transmitted by the sensing device with the system component's requirement data. For example, the system component may compare analogous portions of the identification data and requirement data to determine if the two data sets substantially match. In some embodiments, the system component may only use sensory data transmitted by the sensing devices if the sensing devices transmits identification data that matches the requirement data.
Other operations or series of operations may be used.
It will be appreciated by persons skilled in the art that the present invention is not limited to what has been particularly shown and described hereinabove. Rather the scope of the present invention is defined only by the claims, which follow: