This document relates to an infusion pump system, such as a portable infusion pump system for dispensing insulin or another medicine.
Pump devices are commonly used to deliver one or more fluids to a targeted individual. For example, a medical infusion pump device may be used to deliver a medicine to a patient as part of a medical treatment. The medicine that is delivered by the infusion pump device can depend on the condition of the patient and the desired treatment plan. For example, infusion pump devices have been used to deliver insulin to the vasculature of diabetes patients so as to regulate blood-glucose levels.
Users of infusion pump devices often need to communicate with the infusion pump via a user interface to control the operations of the infusion pump in a safe and effective manner. For example, a user may press a series of buttons on the user interface to enter food intake data into the infusion pump, such as a number of grams of carbohydrates that is indicative of a recently or soon-to-be consumed meal. The food intake data can be used in conjunction with other parameters stored by the infusion pump system to calculate a suggested bolus dosage of insulin based on the grams of carbohydrates entered by the user. In another example, a user may enter information into the infusion pump system via a user interface that indicates that the user is going to perform a level of physical exercise. In some circumstances, the infusion pump system may reduce the amount of a planned dispensation of insulin in response to the exercise information entered by the user.
Some embodiments of an infusion pump system may be configured to receive user input at the infusion pump system using voice input. Some such embodiments can interpret the user's voice input using speech recognition technology, and in response to the user's voice input, the infusion pump system can automatically perform one or more tasks (e.g., without additional user intervention). By incorporating speech recognition equipment within the infusion pump system, user communications with the pump system can be enhanced and simplified. In particular embodiments, the infusion pump system may further include a capability to perform natural language processing of the user's voice input, thereby providing an infusion pump system configured to correlate any one of a number of spoken phrases into selected tasks. In addition or in the alternative, some embodiments of an infusion pump system may be configured to allow the user to provide input to the infusion pump system using photographic images. For example, the user may take a photo of a soon-to-be-consumed meal, and the photo may be provided as food intake data that is input to the infusion pump system for purposes of performing one or more tasks by the infusion pump system. In response, the infusion pump system may, for example, use image recognition technology to estimate the carbohydrate and other nutritional contents of the food depicted in the photo and then suggest a particular bolus dosage of insulin (or other medicine) corresponding to the food in the photo.
In particular embodiments described herein, a medical infusion pump system may include a portable housing that defines a space to receive a medicine. The system may also include a pump drive system to dispense medicine from the portable housing when the medicine is received in the space. In some embodiments, the system may also include control circuitry that communicates control signals to the pump drive system to control dispensation of the medicine from the portable housing. The system may also include a speech recognition system that is in communication with the control circuitry. The control circuitry may select one or more tasks to be performed by the infusion pump system in response to the speech recognition system receiving a user's voice input.
In some embodiments of the medical infusion pump system that includes the speech recognition system, at least a portion of the speech recognition system may be stored in one or more computer-readable memory devices at a remote server system, and the control circuitry may be configured to communicate with the remote server system to use the speech recognition system. Optionally, at least a portion of the speech recognition system may be disposed in the portable housing. Further, the control circuitry may be housed in a controller housing that is removably attachable to the portable housing, and at least a portion of the speech recognition system may be disposed in the controller housing. In some embodiments, the speech recognition system may optionally comprise a first subsystem and a second subsystem. At least a portion of the first subsystem may be stored in one or more computer-readable memory devices at a remote server system that communicates with the control circuitry. In addition, at least a portion of the second subsystem may be stored in one or more computer-readable memory devices in the portable housing or in a controller device housing in which the control circuitry is housed and that is removably attachable to the portable housing. In some embodiments, the medical infusion pump system may also include a voice synthesizer for outputting audible human language communications from the infusion pump system. In particular embodiments, the medical infusion pump system may include a remote control device that is separate from the portable housing and that houses the control circuitry. The remote control device may be configured to wirelessly communicate with a wireless communication device housed in the portable housing, and the remote control device may include a microphone for receiving the voice input. Further, in some embodiments the medical infusion pump optionally includes a voice synthesizer for outputting audible human language communications from the remote control device.
In particular embodiments described herein, a medical infusion pump system may include a portable housing that defines a space to receive a medicine. The system may also include a pump drive system to dispense medicine from the portable housing when the medicine is received in the space. In some embodiments, the system may also include control circuitry that communicates control signals to the pump drive system to control dispensation of the medicine from the portable housing. The system may also include an image recognition system in communication with the control circuitry. The control circuitry may select one or more tasks to be performed by the infusion pump system in response to the image recognition system receiving user input comprising a user-provided digital image.
In some embodiments of the medical infusion pump system that includes the image recognition system, at least a portion of the image recognition system may be stored in one or more computer-readable memory devices at a remote server system, and the control circuitry may be configured to communicate with the remote server system to use the image recognition system. In particular embodiments, at least a portion of the image recognition system may be disposed in the portable housing. Optionally, the control circuitry may be housed in a controller housing that is removably attachable to the portable housing, and at least a portion of the image recognition system may be disposed in the controller housing. In some embodiments, the image recognition system may comprise a first subsystem and a second subsystem. At least a portion of the first subsystem may be stored in one or more computer-readable memory devices at a remote server system that communicates with the control circuitry, and at least a portion of the second subsystem may be stored in one or more computer-readable memory devices in the portable housing or in a controller device housing in which the control circuitry is housed and that is removably attachable to the portable housing. Optionally, the medical infusion pump system that includes the image recognition system may include a voice synthesizer for outputting audible human language communications from the infusion pump system. The system may also optionally include a remote control device that is separate from the portable housing and that houses the control circuitry. The remote control device may be configured to wirelessly communicate with a wireless communication device housed in the portable housing, and the remote control device may include a camera device for receiving the digital image. Some such embodiments may include a voice synthesizer for outputting audible human language communications from the remote control device.
Some embodiments described herein may include a method of controlling a portable infusion pump system. The method may include receiving a user's voice input that is indicative of a task associated with using a portable infusion pump system, and controlling the portable infusion pump system to change an operation of the portable infusion pump system based upon the user's voice input. The method may optionally include, prompting a user via a user interface display to confirm the operation change of the portable infusion pump system in response to receiving the user's voice input. In some embodiments, the operation change may optionally comprise calculating or initiating a bolus dispensation of a medicine from the portable infusion pump system.
Some embodiments described herein may include another method of controlling a portable infusion pump system. The method may include receiving user input comprising a digital image that is indicative of a food item consumed or to be consumed by the user of the portable infusion pump system, and controlling the portable infusion pump system to change an operation of the portable infusion pump system based upon the user input comprising the digital image. The method may optionally include, prompting a user via a user interface display to confirm the operation change of the portable infusion pump system in response to receiving the user input comprising the digital image. In some embodiments, the operation change may optionally comprise calculating or initiating a bolus dispensation of a medicine from the portable infusion pump system.
Some or all of the embodiments described herein may provide one or more of the following advantages. First, some embodiments of the infusion pump system may be configured to receive user input via speech recognition technology. Second, some embodiments of the infusion pump system may be configured to receive user input via image recognition technology. Third, some embodiments of an infusion pump system equipped with speech or image recognition technology may facilitate convenient user input of information to the infusion pump system. Third, the safety and efficacy of an infusion pump system may be enhanced because the convenient manner of inputting data to the infusion pump using speech or image recognition may facilitate more timely and complete data entry by the user. Fourth, in some circumstances, some users who may be unable (mentally or physically) to reliably operate a conventional push-button user interface of an infusion pump system may instead be served by embodiments of the system described herein, which can permit such users to reliably input data to an infusion pump system using the speech or image recognition communication interface. Fifth, the infusion pump system equipped with speech or image recognition capabilities may be configured to be portable, wearable, and (in some circumstances) concealable. For example, a user can conveniently wear the infusion pump system on the user's skin under clothing or can carry the pump system in the user's pocket (or other portable location) while receiving the medicine dispensed from the pump device.
The details of one or more embodiments of the invention are set forth in the accompanying drawings and the description below. Other features, objects, and advantages of the invention will be apparent from the description and drawings, and from the claims.
Like reference symbols in the various drawings indicate like elements.
Referring to
In this embodiment, the infusion pump system 10 includes the remote control device 50 in communication with the portable pump 60, which is used to dispense insulin or another medication to a user 15 via an infusion set 70 attached to and penetrating the user's skin 20. In some embodiments, the portable pump 60 optionally includes a user interface 62 comprised of input devices such as buttons 63a, 63b, 64a, 64b, 64c and output devices such as display 65. In addition, in this embodiment the user 15 can communicate with the infusion pump system 10 by providing voice input, such as the example verbal statement 16 depicted in
The infusion pump system 10 is configured to controllably dispense a medicine to be infused into the tissue or vasculature of a targeted individual, such as a human or animal patient. In some embodiments, the portable pump 60 includes the housing structure 66 that defines a cavity in which a fluid cartridge (not shown) can be slidably received. For example, the fluid cartridge can be a carpule that is either user-fillable or is preloaded with insulin or another medicine for use in the treatment of Diabetes (e.g., Byetta®, Symlin®, or others). Such a cartridge may be supplied, for example, by Eli Lilly and Co. of Indianapolis, Ind. Other examples of medicines that can be contained in the fluid cartridge include: pain relief drugs, hormone therapy, blood pressure treatments, anti-emetics, osteoporosis treatments, or other injectable medicines. The fluid cartridge may have other configurations. For example, in some embodiments the fluid cartridge may comprise a reservoir that is integral with the pump housing structure 66 (e.g., the fluid cartridge can be defined by one or more walls of the pump housing structure 66 that surround a plunger to define a reservoir in which the medicine is injected or otherwise received).
Still referring to
In some embodiments, the infusion pump system 10 can be configured to supply scheduled basal dosages of insulin (or another medication) along with user-selected bolus dosages. The basal delivery rate can be selected to maintain a user's blood glucose level in a targeted range during normal activity throughout the day. The user-selected bolus deliveries may provide substantially larger amounts of insulin in particular circumstances in which the user consumed (or will consume) carbohydrates (e.g., during a meal) or in which the user's blood glucose level requires a significant downward correction. In some embodiments, the infusion pump system 10 can suggest a bolus dosage to the user in a manner that accounts for the user's food intake, the user's recent blood glucose level (e.g., manually input into the portable pump 60 by the user, detected from an integral blood test strip analyzer, wirelessly transmitted to the portable pump 60 from an external blood strip reader device, wirelessly transmitted to the portable pump 60 from an body-worn continuous glucose monitoring device, or the like), the rate of change in the user's blood glucose level, and previously delivered insulin that has not acted on the user. For example, a user can enter a carbohydrate value indicative of a meal into the portable pump 60, and in response thereto, the portable pump 60 can output a suggested bolus dosage to the user. In another example, as will be described further below, the user can provide a voice input that identifies food items that the user will consume, and the infusion pump system 10 can use speech recognition technology to determine a suggested bolus dosage that corresponds to the food items.
In some embodiments, the infusion pump system 10 may modify a bolus suggestion (e.g., a bolus delivery in conjunction with a meal) in response to certain circumstances. For example, the infusion pump system 10 may decrease or otherwise modify a post-meal bolus delivery based on a rapidly falling blood glucose level, a current blood glucose level that is below a threshold limit, based on an increased level of physical activity, or the like.
The infusion pump system 10 can be configured to be portable and can be wearable and concealable. For example, a user can conveniently wear some or all of the infusion pump system 10 on the user's skin (e.g., using skin adhesive) underneath the user's clothing or carry the portable pump 60 or remote control device 50 in the user's pocket (or other portable location) while receiving the medicine dispensed from the infusion pump system 10. As such, the pump system 10 can be used to deliver medicine to the tissues or vasculature of the user in a portable, concealable, and discrete manner.
Still referring to
In some embodiments, the user can adjust the settings or otherwise program the portable pump 60 by pressing one or more buttons 63a, 63b, 64a, 64b, and 64c of the user interface 62. For example, in embodiments of the infusion pump system 10 configured to dispense insulin, the user may press one or more of the buttons 63a, 63b, 64a, 64b, and 64c to change the dispensation rate of insulin or to request that a bolus of insulin be dispensed immediately, at a scheduled later time, over a period of time, or following a particular time-based profile. In another example, the user may use the buttons 63a, 63b, 64a, 64b, and 64c to manually input information such as the user's current blood glucose level (e.g., as measured by an external blood glucose meter), the current rate of change in the user's blood glucose level, or the like into the portable pump 60.
As an alternative to, or in conjunction with, pressing one or more buttons 63a, 63b, 64a, 64b, and 64c of the user interface 62 to adjust or program the infusion pump system 10, the example infusion pump system 10 can receive voice input from the user. The use of speech recognition equipment (housed in the control device 50, in the pump device 60, or both) provides an additional functionality that can enhance and simplify user interactions with the portable pump 60. For instance, using speech recognition, the need to manually actuate multiple buttons 63a, 63b, 64a, 64b, and 64c in a specific order for purposes shuffling through menus may be eliminated or otherwise reduced in some circumstances. In one example, as depicted in
Still referring to
In this example, the user 15 has made a statement 16 that identifies types of food that will be consumed, but it should be understood from the description herein that many other types of statements corresponding to other infusion pump tasks can be similarly initiated using voice input. For instance, in other non-limiting examples of the types of statements that can be made to initiate tasks, the user 15 may speak a command to “stop the pump,” “start the pump,” or “stop the bolus.” Further, the user 15 may speak a command to “start a temporary basal rate of 50% for 3 hours,” or “I am going to exercise for 1 hour.” In still further examples, the user 15 may speak commands such as: “prime the infusion set,” “my blood glucose level is 130,” “I am going to sleep now,” “display estimated battery life,” “display estimated medicine expiration time,” “snooze all alerts for 30 minutes,” “how much insulin do I have on board,” “how long have I been using this infusion set,” “what time is it,” “change to basal pattern B,” “change to my weekend basal pattern,” “soccer practice starts in 30 minutes” (which would be linked to a particular pre-programmed temporary basal pattern), “give that bolus as a square-wave bolus,” “give that bolus as a 4-hour combo bolus,” “remind me to bolus in an hour,” “remind me to check my blood sugar in an hour,” “remind me to eat lunch at 11:30,” “blocked set alarm acknowledged,” and the like. It should be recognized that the user 15 can provide a wide variety of types of statements to initiate a wide variety of tasks by the infusion pump system 10, and that the examples provided here are merely illustrative. In some embodiments, as will be described further in reference to
Still referring to
In this embodiment the control device 50 is depicted as a smart phone device, but it should be understood from the description herein that, in other embodiments, the control device 50 can be implemented in the form of devices other than a smart phone device. Some other example devices that can be used similarly to the remote control device 50 can include, but are not limited to, a personal computer, a tablet computing device, a blood glucose meter device (e.g., an external blood strip reader), a continuous glucose meter device, a wearable computing device, a PDA, or a custom remote device. In still other embodiments, the control device is not a remote device, but instead is included as part of, or mechanically attached together with, the pump device. For instance, in such embodiments the pump device of the infusion pump system can be equipped with the capabilities to perform the functions described herein in regard to the remote control device 50. Further, in some embodiments certain operations or parts of certain operations may be performed at a remote server system, including a cloud-based server system, rather than completely on a personal computing device such as the remote control device 50. Accordingly, the remote control device 50, or equivalent, can be connected to a network such as the internet or an intranet system. Such a division of tasks may provide better process optimization, computational efficiency, and response time.
Still referring to
Using an ASR system in the remote control device 50, or remotely located at a server in communication with the remote control device 50, or in a combination of tasks among the remote control device 50 and at a remote server, the audio signal from the voice input 16 can be transcribed to one or more candidate text transcriptions correlating to the audio signal of statement 16. In some embodiments, the control device 50 can generate speech recognition confidence values for the candidate transcriptions that are generated. In particular embodiments, the transcription with the highest confidence value may be selected by the ASR system as the designated transcription. Other techniques may also be used to create transcription(s) in response to the voice input 16, and to select which candidate transcription to use as the designated transcription. In some circumstances, no candidate transcription having a confidence value that surpasses a threshold confidence level is identified. In some such circumstances, the control device 50 may request clarification from the user 15, or may request more information from the user 15. Such requests may be presented to the user 15 audibly using voice synthesis at the remote control device 50, or visually by presenting an indication on the display of the remote control device 50, or by a combination of audible and visual indicators.
Still referring to
Still referring to
Still referring to
Still referring to
Referring now to
In operation 205, the control circuitry of a medical device can receive voice input from a vocal utterance spoken by a user of the medical device. The voice input can be indicative of a task associated with using the medical device. One example of a medical device to perform operation 205 is depicted in
In operation 210, the voice input is coded to digital format (e.g., an “audio file,” “waveform,” “sample,” and the like) by the control circuitry of the medical device and saved in memory of the medical device. For example, in the context of the infusion pump system 10 of
In operation 215, the digitized voice input is analyzed by the control circuitry of the medical device to determine one or more candidate textual transcriptions corresponding to the voice input. This step of the process 200 can be optionally performed using an ASR system, as explained above in regard to
In operation 220, the control circuitry of the medical device compares the textual transcription(s) from operation 215 to tasks pertaining to the medical device and that are stored in a task database. In some embodiments, the task database is stored in the memory of the medical device. In alternative embodiments, the task database is stored at a remote server system that is accessible by the medical device over a network such as the internet. One or more tasks that are stored in the task database can be identified as candidates to have a correspondence to the textual transcription(s). A statistical confidence level can be generated in regard to the correspondence between the textual transcription(s) and the candidate task(s).
In operation 225, the control circuitry of the medical device compares the statistical confidence level(s) generated in operation 220 to a predetermined threshold confidence level. If one and only one particular task has a statistical confidence level that surpasses the threshold confidence level, that particular task is selected as the task to present to the user, and the process 200 moves to operation 235. However, if no particular task has a statistical confidence level that surpasses the threshold confidence level, or if multiple tasks have statistical confidence level(s) that surpass the threshold confidence level, then the process 200 moves to operation 230. In alternative embodiments, if multiple tasks have statistical confidence level(s) that surpass the threshold confidence level, then the task with the highest confidence level is selected as the task to present to the user. In some such alternative embodiments, the task with the highest confidence level is only selected if the confidence level of the task is greater than the next highest confidence level by more than a predetermined differential threshold value.
In operation 230, the control circuitry of the medical device requests user clarification in regard to the voice input that was previously provided by the user in operation 205. The request for user clarification can be presented audibly to the user by voice synthesis via the medical device, by displaying information on the user interface display of the medical device, or both. In some circumstances, the clarification requested may be in relation to a candidate task that had a statistical confidence level that was determined to be less than the threshold confidence level. For instance, such a clarification request could be, “Do you want to stop the pump?” In another circumstance, the clarification requested may be general, rather than in relation to a candidate task. For example, in that circumstance the clarification request could be, “Your input was not understood—please try again,” or another indication that the voice input should be restated. After requesting user clarification, the process 200 returns to operation 205 and waits for further voice input from the user.
In operation 235, after selecting a task in operation 225, the control circuitry of the medical device characterizes the selected task, as needed, in preparation for presentation to the user for confirmation of the task. For example, some data may need to be obtained and some calculations may need to be performed to prepare the task for presentation to the user. To provide a more specific example, as described in the context of the infusion pump system 10 of
In operation 240, the control circuitry of the medical device presents the task to the user for user confirmation. The task can be presented audibly to the user by voice synthesis via the medical device, by displaying information on the user interface display of the medical device, or both. As described in relation to the infusion pump system 10 of
In operation 250, the control circuitry of the medical device determines whether the user input received in operation 245 was a confirmation or a denial of the task that was presented to the user. If the user input was a denial of the task that was presented to the user, the process 200 proceeds to operation 230 where user clarification is requested as described above. If the user input was a confirmation of the task that was presented to the user, the process 200 proceeds to operation 255 where the control circuitry of the medical device communicates the task to other portions of the device so as to implement the task. In this embodiment, the control circuitry communicates the task to the pump controller to implement the task. In the context of the infusion pump system 10 of
Now referring to
The pump controller device 360 includes a control module 361 that can be made up of one or more components. In this embodiment, the control module 361 is configured to communicate control or power signals to the other components of the infusion pump system 300, and to receive inputs and signals therefrom. In some embodiments, the control circuitry can include a main processor board that is in communication with a power supply board. The control circuitry can include at least one processor that coordinates the electrical communication to and from the control module 361 and other components of the infusion pump system 300. For example, the user interface 362 of the pump controller device 360 can include input components (e.g., buttons, touchscreen, microphone, or a combination thereof) and output components (e.g., display screen, speaker, vibratory device, or a combination thereof) that are electrically connected to the control circuitry of the control module 361. In some embodiments, the control module 361 can receive input commands from a user's button selections (e.g., buttons as shown in
The processor of the control module 361 can be arranged on a main processor circuit board of the control module 361 along with a number of other electrical components such as computer-readable memory devices. The control circuitry can be programmable in that the user or a clinician may provide one or more instructions to adjust a number of settings for the operation of the infusion pump system 300. Such settings may be stored in the memory devices of the control module 361. Furthermore, the control module 361 may include one or more dedicated memory devices that store executable software instructions for the processor. The control module 361 may include other components, such as sensors, that are electrically connected to the main processor board. A rechargeable battery pack (not shown) may provide electrical energy to the control module 361, and to other components of the pump controller device 360 (e.g., user interface 362, speech recognition module 363, and others).
Still referring to
The user interface 362 can also include components that facilitate voice communications between the pump controller device 360 and a user. In some embodiments, the user interface 362 includes a microphone (refer, for example, to microphone 51 or microphone 61 in
Still referring to
In some embodiments of the infusion pump system 300, the pump controller device 360 includes the speech recognition module 363, task database 364, and food and activity database 365, while the remote server 310 may not have analogous sub-systems. In such embodiments, the speech recognition process and other operations for facilitating voice communications between the pump controller device 360 and a user are performed entirely at the pump controller device 360. In alternative embodiments of the infusion pump system 300, the remote server 310 includes speech recognition module 313, task database 314, and food and activity database 315, while the pump controller device 360 does not have analogous sub-systems. In such embodiments, the speech recognition process and other operations for facilitating voice communications between the pump controller device 360 and a user are performed by the remote server 310.
In particular embodiments, both the pump controller device 360 and the remote server 310 include the sub-systems for performing speech recognition and other operations for facilitating voice communications between the pump controller device 360 and a user. That is, the pump controller device 360 includes the speech recognition module 363, the task database 364, and the food and activity database 365; and in addition the remote server 310 includes the speech recognition module 313, the task database 314, and the food and activity database 315. In alternative embodiments, one or more of the sub-systems are located in both the pump controller device 360 and the remote server 310, but not all of the sub-systems are located in both.
Various techniques can be used to coordinate the activities between the pump controller device 360 and the remote server 310 when some or all of the sub-systems are arranged in both the pump controller device 360 and the remote server 310. For example, in some embodiments the processing can be initiated locally at the pump controller device 360, and if the pump controller device 360 is unable to attain the threshold statistical confidence levels for the textual transcription of the voice signal or the task matching (refer to
The speech recognition modules 363 and 313 are electrical communication with the control module 361. Optionally, the speech recognition modules 363 and 313 can facilitate the operations of an ASR (“automated speech recognition”) system. Using the ASR system, a digitized audio signal of a user voice input can be transcribed to one or more candidate text transcriptions that are correlated to the audio signal. In some embodiments, statistical confidence values for the candidate transcriptions are generated. In particular embodiments, the transcription with the highest confidence value may be selected as the designated transcription by the ASR system. Other ASR techniques may also be used to create transcription(s), and to select which candidate transcription to use as the designated transcription.
The task databases 364 and 314 are electrical communication with the control module 361. The task databases 364 and 314 are data repositories containing textual tasks and code that relate to the operation of the infusion pump system 300. The textual tasks contained in the task databases 364 and 314 can be compared to the textual transcriptions provided from the ASR system in operation in the speech recognition modules 363 and 313. Accordingly, candidate tasks can be identified as matches with voice inputs provided by a user of the infusion pump system 300. In some embodiments, when no matching task is determined that surpasses a statistical confidence threshold value, the infusion pump system 300 may prompt the user for clarification of the voice input.
The food and activity databases 365 and 315 are electrical communication with the control module 361. The food and activity databases 365 and 315 are data repositories containing data and other types of information that can be used to pre-process a task in preparation presentation to the user and in preparation for implementation of the task. For example, in the infusion pump system 10 of
Still referring to
Still referring to
In brief, in embodiments of the infusion pump system 300 that include the separate pump device 370, the pump controller device 360 may be configured as a reusable component that provides electronics and a user interface to control the operation of the infusion pump system 300, and the separate pump device 370 can be a disposable component that is discarded after a single use. For example, the separate pump device 370 can be a “one time use” component that is thrown away after the fluid cartridge therein is exhausted. Thereafter, the user can wirelessly connect or removably mount a new separate pump device 370 to the reusable pump controller device 360 for the dispensation of a new supply of medicine from the new pump device 370. Accordingly, the user is permitted to reuse the pump controller device 360 (which may include complex or valuable electronics) while disposing of the relatively low-cost separate pump device 370 after each use. Such an infusion pump system 300 can provide enhanced user safety as a new separate pump device 370 is employed with each new fluid cartridge.
Still referring to
Optionally, as shown in
Referring now to
In this example, the infusion pump system 400 can be configured to perform a series of steps A′ through E′ are illustrated that describe operations of an example infusion pump system 400 including with image recognition equipment. By incorporating image recognition capabilities within the infusion pump system 400, user communications with a portable pump 60 can be enhanced and simplified. As a result, the accuracy and completeness of the data entered by the user into the portable pump 60 can be improved, and the user can experience greater convenience and time efficiency. In some embodiments of the infusion pump system 400, speech recognition capabilities (e.g., as described in reference to
As previously described, the infusion pump system 400 can include the remote control device 450 in electrical communication with the portable pump 60, which is used to supply insulin or another medication to a user via an infusion set 70 attached to and penetrating the user's skin 20. In some embodiments, the portable pump 60 includes the user interface 62 comprised of input devices such as buttons 63a, 63b, 64a, 64b, 64c and output devices such as display 65. In addition, in this embodiment the user can communicate with the infusion pump system 400 by providing image input, such as example digital image 440 of a bagel 444 and a serving of orange juice 442. In particular embodiments, the portable pump 60 may also include the wireless communications circuit 40 that facilitates short-range wireless communications 45 between the internal control circuitry of the portable pump 60 and the external remote control device 450. As with the previously described system 10 of
Still referring to
In this embodiment, the image 440 is received by a digital camera system 459 housed in the remote control device 450. The remote control device 450 includes electronic circuitry for digitizing the image 440 into pixels. The digitized image can be stored (permanently or temporarily) in a computer-readable memory device of the remote control device 450. In other embodiments, the image 440 can be received by a digital camera system 69 housed in the pump device 60, and the image can be stored in a computer-readable memory device of the remote control device 50.
In this embodiment the control device 450 is depicted as a smart phone device, but it should be understood from the description herein that, in other embodiments, the control device 450 can be implemented in the form of devices other than a smart phone device. Some other example devices that can be used similarly to the remote control device 450 can include, but are not limited to, a personal computer, a tablet computing device, a blood glucose meter device (e.g., an external blood strip reader), a continuous glucose meter device, a wearable computing device (e.g., glasses equipped with a camera and computer network connectivity), a PDA, a digital camera, or a custom remote device. In still other embodiments, the control device is not a remote device, but instead is included as part of, or mechanically attached together with, the pump device. For instance, in such embodiments the pump device of the infusion pump system can be equipped with the capabilities to perform the functions described herein in regard to the remote control device 450. Further, in some embodiments certain operations or parts of certain operations may be performed at a remote server system, including a cloud-based server system, rather than completely on a personal computing device such as the remote control device 450. Accordingly, the remote control device 450, or equivalent, can be connected to a network such as the internet or an intranet system. Such a division of tasks may provide better process optimization, computational efficiency, and response time.
Still referring to
In this embodiment, the control device 450 is equipped with an image recognition module (refer, for example, to element 369 in
Still referring to
In some embodiments, step C′ can be performed as follows. The candidate matching images selected from the image database 480 as determined by the image recognition process of step B′ can have metadata associated therewith. The metadata can identify the type of food in the image(s) (e.g., a bagel and a serving of orange juice). Using such metadata, the nutritional information for the food types can be queried from the database 490. The nutritional information obtained from the database 490 can be used in computations—along with other parameters such as the user's most current blood glucose reading, an insulin-on-board level, an insulin sensitivity factor for the user, and the like—to determine a recommended bolus dispensation. The descriptions of the food items identified as matching the image 440, and the recommended associated bolus can then be characterized in preparation for presentation to the user (for subsequent user confirmation).
Still referring to
In this example at step E′, the remote control device 450 communicates the task to the portable pump 60 for activation of the portable pump 60 in accordance with the task confirmed by the user (e.g., after the user selected the “YES” button 452). In the example, the display 65 of the portable pump 60 indicates that a bolus dispensation of 4.9 Units has been initiated. In this embodiment, communications between the remote control device 450 and the portable pump 60 are conducted by short-range wireless technologies such as, but not limited to, RF, Bluetooth, NFC, IR, and the like. Accordingly, the portable pump 60 can include a wireless communication circuit 40 that sends and receives data in cooperation with the remote control device 450. In alternative embodiments, the communications between the remote control device 450 and the portable pump 60 can be via a hardwired connection therebetween.
In another embodiment, rather than (or in addition to) using photographic image recognition to ascertain nutritional information for food to be consumed, a portable spectroscope scanner system can be used to ascertain nutritional information for food to be consumed. In this technique, a user can scan food items to be consumed using a portable spectroscope scanner. The spectroscope scanner will create a spectrograph of the food items that can be analyzed to determine nutritional information of the food items. Some spectroscope scanner systems may utilize a reference material placed next to the food for calibration as part of routine use or occasionally.
In some embodiments, the spectroscope scanner transmits the spectrograph data to another processing device that operates a spectrograph analysis application that can be run to determine the nutritional information of the food that was scanned. Such processing devices can include a cloud-based computer system or a local computing device, such as a smartphone, tablet PC, desktop PC, an infusion pump, and the like. In some embodiments, the spectroscope scanner may be able to determine the nutritional information of the food that was scanned without the assistance of another processing device. In particular embodiments, as part of the analysis of the spectrograph, statistical priority can be given to foods that the user has previously utilized the spectrograph analysis technique to identify. The processing device that analyzes the spectrograph can determine the nutritional information and then transmit the nutritional information to the remote control device 450. The remote control device 450 can display the nutritional information to the user, and display a prompt by which the user can initiate a corresponding bolus dispensation via the portable pump device 60, in a manner analogous to that described above.
Referring now to
In this example, the infusion pump system 400 can be configured to perform a series of steps A″ through G″ are illustrated that describe operations of an example infusion pump system 500 equipped with natural language processing (“NLP”) technology. Using NLP, the infusion pump system 500 is capable of receiving instructions from a user 515 via natural language input. One or more NLP algorithms can be stored in the computer-readable memory device in as part of a speech recognition module (refer, for example, to module 363 in
Similar to previously described embodiments, the infusion pump system 500 can include the remote control device 550 in electrical communication with the portable pump 60 that is used to supply insulin or another medication to a user 515 via an infusion set 70 attached to and penetrating the user's skin 20. In particular embodiments, the portable pump 60 may also include the wireless communications circuit 40 that facilitates short-range wireless communications 545 between the internal control circuitry of the portable pump 60 and the external remote control device 550.
As an alternative to, or in conjunction with, pressing one or more buttons 63a, 63b, 64a, 64b, and 64c of the user interface 62 to communicate with the infusion pump system 500, the example infusion pump system 500 can receive natural language voice input from the user 515. The use of NLP technology provides an additional functionality that can enhance and simplify user 515 interactions with the portable pump 60. For instance, using natural language equipment (which may optionally a microphone 551 or 61 and a corresponding NLP software program implemented by the system 500), the need for user activation of multiple buttons 63a, 63b, 64a, 64b, and 64c for shuffling through menus may be eliminated or otherwise reduced in some circumstances. In addition, using NLP equipment, the capabilities of the infusion pump system 500 can extend beyond those that are accessible via the user interface 62. In one such example, as depicted in
Still referring to
In this example, the user 515 has made a statement 516 that identifies the user's 515 blood glucose level, but many other types of statements corresponding to other tasks, questions, or concerns can be similarly initiated using natural language voice input. For instance, in other non-limiting examples such statements can include “I am going for a 45 minute jog,” “tell me about my last bolus,” “how long have I been wearing this infusion set,” “what do I do about the current alarm,” or “how much insulin is left in my reservoir?” It should be recognized from the description herein that the user 515 can provide a wide variety of types of statements to initiate a wide variety of responses by the infusion pump system 500, and that the examples provided here are merely illustrative.
The natural language statement 516 is received by the microphone 551 of the control device 550. The remote control device 550 can include electronic circuitry for converting the statement 516 to an audio signal (e.g., an “audio file,” “waveform,” or “sample”) that corresponds to the statement 516. The audio signal corresponding to the statement 516 can be saved in the memory of the remote control device 550. In other embodiments, the natural language statement can be received by the microphone 61 housed in the pump device 60.
In this embodiment the control device 550 is depicted as a smart phone device, but it should be understood from the description herein that, in other embodiments, the control device 550 can be implemented in the form of devices other than a smart phone device. Some other example devices that can be used similarly to the remote control device 550 can include, but are not limited to, a personal computer, a tablet computing device, a blood glucose meter device (e.g., an external blood strip reader), a continuous glucose meter device, a wearable computing device, a PDA, or a custom remote device. In still other embodiments, the control device is not a remote device, but instead is included as part of, or mechanically attached together with, the pump device. For instance, in such embodiments the pump device of the infusion pump system can be equipped with the capabilities to perform the functions described herein in regard to the remote control device 550. Further, in some embodiments certain NLP operations or parts of certain NLP operations may be performed at a remote server system, including a cloud-based server system, rather than completely on a personal computing device such as the remote control device 550. Accordingly, the remote control device 550, or equivalent, can be connected to a network such as the internet or an intranet system. Such a division of tasks may provide better process optimization, computational efficiency, and response time.
Still referring to
In this example at step C″, the text transcription(s) of the speech recognition process from step B″ is processed using a NLP program executed by the control device 550, the pump device 60, the remote server system, or a combination thereof to determine the likely meaning of the statement 516 and how the infusion pump system 500 should respond. In some cases, in addition to processing the text transcription(s) using NLP, the text transcription(s) is compared to a compilation of tasks or queries in a natural language search engine database 580 to determine the task most likely represented by the statement 516. In some embodiments, the natural language search engine database 580 is stored in the computer-readable memory device of the remote control device 550. However, the natural language search engine database 580 can also be stored in the computer-readable memory device in the portable pump 60, stored in the computer-readable memory device of a remote server system in communication with the remote control device 550 or the portable pump 60, or stored in computer-readable memory devices at a combination of such locations. In this embodiment, the natural language search engine database 580 is a storage repository that is programmed to contain an extensive number of tasks and queries that correspond to a variety of types of user voice input statements, such as statement 516. The transcription(s) of the voice input from step B″ can be compared to the tasks stored in the natural language search engine database 580 to find matching tasks or queries. In some embodiments, a confidence level for the match between the transcription(s) and the task(s) or queries can be determined. The task or query with the highest confidence level can be selected. In particular embodiments, if no task query has such a confidence level that surpasses a threshold level, or if multiple tasks or queries have confidence levels that are within a differentiation threshold level of each other, the user 515 is presented with a request for clarification or more information as described above. In some such cases, the user 515 may be presented with the task or query having the highest confidence level and the user 515 may be asked whether that task is what the user 515 wants the infusion pump system 500 to perform.
Still referring to
In this example at step E″, the task or query is presented to the user 515 for confirmation that the task or query is what the user 515 desires. The task or query may be presented to the user 515 audibly using voice synthesis at the remote control device 550, or visually by presenting an indication on the display of the remote control device 550, or by a combination of audible and visual indicators. For example, in response to the statement 516, the user 515 is presented with information indicating that the infusion pump system 500 recommends a correction bolus dispensation of 5.5 Units of insulin. To confirm that task, the user 515 can select “YES” 552 on the remote control device 550. In response to a selection of the “YES” button 552, the control device 550 can communicate with the pump device 60 so as to initiate the dispensation of the bolus dosage (e.g., 4.9 Units in this example), as described below. Or to deny that task, the user 515 can select “NO” 554 on the remote control device 550. Optionally, in response to a selection of the “NO” button 554, the control device 550 can present the user with an option to manually input or verbally speak a specific number for a bolus dosage that is different from the suggested dosage displayed on the screen at step E. Alternatively, or in addition to, the manual selection of “YES” 552 or “NO” 554, the user 515 may speak “yes” or “no” to the remote control device 550 to confirm or deny the task presented. At step F″, the remote control device 550 receives such user confirmation.
At step G″, the remote control device 550 communicates the task to the portable pump 60 for activation of the portable pump 60 in accordance with the task confirmed by the user 515 (e.g., after the user selected the “YES” button 552). In the example, the display 65 of the portable pump 60 indicates that a bolus dispensation of 5.5 Units has been initiated. In this embodiment, communications between the remote control device 550 and the portable pump 60 are conducted by short-range wireless technologies such as, but not limited to, RF, Bluetooth, NFC, IR, and the like. Accordingly, the portable pump 60 can include a wireless communication circuit 40 that sends and receives data in cooperation with the remote control device 550. In alternative embodiments, the communications between the remote control device 550 and the portable pump 60 can be via a hardwired connection therebetween.
A number of embodiments of the invention have been described. Nevertheless, it will be understood that various modifications may be made without departing from the spirit and scope of the invention. Accordingly, other embodiments are within the scope of the following claims.
This application is a continuation of U.S. application Ser. No. 14/094,185, filed Dec. 2, 2013, now U.S. Pat. No. 10,569,015, issued Feb. 25, 2020, the entire contents of which are hereby incorporated by reference.
Number | Name | Date | Kind |
---|---|---|---|
2605765 | Kollsman | Aug 1952 | A |
3688764 | Owen | Sep 1972 | A |
3886938 | Szabo et al. | Jun 1975 | A |
4077405 | Haerten et al. | Mar 1978 | A |
4151845 | Clemens | May 1979 | A |
4231368 | Becker | Nov 1980 | A |
4235234 | Martin et al. | Nov 1980 | A |
4265241 | Portner et al. | May 1981 | A |
4300554 | Hessberg et al. | Nov 1981 | A |
4313439 | Babb et al. | Feb 1982 | A |
4373527 | Fischell | Feb 1983 | A |
4398908 | Siposs | Aug 1983 | A |
4435173 | Siposs et al. | Mar 1984 | A |
4443218 | Decant et al. | Apr 1984 | A |
4475901 | Kraegen et al. | Oct 1984 | A |
4493704 | Beard et al. | Jan 1985 | A |
4529401 | Leslie et al. | Jul 1985 | A |
4552561 | Eckenhoff et al. | Nov 1985 | A |
4564054 | Gustavsson | Jan 1986 | A |
4652260 | Fenton et al. | Mar 1987 | A |
4668220 | Hawrylenko | May 1987 | A |
4681569 | Coble et al. | Jul 1987 | A |
4731726 | Allen, III | Mar 1988 | A |
4734092 | Millerd | Mar 1988 | A |
4749109 | Kamen | Jun 1988 | A |
4768506 | Parker et al. | Sep 1988 | A |
4838857 | Strowe et al. | Jun 1989 | A |
4850817 | Nason et al. | Jul 1989 | A |
4902278 | Maget et al. | Feb 1990 | A |
5029591 | Teves | Jul 1991 | A |
5045064 | Idriss | Sep 1991 | A |
5088981 | Howson et al. | Feb 1992 | A |
5088990 | Hivale et al. | Feb 1992 | A |
5176632 | Bernardi | Jan 1993 | A |
5190522 | Wojcicki et al. | Mar 1993 | A |
5209230 | Swedlow et al. | May 1993 | A |
5225763 | Krohn et al. | Jul 1993 | A |
5250027 | Lewis et al. | Oct 1993 | A |
5261882 | Sealfon | Nov 1993 | A |
5314412 | Rex | May 1994 | A |
5335994 | Weynant nee Girones | Aug 1994 | A |
5338157 | Blomquist | Aug 1994 | A |
5342180 | Daoud | Aug 1994 | A |
5349575 | Park | Sep 1994 | A |
5389078 | Zalesky et al. | Feb 1995 | A |
5395340 | Lee | Mar 1995 | A |
5411487 | Castagna | May 1995 | A |
5545143 | Fischell et al. | Aug 1996 | A |
5551850 | Williamson et al. | Sep 1996 | A |
5554123 | Herskowitz | Sep 1996 | A |
5569186 | Lord et al. | Oct 1996 | A |
5626566 | Petersen et al. | May 1997 | A |
5637095 | Nason et al. | Jun 1997 | A |
5656032 | Kriesel et al. | Aug 1997 | A |
5665065 | Colman et al. | Sep 1997 | A |
5672167 | Athayde et al. | Sep 1997 | A |
5693018 | Kriesel et al. | Dec 1997 | A |
5718562 | Lawless | Feb 1998 | A |
5741216 | Hemmingsen et al. | Apr 1998 | A |
5766155 | Hyman et al. | Jun 1998 | A |
5772635 | Dastur et al. | Jun 1998 | A |
5800420 | Grose et al. | Sep 1998 | A |
5816306 | Giacomel | Oct 1998 | A |
5822715 | Worthington et al. | Oct 1998 | A |
5851197 | Marano et al. | Dec 1998 | A |
5852803 | Ashby et al. | Dec 1998 | A |
5858001 | Tsals et al. | Jan 1999 | A |
5858005 | Kriesel | Jan 1999 | A |
5873731 | Prendergast | Feb 1999 | A |
5893838 | Daoud et al. | Apr 1999 | A |
5914941 | Janky | Jun 1999 | A |
5919167 | Mulhauser et al. | Jul 1999 | A |
5925018 | Ungerstedt | Jul 1999 | A |
5928201 | Poulsen et al. | Jul 1999 | A |
5947934 | Hansen et al. | Sep 1999 | A |
5951530 | Steengaard et al. | Sep 1999 | A |
5957889 | Poulsen et al. | Sep 1999 | A |
5984894 | Poulsen et al. | Nov 1999 | A |
5984897 | Petersen et al. | Nov 1999 | A |
5997475 | Bortz | Dec 1999 | A |
6003736 | Ljunggren | Dec 1999 | A |
6010485 | Buch-Rasmussen et al. | Jan 2000 | A |
6033377 | Rasmussen et al. | Mar 2000 | A |
6045537 | Klitmose | Apr 2000 | A |
6056728 | Von Schuckmann | May 2000 | A |
6074372 | Hansen | Jun 2000 | A |
6106498 | Friedli | Aug 2000 | A |
6110149 | Klitgaard et al. | Aug 2000 | A |
6126595 | Amano et al. | Oct 2000 | A |
6127061 | Shun et al. | Oct 2000 | A |
6156014 | Petersen et al. | Dec 2000 | A |
6171276 | Lippe et al. | Jan 2001 | B1 |
6231540 | Smedegaard | May 2001 | B1 |
6233471 | Berner et al. | May 2001 | B1 |
6248067 | Causey, III et al. | Jun 2001 | B1 |
6248090 | Jensen et al. | Jun 2001 | B1 |
6248093 | Moberg | Jun 2001 | B1 |
6251113 | Appelbaum et al. | Jun 2001 | B1 |
6269340 | Ford et al. | Jul 2001 | B1 |
6277098 | Klitmose et al. | Aug 2001 | B1 |
6292440 | Lee | Sep 2001 | B1 |
6302855 | Lav et al. | Oct 2001 | B1 |
6302869 | Klitgaard | Oct 2001 | B1 |
6368314 | Kipfer et al. | Apr 2002 | B1 |
6375638 | Nason et al. | Apr 2002 | B2 |
6379301 | Worthington et al. | Apr 2002 | B1 |
6379339 | Klitgaard et al. | Apr 2002 | B1 |
6381496 | Meadows et al. | Apr 2002 | B1 |
6397098 | Uber et al. | May 2002 | B1 |
6404098 | Kayama et al. | Jun 2002 | B1 |
6427088 | Bowman, IV et al. | Jul 2002 | B1 |
D461241 | Moberg et al. | Aug 2002 | S |
D461891 | Moberg | Aug 2002 | S |
6434528 | Sanders | Aug 2002 | B1 |
6436072 | Kullas et al. | Aug 2002 | B1 |
6461329 | Van et al. | Oct 2002 | B1 |
6461331 | Van Antwerp | Oct 2002 | B1 |
6474219 | Klitmose et al. | Nov 2002 | B2 |
6485461 | Mason et al. | Nov 2002 | B1 |
6491684 | Joshi et al. | Dec 2002 | B1 |
6508788 | Preuthun | Jan 2003 | B2 |
6524280 | Hansen et al. | Feb 2003 | B2 |
6533183 | Aasmul et al. | Mar 2003 | B2 |
6537251 | Klitmose | Mar 2003 | B2 |
6537268 | Gibson et al. | Mar 2003 | B1 |
6540672 | Simonsen et al. | Apr 2003 | B1 |
6544229 | Danby et al. | Apr 2003 | B1 |
6547764 | Larsen et al. | Apr 2003 | B2 |
6551276 | Mann et al. | Apr 2003 | B1 |
6554798 | Mann et al. | Apr 2003 | B1 |
6554800 | Nezhadian et al. | Apr 2003 | B1 |
6558320 | Causey, III et al. | May 2003 | B1 |
6558345 | Houben et al. | May 2003 | B1 |
6558351 | Steil et al. | May 2003 | B1 |
6562001 | Lebel et al. | May 2003 | B2 |
6562011 | Buch-Rasmussen | May 2003 | B1 |
6564105 | Starkweather et al. | May 2003 | B2 |
6569126 | Poulsen et al. | May 2003 | B1 |
6571128 | Lebel et al. | May 2003 | B2 |
6572542 | Houben et al. | Jun 2003 | B1 |
6572545 | Knobbe et al. | Jun 2003 | B2 |
6577899 | Lebel et al. | Jun 2003 | B2 |
6582404 | Klitgaard et al. | Jun 2003 | B1 |
6585644 | Lebel et al. | Jul 2003 | B2 |
6585699 | Ljunggreen et al. | Jul 2003 | B2 |
6587199 | Luu | Jul 2003 | B1 |
6589229 | Connelly et al. | Jul 2003 | B1 |
6599281 | Struys et al. | Jul 2003 | B1 |
6605067 | Larsen | Aug 2003 | B1 |
6605072 | Struys et al. | Aug 2003 | B2 |
6613019 | Munk | Sep 2003 | B2 |
6641533 | Causey, III et al. | Nov 2003 | B2 |
6648821 | Lebel et al. | Nov 2003 | B2 |
6650951 | Jones et al. | Nov 2003 | B1 |
6656158 | Mahoney et al. | Dec 2003 | B2 |
6656159 | Flaherty | Dec 2003 | B2 |
6659948 | Lebel et al. | Dec 2003 | B2 |
6659978 | Kasuga et al. | Dec 2003 | B1 |
6659980 | Moberg et al. | Dec 2003 | B2 |
6663602 | Møller | Dec 2003 | B2 |
6668196 | Villegas et al. | Dec 2003 | B1 |
6669668 | Kleeman et al. | Dec 2003 | B1 |
6669669 | Flaherty et al. | Dec 2003 | B2 |
6685674 | Douglas et al. | Feb 2004 | B2 |
6687546 | Lebel et al. | Feb 2004 | B2 |
6689100 | Connelly et al. | Feb 2004 | B2 |
6690192 | Wing | Feb 2004 | B1 |
6691043 | Ribeiro, Jr. | Feb 2004 | B2 |
6692457 | Flaherty | Feb 2004 | B2 |
6692472 | Hansen et al. | Feb 2004 | B2 |
6694191 | Starkweather et al. | Feb 2004 | B2 |
6699218 | Flaherty et al. | Mar 2004 | B2 |
6702779 | Connelly et al. | Mar 2004 | B2 |
6715516 | Ohms et al. | Apr 2004 | B2 |
6716198 | Larsen | Apr 2004 | B2 |
6723072 | Flaherty et al. | Apr 2004 | B2 |
6723077 | Pickup et al. | Apr 2004 | B2 |
6733446 | Lebel et al. | May 2004 | B2 |
6736796 | Shekalim | May 2004 | B2 |
6740059 | Flaherty | May 2004 | B2 |
6740072 | Starkweather et al. | May 2004 | B2 |
6740075 | Lebel et al. | May 2004 | B2 |
6744350 | Blomquist | Jun 2004 | B2 |
6749587 | Flaherty | Jun 2004 | B2 |
6752785 | Van et al. | Jun 2004 | B2 |
6752787 | Causey, III et al. | Jun 2004 | B1 |
6758810 | Lebel et al. | Jul 2004 | B2 |
6761286 | Py et al. | Jul 2004 | B2 |
6768425 | Flaherty et al. | Jul 2004 | B2 |
6780156 | Haueter et al. | Aug 2004 | B2 |
6786246 | Ohms et al. | Sep 2004 | B2 |
6786890 | Preuthun et al. | Sep 2004 | B2 |
6796957 | Carpenter et al. | Sep 2004 | B2 |
6796970 | Klitmose et al. | Sep 2004 | B1 |
6799149 | Hartlaub | Sep 2004 | B2 |
6809653 | Mann et al. | Oct 2004 | B1 |
6810290 | Lebel et al. | Oct 2004 | B2 |
6811533 | Lebel et al. | Nov 2004 | B2 |
6811534 | Bowman, IV et al. | Nov 2004 | B2 |
6813519 | Lebel et al. | Nov 2004 | B2 |
6827702 | Lebel et al. | Dec 2004 | B2 |
6830558 | Flaherty et al. | Dec 2004 | B2 |
6852104 | Blomquist | Feb 2005 | B2 |
6854620 | Ramey | Feb 2005 | B2 |
6854653 | Eilersen | Feb 2005 | B2 |
6855129 | Jensen et al. | Feb 2005 | B2 |
6872200 | Mann et al. | Mar 2005 | B2 |
6873268 | Lebel et al. | Mar 2005 | B2 |
6878132 | Kipfer | Apr 2005 | B2 |
6893415 | Madsen et al. | May 2005 | B2 |
6899695 | Herrera | May 2005 | B2 |
6899699 | Enggaard | May 2005 | B2 |
6922590 | Whitehurst | Jul 2005 | B1 |
6923763 | Kovatchev et al. | Aug 2005 | B1 |
6925393 | Kalatz et al. | Aug 2005 | B1 |
6936006 | Sabra | Aug 2005 | B2 |
6936029 | Maim et al. | Aug 2005 | B2 |
6945961 | Miller et al. | Sep 2005 | B2 |
6948918 | Hansen | Sep 2005 | B2 |
6950708 | Bowman, IV et al. | Sep 2005 | B2 |
6960192 | Flaherty et al. | Nov 2005 | B1 |
6979326 | Mann et al. | Dec 2005 | B2 |
6997911 | Klitmose | Feb 2006 | B2 |
6997920 | Mann et al. | Feb 2006 | B2 |
7005078 | Van Lintel et al. | Feb 2006 | B2 |
7008399 | Larson et al. | Mar 2006 | B2 |
7014625 | Bengtsson | Mar 2006 | B2 |
7018360 | Flaherty et al. | Mar 2006 | B2 |
7025743 | Mann | Apr 2006 | B2 |
7029455 | Flaherty | Apr 2006 | B2 |
7054836 | Christensen et al. | May 2006 | B2 |
7060059 | Keith et al. | Jun 2006 | B2 |
7066910 | Bauhahn et al. | Jun 2006 | B2 |
7070580 | Nielsen | Jul 2006 | B2 |
7098803 | Mann et al. | Aug 2006 | B2 |
7104972 | Møller et al. | Sep 2006 | B2 |
7109878 | Mann et al. | Sep 2006 | B2 |
7123964 | Betzold et al. | Oct 2006 | B2 |
7128727 | Flaherty et al. | Oct 2006 | B2 |
7133329 | Skyggebjerg et al. | Nov 2006 | B2 |
7172572 | Diamond et al. | Feb 2007 | B2 |
7179226 | Crothall et al. | Feb 2007 | B2 |
7204823 | Estes et al. | Apr 2007 | B2 |
7220240 | Struys et al. | May 2007 | B2 |
7232423 | Mernoe | Jun 2007 | B2 |
7267665 | Steil et al. | Sep 2007 | B2 |
7291107 | Hellwig et al. | Nov 2007 | B2 |
7354420 | Steil et al. | Apr 2008 | B2 |
7402153 | Steil et al. | Jul 2008 | B2 |
7404796 | Ginsberg | Jul 2008 | B2 |
7429255 | Thompson | Sep 2008 | B2 |
7491187 | Van et al. | Feb 2009 | B2 |
7494481 | Moberg et al. | Feb 2009 | B2 |
7547281 | Hayes et al. | Jun 2009 | B2 |
7553281 | Hellwig | Jun 2009 | B2 |
7569030 | Lebel et al. | Aug 2009 | B2 |
7569050 | Moberg et al. | Aug 2009 | B2 |
7570980 | Ginsberg | Aug 2009 | B2 |
7591801 | Brauker et al. | Sep 2009 | B2 |
7597682 | Moberg | Oct 2009 | B2 |
7641649 | Moberg et al. | Jan 2010 | B2 |
7651845 | Doyle et al. | Jan 2010 | B2 |
7654982 | Carlisle et al. | Feb 2010 | B2 |
7670288 | Sher | Mar 2010 | B2 |
7708717 | Estes et al. | May 2010 | B2 |
7734323 | Blomquist et al. | Jun 2010 | B2 |
7785313 | Mastrototaro | Aug 2010 | B2 |
7789859 | Estes et al. | Sep 2010 | B2 |
7794426 | Briones et al. | Sep 2010 | B2 |
7806853 | Wittmann et al. | Oct 2010 | B2 |
7806854 | Damiano et al. | Oct 2010 | B2 |
7815602 | Mann et al. | Oct 2010 | B2 |
7819843 | Mann et al. | Oct 2010 | B2 |
7828528 | Estes et al. | Nov 2010 | B2 |
7850641 | Lebel et al. | Dec 2010 | B2 |
7875022 | Wenger et al. | Jan 2011 | B2 |
7904061 | Zaffino et al. | Mar 2011 | B1 |
7938797 | Estes | May 2011 | B2 |
7938801 | Hawkins et al. | May 2011 | B2 |
7946985 | Mastrototaro et al. | May 2011 | B2 |
7959598 | Estes | Jun 2011 | B2 |
7967812 | Jasperson et al. | Jun 2011 | B2 |
7976492 | Brauker et al. | Jul 2011 | B2 |
8029459 | Rush et al. | Oct 2011 | B2 |
8062249 | Wilinska et al. | Nov 2011 | B2 |
8105268 | Lebel et al. | Jan 2012 | B2 |
8114023 | Ward et al. | Feb 2012 | B2 |
8152789 | Starkweather et al. | Apr 2012 | B2 |
8206296 | Jennewine | Jun 2012 | B2 |
8206350 | Mann et al. | Jun 2012 | B2 |
8208984 | Blomquist et al. | Jun 2012 | B2 |
8221385 | Estes | Jul 2012 | B2 |
8226556 | Hayes et al. | Jul 2012 | B2 |
8246540 | Ginsberg | Aug 2012 | B2 |
8262616 | Grant et al. | Sep 2012 | B2 |
8267893 | Moberg et al. | Sep 2012 | B2 |
8273052 | Damiano et al. | Sep 2012 | B2 |
D669165 | Sims et al. | Oct 2012 | S |
8282626 | Wenger et al. | Oct 2012 | B2 |
8287487 | Estes | Oct 2012 | B2 |
8318154 | Frost et al. | Nov 2012 | B2 |
8348844 | Kunjan et al. | Jan 2013 | B2 |
8348886 | Kanderian et al. | Jan 2013 | B2 |
8348923 | Kanderian et al. | Jan 2013 | B2 |
8352011 | Van et al. | Jan 2013 | B2 |
8372039 | Mernoe et al. | Feb 2013 | B2 |
8417311 | Rule | Apr 2013 | B2 |
8430847 | Mernoe et al. | Apr 2013 | B2 |
8439834 | Schmelzeisen-Redeker et al. | May 2013 | B2 |
8439897 | Yodfat et al. | May 2013 | B2 |
8454576 | Mastrototaro et al. | Jun 2013 | B2 |
8460231 | Brauker et al. | Jun 2013 | B2 |
8467972 | Rush | Jun 2013 | B2 |
8475409 | Tsoukalis | Jul 2013 | B2 |
8480655 | Jasperson et al. | Jul 2013 | B2 |
8548544 | Kircher et al. | Oct 2013 | B2 |
8548552 | Tsoukalis | Oct 2013 | B2 |
8551045 | Sie et al. | Oct 2013 | B2 |
8560082 | Wei | Oct 2013 | B2 |
8560131 | Haueter et al. | Oct 2013 | B2 |
8562558 | Kamath et al. | Oct 2013 | B2 |
8562587 | Kovatchev et al. | Oct 2013 | B2 |
8568713 | Frost et al. | Oct 2013 | B2 |
8579854 | Budiman et al. | Nov 2013 | B2 |
8579879 | Palerm et al. | Nov 2013 | B2 |
8585591 | Sloan et al. | Nov 2013 | B2 |
8585593 | Kovatchev et al. | Nov 2013 | B2 |
8585637 | Wilinska et al. | Nov 2013 | B2 |
8585638 | Blomquist | Nov 2013 | B2 |
8597274 | Sloan et al. | Dec 2013 | B2 |
8615366 | Galley et al. | Dec 2013 | B2 |
8622988 | Hayter | Jan 2014 | B2 |
8679016 | Mastrototaro et al. | Mar 2014 | B2 |
8679060 | Mernoe et al. | Mar 2014 | B2 |
8690820 | Cinar et al. | Apr 2014 | B2 |
8694115 | Goetz et al. | Apr 2014 | B2 |
8706691 | Mcdaniel et al. | Apr 2014 | B2 |
8718949 | Blomquist et al. | May 2014 | B2 |
8721585 | Brauker et al. | May 2014 | B2 |
8727982 | Jennewine | May 2014 | B2 |
8734422 | Hayter | May 2014 | B2 |
8734428 | Blomquist | May 2014 | B2 |
8747315 | Brauker et al. | Jun 2014 | B2 |
8762070 | Doyle et al. | Jun 2014 | B2 |
8771222 | Kanderian et al. | Jul 2014 | B2 |
8777896 | Starkweather et al. | Jul 2014 | B2 |
8777924 | Kanderian et al. | Jul 2014 | B2 |
8784364 | Kamen et al. | Jul 2014 | B2 |
8784369 | Starkweather et al. | Jul 2014 | B2 |
8784370 | Lebel et al. | Jul 2014 | B2 |
8795224 | Starkweather et al. | Aug 2014 | B2 |
8795252 | Hayter | Aug 2014 | B2 |
8808230 | Rotstein | Aug 2014 | B2 |
8876755 | Taub et al. | Nov 2014 | B2 |
8882741 | Brauker et al. | Nov 2014 | B2 |
8903501 | Perryman | Dec 2014 | B2 |
8919180 | Gottlieb et al. | Dec 2014 | B2 |
8920401 | Brauker et al. | Dec 2014 | B2 |
8926585 | Brauker et al. | Jan 2015 | B2 |
8945094 | Nordh | Feb 2015 | B2 |
8956291 | Valk et al. | Feb 2015 | B2 |
8956321 | Dejournett | Feb 2015 | B2 |
8977504 | Hovorka | Mar 2015 | B2 |
8992475 | Mann et al. | Mar 2015 | B2 |
9034323 | Frost et al. | May 2015 | B2 |
9050413 | Brauker et al. | Jun 2015 | B2 |
9056165 | Steil et al. | Jun 2015 | B2 |
9056168 | Kircher et al. | Jun 2015 | B2 |
9078963 | Estes | Jul 2015 | B2 |
9089305 | Hovorka | Jul 2015 | B2 |
9149233 | Kamath et al. | Oct 2015 | B2 |
9155843 | Brauker et al. | Oct 2015 | B2 |
9247901 | Kamath et al. | Feb 2016 | B2 |
9314566 | Wenger et al. | Apr 2016 | B2 |
9320471 | Hayes et al. | Apr 2016 | B2 |
9333298 | Kim et al. | May 2016 | B2 |
9415157 | Mann et al. | Aug 2016 | B2 |
9474855 | Mccann et al. | Oct 2016 | B2 |
9480796 | Starkweather et al. | Nov 2016 | B2 |
9486172 | Cobelli et al. | Nov 2016 | B2 |
9486578 | Finan et al. | Nov 2016 | B2 |
9561324 | Estes | Feb 2017 | B2 |
9878097 | Estes | Jan 2018 | B2 |
9968729 | Estes | May 2018 | B2 |
10449294 | Estes | Oct 2019 | B1 |
10569015 | Estes | Feb 2020 | B2 |
20010003542 | Kita | Jun 2001 | A1 |
20010041869 | Causey, III et al. | Nov 2001 | A1 |
20010056262 | Cabiri | Dec 2001 | A1 |
20020002326 | Causey, III et al. | Jan 2002 | A1 |
20020004651 | Ljndggreen et al. | Jan 2002 | A1 |
20020007154 | Hansen et al. | Jan 2002 | A1 |
20020013784 | Swanson | Jan 2002 | A1 |
20020016534 | Trepagnier et al. | Feb 2002 | A1 |
20020016568 | Lebel | Feb 2002 | A1 |
20020032402 | Daoud et al. | Mar 2002 | A1 |
20020040208 | Flaherty et al. | Apr 2002 | A1 |
20020046315 | Miller et al. | Apr 2002 | A1 |
20020055845 | Ueda et al. | May 2002 | A1 |
20020072720 | Hague et al. | Jun 2002 | A1 |
20020091358 | Klitmose | Jul 2002 | A1 |
20020107476 | Mann | Aug 2002 | A1 |
20020126036 | Flaherty et al. | Sep 2002 | A1 |
20020156462 | Stultz | Oct 2002 | A1 |
20020164973 | Janik et al. | Nov 2002 | A1 |
20030028089 | Galley | Feb 2003 | A1 |
20030055380 | Flaherty | Mar 2003 | A1 |
20030065308 | Lebel et al. | Apr 2003 | A1 |
20030088238 | Poulsen | May 2003 | A1 |
20030104982 | Wittmann et al. | Jun 2003 | A1 |
20030121055 | Kaminski et al. | Jun 2003 | A1 |
20030125672 | Adair et al. | Jul 2003 | A1 |
20030161744 | Vilks et al. | Aug 2003 | A1 |
20030198558 | Nason et al. | Oct 2003 | A1 |
20030199825 | Flaherty | Oct 2003 | A1 |
20030208113 | Mault et al. | Nov 2003 | A1 |
20030216683 | Shekalim | Nov 2003 | A1 |
20030216686 | Lynch et al. | Nov 2003 | A1 |
20040006316 | Patton | Jan 2004 | A1 |
20040010207 | Flaherty et al. | Jan 2004 | A1 |
20040019325 | Shekalim | Jan 2004 | A1 |
20040064088 | Gorman et al. | Apr 2004 | A1 |
20040064096 | Flaherty et al. | Apr 2004 | A1 |
20040068230 | Estes et al. | Apr 2004 | A1 |
20040078028 | Flaherty et al. | Apr 2004 | A1 |
20040087894 | Flaherty | May 2004 | A1 |
20040092865 | Flaherty et al. | May 2004 | A1 |
20040092878 | Flaherty | May 2004 | A1 |
20040115068 | Hansen et al. | Jun 2004 | A1 |
20040116866 | Gorman et al. | Jun 2004 | A1 |
20040127844 | Flaherty | Jul 2004 | A1 |
20040153032 | Garribotto et al. | Aug 2004 | A1 |
20040158207 | Hunn et al. | Aug 2004 | A1 |
20040167464 | Ireland et al. | Aug 2004 | A1 |
20040171983 | Sparks et al. | Sep 2004 | A1 |
20040176720 | Kipfer | Sep 2004 | A1 |
20040176727 | Shekalim | Sep 2004 | A1 |
20040187952 | Jones | Sep 2004 | A1 |
20040204673 | Flaherty | Oct 2004 | A1 |
20040204744 | Penner et al. | Oct 2004 | A1 |
20040220551 | Flaherty et al. | Nov 2004 | A1 |
20040235446 | Flaherty et al. | Nov 2004 | A1 |
20040260233 | Garibotto et al. | Dec 2004 | A1 |
20050010165 | Hickle | Jan 2005 | A1 |
20050021005 | Flaherty et al. | Jan 2005 | A1 |
20050021104 | DiLorenzo | Jan 2005 | A1 |
20050022274 | Campbell et al. | Jan 2005 | A1 |
20050033223 | Herrera | Feb 2005 | A1 |
20050038332 | Saidara et al. | Feb 2005 | A1 |
20050049179 | Davidson et al. | Mar 2005 | A1 |
20050065760 | Murtfeldt et al. | Mar 2005 | A1 |
20050090808 | Malave et al. | Apr 2005 | A1 |
20050090851 | Devlin | Apr 2005 | A1 |
20050095063 | Fathallah | May 2005 | A1 |
20050101933 | Marrs et al. | May 2005 | A1 |
20050107743 | Fangrow, Jr. | May 2005 | A1 |
20050113745 | Stultz | May 2005 | A1 |
20050124866 | Elaz et al. | Jun 2005 | A1 |
20050137530 | Campbell et al. | Jun 2005 | A1 |
20050160858 | Mernoe | Jul 2005 | A1 |
20050171512 | Flaherty | Aug 2005 | A1 |
20050182366 | Vogt et al. | Aug 2005 | A1 |
20050192561 | Mernoe | Sep 2005 | A1 |
20050203461 | Flaherty et al. | Sep 2005 | A1 |
20050215982 | Malave et al. | Sep 2005 | A1 |
20050222645 | Malave et al. | Oct 2005 | A1 |
20050234404 | Vilks et al. | Oct 2005 | A1 |
20050238507 | DiIanni et al. | Oct 2005 | A1 |
20050240544 | Kil et al. | Oct 2005 | A1 |
20050245878 | Mernoe et al. | Nov 2005 | A1 |
20050251097 | Mernoe | Nov 2005 | A1 |
20050267402 | Stewart et al. | Dec 2005 | A1 |
20050273059 | Mernoe et al. | Dec 2005 | A1 |
20050277890 | Stewart et al. | Dec 2005 | A1 |
20060036214 | Mogensen et al. | Feb 2006 | A1 |
20060041229 | Garibotto et al. | Feb 2006 | A1 |
20060042633 | Bishop et al. | Mar 2006 | A1 |
20060069351 | Safabash et al. | Mar 2006 | A9 |
20060069382 | Pedersen | Mar 2006 | A1 |
20060074381 | Malave et al. | Apr 2006 | A1 |
20060075269 | Liong et al. | Apr 2006 | A1 |
20060095014 | Ethelfeld | May 2006 | A1 |
20060125654 | Liao et al. | Jun 2006 | A1 |
20060135913 | Ethelfeld | Jun 2006 | A1 |
20060142698 | Ethelfeld | Jun 2006 | A1 |
20060151545 | Imhof et al. | Jul 2006 | A1 |
20060173406 | Hayes | Aug 2006 | A1 |
20060173410 | Moberg et al. | Aug 2006 | A1 |
20060178633 | Garibotto et al. | Aug 2006 | A1 |
20060184104 | Cheney et al. | Aug 2006 | A1 |
20060184119 | Remde et al. | Aug 2006 | A1 |
20060200073 | Radmer et al. | Sep 2006 | A1 |
20060206054 | Shekalim | Sep 2006 | A1 |
20060214511 | Dayan | Sep 2006 | A1 |
20060247574 | Maule et al. | Nov 2006 | A1 |
20060247581 | Pedersen et al. | Nov 2006 | A1 |
20060253086 | Moberg et al. | Nov 2006 | A1 |
20060258973 | Volt | Nov 2006 | A1 |
20060258976 | Shturman et al. | Nov 2006 | A1 |
20060264835 | Nielsen et al. | Nov 2006 | A1 |
20060264890 | Moberg et al. | Nov 2006 | A1 |
20060264894 | Moberg et al. | Nov 2006 | A1 |
20070016127 | Staib et al. | Jan 2007 | A1 |
20070060870 | Tolle et al. | Mar 2007 | A1 |
20070073228 | Mernoe et al. | Mar 2007 | A1 |
20070073235 | Estes | Mar 2007 | A1 |
20070073236 | Mernoe et al. | Mar 2007 | A1 |
20070078818 | Zivitz et al. | Apr 2007 | A1 |
20070079836 | Reghabi et al. | Apr 2007 | A1 |
20070088271 | Richards | Apr 2007 | A1 |
20070093750 | Jan et al. | Apr 2007 | A1 |
20070093786 | Goldsmith et al. | Apr 2007 | A1 |
20070100222 | Mastrototaro et al. | May 2007 | A1 |
20070106218 | Yodfat et al. | May 2007 | A1 |
20070112298 | Mueller et al. | May 2007 | A1 |
20070118364 | Wise | May 2007 | A1 |
20070118405 | Campbell et al. | May 2007 | A1 |
20070123819 | Mernoe et al. | May 2007 | A1 |
20070124002 | Estes et al. | May 2007 | A1 |
20070142776 | Kovelman et al. | Jun 2007 | A9 |
20070155307 | Ng et al. | Jul 2007 | A1 |
20070156092 | Estes et al. | Jul 2007 | A1 |
20070156094 | Safabash et al. | Jul 2007 | A1 |
20070166170 | Nason et al. | Jul 2007 | A1 |
20070167905 | Estes et al. | Jul 2007 | A1 |
20070167912 | Causey et al. | Jul 2007 | A1 |
20070169607 | Keller et al. | Jul 2007 | A1 |
20070173761 | Kanderian et al. | Jul 2007 | A1 |
20070179444 | Causey et al. | Aug 2007 | A1 |
20070191702 | Yodfat et al. | Aug 2007 | A1 |
20070219432 | Thompson | Sep 2007 | A1 |
20070219480 | Kamen et al. | Sep 2007 | A1 |
20070233521 | Wehba et al. | Oct 2007 | A1 |
20070239116 | Follman et al. | Oct 2007 | A1 |
20070248238 | Abreu | Oct 2007 | A1 |
20070252774 | Qi et al. | Nov 2007 | A1 |
20070255250 | Moberg et al. | Nov 2007 | A1 |
20070282299 | Hellwig | Dec 2007 | A1 |
20070287931 | Dilorenzo | Dec 2007 | A1 |
20080009824 | Moberg et al. | Jan 2008 | A1 |
20080015422 | Wessel | Jan 2008 | A1 |
20080027574 | Thomas | Jan 2008 | A1 |
20080031481 | Warren et al. | Feb 2008 | A1 |
20080045891 | Maule et al. | Feb 2008 | A1 |
20080051697 | Mounce et al. | Feb 2008 | A1 |
20080051698 | Mounce et al. | Feb 2008 | A1 |
20080051714 | Moberg et al. | Feb 2008 | A1 |
20080051716 | Stutz | Feb 2008 | A1 |
20080051730 | Bikovsky | Feb 2008 | A1 |
20080051738 | Griffin | Feb 2008 | A1 |
20080077081 | Mounce et al. | Mar 2008 | A1 |
20080097326 | Moberg et al. | Apr 2008 | A1 |
20080097375 | Bikovsky | Apr 2008 | A1 |
20080097381 | Moberg et al. | Apr 2008 | A1 |
20080103022 | Dvorak et al. | May 2008 | A1 |
20080109050 | John | May 2008 | A1 |
20080125700 | Moberg et al. | May 2008 | A1 |
20080125701 | Moberg et al. | May 2008 | A1 |
20080129535 | Thompson et al. | Jun 2008 | A1 |
20080172027 | Blomquist | Jul 2008 | A1 |
20080177149 | Weinert | Jul 2008 | A1 |
20080183060 | Steil et al. | Jul 2008 | A1 |
20080188796 | Steil et al. | Aug 2008 | A1 |
20080198012 | Kamen | Aug 2008 | A1 |
20080201325 | Doniger et al. | Aug 2008 | A1 |
20080208627 | Skyggebjerg | Aug 2008 | A1 |
20080215035 | Yodfat et al. | Sep 2008 | A1 |
20080234630 | Iddan et al. | Sep 2008 | A1 |
20080255516 | Yodfat et al. | Oct 2008 | A1 |
20080269683 | Bikovsky | Oct 2008 | A1 |
20080269687 | Chong et al. | Oct 2008 | A1 |
20080269714 | Mastrototaro et al. | Oct 2008 | A1 |
20080269723 | Mastrototaro et al. | Oct 2008 | A1 |
20080294094 | Mhatre et al. | Nov 2008 | A1 |
20080294109 | Estes et al. | Nov 2008 | A1 |
20080294142 | Patel et al. | Nov 2008 | A1 |
20080300572 | Rankers et al. | Dec 2008 | A1 |
20080306434 | Dobbles et al. | Dec 2008 | A1 |
20080306444 | Brister et al. | Dec 2008 | A1 |
20080312512 | Brukalo et al. | Dec 2008 | A1 |
20080312634 | Helmerson et al. | Dec 2008 | A1 |
20080319381 | Yodfat et al. | Dec 2008 | A1 |
20080319383 | Byland et al. | Dec 2008 | A1 |
20080319384 | Yodfat et al. | Dec 2008 | A1 |
20080319394 | Yodfat et al. | Dec 2008 | A1 |
20080319414 | Yodfat et al. | Dec 2008 | A1 |
20080319416 | Yodfat et al. | Dec 2008 | A1 |
20090036760 | Hayter | Feb 2009 | A1 |
20090036870 | Mounce et al. | Feb 2009 | A1 |
20090043291 | Thompson | Feb 2009 | A1 |
20090048584 | Thompson | Feb 2009 | A1 |
20090069745 | Estes et al. | Mar 2009 | A1 |
20090069746 | Miller et al. | Mar 2009 | A1 |
20090069749 | Miller et al. | Mar 2009 | A1 |
20090069784 | Estes et al. | Mar 2009 | A1 |
20090069785 | Miller et al. | Mar 2009 | A1 |
20090069787 | Estes et al. | Mar 2009 | A1 |
20090076453 | Mejlhede et al. | Mar 2009 | A1 |
20090076849 | Diller | Mar 2009 | A1 |
20090082728 | Bikovsky | Mar 2009 | A1 |
20090093756 | Minaie et al. | Apr 2009 | A1 |
20090099507 | Koops | Apr 2009 | A1 |
20090105636 | Hayter et al. | Apr 2009 | A1 |
20090112333 | Sahai | Apr 2009 | A1 |
20090118664 | Estes et al. | May 2009 | A1 |
20090143916 | Boll et al. | Jun 2009 | A1 |
20090164190 | Hayter | Jun 2009 | A1 |
20090177142 | Blomquist et al. | Jul 2009 | A1 |
20090177154 | Blomquist | Jul 2009 | A1 |
20090192722 | Shariati et al. | Jul 2009 | A1 |
20090198191 | Chong et al. | Aug 2009 | A1 |
20090198215 | Chong et al. | Aug 2009 | A1 |
20090221890 | Saffer et al. | Sep 2009 | A1 |
20100010330 | Rankers et al. | Jan 2010 | A1 |
20100049164 | Estes | Feb 2010 | A1 |
20100057040 | Hayter | Mar 2010 | A1 |
20100057041 | Hayter | Mar 2010 | A1 |
20100094078 | Weston | Apr 2010 | A1 |
20100121167 | Mcgarraugh | May 2010 | A1 |
20100165795 | Elder et al. | Jul 2010 | A1 |
20100168538 | Keenan et al. | Jul 2010 | A1 |
20100168820 | Maniak et al. | Jul 2010 | A1 |
20100174229 | Hsu | Jul 2010 | A1 |
20100174266 | Estes | Jul 2010 | A1 |
20100179409 | Kamath et al. | Jul 2010 | A1 |
20100211005 | Edwards et al. | Aug 2010 | A1 |
20100249530 | Rankers | Sep 2010 | A1 |
20100273738 | Valcke et al. | Oct 2010 | A1 |
20100286601 | Yodfat et al. | Nov 2010 | A1 |
20100286653 | Kubel et al. | Nov 2010 | A1 |
20100298685 | Hayter | Nov 2010 | A1 |
20100298765 | Budiman et al. | Nov 2010 | A1 |
20100324382 | Cantwell et al. | Dec 2010 | A1 |
20100324977 | Dragt | Dec 2010 | A1 |
20100325864 | Briones et al. | Dec 2010 | A1 |
20110009813 | Rankers | Jan 2011 | A1 |
20110015511 | Bousamra et al. | Jan 2011 | A1 |
20110040247 | Mandro et al. | Feb 2011 | A1 |
20110071464 | Palerm | Mar 2011 | A1 |
20110098637 | Hill | Apr 2011 | A1 |
20110098674 | Vicente et al. | Apr 2011 | A1 |
20110105955 | Yudovsky et al. | May 2011 | A1 |
20110106050 | Yodfat et al. | May 2011 | A1 |
20110118699 | Yodfat et al. | May 2011 | A1 |
20110124996 | Reinke | May 2011 | A1 |
20110130716 | Estes et al. | Jun 2011 | A1 |
20110163880 | Halff et al. | Jul 2011 | A1 |
20110199194 | Waldock et al. | Aug 2011 | A1 |
20110224523 | Budiman | Sep 2011 | A1 |
20110313390 | Roy et al. | Dec 2011 | A1 |
20110319813 | Kamen et al. | Dec 2011 | A1 |
20120016304 | Patel et al. | Jan 2012 | A1 |
20120029468 | Diperna et al. | Feb 2012 | A1 |
20120046606 | Arefieg | Feb 2012 | A1 |
20120065894 | Tubb et al. | Mar 2012 | A1 |
20120078067 | Kovatchev et al. | Mar 2012 | A1 |
20120123234 | Atlas et al. | May 2012 | A1 |
20120150556 | Galasso et al. | Jun 2012 | A1 |
20120172694 | Desborough et al. | Jul 2012 | A1 |
20120172802 | Blomquist | Jul 2012 | A1 |
20120185267 | Kamen et al. | Jul 2012 | A1 |
20120197207 | Stefanski | Aug 2012 | A1 |
20120203467 | Kamath et al. | Aug 2012 | A1 |
20120209208 | Stefanski | Aug 2012 | A1 |
20120227737 | Mastrototaro et al. | Sep 2012 | A1 |
20120238851 | Kamen et al. | Sep 2012 | A1 |
20120238853 | Arefieg | Sep 2012 | A1 |
20120238999 | Estes et al. | Sep 2012 | A1 |
20120245448 | Shariati et al. | Sep 2012 | A1 |
20120245556 | Kovatchev et al. | Sep 2012 | A1 |
20120245855 | Kamath et al. | Sep 2012 | A1 |
20120246106 | Atlas et al. | Sep 2012 | A1 |
20120259191 | Shariati et al. | Oct 2012 | A1 |
20120277723 | Skladnev et al. | Nov 2012 | A1 |
20120283694 | Yodfat et al. | Nov 2012 | A1 |
20120289931 | Robinson et al. | Nov 2012 | A1 |
20120302991 | Blomquist et al. | Nov 2012 | A1 |
20120323590 | Udani | Dec 2012 | A1 |
20120330270 | Colton | Dec 2012 | A1 |
20130046281 | Javitt | Feb 2013 | A1 |
20130053818 | Estes | Feb 2013 | A1 |
20130053819 | Estes | Feb 2013 | A1 |
20130053820 | Estes et al. | Feb 2013 | A1 |
20130102867 | Desborough et al. | Apr 2013 | A1 |
20130116649 | Breton et al. | May 2013 | A1 |
20130138205 | Kushwaha et al. | May 2013 | A1 |
20130159456 | Daoud et al. | Jun 2013 | A1 |
20130165041 | Bukovjan et al. | Jun 2013 | A1 |
20130204186 | Moore et al. | Aug 2013 | A1 |
20130204202 | Trombly et al. | Aug 2013 | A1 |
20130218126 | Hayter et al. | Aug 2013 | A1 |
20130237932 | Thueer et al. | Sep 2013 | A1 |
20130245563 | Mercer et al. | Sep 2013 | A1 |
20130245604 | Kouyoumjian et al. | Sep 2013 | A1 |
20130253418 | Kamath et al. | Sep 2013 | A1 |
20130275139 | Coleman | Oct 2013 | A1 |
20130281965 | Kamen et al. | Oct 2013 | A1 |
20130297334 | Galasso et al. | Nov 2013 | A1 |
20130338629 | Agrawal et al. | Dec 2013 | A1 |
20130338630 | Agrawal et al. | Dec 2013 | A1 |
20130345663 | Agrawal et al. | Dec 2013 | A1 |
20140005633 | Finan | Jan 2014 | A1 |
20140025015 | Cross et al. | Jan 2014 | A1 |
20140031759 | Kouyoumjian et al. | Jan 2014 | A1 |
20140039383 | Dobbles et al. | Feb 2014 | A1 |
20140052091 | Dobbles et al. | Feb 2014 | A1 |
20140052092 | Dobbles et al. | Feb 2014 | A1 |
20140052093 | Dobbles et al. | Feb 2014 | A1 |
20140052094 | Dobbles et al. | Feb 2014 | A1 |
20140052095 | Dobbles et al. | Feb 2014 | A1 |
20140066884 | Keenan et al. | Mar 2014 | A1 |
20140066885 | Keenan et al. | Mar 2014 | A1 |
20140066886 | Roy et al. | Mar 2014 | A1 |
20140066887 | Mastrototaro et al. | Mar 2014 | A1 |
20140066888 | Parikh et al. | Mar 2014 | A1 |
20140066889 | Grosman et al. | Mar 2014 | A1 |
20140066890 | Sloan et al. | Mar 2014 | A1 |
20140066892 | Keenan et al. | Mar 2014 | A1 |
20140088557 | Mernoe et al. | Mar 2014 | A1 |
20140094766 | Estes et al. | Apr 2014 | A1 |
20140107607 | Estes | Apr 2014 | A1 |
20140114278 | Dobbles et al. | Apr 2014 | A1 |
20140121635 | Hayter | May 2014 | A1 |
20140128705 | Mazlish | May 2014 | A1 |
20140128803 | Dobbles et al. | May 2014 | A1 |
20140180203 | Budiman et al. | Jun 2014 | A1 |
20140180240 | Finan et al. | Jun 2014 | A1 |
20140228627 | Soffer et al. | Aug 2014 | A1 |
20140228668 | Wakizaka et al. | Aug 2014 | A1 |
20140235981 | Hayter | Aug 2014 | A1 |
20140249500 | Estes | Sep 2014 | A1 |
20140276553 | Rosinko et al. | Sep 2014 | A1 |
20140276554 | Finan et al. | Sep 2014 | A1 |
20140276555 | Morales | Sep 2014 | A1 |
20140276583 | Chen et al. | Sep 2014 | A1 |
20140309615 | Mazlish | Oct 2014 | A1 |
20140323959 | Lebel et al. | Oct 2014 | A1 |
20150018633 | Kovachev et al. | Jan 2015 | A1 |
20150025471 | Enggaard | Jan 2015 | A1 |
20150025495 | Peyser | Jan 2015 | A1 |
20150030641 | Anderson et al. | Jan 2015 | A1 |
20150045737 | Stefanski | Feb 2015 | A1 |
20150073337 | Saint et al. | Mar 2015 | A1 |
20150080789 | Estes et al. | Mar 2015 | A1 |
20150120323 | Galasso et al. | Apr 2015 | A1 |
20150136336 | Huang | May 2015 | A1 |
20150148774 | Yao | May 2015 | A1 |
20150157794 | Roy et al. | Jun 2015 | A1 |
20150164414 | Matthews | Jun 2015 | A1 |
20150165119 | Palerm et al. | Jun 2015 | A1 |
20150217051 | Mastrototaro et al. | Aug 2015 | A1 |
20150217052 | Keenan et al. | Aug 2015 | A1 |
20150265767 | Vazquez et al. | Sep 2015 | A1 |
20150265768 | Vazquez et al. | Sep 2015 | A1 |
20150314062 | Blomquist et al. | Nov 2015 | A1 |
20150320933 | Estes | Nov 2015 | A1 |
20150328402 | Nogueira et al. | Nov 2015 | A1 |
20150351683 | Brauker et al. | Dec 2015 | A1 |
20150352282 | Mazlish | Dec 2015 | A1 |
20150352283 | Galasso | Dec 2015 | A1 |
20160000998 | Estes | Jan 2016 | A1 |
20160030669 | Harris et al. | Feb 2016 | A1 |
20160038673 | Morales | Feb 2016 | A1 |
20160082187 | Schaible et al. | Mar 2016 | A1 |
20160082188 | Blomquist et al. | Mar 2016 | A1 |
20160158438 | Monirabbasi et al. | Jun 2016 | A1 |
20160162662 | Monirabbasi et al. | Jun 2016 | A1 |
20160213841 | Geismar et al. | Jul 2016 | A1 |
20160256629 | Grosman et al. | Sep 2016 | A1 |
20170182248 | Rosinko | Jun 2017 | A1 |
20170189614 | Mazlish et al. | Jul 2017 | A1 |
20170203036 | Mazlish et al. | Jul 2017 | A1 |
Number | Date | Country |
---|---|---|
2543545 | May 2005 | CA |
19627619 | Jan 1998 | DE |
10236669 | Feb 2004 | DE |
PA200401893 | Dec 2004 | DK |
PA 200401893 | Dec 2004 | DK |
0062974 | Oct 1982 | EP |
0275213 | Jul 1988 | EP |
0496141 | Jul 1992 | EP |
0612004 | Aug 1994 | EP |
0580723 | Oct 1995 | EP |
1045146 | Oct 2000 | EP |
1136698 | Sep 2001 | EP |
1177802 | Feb 2002 | EP |
0721358 | May 2002 | EP |
1495775 | Jan 2005 | EP |
1527792 | May 2005 | EP |
1754498 | Feb 2007 | EP |
1818664 | Aug 2007 | EP |
2585252 | Jan 1987 | FR |
0747701 | Apr 1956 | GB |
2218831 | Nov 1989 | GB |
H09504974 | May 1997 | JP |
2000513974 | Oct 2000 | JP |
2002-085556 | Mar 2002 | JP |
2002507459 | Mar 2002 | JP |
2002523149 | Jul 2002 | JP |
2003-531691 | Oct 2003 | JP |
2010-502361 | Jan 2010 | JP |
2010-524639 | Jul 2010 | JP |
WO 1990015928 | Dec 1990 | WO |
WO 1997021457 | Jun 1997 | WO |
WO 1998004301 | Feb 1998 | WO |
WO 1998011927 | Mar 1998 | WO |
WO 1998057683 | Dec 1998 | WO |
WO 1999021596 | May 1999 | WO |
WO 1999039118 | Aug 1999 | WO |
WO 1999048546 | Sep 1999 | WO |
WO 2001054753 | Aug 2001 | WO |
WO 2001072360 | Oct 2001 | WO |
WO 2001091822 | Dec 2001 | WO |
WO 2001091833 | Dec 2001 | WO |
WO 2002040083 | May 2002 | WO |
WO 2002057627 | Jul 2002 | WO |
WO 2002068015 | Sep 2002 | WO |
WO 2002084336 | Oct 2002 | WO |
WO 2002100469 | Dec 2002 | WO |
WO 2003026726 | Apr 2003 | WO |
WO 2003103763 | Dec 2003 | WO |
WO 2004056412 | Jul 2004 | WO |
WO 2004110526 | Dec 2004 | WO |
WO 2005002652 | Jan 2005 | WO |
WO 2005039673 | May 2005 | WO |
WO 2005072794 | Aug 2005 | WO |
WO 2005072795 | Aug 2005 | WO |
WO 2006067217 | Jun 2006 | WO |
WO 2006097453 | Sep 2006 | WO |
WO 2006105792 | Oct 2006 | WO |
WO 2006105793 | Oct 2006 | WO |
WO 2006105794 | Oct 2006 | WO |
WO 2007141786 | Dec 2007 | WO |
2008134146 | Nov 2008 | WO |
Entry |
---|
Accu-Chek Spirit, “Pump Therapy Made for You,” Roche, 2006, 6 pages. |
Asante Solutions Pearl User Manual, Asante Inc., 2012, 180 pages. |
Collins and Lee, “Microfluidic flow transducer based on the measurement of electrical admittance,” Lab Chip, 2004 4 pages. |
Debiotech News Release, “Debiotech reveals its new miniaturized Disposable Insulin Nanopump™ for Diabetes therapy,” available at http://www.debiotech.com/news/nw_159.html Apr. 24, 2006, 3 pages. |
International Search Report and Written Opinion in International Application No. PCT/US2014/67665, dated Apr. 21, 2015, 13 pages. |
Medtronic News Release, “Medtronic Receives FDA Approval for World's First Insulin Pump with Real-time Continuous Glucose Monitoring,” Apr. 13, 2006, 3 pages. |
OmniPod Insulin Management System—Investor Relations—Press Release, Feb. 1, 2005, http://investors.insulet.com/phoenix.zhtml?c=209336&p=irol-newsArticle&ID=988708&highlight= 1 page. |
OmniPod Quick Start Guide, 2007, 2 pages. |
Patent Abstracts of Japan, vol. 1999, No. 04, and JP 11 010036, Apr. 30, 1999 and Jan. 19, 1999, Toray Ind. Inc. |
The Medtronic Diabetes Connection, 2006, 6 pages. |
Xilas Temp Touch, “The latest in high-tech and convenient devices,” DOCNews, vol. 2, No. 7, Jul. 1, 2005, http://docnews.diabetesjournals.ord/cgi/content/full/2/7/13, 3 pages. |
Number | Date | Country | |
---|---|---|---|
20200147305 A1 | May 2020 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 14094185 | Dec 2013 | US |
Child | 16743332 | US |