Conventional surgical implants come in a variety of discrete sizes, but generally lack patient-specific sizing or other customization. Some surgical implants include patient-specific surfaces, but such surfaces may be based on incomplete patient anatomy data. Generally speaking, existing surgical implant production and analysis systems lack the ability to efficiently and dynamically evaluate a compilation of an entire complex data model to design and build a personalized reference model. Thus, existing systems and processes may produce inferior surgical devices with inaccurate surface structures based on these less than ideal personalized reference models.
Therefore, there is an unmet and unsolved need for patient-specific devices with higher accuracy patient-specific surfaces (and other features) and systems and processes for producing such patient-specific devices.
Briefly described, and according to one embodiment, aspects of the present disclosure generally relate to patient-specific implants and systems and processes for producing the same. In at least one embodiment, a patient-specific device can include an implant, instrument, surgical guide, cut guide, tools, or any other patient, consumer, or animal-specific use-case. In various embodiments, the patient-specific device may be a 3D printed medical device. In some embodiments, the patient-specific device can include one or more surfaces, features, and other parameters that are customized and tailored for a patient's anatomy to improve mechanical performance and surgical outcomes. In at least this way, aspects of the present disclosure provide an implant device, a cut guide, or similar medical device, that is contoured to accurately align with the structures and surface textures of a patient's existing anatomy to provide better fit and function of the surgical implant or device. In some embodiments, the patient-specific device can include at least one porous area comprising a lattice structure configured to promote osseointegration. In some embodiments, the patient-specific device can include a cut guide with at least one textured area comprising a lattice structure configured to increase friction and improve surgical outcomes.
The patient-specific devices discussed herein may include at least one patient-specific, artificial intelligence derived surface. In at least one embodiment, patient-specific devices discussed herein include a patient-specific surface for mating with a portion of the patient's bone anatomy and a non-patient specific surface for articulation, mating with another portion of the patient's anatomy, etc. In one or more embodiments, the patient-specific device includes only patient-specific (e.g., custom to the patient) surfaces.
According to particular embodiments, a patient-specific device (e.g., implant, instrument, trial, surgical guide, etc.) may be designed and manufactured via one or more systems and processes that leverage advanced surgical planning, implant design, and additive manufacturing/3D-printing. In various embodiments, the disclosed processes and systems automatically or semi-automatically design patient-specific devices based on input/received data. Such as, for example, leveraging artificial intelligence techniques to design a patient anatomy model based on received 2D patient images and then using the patient anatomy model to generate, design, or produce one or more surfaces on a patient-specific device that corresponds to the anatomy of the patient anatomy model.
According to a first aspect, the present disclosure includes a medical device production process comprising: populating a database with: a plurality of 2D image packages, each of the plurality of 2D image packages comprising a series of 2D image files of patient-specific anatomy; a plurality of patient anatomical representations, each of the plurality of patient anatomical representations associated with at least one of the plurality of 2D image packages; a plurality of patient-specific device files, each of the plurality of patient-specific device files associated with at least one of the plurality of patient anatomical representations; receiving a 2D image package associated with a particular patient and including a series of 2D images files associated with an anatomical feature of the particular patient; extracting one or more aspects of the anatomical feature of the particular patient from the 2D image package; processing the one or more aspects of the anatomical feature via an artificial intelligence model to create a non-rigid shape reference based upon the plurality of patient anatomical representations and patient-specific device files; and generating a digital representation of at least one patient-specific device based on the non-rigid shape reference of the anatomical feature of the particular patient and the database.
In a second aspect of the process of the first aspect or any other aspect, wherein the patient-specific device comprises a patient-specific implant.
In a third aspect of the process of the first aspect or any other aspect, wherein the patient-specific device comprises a patient-specific instrument.
In a fourth aspect of the process of the first aspect or any other aspect, wherein the artificial intelligence model is a particular artificial intelligence model; and generating the digital representation of the at least one patient-specific device is based on: the 3D digital representation of the anatomical feature of the particular patient; the plurality of patient-specific device files; and processing the one or more aspects of the anatomical feature via a specific artificial intelligence model.
In a fifth aspect of the process of the fourth aspect or any other aspect, wherein the specific artificial intelligence model outputs a patient-specific device file based on a compilation of the plurality of patient-specific device files.
In a sixth aspect of the process of the fifth aspect or any other aspect, wherein generating the digital representation of the at least one patient-specific device comprises processing the one or more aspects of the anatomical feature, the 3D digital representation of the anatomical feature of the particular patient, and the plurality of patient-specific device files via the artificial intelligence.
In a seventh aspect of the process of the fourth aspect or any other aspect, wherein the specific artificial intelligence model is the particular artificial intelligence model.
According to an eighth aspect, a process for training an artificial intelligence model comprising: receiving a 2D image file package comprising patient anatomy data; extracting one or more aspects from the patient anatomy data; processing the one or more aspects of the patient anatomy data via an artificial intelligence model to generate a 3D digital representation of at least a portion of the patient anatomy data; comparing the 3D digital representation to a known 3D digital representation; updating one or more emphasis guidelines of the artificial intelligence model based on the comparing step; and upon a determination that at least a portion of the 3D digital representation and at least a portion of the known 3D digital representation are within a preconfigured threshold, saving the updated one or more emphasis guidelines.
In a ninth aspect of the process of the eighth aspect or any other aspect, further comprising: receiving a second 2D image file package comprising second patient anatomy data; extracting one or more second aspects from the second patient anatomy data; processing the one or more second aspects from the second patient anatomy data via the artificial intelligence model using the updated one or more emphasis guidelines, and generating a second 3D digital representation of at least a portion of the second patient anatomy data.
In a tenth aspect of the process of the ninth aspect or any other aspect, further comprising generating a CAD file for at least one patient-specific implant based on the 3D digital representation.
In an eleventh aspect of the process of the ninth aspect or any other aspect, further comprising generating a CAD file for at least one patient-specific instrument based on the 3D digital representation.
In a twelfth aspect of the process of the eighth aspect or any other aspect, further comprising: generating a CAD file for at least one patient-specific device; wherein the artificial intelligence model is a particular artificial intelligence model; and generating the CAD file for the at least one patient-specific device is based on: the 3D digital representation of the at least one aspects of the patient anatomy data; the 2D image file package; and processing the one or more aspects of the patient anatomy data via a specific artificial intelligence model.
In a thirteenth aspect of the process of the twelfth aspect or any other aspect, wherein the specific artificial intelligence model outputs a patient-specific device file based on a compilation of a plurality of patient-specific device files included in the 2D image file package. In a fourteenth aspect of the process of the twelfth aspect or any other aspect, wherein generating the CAD file for the at least one patient-specific device comprises processing the at least one aspect of the patient anatomy data, the 3D digital representation of at least a portion of the patient anatomy data, and the 2D image file package via the artificial intelligence model.
According to a fifteenth aspect, a process for training an artificial intelligence model comprising: receiving a DICOM file comprising patient anatomy data; extracting one or more aspects from the patient anatomy data; processing the one or more aspects of the patient anatomy data via an artificial intelligence model to generate a 3D device model for replacing a portion of the patient anatomy; comparing a portion of the 3D device model of a portion of a known 3D device model; updating one or more emphasis guidelines of the artificial intelligence model based on the comparing step, the one or more emphasis guidelines associated with one or more of a 3D digital representation of healthy anatomy, an device model, and previously stored patient anatomy; and upon a determination that the portion of the 3D device model and the portion of the known 3D device model is within a preconfigured threshold, saving the updated one or more emphasis guidelines.
In a sixteenth aspect of the process of the fifteenth aspect or any other aspect, further comprising: receiving a second DICOM file package comprising second patient anatomy data; extracting one or more second aspects from the second patient anatomy data; processing the one or more second aspects from the second patient anatomy data via the artificial intelligence model using the updated one or more emphasis guidelines; and generating a second 3D device model of at least a portion of the second patient anatomy data.
In a seventeenth aspect of the process of the fifteenth aspect or any other aspect, further comprising: generating a CAD file for at least one patient-specific device; wherein the artificial intelligence model is a particular artificial intelligence model; and generating the CAD file for the at least one patient-specific device is based on: the 3D digital representation of the at least one or more aspects of the patient anatomy data; the DICOM file; and processing the one or more aspects of the patient anatomy data via a specific artificial intelligence model.
In an eighteenth aspect of the process of the seventeenth aspect or any other aspect, wherein the patient-specific device comprises a patient-specific implant.
In a nineteenth aspect of the process of the seventeenth aspect or any other aspect, wherein the patient-specific device comprises a patient-specific instrument.
In a twentieth aspect of the process of the seventeenth aspect or any other aspect, wherein generating the CAD file for the at least one patient-specific device comprises processing the at least one or more aspects of the patient anatomy data, the 3D digital representation of at least a portion of the patient anatomy data, and the DICOM file via the artificial intelligence model.
According to a twenty-first aspect, a medical device comprising: a 3D-printed body, the 3D-printed body defining at least one opening for receiving a fixation device and comprising: a metallic material, a patient-specific surface derived from a non-rigid model representing patient anatomy, the non-rigid model produced at least in part by an artificial intelligence model based on one or more feature extracted from a 2D image package comprising images of the patient anatomy, and at least one porous area comprising a lattice structure and configured to promote osseointegration with the at least one porous area.
In a twenty-second aspect of the medical device of the twenty-first aspect or any other aspect, wherein the non-rigid model comprises one or more points excluded from the one or more features.
In a twenty-third aspect of the medical device of the twenty-second aspect or any other aspect, wherein the one or more points are excluded from the 2D image package.
In a twenty-fourth aspect of the medical device of the twenty-third aspect or any other aspect, wherein the one or more points are produced by the artificial intelligence model.
In a twenty-fifth aspect of the medical device of the twenty-fourth aspect or any other aspect, wherein the patient-specific surface includes a surface feature mirroring the one or more points.
In a twenty-sixth aspect of the medical device of the twenty-fifth aspect or any other aspect, wherein the at least one porous area comprises the patient-specific surface.
In a twenty-seventh aspect of the medical device of the twenty-sixth aspect or any other aspect, wherein the 3D-printed body comprises at least one non-porous area.
According to a twenty-eighth aspect, an orthopedic medical device comprising: a 3D-printed body, the 3D-printed body comprising: a metallic material; a patient-specific surface derived from a non-rigid model representing patient anatomy, the non-rigid model produced at least in part by an artificial intelligence model based on one or more features extracted from a 2D image package comprising images of the patient anatomy; and at least one porous area comprising a lattice structure and configured to promote tissue integration.
These and other aspects, features, and benefits of the systems and processes described herein will become apparent from the following detailed written description taken in conjunction with the following drawings, although variations and modifications thereto may be affected without departing from the spirit and scope of the novel concepts of the disclosure.
The accompanying drawings illustrate one or more embodiments and/or aspects of the disclosure and, together with the written description, serve to explain the principles of the disclosure. Wherever possible, the same reference numbers are used throughout the drawings to refer to the same or like elements of an embodiment, and wherein:
While the disclosure is susceptible to various modifications and alternative forms, specific embodiments thereof are shown by way of example in the drawings and will herein be described in detail. It should be understood, however, that the drawings and detailed description presented herein are not intended to limit the disclosure to the particular embodiment disclosed, but on the contrary, the intention is to cover all modifications, equivalents, and alternatives falling within the spirit and scope of the present disclosure.
To promote an understanding of the principles of the present disclosure, reference will now be made to the embodiments illustrated in the drawings and specific language will be used to describe the same. It will, nevertheless, be understood that no limitation of the scope of the disclosure is thereby intended; any alterations and further modifications of the described or illustrated embodiments, and any further applications of the principles of the disclosure as illustrated therein are contemplated as would normally occur to one skilled in the art to which the disclosure relates. All limitations of scope should be determined by and as expressed in the claims.
Whether a term is capitalized is not considered definitive or limiting of the meaning of a term. As used in this document, a capitalized term shall have the same meaning as an uncapitalized term, unless the context of the usage specifically indicates that a more restrictive meaning for the capitalized term is intended. However, the capitalization or lack thereof within the remainder of this document is not intended to be necessarily limiting unless the context clearly indicates that such limitation is intended.
In various embodiments, aspects of the present disclosure generally relate to systems and processes for creating and generating a personalized surgical implant using a plurality of patient files and associated data elements. The system may leverage a plurality of advanced data transformation processes to produce personalized surgical devices with one or more patient-specific surfaces using a dynamic and scalable analytics system to promote osseointegration and/or cut guide friction and improve surgical outcomes for a particular patient.
In some embodiments, the system is a platform for extracting and evaluating patient-specific anatomical features from a plurality of patient data files and associated data elements, and other parameters to create an advanced customized device generation process. For example, in at least one embodiment, the system leverages a patient-specific surface design to create a surgical implant device with a customized porous architecture designed for osseointegration, a fully interconnected lattice structure, and a maximized surface area to provide a patient-specific implant surface. In another non-limiting embodiment, the system leverages a patient-specific design to create a contoured surgical cut guide device with a customized shape and texture for increased friction to provide a patient-specific cut guide surface to improve surgical procedures for both the provider and the patient.
For the purposes of example and explanation of the fundamental processes and components of the disclosed systems and processes, reference is made to the figures. In general, the figures illustrate patient-specific devices produced by processes discussed herein and exemplary systems and processes for producing such devices.
Exemplary Patient-Specific Devices
As will be understood from discussions herein, exemplary implant 100 may be produced from patient images and information, which may be saved in memory. It will be appreciated that
Exemplary Patient-Specific Device Production Systems and Processes
In at least one embodiment, the system may process the patient population data 715 to extract anatomy data 720. In some embodiments, the anatomy data 720 can include data elements and information associated with specific patients, specific patient characteristics (e.g., age, medical condition, etc.), specific anatomical features (e.g., body parts), specific anatomy characteristics (e.g., diseased, healthy, etc.), and other parameters. As will be understood, the anatomy data 720 may include one or more features extracted from the patient population data 715 and/or the patient-specific imagine files 725.
The system, in at least one embodiment, may construct and/or train an initial artificial intelligence model 735 based on the received patient population data 715, including, but not limited to, image data and/or extracted feature information from memory 730 (see
The system can train a particular artificial intelligence model based on the anatomical representations of the image files of existing patient population 715 and anatomy data 720 to generate a non-rigid shape reference to create a 3D digital representation 740 within preconfigured tolerances. The artificial intelligence model training aspects of the system and processes are described in more detail in connection with
The system, in at least one embodiment, leverages the trained particular artificial intelligence model to produce one or more 3D digital representations 740 of patient anatomy based on patient image files (e.g., patient-specific image files 725) and anatomy data (e.g., anatomy data 720).
In particular embodiments, the system receives 2D patient-specific image files 725 and anatomy data 720 that is specific to the patient, specific patient anatomy, or a particular procedure (e.g., ankle arthroplasty). The system, in at least one embodiment, transforms the 2D patient-specific image files 725 into a digital representation, which is supplied as a data input into the artificial intelligence model. In some embodiments, the digital representation includes a 3D representation. In some embodiments, the user interface 710 is designed to receive data inputs from a user. For example, the user interface 710 can be used to upload patient files for a new patient. Further examples of the user interface 710, in at least some embodiments, are shown in
In one or more embodiments, the system uses the 2D patient-specific image files 725 and anatomy data 720 as inputs (along with, in at least one embodiment, other data elements) to the trained particular artificial intelligence model 735. In this example, the particular artificial intelligence model 735 extracts one or more features from the 2D patient-specific image files 725 and uses a non-rigid shape reference construct to “fill in” data points not captured in the 2D patient-specific image files 725.
In some embodiments, the trained artificial intelligence model 735 can analyze the contours and/or features of the patient image files 725 and compare those contours and/or features to contours and/or features from image files stored in memory 730 and/or associated with the population data 715 used to train the artificial intelligence model 735. When similar features, aspects, and/or patient-related data elements (e.g., similar injury and/or implant recommendation in physician notes) are identified, the particular training model 735 can use the non-rigid shape references to generate a patient-specific non-rigid shape reference for the particular patient being evaluated. In at least this way, the artificial intelligence model 735 generates a non-rigid shape reference that accurately and efficiently transforms the 2D images into 3D digital representations or anatomy models by filling in gaps in the data elements associated with the patient-specific anatomy that is not shown in the patient-specific image files 725 and may create one or more patient-specific surfaces (e.g., a surface that is not displayed due to the angle or resolution of the image files). In some embodiments, a first aspect may include the anatomy data related to the particular joint, while a second aspect may include patient-related data elements. Other configurations of first and second aspects (and additional aspects) of the patient anatomy data are contemplated within the scope of the present disclosure. In non-limiting embodiments, the system analyzes all relevant data elements associated with an anatomical feature to generate a non-rigid shape reference used with the particular artificial intelligence model to create a 3D digital representations 740 that is patient-specific and fine-tuned based on patient-specific information that may be indirectly related to the particular anatomical feature for which the device is being generated (e.g., medical history, accidents, injuries, medical conditions, etc.). The data analysis and implant generation technique(s) described herein are an improvement over the prior art for many reasons, but at least in that the embodiments described herein do not limit data analysis to the best fit or nearest neighbor techniques and instead intelligently considers all relevant data elements associated with an anatomical feature, including those not directly related in order to design and generate a patient-specific surgical device. It will be appreciated that the data analysis and modeling techniques are not limited to those described in the embodiment described in connection with
The generated patient-specific 3D digital representation 740 may represent a calculated approximation at a data representation of the patient's anatomy based on the patient-specific image files 725 and other patient-specific data. In at least one embodiment, the system may generate the 3D digital representation 740, which may be an anatomy model based on a patient-specific surface or anatomy data representation including the non-rigid shape reference. For example, in a certain embodiment, the system may combine a surface from the non-rigid shape reference artificial intelligence model 735 that corresponds to a surface of patient anatomy (e.g., a surface of a patient's tibia) with a general anatomical feature shape. As shown in the embodiment in
In various embodiments, the system may automatically (or semi-automatically) adjust the patient-specific surface from the non-rigid shape reference created by the particular artificial intelligence model 735 for the 3D digital representation 740. In one embodiment, the system may automatically add an offset or depth to the patient-specific surface (e.g., a depth for porosity or the like) for the 3D digital representation 740. As such, as will be understood from discussions herein, patient-specific surfaces of 3D device models may or may not exactly match patient surfaces determined by the non-rigid shape model 735 or the patient-specific image files 725. In various embodiments, the created 3D digital representation 740 can serve as the basis for a CAD implant file 760 and a 3D device model 765.
In at least one embodiment, the system uses the patient-specific anatomy model 740 or an extracted portion thereof, such as, for example, a patient-specific surface, as an input into a second artificial intelligence model (sometimes referred to herein as a “specific” artificial intelligence model) for creating a patient-specific device (e.g., implant, surgical guide, instrument, etc.). In some embodiments, in system leverages the patient-specific 3D digital representation 740 and other patient data (e.g., medical history) is an input to the trained specific artificial intelligence model to generate a device CAD file 760. In some embodiments, the artificial intelligence model generates a digital representation of the patient-specific device. In at least one embodiment, the iteratively trained learning model 750 is trained on implant population data 755 including data elements associated with implants designed and created for numerous patients, as stored in memory 730 and/or in the patient population data 715, as described in more detail in connection with
It will be understood that printing an implant/device 770 can include any patient-specific instrument, including but not limited to a 3D-printed body. In some embodiments, the system can 3D print the implant/device or may use other manufacturing techniques to create the implant/device. In some embodiments, the output of the implant generation process 700 can include non-surgical implants, medical devices, or other objects or materials. In various embodiments, the 3D-printed body can be constructed from a metallic material (e.g., medical grade titanium alloy), although other constructions and material types are contemplated within the scope of the present disclosure. Additional details associated with embodiments of the implant generation process 700 are described in connection with the flow diagrams of
A non-limiting illustrative example of the above-described process overview shown in
In various embodiments, the system can process the 2D patient-specific image files 725 using a segmentation process to generate the 3D digital representation 740 of the patient anatomy. In some embodiments, the segmentation process can be executed using the one or more trained artificial intelligence model 735. In some embodiments, the 3D digital representation 740 is transformed into a point cloud space and is input into the trained artificial intelligence model 735 to perform a registration and/or alignment of the point cloud space (e.g., a non-rigid or rigid point cloud registration). In various embodiments, the registration and/or alignment of the point cloud space can be input into the trained artificial intelligence model 735 with patient image files 725 and/or other patient population data 715.
In some embodiments, the output of the artificial intelligence model 735 can include a lower dimensional representation of the patient anatomy, which is combined with patient-specific surfaces to create a patient-specific device 770. In some embodiments, the patient-specific surfaces generated are based on the output of the trained artificial model 735 applied to the lower dimensional representation of the patient anatomy. In some embodiments, the process of generating patient-specific surfaces can be performed by a separate trained artificial intelligence model.
Further, in the non-limiting illustrative example, a second artificial intelligence model 750, a specific artificial intelligence model in some embodiments, is used to create a CAD file 760 (or other digital file) for a custom implant (in some embodiments, this can be integrated into the same artificial intelligence model 735 used to generate the 3D digital representation 740). In some embodiments, the specific artificial intelligence model 750 is trained based on implant population data 755, including details and designs for many (e.g., hundreds or thousands of) surgical implants and devices created for a plurality of patients, including details related specifically to ankle implants, in this non-limiting example. The patient-specific 3D digital representation 740 of the patient's ankle, including the patient-specific surfaces generated based on the patient image files 725 is an input to the trained implant/device artificial intelligence model 750 to develop an implant personalized to the patient's ankle anatomy. In at least this way, the specific artificial intelligence model 750 takes population data of all ankle implants and ankle implant-related devices (e.g., cut guides) designed for a plurality of patients and generates an implant design customized for the patient using all relevant data elements (e.g., surgeries, disorders, similar features/parameters/aspects of the population data information and the patient-specific information) to generate a CAD file 760. In some embodiments, the system creates a 3D implant model 765 from the generated CAD file 760 and compares the 3D implant model 765 to the patient's 3D digital representation 740 before printing (or otherwise generating) the patient-specific implant/device 770.
In this illustrative example, the system designs a custom ankle implant, and then the 3D device model of the ankle implant is generated and digitally modeled (or physically modeled) in the patient's anatomy model to ensure an accurate fit. After a design engineer and/or provider has approved the proposed design, or the advanced model has performed a verification/validation process, the system prints or otherwise creates the custom implant and/or device for the patient's ankle surgery. It will be appreciated that this example is merely illustrative, and the system and processes described herein can be applied to other anatomical features, surgical procedures, devices, etc.
In some embodiments, the implant generation process 700 can be executed via a computing environment (not shown), which can be provided in the form of one or more computing devices, server banks, computer banks, or other arrangements. For example, the computing environment can include one or more computing devices that together may include a hosted computing resource, a grid computing resource, and/or any other distributed computing arrangement. The computing environment can further include one or more routers. In some cases, the computing environment can correspond to an elastic computing resource where the allotted capacity of processing, network, storage, or other computing-related resources may vary over time. In some embodiments, the computing environment can also include an elastic cache designed to scale and adapt based on system demands and available capacity.
In some embodiments, the dashboard 810 can include a list of action items 815 assigned to the particular user 840 logged into the dashboard 810. In some embodiments, for example, if the user is a system administrator, team lead, or similar role, the dashboard may include all action items 815 assigned to all users with access to the system or platform. In some embodiments, the action items 815 and associated fields can be edited from the dashboard 810. In some embodiments, the dashboard 810 can further include a list of created events 820. For example, the name field associated with an event can include one or more fields for the patient name, surgeon name, the name of the user who created the event, the user assigned to the next steps for the event, etc. In some embodiments, the created event 820 list includes new implant requests that have been created but have not been submitted to the design team.
In some embodiments, after a new implant request or other created event 820 has been submitted for design, a design call 825 is generated. In one non-limiting example, the design call 825 can include a schedule function to coordinate a meeting between the provider and one or more design engineers to review the details, recommendations, and other strategy items or considerations before the design team creates a design proposal. In some embodiments, after the design call, 825 is complete, the system can generate a design 830 event until the design engineers have submitted a design proposal for the implant. In some embodiments, there may be a history 835 window that shows a user's recently viewed events. In some embodiments, the history 835 window is updated according to when another user accesses an event.
In some embodiments, the display configuration of the user interface 710 may be modified based on a particular user's function or access level. For example, the dashboard 810, screens, windows, options, and accessible data may populate differently for a user who is logged in as an external provider than it would appear for a user logged in as an internal design engineer. In this embodiment, the internal design engineer is used as a non-limiting example but can include other internal-facing user roles. Other internal-facing user roles can include but are not limited to, the commercial team, customer service, design engineers, design leads, regulatory, and system administrators. As used throughout the disclosure, the “provider” can include but is not limited to, a clinical staffer, physician, members of the surgical team, etc. There may also be additional user roles, functions, or permission levels based on required training (e.g., HIPAA training). For example, users that have received HIPAA training may be able to access and view all details of patient data, whereas someone without HIPAA training would only have access to view de-identified patient data and image files, including where all patient information and personal identifying information has been removed.
In some embodiments, the user interface 710 can include one or more interactive interfaces to display customized information and requests based on the permissions and access level of a particular user. When a particular user logs in via an authentication process, the system can identify the access level associated with the user. In some embodiments, the user interface 710 can include a plurality of tabs, screens, windows, menus, and other options designed to receive input. In one example, the user interface 710 can include, but is not limited to: an interactive dashboard with pre-configured fields to accept a new case with an implant request, modify the request submissions, view request submissions and status, input patient data and upload patient files, and generate design proposals and manufacturing requests. Continuing with this particular example, the user interface 710 communicates with the memory 730 to retrieve and package a plurality of data elements via an API associated with the particular access level associated with the user. Examples of the plurality of user interfaces 710 associated with the different access levels may include but are not limited to, the interface graphics as seen and described in connection to
It will be understood that the user interface 710 can be used for other interactions with patient population data 715, patient-specific image files 725, data elements, and similar aspects of the disclosed system and processes including, but not limited to importing, editing, or modifying, viewing, exporting, linking, or other interactions. The user interface 710 can be integrated with a plurality of third-party applications and utilize a network 1170 or cloud-based servers to provide seamless synchronization among the components of the system throughout the implant generation process 700. In one example, if a user starts a new request using a user interface 710 that is web-based, and then wants to complete the request using a user interface 710 on a mobile device, the system can utilize a synchronized data storage solution to synchronize the user interfaces 710. In some embodiments, the authentication process can identify that the same user is accessing a request, and the system can retrieve the data elements associated with the request and send the plurality of data elements via an API to the user interface 710. The authentication process is described in more detail in connection with
In some embodiments, the event visual graphic 1000 may include a timeline 1010 to track the status of the event. In this example, different graphics or visual indicators 1015 can be used for events that are complete, in progress, or completed. In some embodiments, the timeline 1010 may include information as to the user who has been assigned a task associated with the event. In some embodiments, the event can include a design proposal file upload 1020. In some embodiments, the design proposal may include a personalized implant design generated by the processes and systems described herein and submitted for provider approval. Additional file upload options 1025 can also be included on the visual graphic 1000.
The event visual graphic 1000 can further include event details 1030. The event details 1030 can include but are not limited to, patient data, anatomy data 720, image data 1150, implant data 1150, CAD model data 1160, and other data, like the provider information, the facility information, the proposed surgery date, etc.
The event visual graphic 1000 can also include patient data, depending on the access level associated with a user 840. There can also be a comment field 1050 where one or more users interacting with the event can add remarks or notes. As described in connection with
It will be appreciated that the fields and options associated with each aspect of the visual graphics 800, 900, and 1000 are not limited to those shown in the visual graphics 800, 900, and 1000. In at least this way, the system can customize user interface(s) 110 according to the particular user, system, process, application, etc.
While the system components described include a server 1105 and a network interface 1170, it will be recognized by one skilled in the art that the system configuration could include a plurality of computing devices in various configurations. The computing elements of the system 1100 can be provided via one or more computing devices that may be arranged, for example, in one or more server banks, computer banks, or other arrangements. Such computing devices can be located in a single installation or may be distributed among many different geographical locations. For example, the system 1100 can include computing devices that together may include a hosted computing resource, a grid computing resource, and/or any other distributed computing arrangement. The computing devices can further include one or more routers. In some cases, system 1100 can correspond to a dynamic computing resource where the allotted capacity of processing, network, storage, or other computing-related resources may vary over time.
In one embodiment, the implant generation system 1100 is designed to process data elements using a plurality of data processing modules within the server 1105. The system 1100 server 1105 can include: an advanced analytics module 1110, a processing module 1115, a training module 1120, a feedback module 1125, and a generation module 1130. In some embodiments, the system 1100 may include additional modules for processing and analyzing different data elements and data sets. As will be understood from discussions herein, the plurality of data processing modules shown in
In some embodiments, the data elements can include patient population data 715, patient-specific data (not shown), anatomy data 720, image data 1150, implant data 1155, and CAD model data 1160. It will be understood that the data elements can include other data and are not limited to the embodiment(s) shown or described in connection with
In one or more embodiments, the system 1100 received transmitted data including patient data, which, as described in connection with
In some embodiments, the system 1100 receives transmitted or extracted data including anatomy data 720. As described in connection with
In some embodiments, the anatomy data 720 associated with specific anatomical features can include data about a particular body part (e.g., left foot) or portion thereof (e.g., left ankle, or a segment of the left ankle joint). The anatomy data 720 associated with specific anatomical features can include data elements and information about the specific anatomical feature being evaluated from not only the particular patient being considered but also the specific anatomical feature from the entire patient population data 715 and anatomy data 720 database. Similar to the example above, this information can include data elements related to other anatomical features that may impact the specific anatomical feature being replaced or modified with a surgical implant.
In some embodiments, the anatomy data 720 associated with specific characteristics can include healthy, diseased, deformed, injured, or similar characteristics that may impact the size, shape, measurement, surface, mobility, etc. of a particular anatomical feature. In some embodiments, the specific characteristics can include a plurality of disease states, the change to a specific anatomical feature over a healing process or over time as a patient ages, or similar variations of a specific anatomical feature over a period of time. The anatomy data 720 can further include other parameters or metrics associated with the anatomical feature(s) of a plurality of patients and for the particular patient being evaluated.
In some embodiments, the system 1100 receives transmitted or extracted data including image data 1150. In some embodiments, the image data 1150 can include the patient-specific image files 725 described in connection with
In some embodiments, the system 1100 receives transmitted or extracted data including implant data 1155. The implant data 1155 can include any details needed to generate the personalized implant. In some embodiments, the implant data 1155 can include one or more implant files associated with a particular patient or plurality of patients, including but not limited to the implant population data 755, described in connection with
In some embodiments, the system 1100 receives transmitted or extracted data including CAD model data 1160. The CAD model data 1160 can include but is not limited to the technical drawings generated that are used to print, manufacture, or otherwise generate the surgical implant. In some embodiments, the CAD model data 1160 may include information and data from the implant data 1155. The CAD model data 1160 may also include product code(s), part number(s), stock keeping number(s) (SKU), text to be printed on the implant, project information, other patient information, tolerances, specifications, measurements, notices, instructions, manufacturing directions, provider notes, design engineer remarks, and other details to represent the implant to be printed, created, manufactured, or otherwise generated. In some embodiments, the memory 730 can include other data elements than those shown in
In various embodiments, the implant generation process 700 and the implant generation system 1100 can include one or more processes that may be executed via a network interface 1170 designed to perform one or more processes for advanced data processing and transforming data for executing a generation model to create one or more models and/or design files. The networked interface 1170 may include a networked system provided in the form of a computer architecture system provided in the form of various computing components designed to communicate over a network. The network interface 1170 includes, for example, the Internet, intranets, extranets, wide area networks (“WANs”), local area networks (“LANs”), wired networks, wireless networks, cloud networks, or other suitable networks, or any combination of two or more such networks. For example, the network 1170 can include satellite networks, cable networks, Ethernet networks, and other types of networks. In one embodiment, the network interface 1170 is an isolated private network utilizing a private IP address and limiting access to the network.
In one or more embodiments, the system 1100 may also include an authentication service, a notification queue, and a regulatory assessment. The system 1100 can be designed to communicate with and integrate with a plurality of third-party applications. In some embodiments, the authentication service can confirm the identity and user by one or more validation techniques. In one or more embodiments, the authentication service may use, as non-limiting examples: a username and password, personalized token, personal identification number, or biometric data to validate system access associated with a particular account. The authentication service can also be designed to associate a unique identifier for each user, wherein the unique identifier can be used throughout the systems and processes described herein. The unique identifier can be used to indicate the specific access level assigned to a particular user.
The notification queue can be an application for creating, modifying, and logging events, requests, tasks, and action items. The notification queue can also generate, update, and push notifications to one or more devices or applications, including the user interface 710, in some embodiments. In various embodiments, the notification queue is designed to be dynamic, scalable, and integrated with a plurality of third-party applications. In some embodiments, the notification queue is designed to receive data elements from a plurality of data sources or applications and handle the notification generation and management for the plurality of data sources and/or applications. The notification queue can be further designed to create and monitor new events, wherein each event is assigned a unique identification tag and time stamp. Each event can have one or more tasks assigned, each with an individual status, time stamp, and identification. A notification status can be updated automatically in response to receiving user input or other indications that an event has been modified, completed, or assigned to another user.
In some embodiments, the systems and processes herein store and index data from data sources, in a variety of formats, and then use processes for validating and transforming the data to generate CAD files or 3D device models and personalized surgical implants.
In at least one embodiment, the system is designed to automatically (or in response to an input) collect, retrieve, or access a plurality of data elements, including patient data and image files, from one or more patient databases. In some embodiments, the plurality of data elements can include data from a large number of sources and patients with varying data types and formats. In various embodiments, the system is designed to automatically analyze, validate, and index data to facilitate evaluating a plurality of aspects and/or characteristics between a plurality of data elements and data elements associated with a particular patient.
The plurality of data elements may be stored in a dynamic database that utilizes a plurality of modules. In some embodiments, the dynamic database can be provided in the form of the memory 730, be integrated with the memory 730, and/or be connected to the memory 730. The system can be designed to request and synchronize data among and between a plurality of modules. In at least this way, the system is flexible, scalable, and efficient at processing a growing number of data elements related to one or more aspects and/or characteristics. In some embodiments, the system utilizes an advanced generation model to process the output of one or more iteratively trained learning models and create a 3D digital representation, a 3D device model, a CAD file, and/or a surgical implant. In one or more embodiments, the present system can create or otherwise generate a personalized surgical implant using 3D printing techniques.
In one or more embodiments, the present system may transform the plurality of data elements for analysis via the training model processes and other techniques described herein. In at least one embodiment, the present system may clean and transform data to remove, impute, or otherwise modify missing, null, or erroneous data values. In various embodiments, the present system may remove identifying information in order to anonymize and remove any correlated data. Similarly, the system may index and correlate specific data elements, data types, and data sets to facilitate the implant generation process. In at least one embodiment, the system may include data elements from a plurality of patients to create a dynamic database and one or more trained artificial intelligence training models. In certain embodiments, the system may include one or more algorithms to automatically update and train the database and/or training models. For example, in some embodiments, data corresponding to the various aspects of an anatomical feature can be processed with the one or more algorithms to generate parameter values determined from analyzing all the data elements within a database. In various embodiments, the system may include an interface for operating and controlling the various facets of the system and processes as described herein.
As discussed herein, the system may receive the patient-specific image packages from one or more sources. In at least one embodiment, the system receives the patient-specific image packages from a user interface 710, such as the user interface 710 discussed in relation to
At step 1210, the system can associate an anatomical representation with the image packages received at step 1210. In some embodiments, the anatomical representation may include a digital twin or a genetic digital twin. In one non-limiting example, the digital twin can be updated and modified with data over a period of time or as the patient ages. In at least this way, the anatomical representation associated with the image packages can include and incorporate data elements associated with patient population data 715, patient-specific patient data, and anatomy data 720, as described in connection in more detail with
At step 1215, the system can extract data associated with one or more anatomical features corresponding to a particular patient, including but not limited to the anatomy data 720. At step 1215, the process can further include extracting measurements or other information from the image package(s). In some embodiments, the information extracted at step 1215 may include information related to anatomical features on the other side of the patient or with other patients. For example, at step 1215, the process may include extracting information regarding a patient's left knee, which is scheduled to be replaced. In this non-limiting example, at step 1215, the process could also extract information from the left knee of other patients as stored in the image files of the patient population data 715. In some embodiments, the system trains the particular artificial intelligence model using patient population data 715. In one non-limiting example, the system creates a non-rigid shape reference based at least on portions of extracted anatomical left knee data and information from the patient population data 715. In this non-limiting example, the trained particular artificial intelligence model extracts information related to the patient's knee from the patient-specific image files 725 and uses the non-rigid shape reference generated with the particular artificial intelligence model 735 to create a patient-specific 3D digital representation 740 from the patient-specific 2D image files 725.
Additionally, the system could also extract information related to the patient's right knee and information related to the right knees of all other patients in the patient population database 715. In at least this way, the system and processes described herein extract and evaluate all known data elements that may be relevant to the design of the patient's 3D digital representation and/or surgical implant model and/or 3D device model.
At step 1220, the data elements can be processed using advanced artificial intelligence processes. In some embodiments, there may be multiple artificial intelligence models utilized in step 1220. In one non-limiting example, each anatomical feature may be processed by a different machine-learning model. At step 1220 the process can also be configured to generate, train, and execute a plurality of nodes, neural networks, gradient boosting algorithms, mutual information classifiers, random forest classifications, and other machine learning and artificial intelligence related algorithms to process the data elements. In some embodiments, the one or more artificial intelligence models and/or trained learning models can include deep learning, machine learning, neural networks, computer vision and similar advanced artificial intelligence-based technologies.
In some embodiments, a second aspect of the anatomical features and/or representations can be processed via one or more specific artificial intelligence models at step 1220. In some embodiments, the second aspect of the anatomical features can be processed during step 1220 while the first aspect is being processed. It will be appreciated by one skilled in the art that the system and process described is not limited to only a first and second aspect of the anatomical features being evaluated, but instead can include the evaluation and processing of a plurality of aspects of the anatomical features. At step 1220, the process evaluates all data elements extracted and saved in the memory 730 to generate a non-rigid shape reference in step 1225 based on an evaluation of all accumulated data elements.
At step 1225, the system uses the artificial intelligence model(s) to create a non-rigid shape reference based on one or more feature extracted from the patient-specific 2D image file 725 package at step 1205, which is then used, in some embodiments, to transform contours and surface features of the 2D image patient image files 725 into a 3D digital representation of the patient anatomy, including one or more patient-specific surface(s). In some embodiments, the patient-specific surfaces can include one or more porous areas, smooth or textured surface features, features of different thicknesses, and/or other surface features. It will be appreciated that other patient-specific configurations and surfaces are contemplated.
At step 1230, a 3D device model of the anatomical feature is created based on the non-rigid shape reference generated. In some embodiments, the data elements are transformed at step 1230 from a 2D representation to a 3D digital representation and then the associated point cloud space is aligned with an artificial intelligence model trained based on or more aspects of patient data to generate a lower-dimensional representation model of the patient anatomy. At step 1235, a CAD file for at least one patient-specific implant is generated. As described above in connection with
At step 1235, the process generates and/or retrieves the CAD model data 1260, described in connection with
Using the generated CAD file from step 1235, the system can then create or generate the personalized patient-specific implant at step 1240. In some embodiments, a plurality of implants may be generated at step 1240. For example, the system can create a test implant in addition to the implant to be used in the patient, for a provider to practice before the surgical procedure. In another non-limiting embodiment, the process 1200 may generate a plurality of implants in various sizes at step 1240. In this example, the provider can field-test the appropriate fit of the implant as needed during the operation without requiring any modifications to the implant. In some embodiments, the test implant can be compared to the 3D digital representation, either physically or digitally, to facilitate the field-test, as described in more detail in connection with
At step 1305, the system receives (or retrieves from the memory 730) an image file package including patient population data 715, anatomy data 720, a series of 2D images associated with a particular anatomical feature, and other patient data (e.g., medical history). In some embodiments, the step 1305 described above can apply to training the particular artificial intelligence model 735. In some embodiments, the system is designed to process patient population data 715 including information from thousands of patients (e.g., medical history, injuries, surgeries, diseases, disorders, allergies, height, weight, etc.). The patient population data 715 can be received by a user interface 710 and/or retrieved from memory 730.
In some embodiments, the step 1305 may instead include receiving implant population data 755 for training the specific artificial intelligence model 750. In some embodiments, the step 1305 can include receiving both patient population data 715 and implant population data 755 for training an ensemble artificial intelligence model comprising both the particular artificial intelligence model 735 and the specific artificial intelligence model 750. In one non-limiting embodiment, the system can identify that the patient needs ankle surgery and retrieve anatomy data 720 related to the ankle from the patient population data 715 in order to train the artificial intelligence model. The system can train the implant/device model 750 based on a specific type of implant (e.g., tibial), implants for a single anatomical feature (e.g., ankle or foot), or for all implants and fine-tuned or with updated emphasis guidelines based on the particular procedure to be performed.
At step 1310, the system can extract aspects from the image file package received at step 1305, which includes known parameters used to iteratively train the one or more artificial intelligence training models. The system in step 1310, can extract and process a plurality of input data sets associated with a specific data element type, wherein each of the plurality of input data sets can have multiple data types. In one embodiment, a specific anatomical feature of the process 1200 may include multiple associated input data sets.
At step 1315, the system can input the data elements corresponding to the extracted aspects as a training input data set into one or more artificial intelligence training models. In one non-limiting example, the system can iteratively train the artificial intelligence training models based on multiple training input data sets of different data types, including data elements associated with a particular anatomical feature, particular implant, parameter, characteristic, or other aspects.
The system, in step 1320, receives a plurality of parameter values as outputs from the plurality of trained learning models. In some embodiments, the system can use the plurality of output parameter values to generate the non-rigid shape references and/or the 3D digital representation of the anatomical feature. In at least this way, the system can utilize a plurality of trained learning models to output specific parameters or metrics tailored to certain design details of the anatomical feature. In one example, if an anatomical feature has anatomy data 720 associated with a specific deformity, the system can use a artificial intelligence training model based primarily on data analysis of patient population data 715 and anatomy data 720 related to the deformity, or other parameters that may influence the deformity. Alternatively, the system could also utilize a combination of multiple training models where the deformity is one of a plurality of parameters, in addition to the surface, texture, etc., wherein the process 1300 can recognize and provide a personalized implant designed for the specific patient based on data elements from all available patient records associated with patient population data 715, anatomy data 720, image data 1150, implant data 1155, CAD model data 1160, and other data. It will be appreciated by one skilled in the art that a combination of multiple parameters can be used in a single training loop to provide a personalized implant based on a high level of certainty.
In some embodiments, the artificial intelligence model 735 is a particular artificial intelligence model wherein the system includes training the model to create a non-rigid shape reference to transform 2D anatomical features into 3D data representations of the anatomy in order to generate a 3D digital representation 740 for the patient-specific anatomy. For example, in one embodiment, the system may extract initial data and features from image files associated with many thousands of patients as it relates to a particular type of anatomical feature and/or procedure and/or implant. In one embodiment, the system may extract initial data and features from the patient-specific image files 725 to construct a data representation of patient anatomy (e.g., a diseased ankle of a particular patient).
Additionally, at step 1320, the system can generate a 3D device model from the output of the one or more artificial intelligence training models. In some embodiments, when training the particular artificial intelligence model 735, the 3D device model may be a 3D digital representation 740 generated based on a non-rigid shape reference created as an output of the one or more artificial intelligence training models. In some embodiments, the 3D model may be a 3D device model 765 to compare to the anatomy model 740, for example when training the specific artificial intelligence model 750. In some embodiments, the system can train a specific artificial intelligence model 750 using implant population data 755 including implant information and data related to a plurality of implants designed for a plurality of patients. In some embodiments, the system can train the specific artificial intelligence model 750 based on implant population data 755 for a single anatomical feature or for a plurality of anatomical features with fine-tuned emphasis guidelines for specific anatomical features to provide a trained specific artificial intelligence model 750 designed to generate a CAD file 760 and/or 3D device model 765 for a customized implant 770.
In some embodiments, the system can compare the 3D digital representation 740 to a known 3D digital representation and the training process can include updating one or more emphasis guidelines 1330 of the artificial intelligence training model in order to update the model based on patient-specific surfaces. In some embodiments, system can update one or more aspects of the 3D digital representation, rerunning the artificial intelligence training model to determine if the output of the comparing step 1325 the generated 3D digital representation to the known 3D digital representation is within a preconfigured threshold 1335. If the output of the comparing step 1325 is within the preconfigured threshold as determined by step 1335, the updated emphasis guidelines may be saved 1340, and the system can save the updated artificial intelligence training model as the trained artificial intelligence model.
At step 1325, the generated 3D digital representation can then be compared to the known parameters for the training input data set. At step 1330 if the output of compare step of 1325 is not within a preconfigured threshold of the known parameters, one or more emphasis guidelines can be updated for a plurality of nodes within the artificial intelligence training model. In some embodiments, for example while training the particular artificial intelligence model 735, the comparison step 1325 can include comparing the generated 3D digital representation to a known 3D digital representation in the patient population data 715. In some embodiments, for example while training the specific artificial intelligence model 750, the comparison step 1325 can include comparing the generated 3D device model/file to a known 3D device model/file in the implant/device population data 755.
In some embodiments, the comparison step 1325 can be performed by one or more artificial intelligence models, including but not limited to the particular artificial intelligence model 735 and/or a specific artificial intelligence model 750. In various embodiments, one or more artificial intelligence models, including but not limited to the particular artificial intelligence model and/or the specific artificial intelligence model, can be trained based on a specific anatomical feature, surface topography, lattice structure, or aspects thereof. It will be appreciated by one skilled in the art that various artificial intelligence models can include models trained for a specific patient, anatomical feature, surgery, or disease state, based on one or more patient-specific files and/or implant parameters. In some embodiments, the artificial intelligence model compares an output to a known 3D digital representation stored in memory 730 and or from the patient population data 715 to iteratively train the artificial intelligence model and adjust emphasis guidelines. In some embodiments, the one or more artificial intelligence model(s) is provided in the form of an ensemble artificial intelligence model comprising the particular artificial intelligence model 735 and the specific artificial intelligence model 750. It will be understood by one skilled in the art that the combination and configuration of the one or more artificial intelligence model(s) is not limited to those described in connection with
The system may then, in some embodiments, adjust, change, or add data to the 3D digital representation 740 based on other patient population data 715 (e.g., anatomy aspects and feature from patients with similar data elements including similar images, similar size, similar injuries, etc.). The system can further adjust a 3D digital representation 740 based on one or more patient-specific data elements (e.g., provider notes, other images, etc.). In some embodiments, the system can fine-tune or otherwise adjust the generated 3D digital representation 740 by adjusting the emphasis guidelines for the training model, or by directly editing features of the generated 3D digital representation 740 without adjusting the training model 735.
As used throughout the present disclosure, “emphasis guidelines” can include weights, ranks, or other metrics associated with iteratively training an artificial intelligence or machine learning training model to improve the accuracy and relevance of generated results.
The system can update the one or more emphasis guidelines based on the results of the step 1325, to iteratively train and improve the training model in step 1330. The one or more emphasis guidelines of the system can be updated for a plurality of nodes within the raw training models based on the results of the comparing step, to iteratively train and improve the training models. When the emphasis guidelines are updated at step 1330, the artificial intelligence training model can be rerun, re-executed, or the data otherwise processed again through the algorithm(s) to determine if the generated 3D digital representation 740 is within a pre-defined threshold of the known 3D digital representation, based on the output of the comparing step 1325.
In some embodiments, the predefined threshold can include various tolerance levels, which can be adjusted based on specific parameters. In one non-limiting example, a size of an anatomical feature relative to another anatomical feature may have a smaller threshold tolerance level than surface structure deviations. In further embodiments, the predefined threshold can be modified dependent on the particular implant to be generated, on specific characteristics of the anatomical feature being modeled, or on any number of other factors.
In some embodiments, the system can further use the artificial intelligence training model that has been updated with the saved emphasis guidelines to further process input data as the iteratively trained artificial intelligence training model. In some embodiments, the iteratively trained artificial intelligence training model can include the particular artificial intelligence model 735 and/or the specific artificial intelligence model 750 and/or the ensemble artificial intelligence model comprising both the particular 735 and specific 750 artificial intelligence models. The trained artificial intelligence model(s) can continue to be updated as additional patient population data 715 is received and processed by the system.
At step 1335, when the generated 3D digital representation output of the artificial intelligence training model is within a preconfigured threshold of known parameters for the training input data sets, as determined by the comparing step of 1325, the system can save the plurality of artificial intelligence training models as trained artificial intelligence models and save the one or more updated emphasis guidelines from step 1330 in the memory 730 in step 1340.
Also, the system can include one or more secondary metrics as parameters in one or more processes to iteratively train a training model or a plurality of artificial intelligence training models. When used throughout the present disclosure, one skilled in the art will understand that processes for “iteratively training the artificial intelligence training model” can include machine learning processes and other advanced artificial intelligence processes. For example, the system and processes of the present disclosure can perform diagnostics, and image analysis, generate tasks or action items, provide customized recommendations according to user settings and preferences, generate 3D device models, CAD files, and personalized implant designs, generate notifications, and similar processes. In some embodiments, the system may use additional inputs and/or feedback loops to an iterative training process for a personalized implant generation process based on a plurality of parameters and adjustable metric values.
In some embodiments, the system can include a process for iteratively training an artificial intelligence model (or multiple artificial intelligence models) according to some embodiments of the present disclosure. In various embodiments, the iteratively trained artificial intelligence model(s) can be designed to perform various advanced data analysis and modeling processes. In one example, the system generates and iteratively trains the artificial intelligence training model to provide dynamic data analysis, and the advanced analytics can be used to generate a 3D device model, and CAD file, and print an implant/device. In some embodiments, these processes can be performed by multiple artificial intelligence models, or multiple aspects of a single artificial intelligence model (e.g., an ensemble model), or a combination thereof. In one non-limiting embodiment, the artificial intelligence training model(s) can be designed to generate, train, and execute a plurality of nodes, neural networks, gradient boosting algorithms, mutual information classifiers, random forest classifications, and other machine learning and artificial intelligence-related algorithms.
It will be appreciated by one skilled in the art that the system and processes described herein can include different and/or additional details, data elements, measurements, parameters, metrics, and implants than those described in the specific embodiments herein.
Additionally, the embodiments described herein are further directed to an improved system and methodology for using artificial intelligence to analyze and process patient data from a dynamic database to generate a personalized surgical implant/device 770. These embodiments represent an advancement over the prior art for several reasons, including that the systems and methods are capable of complex and efficient data analysis related to a plurality of different anatomical feature aspects and parameters. While the prior art may utilize a best fit or nearest neighbor to develop surgical implants, the embodiments described herein do not limit data analysis to the best fit or nearest neighbor techniques and instead intelligently considers all relevant data elements associated with an anatomical feature, including those not directly related. In a non-limiting example, the system and processes described herein can evaluate anatomy data 720 related to a joint other than the one being replaced in surgery, because the system recognizes that injuries to a different joint may impact the particular joint being replaced, for which the implant is being generated. The embodiments can include measuring the changes to anatomical features over time, accumulating these measurements across a memory 730 of population patient data 715 and/or patient-specific data received or extracted, performing statistical analysis on the accumulated measurements, performing advanced diagnostics, and producing one or more customized outputs.
According to particular embodiments, the system may automatically link or associate data included in the 3D digital representation 740 to one or more parameters according to the advanced analytic processes described herein. In some embodiments, the linked data can include information related to joint replacement, including anatomical features of common joints. Non-limiting examples can include foot, ankle, shoulder, elbow, hips, knees, tracheal stents, etc. It will be understood by those skilled in the art that the embodiments described herein are not limited to a joint replacement but can be applied to the generation of other implants, medical devices, instruments, or other objects. In some embodiments, the implant generation process 700 can evaluate the anatomy data 720 in light of the patient population data 715, the patient's image files 725, and a patient's personal information or file history to further personalize the generated implant 770.
While the particular embodiments described herein relate to healthcare-based applications. It will be recognized and appreciated that the systems and processes described herein are also applicable to at least, but not including, other types of design services. For example, the systems and methods described herein could be used to analyze and generate custom solutions related to architecture, construction, engineering, repairs, art restoration, hobby electronics, advanced computing, etc.
It will be appreciated by one skilled in the art that this embodiment is only a non-limiting example used to illustrate that the system and processes are designed to dynamically provide an advanced artificial intelligence model for evaluating a plurality of data elements associated with an anatomical feature, including data elements directly related to other anatomical features but that may have an impact on the anatomical feature being evaluated. Additionally, the non-limiting example can be used for personalized implant generation requests and advanced data processing techniques across a plurality of systems and applications and can apply to a plurality of different services and industries.
Computer program code that implements the functionality described herein typically comprises one or more program modules that may be stored on a data storage device. This program code, as is known to those skilled in the art, usually includes an operating system, one or more application programs, other program modules, and program data. A user may enter commands and information into to the computer through a keyboard, touch screen, pointing device, a script containing computer program code written in a scripting language, or other input devices, such as a microphone, etc. These and other input devices are often connected to the processing unit through known electrical, optical, or wireless connections.
The computer that affects many aspects of the described processes will typically operate in a networked environment using logical connections to one or more remote computers or data sources, which are described further below. Remote computers may be another personal computer, a server, a router, a network PC, or other common network nodes common network node, and typically include many or all of the elements described above relative to the main computer system in which the systems are embodied. The logical connections between computers include a LAN, a WAN, virtual networks (WAN or LAN), and wireless LAN (“WLAN”) that are presented here by way of example and not limitation. Such networking environments are commonplace in office-wide or enterprise-wide computer networks, intranets, and the Internet.
When used in a LAN or WLAN networking environment, a computer system implementing aspects of the system is connected to the local network 1170 through a network interface or adapter. When used in a WAN or WLAN networking environment, the computer may include a modem, a wireless link, or other mechanisms for establishing communications over the WAN, such as the Internet. In a networked environment, program modules depicted relative to the computer, or portions thereof may be stored in a remote data storage device. It will be appreciated that the network connections described or shown are non-limiting examples and other mechanisms of establishing communications over WAN or the Internet may be used.
Additional aspects, features, and processes of the claimed systems will be readily discernible from the description herein, by those of ordinary skill in the art. Many embodiments and adaptations of the disclosure and claimed systems other than those herein described, as well as many variations, modifications, and equivalent arrangements and processes, will be apparent from or reasonably suggested by the disclosure and the description thereof, without departing from the substance or scope of the claims. Furthermore, any sequence(s) and/or temporal order of steps of various processes described and claimed herein are those considered to be the best mode contemplated for carrying out the claimed systems. It should also be understood that, although steps of various processes may be shown and described as being in a preferred sequence or temporal order, the steps of any such processes are not limited to being carried out in any particular sequence or order, absent a specific indication of such to achieve a particular intended result. In most cases, the steps of such processes may be carried out in a variety of different sequences and orders, while still falling within the scope of the claimed systems. In addition, some steps may be carried out simultaneously, contemporaneously, or in synchronization with other steps.
Aspects, features, and benefits of the claimed devices and processes for using the same will become apparent from the information disclosed in the exhibits and the other applications as incorporated by reference. Variations and modifications to the disclosed systems and processes may be affected without departing from the spirit and scope of the novel concepts of the disclosure.
It will, nevertheless, be understood that no limitation of the scope of the disclosure is intended by the information disclosed in the exhibits or the applications incorporated by reference; any alterations and further modifications of the described or illustrated embodiments, and any further applications of the principles of the disclosure as illustrated therein are contemplated as would normally occur to one skilled in the art to which the disclosure relates. The description of the disclosed embodiments has been for illustration and description and is not intended to be exhaustive or to limit the devices and processes for using the same to the precise forms disclosed. Many modifications and variations are possible in light of the above teaching.
The embodiments were chosen and described to explain the principles of the devices and processes for using the same and their practical application to enable others skilled in the art to utilize the devices and processes for using the same and various embodiments and with various modifications as are suited to the particular use contemplated. Alternative embodiments will become apparent to those skilled in the art to which the present devices and processes for using the same pertain without departing from their spirit and scope. Accordingly, the scope of the present devices and processes for using the same is defined by the appended claims rather than the description and the embodiments described therein.
Number | Name | Date | Kind |
---|---|---|---|
3872519 | Giannestras | Mar 1975 | A |
4440835 | Mgnaud | Apr 1984 | A |
4588574 | Felder et al. | May 1986 | A |
4829152 | Rostoker | May 1989 | A |
5248456 | Evans, Jr. et al. | Sep 1993 | A |
6183519 | Bonnin | Feb 2001 | B1 |
6419491 | Ricci | Jul 2002 | B1 |
6461358 | Faccioli | Oct 2002 | B1 |
7001672 | Justin et al. | Feb 2006 | B2 |
7468075 | Lang et al. | Dec 2008 | B2 |
7534246 | Reiley | May 2009 | B2 |
7534263 | Burdulis, Jr. et al. | May 2009 | B2 |
7534270 | Ball | May 2009 | B2 |
D595853 | Hanson | Jul 2009 | S |
7618451 | Berez et al. | Nov 2009 | B2 |
7632575 | Justin et al. | Dec 2009 | B2 |
7634119 | Tsougarakis et al. | Dec 2009 | B2 |
7666522 | Justin et al. | Feb 2010 | B2 |
7717956 | Lang | May 2010 | B2 |
D623749 | Horton | Sep 2010 | S |
7796791 | Tsougarakis et al. | Sep 2010 | B2 |
7799077 | Lang et al. | Sep 2010 | B2 |
7981158 | Fitz et al. | Jul 2011 | B2 |
8062302 | Lang et al. | Nov 2011 | B2 |
8066708 | Lang et al. | Nov 2011 | B2 |
8077950 | Tsougarakis et al. | Dec 2011 | B2 |
8083745 | Lang et al. | Dec 2011 | B2 |
8094900 | Steines et al. | Jan 2012 | B2 |
8105330 | Fitz et al. | Jan 2012 | B2 |
8234097 | Steines et al. | Jan 2012 | B2 |
D653756 | Courtney et al. | Feb 2012 | S |
8122582 | Burdulis, Jr. et al. | Feb 2012 | B2 |
8142886 | Noble et al. | Mar 2012 | B2 |
8337501 | Fitz et al. | Dec 2012 | B2 |
8337507 | Lang et al. | Dec 2012 | B2 |
D675320 | Oi | Jan 2013 | S |
8343218 | Lang et al. | Jan 2013 | B2 |
8366771 | Burdulis, Jr. et al. | Feb 2013 | B2 |
8377129 | Fitz et al. | Feb 2013 | B2 |
8382755 | Austin | Feb 2013 | B2 |
8430930 | Hunt | Apr 2013 | B2 |
8439926 | Bojarski et al. | May 2013 | B2 |
8457930 | Schroeder | Jun 2013 | B2 |
8460304 | Fitz et al. | Jun 2013 | B2 |
8480754 | Bojarski et al. | Jul 2013 | B2 |
8485820 | Ali | Jul 2013 | B1 |
8500740 | Bojarski et al. | Aug 2013 | B2 |
8529568 | Bouadi | Sep 2013 | B2 |
8529630 | Bojarski et al. | Sep 2013 | B2 |
D692136 | Tyber | Oct 2013 | S |
8545569 | Fitz et al. | Oct 2013 | B2 |
8551099 | Lang et al. | Oct 2013 | B2 |
8551102 | Fitz et al. | Oct 2013 | B2 |
8551103 | Fitz et al. | Oct 2013 | B2 |
8551169 | Fitz et al. | Oct 2013 | B2 |
8551173 | Lechmann et al. | Oct 2013 | B2 |
8556906 | Fitz et al. | Oct 2013 | B2 |
8556907 | Fitz et al. | Oct 2013 | B2 |
8556971 | Lang | Oct 2013 | B2 |
8556983 | Bojarski et al. | Oct 2013 | B2 |
8561278 | Fitz et al. | Oct 2013 | B2 |
8562611 | Fitz et al. | Oct 2013 | B2 |
8562618 | Fitz et al. | Oct 2013 | B2 |
8568479 | Fitz et al. | Oct 2013 | B2 |
8568480 | Fitz et al. | Oct 2013 | B2 |
8585708 | Fitz et al. | Nov 2013 | B2 |
8617172 | Fitz et al. | Dec 2013 | B2 |
8617242 | Philipp et al. | Dec 2013 | B2 |
8623026 | Wong et al. | Jan 2014 | B2 |
8634617 | Tsougarakis et al. | Jan 2014 | B2 |
8638998 | Steines et al. | Jan 2014 | B2 |
8641716 | Fitz et al. | Feb 2014 | B2 |
8657827 | Fitz et al. | Feb 2014 | B2 |
8682052 | Fitz et al. | Mar 2014 | B2 |
8690945 | Fitz et al. | Apr 2014 | B2 |
8709089 | Lang et al. | Apr 2014 | B2 |
8715362 | Reiley | May 2014 | B2 |
8735773 | Lang | May 2014 | B2 |
D708747 | Curran et al. | Jul 2014 | S |
8768028 | Lang et al. | Jul 2014 | B2 |
8771365 | Bojarski et al. | Jul 2014 | B2 |
8775133 | Schroeder | Jul 2014 | B2 |
8828311 | Medina et al. | Sep 2014 | B2 |
8843229 | Vanasse et al. | Sep 2014 | B2 |
8882847 | Burdulis, Jr. et al. | Nov 2014 | B2 |
8888485 | Ali | Nov 2014 | B2 |
8906107 | Bojarski et al. | Dec 2014 | B2 |
8926706 | Bojarski et al. | Jan 2015 | B2 |
8932363 | Tsougarakis et al. | Jan 2015 | B2 |
D722693 | Kaufmann et al. | Feb 2015 | S |
8945230 | Lang et al. | Feb 2015 | B2 |
8951259 | Fitz et al. | Feb 2015 | B2 |
8951260 | Lang et al. | Feb 2015 | B2 |
8965088 | Tsougarakis et al. | Feb 2015 | B2 |
8974539 | Bojarski et al. | Mar 2015 | B2 |
8998915 | Fitz et al. | Apr 2015 | B2 |
9020788 | Lang et al. | Apr 2015 | B2 |
9023050 | Lang et al. | May 2015 | B2 |
9034237 | Sperry et al. | May 2015 | B2 |
9055953 | Lang et al. | Jun 2015 | B2 |
9066728 | Burdulis, Jr. et al. | Jun 2015 | B2 |
9072531 | Fitz et al. | Jul 2015 | B2 |
9084617 | Lang et al. | Jul 2015 | B2 |
D735860 | Palinchik | Aug 2015 | S |
D736384 | Palinchik | Aug 2015 | S |
9095353 | Burdulis, Jr. et al. | Aug 2015 | B2 |
9107679 | Lang et al. | Aug 2015 | B2 |
9107680 | Fitz et al. | Aug 2015 | B2 |
9113921 | Lang et al. | Aug 2015 | B2 |
9125672 | Fitz et al. | Sep 2015 | B2 |
9125673 | Fitz et al. | Sep 2015 | B2 |
9144500 | Harding | Sep 2015 | B2 |
9180015 | Fitz et al. | Nov 2015 | B2 |
9180029 | Hollister et al. | Nov 2015 | B2 |
9186161 | Lang et al. | Nov 2015 | B2 |
9186254 | Fitz et al. | Nov 2015 | B2 |
9186257 | Geisler et al. | Nov 2015 | B2 |
D745159 | Lin | Dec 2015 | S |
9216025 | Fitz et al. | Dec 2015 | B2 |
9220516 | Lang et al. | Dec 2015 | B2 |
9220517 | Lang et al. | Dec 2015 | B2 |
D747485 | Oi | Jan 2016 | S |
9241724 | Lang et al. | Jan 2016 | B2 |
9241725 | Lang et al. | Jan 2016 | B2 |
9271845 | Hunt et al. | Mar 2016 | B2 |
9295482 | Fitz et al. | Mar 2016 | B2 |
9295562 | Lechmann et al. | Mar 2016 | B2 |
9308005 | Fitz et al. | Apr 2016 | B2 |
9308053 | Bojarski et al. | Apr 2016 | B2 |
9308060 | Ali | Apr 2016 | B2 |
9308091 | Lang | Apr 2016 | B2 |
9314256 | Fitz et al. | Apr 2016 | B2 |
9320620 | Bojarski et al. | Apr 2016 | B2 |
9295481 | Fitz et al. | May 2016 | B2 |
9326780 | Wong et al. | May 2016 | B2 |
9333058 | Krastev | May 2016 | B1 |
9339279 | Dubois et al. | May 2016 | B2 |
9358018 | Fitz et al. | Jun 2016 | B2 |
9364896 | Christensen et al. | Jun 2016 | B2 |
9370426 | Gabbrielli et al. | Jun 2016 | B2 |
9375222 | Fitz et al. | Jun 2016 | B2 |
9387079 | Bojarski et al. | Jun 2016 | B2 |
9381025 | Fitz et al. | Jul 2016 | B2 |
9402726 | Linderman et al. | Aug 2016 | B2 |
9408615 | Fitz et al. | Aug 2016 | B2 |
9408686 | Miller et al. | Aug 2016 | B1 |
9415137 | Meridew | Aug 2016 | B2 |
9421108 | Hunt | Aug 2016 | B2 |
D767137 | Lin | Sep 2016 | S |
9433510 | Lechmann et al. | Sep 2016 | B2 |
9433707 | Swords et al. | Sep 2016 | B2 |
9439767 | Bojarski et al. | Sep 2016 | B2 |
9486226 | Chao | Nov 2016 | B2 |
9495483 | Steines et al. | Nov 2016 | B2 |
9517134 | Lang | Dec 2016 | B2 |
9545317 | Hunt | Jan 2017 | B2 |
9549823 | Hunt et al. | Jan 2017 | B2 |
9561115 | Elahinia et al. | Feb 2017 | B2 |
9572669 | Hunt et al. | Feb 2017 | B2 |
9579110 | Bojarski et al. | Feb 2017 | B2 |
9597197 | Lechmann et al. | Mar 2017 | B2 |
9603711 | Bojarski et al. | Mar 2017 | B2 |
9610168 | Terrill | Apr 2017 | B2 |
9636226 | Hunt | May 2017 | B2 |
9636229 | Lang et al. | May 2017 | B2 |
9649178 | Ali | May 2017 | B2 |
9662157 | Schneider et al. | May 2017 | B2 |
9662226 | Wickham | May 2017 | B2 |
9668863 | Sharp et al. | Jun 2017 | B2 |
9675465 | Padovani et al. | Jun 2017 | B2 |
9675471 | Bojarski et al. | Jun 2017 | B2 |
9681956 | Al Hares et al. | Jun 2017 | B2 |
9687945 | Steines et al. | Jun 2017 | B2 |
9688026 | Ho et al. | Jun 2017 | B2 |
9694541 | Pruett et al. | Jul 2017 | B2 |
9700420 | Fitz et al. | Jul 2017 | B2 |
9700971 | Lang | Jul 2017 | B2 |
9715563 | Schroeder | Jul 2017 | B1 |
9737367 | Steines et al. | Aug 2017 | B2 |
9750613 | Petteys | Sep 2017 | B2 |
9757235 | Hunt et al. | Sep 2017 | B2 |
9757245 | O'Neil et al. | Sep 2017 | B2 |
9775680 | Bojarski et al. | Oct 2017 | B2 |
9782270 | Wickham | Oct 2017 | B2 |
9788972 | Flickinger et al. | Oct 2017 | B2 |
9849019 | Miller et al. | Dec 2017 | B2 |
9872773 | Lang et al. | Jan 2018 | B2 |
9877790 | Bojarski et al. | Jan 2018 | B2 |
D809661 | Mueller et al. | Feb 2018 | S |
9907670 | Deridder et al. | Mar 2018 | B2 |
9910935 | Golway et al. | Mar 2018 | B2 |
9913723 | Fitz et al. | Mar 2018 | B2 |
9918849 | Morris et al. | Mar 2018 | B2 |
9925054 | Siegler | Mar 2018 | B2 |
9943370 | Asseln et al. | Apr 2018 | B2 |
9943627 | Zhou et al. | Apr 2018 | B2 |
9949839 | Sander | Apr 2018 | B2 |
9956047 | Bojarski et al. | May 2018 | B2 |
9956048 | Bojarski et al. | May 2018 | B2 |
D829909 | Horton | Oct 2018 | S |
10085839 | Wong et al. | Oct 2018 | B2 |
D835278 | Gottlieb | Dec 2018 | S |
10183442 | Miller | Jan 2019 | B1 |
10245152 | Kloss | Apr 2019 | B2 |
D849944 | Dacosta | May 2019 | S |
10278823 | Xue | May 2019 | B1 |
D850620 | Tyber | Jun 2019 | S |
D855184 | Predick | Jul 2019 | S |
10357377 | Nyahay | Jul 2019 | B2 |
D858769 | Barela et al. | Sep 2019 | S |
10449051 | Hamzey | Oct 2019 | B2 |
10492686 | Hunter | Dec 2019 | B2 |
D877907 | Linder et al. | Mar 2020 | S |
D878589 | Linder | Mar 2020 | S |
D878590 | Linder et al. | Mar 2020 | S |
D879295 | Abbasi | Mar 2020 | S |
D879961 | Linder et al. | Mar 2020 | S |
D881665 | Zemel et al. | Apr 2020 | S |
10624746 | Jones et al. | Apr 2020 | B2 |
10667924 | Nyahay | Jun 2020 | B2 |
10744001 | Sack | Aug 2020 | B2 |
10772732 | Miller et al. | Sep 2020 | B1 |
D899900 | Blanco | Oct 2020 | S |
10940015 | Sack | Mar 2021 | B2 |
D920515 | Miller | May 2021 | S |
D920516 | Miller | May 2021 | S |
D920517 | Miller | May 2021 | S |
11026798 | Miller | Jun 2021 | B1 |
11033394 | Hamzey | Jun 2021 | B2 |
11135771 | Reith | Oct 2021 | B1 |
D938033 | Dang | Dec 2021 | S |
11324525 | Garvey | May 2022 | B1 |
11353277 | Muceus | Jun 2022 | B2 |
11439726 | Spence | Sep 2022 | B2 |
11471203 | Sutika | Oct 2022 | B2 |
11666452 | Melkent et al. | Jun 2023 | B2 |
11744716 | Jebsen et al. | Sep 2023 | B2 |
20040148032 | Rutter et al. | Jul 2004 | A1 |
20060249875 | Robb et al. | Nov 2006 | A1 |
20070100346 | Wyss | May 2007 | A1 |
20070118243 | Schroeder et al. | May 2007 | A1 |
20080206297 | Roeder et al. | Aug 2008 | A1 |
20090093668 | Marten et al. | Apr 2009 | A1 |
20090182430 | Tyber et al. | Jul 2009 | A1 |
20100137990 | Apatsidis et al. | Jun 2010 | A1 |
20100168798 | Clineff et al. | Jul 2010 | A1 |
20100286791 | Goldsmith | Nov 2010 | A1 |
20110144752 | Defelice et al. | Jun 2011 | A1 |
20110190898 | Lenz | Aug 2011 | A1 |
20110224796 | Weiland et al. | Sep 2011 | A1 |
20110230974 | Musani | Sep 2011 | A1 |
20120064288 | Nakano et al. | Mar 2012 | A1 |
20120215310 | Sharp et al. | Aug 2012 | A1 |
20130068968 | Daniel | Mar 2013 | A1 |
20130123935 | Hunt et al. | May 2013 | A1 |
20130158651 | Hollister et al. | Jun 2013 | A1 |
20130197657 | Anca et al. | Aug 2013 | A1 |
20130218282 | Hunt | Aug 2013 | A1 |
20130274890 | McKay | Oct 2013 | A1 |
20130296874 | Chao | Nov 2013 | A1 |
20140100779 | Tuke | Apr 2014 | A1 |
20140107785 | Geisler et al. | Apr 2014 | A1 |
20140107786 | Geisler et al. | Apr 2014 | A1 |
20140236299 | Roeder et al. | Aug 2014 | A1 |
20140277443 | Fleury et al. | Sep 2014 | A1 |
20140277452 | Skaer | Sep 2014 | A1 |
20140288650 | Hunt | Sep 2014 | A1 |
20140336680 | Medina et al. | Nov 2014 | A1 |
20140350688 | Michel | Nov 2014 | A1 |
20140371863 | Vanasse et al. | Dec 2014 | A1 |
20150025666 | Olivieri | Jan 2015 | A1 |
20150105858 | Papay et al. | Apr 2015 | A1 |
20150282945 | Hunt | Oct 2015 | A1 |
20150282946 | Hunt | Oct 2015 | A1 |
20150320461 | Ehmke | Nov 2015 | A1 |
20150335434 | Patterson et al. | Nov 2015 | A1 |
20150343709 | Gerstle et al. | Dec 2015 | A1 |
20150351915 | Defelice et al. | Dec 2015 | A1 |
20160008139 | Siegler | Jan 2016 | A1 |
20160051371 | Defelice et al. | Feb 2016 | A1 |
20160089138 | Early et al. | Mar 2016 | A1 |
20160151833 | Tsao | Jun 2016 | A1 |
20160193055 | Ries | Jul 2016 | A1 |
20160199193 | Willis et al. | Jul 2016 | A1 |
20160213485 | Schaufler et al. | Jul 2016 | A1 |
20160213486 | Nunley et al. | Jul 2016 | A1 |
20160213487 | Wilson et al. | Jul 2016 | A1 |
20160213488 | Moore et al. | Jul 2016 | A1 |
20160220288 | Dubois et al. | Aug 2016 | A1 |
20160256279 | Sanders et al. | Sep 2016 | A1 |
20160256610 | Zhou et al. | Sep 2016 | A1 |
20160270931 | Trieu | Sep 2016 | A1 |
20160287388 | Hunt et al. | Oct 2016 | A1 |
20160303793 | Ermoshkin et al. | Oct 2016 | A1 |
20160333152 | Cook et al. | Nov 2016 | A1 |
20160374829 | Vogt et al. | Dec 2016 | A1 |
20170014169 | Dean et al. | Jan 2017 | A1 |
20170020685 | Geisler et al. | Jan 2017 | A1 |
20170036403 | Ruff et al. | Feb 2017 | A1 |
20170042697 | McShane, III et al. | Feb 2017 | A1 |
20170056178 | Sharp et al. | Mar 2017 | A1 |
20170056179 | Lorio | Mar 2017 | A1 |
20170066873 | Gardet | Mar 2017 | A1 |
20170105844 | Kuyler et al. | Apr 2017 | A1 |
20170156880 | Halverson et al. | Jun 2017 | A1 |
20170165085 | Lechmann et al. | Jun 2017 | A1 |
20170165790 | McCarthy et al. | Jun 2017 | A1 |
20170172758 | Field et al. | Jun 2017 | A1 |
20170182222 | Paddock et al. | Jun 2017 | A1 |
20170018919 | Reiley | Jul 2017 | A1 |
20170209274 | Beerens et al. | Jul 2017 | A1 |
20170216035 | Hunt | Aug 2017 | A1 |
20170216036 | Cordaro | Aug 2017 | A1 |
20170239054 | Engstrand et al. | Aug 2017 | A1 |
20170239064 | Cordaro | Aug 2017 | A1 |
20170245998 | Padovani et al. | Aug 2017 | A1 |
20170252165 | Sharp et al. | Sep 2017 | A1 |
20170258606 | Afzal | Sep 2017 | A1 |
20170282455 | Defelice et al. | Oct 2017 | A1 |
20170296244 | Schneider et al. | Oct 2017 | A1 |
20170319344 | Hunt | Nov 2017 | A1 |
20170323037 | Schroeder | Nov 2017 | A1 |
20170333205 | Joly et al. | Nov 2017 | A1 |
20170354510 | O'Neil et al. | Dec 2017 | A1 |
20170354513 | Maglaras et al. | Dec 2017 | A1 |
20170355815 | Becker et al. | Dec 2017 | A1 |
20170360488 | Kowalczyk et al. | Dec 2017 | A1 |
20170360563 | Hunt et al. | Dec 2017 | A1 |
20170360578 | Shin et al. | Dec 2017 | A1 |
20170367843 | Eisen et al. | Dec 2017 | A1 |
20170367844 | Eisen et al. | Dec 2017 | A1 |
20170367845 | Eisen et al. | Dec 2017 | A1 |
20180008419 | Tyber et al. | Jan 2018 | A1 |
20180012517 | Ropelato | Jan 2018 | A1 |
20180022017 | Fukumoto et al. | Jan 2018 | A1 |
20180064540 | Hunt | Mar 2018 | A1 |
20180098858 | Valderraband | Apr 2018 | A1 |
20180104063 | Asaad | Apr 2018 | A1 |
20180110593 | Khalil | Apr 2018 | A1 |
20180110626 | McShane, III et al. | Apr 2018 | A1 |
20180110627 | Sack | Apr 2018 | A1 |
20180117219 | Yang et al. | May 2018 | A1 |
20180147319 | Colucci-Mizenko et al. | May 2018 | A1 |
20180256336 | Mueller et al. | Sep 2018 | A1 |
20180280140 | Jones | Oct 2018 | A1 |
20180289380 | Mauldin | Oct 2018 | A1 |
20180289515 | Nemes et al. | Oct 2018 | A1 |
20190167433 | Allen | Jun 2019 | A1 |
20190262101 | Shanjani et al. | Aug 2019 | A1 |
20190302736 | Chanin | Oct 2019 | A1 |
20190343652 | Petersheim et al. | Nov 2019 | A1 |
20200000595 | Jones | Jan 2020 | A1 |
20200030102 | Mullens et al. | Jan 2020 | A1 |
20200046512 | Newman et al. | Feb 2020 | A1 |
20200085452 | Siegler | Mar 2020 | A1 |
20200085585 | Siegler | Mar 2020 | A1 |
20200155321 | Dikovsky | May 2020 | A1 |
20200171752 | Rogren | Jun 2020 | A1 |
20200171753 | Satko | Jun 2020 | A1 |
20200367910 | Hafez et al. | Nov 2020 | A1 |
20210077276 | Garvey et al. | Mar 2021 | A1 |
20210110605 | Haslam | Apr 2021 | A1 |
20210216683 | Rai | Jul 2021 | A1 |
20210298908 | Holmes | Sep 2021 | A1 |
20210340334 | Portela | Nov 2021 | A1 |
20220023048 | Nolens | Jan 2022 | A1 |
20220087670 | Selmoune | Mar 2022 | A1 |
20220134639 | Allen | May 2022 | A1 |
20220142783 | Ahmadi | May 2022 | A1 |
20220168109 | Giordano | Jun 2022 | A1 |
20220296386 | Fang | Sep 2022 | A1 |
Number | Date | Country |
---|---|---|
109567913 | Apr 2019 | CN |
110090096 | Aug 2019 | CN |
69806985 | Jun 2003 | DE |
1180989 | Apr 2006 | EP |
2832321 | Feb 2015 | EP |
2635239 | Jul 2017 | EP |
2913030 | Mar 2018 | EP |
3586800 | Jan 2020 | EP |
3071400 | Mar 2019 | FR |
4840886 | Dec 2011 | JP |
301007894 | May 2019 | KR |
2014020562 | Feb 2014 | WO |
2015054070 | Apr 2015 | WO |
2020123295 | Jun 2020 | WO |
Entry |
---|
Flaxman et al., A review and guide to creating patient specific 3D printed anatomical models from MRI for benign gynecologic surgery, 2021, 3D Printing in Medicine, pp. 1-10. (Year: 2021). |
Extended European Search Report dated Feb. 12, 2021 for European Patent Application No. EP20196410.3. |
Larraona et al., “Radiopaque material for 3D printing scaffolds”, XXXV Confreso Anual de la Sociedad Espanola de Ingenieria Biomedica. Bilbao, Nov. 29-Dec. 1, 2017, p. 451-454 (Year: 2017). |
Rozema et al., The effects of different steam-sterilization programs on material properties of poly(I-lactide), Journal of Applied Biomaterials, vol. 2, 23-28 (1991) (Year: 1991). |
Alt, Sami. “Design for Sterilization Part 1: Steam Sterillization.” Material, Material Technology Blog, Jun. 3, 2016, www.material-technology.com/single-post/2016/05/24/Design-for-Sterilization-part-1-Steam-Sterillization. |
Ducheyne, Paul. “Comprehensive Biomaterials.” Comprehensive Biomaterials, vol. 1, Elsevier, 2011, pp. 135-135. |
Anat Ratnovsky et al., Mechanical Properties of Different Airway Stents, Med. Eng'g. Physics, Mar. 2011, at 408., http://www.medengphys.com/article/S1350-4533(15)00042-9/fulltext. |
Andrew T. Miller et al., Fatigue of Injection Molded and 30 Printed Polycarbonate Urethane in Solution, 108 Polymer 121 (2017). |
Andrew T. Miller et al., Deformation and Fatigue of Tough 30 Printed Elastomer Scaffolds Processed by Fused 3 Deposition Modeling and Continuous Liquid Interface Production, 75 J. Mechanical Behavior Biomedical Materials 1 (2017). |
Ortho Spine News, “SeaSpine Announces 25,000th NanoMetalene Implantation”, first available Dec. 18, 2019. (https://orthospinenews.com/2019/12/18/seaspine-announces-25000th-nanometalene-implantation/) (Year: 2019). |
Restor3d, “Products”, first available Sep. 28, 2020. (https://web.archive.org/web/20200928123335/https:/restor3d.com/products) (Year: 2020). |
Ortho Spine News, “Nvision Biomedical Technologies: First FDA Clearance for Osteotomy Wedge System”, first available Oct. 28, 2020. (https://orthospinenews.com/2020/10/28/nvision-biomedical-technologies-first-fda-clearance-for-osteotomy-wedge-system-made-of-peek-optima-ha-enhanced/) (Year: 2020). |
Sina, “Application logic of triple periodic minimum surface”, first available Oct. 24, 2020. (https://k.sina.com.cnarticle_2422410454_90630cd6001 OOtlbm.html?from=science) (Year: 2020). |
3D Adept Media, “Johnson & Johnson Medical”, first available Sep. 17, 2018. (https://3dadept.com/johnson-johnson-medical-has-acquired-3d-printed-spmplants-special ist-emerging-implant-technologies/) (Year: 2018). |
Additive Orthopaedics, “Additive Orthopaedics 3d Printed Cotton Bone Segment”, first available Sep. 19, 2020. (https://web.archive.org/web/20200919145251/https:/www.additiveorthopaedics.com/our-products/cotton/) (Year: 2020). |
Indiamart, “Anterior Cervical Fusion Cage for Spine Surgery”, first accessed Dec. 9, 2020. (https://www.indiamart.com/ proddetail/anterior-cervical-fusion-cage-12402896897 .html) (Year: 2020). |
Instagram, “restor3d”, first available Jul. 21, 2020. (https://www.instagram.com/p/CC6dztOAKcM/?utm_source=ig_web_link) (Year: 2020). |