In most general purpose computers, an operating system is the primary software that manages access to resources within the computer. The primary resources are the central processing unit (CPU), which executes application programs designed to run on the computer, main memory and storage. In some computer architectures, additional processing units (such as multiple cores in a processor) and/or additional processors, called co-processors, may be present. Examples of such co-processors include a graphic processing unit (GPU) and a digital signal processor (DSP). The operating system also manages access to these resources by multiple processes.
A field programmable gate array (FPGA) is a kind of logic device that is commonly used in specialized computing devices. An FPGA typically is used to perform a specific, dedicated function, for which a gate array is particularly well-suited. FPGAs typically are found in peripheral devices, or other specialized hardware, such as a printed circuit board connected to and accessed through a system bus such as a PCI bus. In general, such devices are programmed once, and used many times. Because these devices are programmable, they have an advantage over other specialized logic devices in that they can be updated in the field.
This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used to limit the scope of the claimed subject matter.
One or more field programmable gate arrays (FPGA) can be used as a shared programmable co-processor resource in a general purpose computing system. An FPGA can be programmed to perform functions, which in turn can be associated with one or more processes. With multiple processes, the FPGA can be shared, and a process is assigned to at least one portion of the FPGA during a time slot in which to access the FPGA. Programs written in a hardware description language for programming the FPGA are made available as a hardware library. The operating system manages allocating the FPGA resources to processes, programming the FPGA in accordance with the functions to be performed by the processes using the FPGA, and scheduling use of the FPGA by these processes.
The use of a hardware library by an application can be explicit or the application code can be analyzed to determine if a hardware library could accelerate its execution. In particular, application code can be analyzed to identify calls to application programming interfaces (APIs) or other functions that have a hardware library implementation. The code can be analyzed to identify the frequency of such calls. Information from the hardware library can indicate characteristics of the library, such as its size, power consumption and FPGA resource usage. Information about the execution pattern of the application code also can be useful. This information, along with information about other concurrent processes using the FPGA resources, can be used to select a hardware library to implement functions called in the application code.
In the following description, reference is made to the accompanying drawings which form a part hereof, and in which are shown, by way of illustration, specific example implementations of this technique. It is understood that other embodiments may be utilized and structural changes may be made without departing from the scope of the disclosure.
The following section provides a brief, general description of an example computing environment in which an operating system for managing use of FPGA resources can be implemented. The system can be implemented with numerous general purpose or special purpose computing devices. Examples of well known computing devices that may be suitable include, but are not limited to, personal computers, server computers, hand-held or laptop devices (for example, media players, notebook computers, cellular phones, personal data assistants, voice recorders), multiprocessor systems, microprocessor-based systems, set top boxes, game consoles, programmable consumer electronics, network PCs, minicomputers, mainframe computers, distributed computing environments that include any of the above systems or devices, and the like.
With reference to
The computing device may include multiple processing units and/or additional co-processing units such as a graphics processing unit (GPU). The computing device also includes one or more field programmable gate arrays (FPGA), denoted as FPGA unit 120 which is available as a shared (among processes running on the computer) co-processing resource. An FPGA may reside in its own CPU socket or on a separate card plugged into an expansion slot, such as a Peripheral Component Interconnect Express (PCI-E) slot. By providing such an FPGA unit, a variety of functions that are well-suited for implementation by a gate array can be implemented with the resulting benefit of hardware acceleration.
Depending on the configuration of the processing unit and the FPGA unit, the unit, or each functional unit within it, has an associated input/output channel for communication with host operating system processes. For example, a memory region dedicated to the functional unit and shared between it and a process using that functional unit can be provided. A sort of request queue and response queue also can be used to enable asynchronous invocation of operations implemented in the FPGA unit. Additionally, state of the functional units in the FPGA unit for a process can be saved to and restored from a memory region for the functional unit and that process. Alternatively other techniques can be used to ensure that the functional unit is in a known state before it is used by its process.
Depending on the configuration and type of computing device, memory 104 may be volatile (such as RAM), non-volatile (such as ROM, flash memory, etc.) or some combination of the two. This configuration of a processing unit, co-processor and memory is illustrated in
Computing device 100 may also have additional resources and devices. For example, computing device 100 may include additional storage (removable and/or non-removable) including, but not limited to, magnetic or optical disks or tape. Such additional storage is illustrated in
Computing device 100 also can include communications connection(s) 112 that allow the device to communicate with other devices over a communication medium. The implementation of the communications connection 112 is dependent on the kind of communication medium being accessed by the computing device, as it provides an interface to such a medium to permit transmission and/or reception of data over the communication medium. A communication medium typically carries computer program instructions, data files, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism and includes any information delivery media. The term “modulated data signal” means a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal. By way of example, and not limitation, communication media includes wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, RF, infrared and other wireless media.
Computing device 100 may have various input device(s) 114 such as a keyboard, mouse, pen, camera, touch input device, and so on. Output device(s) 116 such as a display, speakers, a printer, and so on may also be included. All of these devices are well known in the art and need not be discussed at length here.
Applications executed on a computing device are implemented using computer-executable instructions and/or computer-interpreted instructions, such as program modules, that are processed by the computing device. Generally, program modules include routines, programs, objects, components, data structures, and so on, that, when processed by a processing unit, instruct the processing unit to perform particular tasks or implement particular abstract data types. In a distributed computing environment, such tasks can be performed by remote processing devices that are linked through a communications network. In a distributed computing environment, program modules may be located in both local and remote computer storage media including memory storage devices.
An operating system executed on a computing device manages access to the various resources of the computer device by processes. Typically, running an application on the computer system causes one or more processes to be created, with each process being allocated to different resources over time. If a resource is shared among processes, and if the processes cannot share the resource concurrently, then the operating system schedules access to the resource over time. One of such resources is the FPGA unit 120 of
Referring to
To illustrate how different functional units can be used over time, reference is now made to
To allow such usage of the FPGA resources by different processes over time, the operating system has a scheduler that determines which process has access to the FPGA resources at each scheduling quantum, i.e., time period, and when an FPGA functional unit will be programmed with a hardware library so that the functional unit is available to be used by that process. Thus, an implementation of a scheduler for the FPGA unit is dependent in part on the nature of the FPGA unit and the one or more FPGAs it includes. Factors related to the FPGAs to be considered include, but are not limited to, the following. For example, in some cases an entire FPGA is refreshed to program a functional unit if one functional unit cannot be programmed independently of other functional units. Another consideration is the speed with which a functional unit can be programmed, and whether programming of a functional unit prevents other functional units from being used during that programming phase. Another factor to consider is whether processes can share a hardware library by sharing a functional unit. The scheduler also takes into account such factors as the number of concurrent processes, application performance guarantees, priority of applications, process context switching costs, access to memory and buses, and availability of software libraries if no functional unit is available within the FPGA unit.
There may be other instances where the FPGA unit provides a general purpose facility to applications or the operating system, which therefore are scheduled for the length of an application instantiation. For example, custom network protocols or offloading can be offered as an accelerated service on the FPGA unit. System calls or standard library calls, normally executed in a general purpose CPU, can be accelerated using the FPGA unit instead. Further, the operating system can multiplex the CPU based on preferences for process priority. In another instance, the operating system can use a profile of an application, generated statically or dynamically, to predict the functionality best suited for running on an FPGA unit and then pre-load that functionality so that it is available for scheduling. By using the profile as a guide, the operating system can ensure there is both space and time available on the FPGA unit to accelerate the application. Finally, the operating system can use simple hints from the application to know when to schedule time on the FPGA unit. For example, certain calls into the operating system (system calls) can denote long delays (calls to disk or the network), which provides a hint that the FPGA unit can be free for some amount of time for other threads or processes to use. Therefore, the operating system uses a variety of hints and preferences to create a schedule to multiplex access to the FPGA unit. Because the operating system controls the scheduler, it has detailed knowledge of executing and pending work, available hardware libraries, and time it takes to program an FPGA. Therefore, it can use this knowledge to determine which processes leverage the FPGA during execution.
Having now described a general overview of such computer architecture, an example implementation of code profiling will now be described.
Referring now to
A selection module 510 receives the information 506 about the hardware libraries and the information 508 about the calls made the libraries, to make a selection about which hardware libraries are to be used. The selection module also can receive data 522 from a scheduler 520, for example about other concurrent processes, to assist in this selection. Other data about the execution pattern 524 from prior executions of the application code 502 also can be used to assist in this selection. Given the input information, the selection module provides an output 512 indicating the hardware libraries that have been selected to assist in accelerating processing by the application code 502.
The code analyzer can be implemented in any of a variety of conventional ways for identifying calls to a library. In this case, the calls are being made to a library with a hardware implementation. In one example, the code analyzer searches the application code for references to functions or objects having implementations in the hardware library, and outputs information about those hardware libraries. In another example, the code analyzer identifies traps into the operating system, which denote system calls. If operating system functionality has been offloaded onto an FPGA, the traps can be redirected to different code paths within the operating system by rewriting the registers used to pass data into the kernel with hints or appropriate data for a call to an FPGA.
A flowchart for an example implementation of the selection module will now be described in connection with
Any specific process to implement such selection is dependent on the application and type of FPGA resources being used. In the simplest case, the hardware libraries can be selected for use if there are no conflicts with other processes in using the FPGA resources, and if the projected utilization of the FPGA resources is estimated to provide better acceleration of the application than if a software library were used instead.
The terms “article of manufacture”, “process”, “machine” and “composition of matter” in the preambles of the appended claims are intended to limit the claims to subject matter deemed to fall within the scope of patentable subject matter defined by the use of these terms in 35 U.S.C. §101.
Any or all of the aforementioned alternate embodiments described herein may be used in any combination desired to form additional hybrid embodiments. It should be understood that the subject matter defined in the appended claims is not necessarily limited to the specific implementations described above. The specific implementations described above are disclosed as examples only.
Number | Name | Date | Kind |
---|---|---|---|
5748979 | Trimberger | May 1998 | A |
5752035 | Trimberger | May 1998 | A |
5915025 | Taguchi et al. | Jun 1999 | A |
6557156 | Guccione | Apr 2003 | B1 |
6704816 | Burke | Mar 2004 | B1 |
6823069 | Kitajima | Nov 2004 | B1 |
6907126 | Inada | Jun 2005 | B2 |
6941538 | Hwang et al. | Sep 2005 | B2 |
7028283 | Keller et al. | Apr 2006 | B1 |
7134025 | Trimberger | Nov 2006 | B1 |
7512813 | Goodnow | Mar 2009 | B2 |
7587699 | McCubbrey | Sep 2009 | B2 |
7669168 | Patterson | Feb 2010 | B1 |
7702927 | Devadas et al. | Apr 2010 | B2 |
7711964 | Van Essen et al. | May 2010 | B2 |
7788502 | Donlin et al. | Aug 2010 | B1 |
8065517 | Cizas et al. | Nov 2011 | B2 |
8230374 | McCubbrey | Jul 2012 | B2 |
8369460 | Su | Feb 2013 | B1 |
8417965 | Sundararajan et al. | Apr 2013 | B1 |
8448122 | Suthar et al. | May 2013 | B1 |
8516268 | Woodall | Aug 2013 | B2 |
20010037457 | Inada | Nov 2001 | A1 |
20010043082 | Wittig et al. | Nov 2001 | A1 |
20020055947 | Schultz | May 2002 | A1 |
20030086300 | Noyes et al. | May 2003 | A1 |
20030110463 | Kuhlmann et al. | Jun 2003 | A1 |
20030172303 | Adusumilli | Sep 2003 | A1 |
20040019765 | Klein, Jr. | Jan 2004 | A1 |
20040049672 | Nollet et al. | Mar 2004 | A1 |
20040060032 | McCubbrey | Mar 2004 | A1 |
20040123258 | Butts | Jun 2004 | A1 |
20040230934 | Taylor | Nov 2004 | A1 |
20060015313 | Wang et al. | Jan 2006 | A1 |
20060015862 | Odom et al. | Jan 2006 | A1 |
20060059373 | Fayad et al. | Mar 2006 | A1 |
20060156406 | Bantz et al. | Jul 2006 | A1 |
20070074045 | Van Essen et al. | Mar 2007 | A1 |
20070129818 | Andrade et al. | Jun 2007 | A1 |
20070277161 | Herbordt et al. | Nov 2007 | A1 |
20080104601 | Kaneko et al. | May 2008 | A1 |
20080133899 | Park et al. | Jun 2008 | A1 |
20090119503 | Isaakian et al. | May 2009 | A1 |
20090282386 | Moir et al. | Nov 2009 | A1 |
20090288076 | Johnson et al. | Nov 2009 | A1 |
20090293051 | Krywaniuk | Nov 2009 | A1 |
20100202239 | Moshayedi et al. | Aug 2010 | A1 |
20100293356 | Plunkett et al. | Nov 2010 | A1 |
20110145780 | Chen | Jun 2011 | A1 |
20110153981 | Yancey et al. | Jun 2011 | A1 |
20120047371 | Woodall | Feb 2012 | A1 |
20120117549 | Doyle et al. | May 2012 | A1 |
20120191967 | Lin et al. | Jul 2012 | A1 |
20130346669 | Nightingale et al. | Dec 2013 | A1 |
20130346758 | LaMacchia et al. | Dec 2013 | A1 |
20130346759 | LaMacchia et al. | Dec 2013 | A1 |
20130346985 | Nightingale | Dec 2013 | A1 |
Number | Date | Country |
---|---|---|
1284681 | Feb 2001 | CN |
2650231 | Oct 2004 | CN |
101222743 | Jul 2008 | CN |
201371945 | Dec 2009 | CN |
102087600 | Jun 2011 | CN |
102119390 | Jul 2011 | CN |
102324006 | Jan 2012 | CN |
102377564 | Mar 2012 | CN |
1930834 | Jun 2008 | EP |
Entry |
---|
Frigo et al.; “Evaluation of the Streams-C C-to-FPGA Compiler: An Applications Perspective”, Feb. 13, 2001, Copyright 2001 ACM, (Frigo—2001.pdf; pp. 1-7). |
Huang et al.; “Function-Level Multitasking Interface Design in an Embedded OS with Reconfigurable Hardware”, IFIP International Federation for imformation processing 2007, Yuan Ze University, Taiwan. (Huang—2007.pdf; pp. 1-10). |
Lysecky et al.; “Dynamic FPGA Routing for Just-in-Time FPGA compilation” University of California, Jun. 11, 2004; (Lysecky—2004.pdf; pp. 1-6). |
David Max Meisner; “Design of a shared hardware library for multi-core environments in FPGA fabrics”, Honor Thesis busmitted to Brown University, Apr. 24, 2007; (Meisner—2007.pdf; pp. 1-48). |
Shibamura et al.; “A Dyanamically Reconfigurable Platform using Embedded Processor FPGA”, Kumamoto University, Japan, Copyright 2004 IEEE (Shibamura—2004.pdf; pp. 1-8). |
Jain et al. “Speeding Up Program Execution Using Reconfigurable Hardware and a Hardware Function Library”, IEEE 1997, (Jain—1997.pdf; pp. 1-6). |
Notice of Allowance dated Jul. 29, 2014 cited in U.S. Appl. No. 13/528,438. |
Marescaux, et al., “Run-Time Support for Heterogeneous Multitasking on Reconfigurable SoCs”, In Integration, the VLSI Journal—Special Issue: Networks on Chip and Reconfigurable Fabrics, vol. 38, Issue 1, Oct. 2004, 24 pages. |
Vuletic, et al., “Seamless Hardware-Software Integration in Reconfigurable Computing Systems”, In IEEE Design & Test of Computers, vol. 22, Issue 2, Mar. 2005, 12 pages. |
“International Search Report & Written Opinion for PCT Patent Application No. PCT/US2013/046418”, Mailed Date: Sep. 11, 2013, Filed Date: Jun. 18, 2013, 11 pages. |
“International Search Report & Written Opinion for PCT Patent Application No. PCT/US2013/046719”, Mailed Date: Sep. 11, 2013, Filed Date: Jun. 20, 2013, 10 pages. |
“International Search Report & Written Opinion for PCT patent Application No. PCT/US2013/046881”, Mailed Date: Nov. 29, 2013, Filed Date: Jun. 20, 2013, 9 pages. |
“International Search Report & Written Opinion for PCT Patent Application No. PCT/US2013/046871”, Mailed Date: Oct. 15, 2013, Filed Date: Jun. 20, 2013, 9 pages. |
U.S. Appl. No. 13/528,175, filed Apr. 11, 2014, Office Action. |
U.S. Appl. No. 13/528,329, filed Oct. 16, 2013, Office Action. |
U.S. Appl. No. 13/528,400, filed Sep. 13, 2013, Office Action. |
U.S. Appl. No. 13/528,400, filed Jun. 16, 2014, Office Action. |
U.S. Appl. No. 13/528,438, filed Sep. 16, 2013, Office Action. |
U.S. Appl. No. 13/528,438, Apr. 16, 2014, Notice of Allowance. |
Office Action dated Aug. 15, 2014 cited in U.S. Appl. No. 13/528,329. |
Office Action dated Nov. 7, 2014 cited in U.S. Appl. No. 13/5289,400. |
Office Action dated Oct. 17, 2014 cited in U.S. Appl. No. 13/528,175. |
Office Action dated Feb. 13, 2015 cited in U.S. Appl. No. 13/528,175. |
Office Action dated Feb. 23, 2015 cited in U.S. Appl. No. 13/528,329. |
“First Office Action Issued in Chinese Patent Application No. 201310248192.3, Mailed Date: Oct. 10, 2015, 14 Pages.” |
Notice of Allowance dated Aug. 31, 2015 cited in U.S. Appl. No. 13/528,400. |
Notice of Allowance dated May 13, 2015 cited in U.S. Appl. No. 13/528,400. |
Office Action dated Jun. 4, 2015 cited in U.S. Appl. No. 13/528,329. |
Office Action dated Oct. 26, 2015 U.S. Appl. No. 13/528,175. |
Notice of Allowance dated Nov. 16, 2015 U.S. Appl. No. 13/528,329. |
First Office Action Received in Chinese Patent Application No. 201310245064.3, Mailed Date: Dec. 24, 2015, 14 Pages. |
“First Office Action and Search Report Issued in Chinese Patent Application No. 201310248194.2”, Mailed Date: Jan. 13, 2016, 13 Pages. |
Number | Date | Country | |
---|---|---|---|
20130346979 A1 | Dec 2013 | US |