The present invention relates to automatic memory management of large object areas, and more particularly to a method and system for performing garbage collection in large object areas by remapping virtual memory addresses.
Conventional compaction of free space in garbage collection that uses memory copying has a negative performance impact by requiring large amounts of memory to be shifted. These known memory copying compaction techniques are expensive and cause significant paging activity with associated memory thrashing. Other conventional compaction methods are inefficient, including those that rely on virtual memory to exploit page faults to implement operating system supported read and write barriers. Furthermore, a known compaction method uses virtual memory management to decommit free pages and commit uncommitted pages, but the resulting number of pages needed, the use of three types of pages (i.e., free, live, and uncommitted), and the intermingling of uncommitted pages with live pages indicate deficiencies in complexity, effectiveness, and efficiency (e.g., additional management is required for the interspersed uncommitted pages and the need to reserve the uncommitted pages requires a reduction in the maximum heap size). Thus, there exists a need to overcome at least one of the preceding deficiencies and limitations of the related art.
The present invention may provide a computer-implemented method of batched remapping of virtual memory addresses for garbage collection in large object areas. The method comprises:
determining a mapping from a first table to a second table, wherein the first table includes a plurality of identifiers of a plurality of page-aligned large objects included in a large object area of a virtual memory, a first plurality of virtual memory addresses to which the page-aligned large objects are initially assigned, and a plurality of virtual memory sizes of the page-aligned large objects, wherein the second table includes a second plurality of virtual memory addresses to which the page-aligned large objects are assigned for a compaction of the plurality of page-aligned large objects, and wherein the page-aligned large objects are not contiguous in the large object area;
receiving a request in a single batch, wherein the request includes the second plurality of virtual memory addresses and requests a remapping of the plurality of page-aligned large objects to the second plurality of virtual memory addresses;
validating the second plurality of virtual memory addresses included in the request; and
remapping the plurality of page-aligned large objects to the second plurality of virtual memory addresses according to the request, wherein a result of the remapping is the compaction that includes the plurality of page-aligned large objects being contiguous in the large object area, and wherein the remapping is performed by a processor of a computer system.
A system and computer program product corresponding to the above-summarized method are also described and claimed herein.
One or more embodiments of the present invention advantageously disclose the use of virtual memory management to reduce fragmentation without requiring the copying of data in memory.
One or more embodiments of the present invention provide a compaction method and system that uses virtual memory management to change virtual memory address ranges in a large object area to align contiguous locations of free virtual memory and to align contiguous locations of live virtual memory by adjusting virtual memory mapping on physical memory. Aligning free virtual memory as contiguous locations and aligning live virtual memory as contiguous locations may include using remapping operations to swap free and live pages. The compaction technique disclosed herein may reduce fragmentation without requiring the resource-intensive copying of data in memory. One embodiment of the compaction method and system disclosed herein utilizes a kernel memory management driver (a.k.a. kernel driver or kernel mode driver) so that the virtual memory mapping adjustments are abstracted to applications. The kernel driver may allow multiple virtual memory remappings to be performed in a single kernel transition so that page-aligned large objects become contiguous in a large object area. The kernel driver may have direct access to underlying kernel application programming interfaces (APIs).
Garbage collector 110 performs automatic memory management that includes identifying objects in a program that cannot be accessed in the future by the program and reclaiming the memory used by the identified objects. Every object has a unique identifier and has a contiguous location in virtual memory. The number of bytes required by an object is arbitrary, but the memory manager rounds up the allocation size to a multiple of the virtual memory page size. Therefore, the virtual size of an object is always greater than or equal to the actual size of the object. A memory manager (not shown) page aligns large objects in a large object area (LOA) in preparation for a compaction phase of garbage collection that includes remapping virtual memory addresses of the large objects. As used herein, a large object is defined as any object allocation that exceeds a predefined threshold value. For example, Microsoft® Common Language Runtime (CLR) defines a large object as any object allocation exceeding 85K bytes. As used herein, a large object area is defined as a memory address space reserved for large objects. The LOA has a base virtual address. A LOA may be a portion of a heap or a portion of virtual memory in which large objects are stored and segregated from other objects that are not large objects. The boundaries of a large object area are specified by the garbage collector 110.
System 100 also includes a kernel driver 122. Garbage collector 110 sends a single batched remapping request 124 to kernel driver 122 to perform one or more remappings in a single kernel transition. Request 124 specifies the virtual memory addresses of virtual memory pages on which large objects are aligned, where the virtual memory addresses are to be remapped to perform the compaction of the large objects. Kernel driver 122 validates the virtual memory addresses included in request 124 and remaps the virtual memory pages (i.e., remaps the virtual memory addresses) so that the large objects are contiguous in the large object area. In one embodiment, the remapping performed kernel driver 122 is facilitated by the kernel driver's direct access to underlying kernel APIs (e.g., ZwMapViewOfSection, ZwOpenSection, ZwUnmapViewOfSection, and ZwClose, which are routines in a kernel mode library that supports kernel mode drivers in the Windows® Driver Kit (WDK) offered by Microsoft® Corporation located in Redmond, Wash.). The functionality of the components of system 100 is also described below relative to the discussion of
In step 202, the garbage collector 110 (see
In step 202, the garbage collector 110 (see
The “Assign” step in the algorithm presented above re-assigns large objects from previously assigned virtual addresses (i.e., the virtual addresses associated with the large objects by the object size and address table) to the virtual addresses determined by the algorithm. Furthermore, the garbage collector may store the virtual addresses assigned by the algorithm in a second table (a.k.a. a remapping table). In step 202, the garbage collector 110 (see
In step 204, garbage collector 110 (see
In step 206, kernel driver 122 (see
In step 208, kernel driver 122 (see
In one embodiment, the kernel driver 122 (see
As used herein, a kernel transition is defined as a sequence of steps that are required for a thread executing in user mode space 102 (see
The large object area garbage collection that utilizes virtual memory address remapping ends at step 210.
The post-compaction fix up phase of garbage collection is unaltered by the process of
Table 1 presented below includes the virtual addresses, actual sizes and virtual sizes of four large objects having the identifiers (IDs) 310, 311, 312 and 314 (see the description of the object size and address table stored prior to step 202 in
Garbage collector 110 (see
Virtual memory 308 includes a large object area 316, which stores large objects 318, 320 and 322. The arrows between physical memory locations 304 and virtual memory 308 indicate mappings of large object 310 to large object 318, large object 312 to large object 320, and large object 314 to large object 322. For example, the arrow from object 310 to object 318 indicates a mapping of a range of memory addresses of object 310 to a range of virtual addresses of object 318. The left side of
Before the remapping occurs, the algorithm presented above in the discussion of
VA310 (see Table 1) is re-assigned as the LOA base address VA312 (see Table 1) is re-assigned as the address: LOA base address+VS310
VA314 (see Table 1) is re-assigned as the address: LOA base address+VS310+VS312
After the garbage collector 110 (see
The right side of
Compaction through virtual memory remapping is prototyped in this section to explore its performance characteristics. The C++ code in this section shows a piece of physical memory (page file backed) being allocated through the Windows® API CreateFileMapping. The prototype in this section then maps the allocated memory into the application's virtual address space using MapViewOfFileEx. At any time, the block of memory in the prototype can be unmapped from the address space using UnmapViewOfFile. The code in this section simply tries a few different virtual addresses to map the memory starting at 0x640000.
Implementing a real large object area garbage collection using the prototype in this section would be inefficient because each of the Windows® APIs utilized in the prototype requires a separate kernel transition, thereby making the process too expensive for the remapping of many large objects after a collection.
Memory 404 may comprise any known type of computer data storage and/or transmission media, including bulk storage, magnetic media, optical media, random access memory (RAM), read-only memory (ROM), a data cache, a data object, etc. In one embodiment, cache memory elements of memory 404 provide temporary storage of at least some program code (e.g., code for program 414) in order to reduce the number of times code must be retrieved from bulk storage during execution. Moreover, similar to CPU 402, memory 404 may reside at a single physical location, comprising one or more types of data storage, or be distributed across a plurality of physical systems in various forms. Further, memory 404 can include data distributed across, for example, a local area network (LAN) or a wide area network (WAN).
I/O interface 406 comprises any system for exchanging information to or from an external source. I/O devices 410 comprise any known type of external device, including a display device (e.g., monitor), keyboard, mouse, printer, speakers, handheld device, facsimile, etc. Bus 408 provides a communication link between each of the components in computer system 400, and may comprise any type of transmission link, including electrical, optical, wireless, etc.
I/O interface 406 also allows computer system 400 to store and retrieve information (e.g., data or program instructions such as code of program 414) from an auxiliary storage device such as computer data storage unit 412 or another computer data storage unit (not shown). Computer data storage unit 412 may be a non-volatile storage device, such as a magnetic disk drive (i.e., hard disk drive) or an optical disc drive (e.g., a CD-ROM drive which receives a CD-ROM disk).
Memory 404 includes computer program code for the program 414 for large object area garbage collection by virtual memory address remapping (e.g., logic for the process of
Memory 404, storage unit 412, and/or one or more other computer data storage units (not shown) that are operatively coupled to computer system 400 may store the addresses of large objects in large object area 114 (see
As will be appreciated by one skilled in the art, the present invention may be embodied as a system, method or computer program product. Accordingly, an embodiment of the present invention may be an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “system” (e.g., system 100 in
Any combination of one or more computer-usable or computer-readable medium(s) (e.g., memory 404 and/or computer data storage unit 412) may be utilized. The computer-usable or computer-readable medium may be, for example but not limited to, an electronic, magnetic, optical, electromagnetic, infrared or semiconductor system, apparatus, device or propagation medium. A non-exhaustive list of more specific examples of the computer-readable medium includes: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a transmission media such as those supporting the Internet or an intranet, or a magnetic storage device. Note that the computer-usable or computer-readable medium could even be paper or another suitable medium upon which the program 414 is printed, as the program 414 can be electronically captured via, for instance, optical scanning of the paper or other medium, then compiled, interpreted, or otherwise processed in a suitable manner, if necessary, and then stored, respectively, in a computer memory 404. In the context of this document, a computer-usable or computer-readable medium may be any medium that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device. The computer-usable medium may include a propagated data signal with the computer-usable program code (e.g., program 414) embodied therewith, either in baseband or as part of a carrier wave. The computer-usable program code may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc.
Computer program code (e.g., code of program 414) for carrying out operations of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language such as Java®, Smalltalk, C++ or the like and conventional procedural programming languages, such as the “C” programming language or similar programming languages. The program code may execute entirely on a user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. Any one of the aforementioned computers or servers may be computer system 400. In the latter scenario, the remote computer may be connected to the user's computer through any type of network (not shown), including a LAN, a WAN, or the connection may be made to an external computer (e.g., through the Internet using an Internet Service Provider).
The present invention is described herein with reference to flowchart illustrations (e.g.,
These computer program instructions may also be stored in a computer-readable medium (e.g., memory 404 or computer data storage unit 412) that can direct a computer (e.g., computer system 400) or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable medium produce an article of manufacture including instruction means which implement the function/act specified in the flowchart and/or block diagram block or blocks.
The computer program instructions may also be loaded onto a computer (e.g., computer system 400) or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer-implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
Any of the components of an embodiment of the present invention can be deployed, managed, serviced, etc. by a service provider that offers to deploy or integrate computing infrastructure with respect to the process for batched remapping of virtual memory addresses for garbage collection in large object areas. Thus, an embodiment of the present invention discloses a process for supporting computer infrastructure, comprising integrating, hosting, maintaining and deploying computer-readable code (e.g., code of program 414) into a computer system (e.g., computer system 400), wherein the code in combination with the computer system is capable of performing the process of batched remapping of virtual memory addresses for garbage collection in large object areas.
In another embodiment, the invention provides a business method that performs the process steps of the invention on a subscription, advertising and/or fee basis. That is, a service provider, such as a Solution Integrator, can offer to create, maintain, support, etc. processes for batched remapping of virtual memory addresses for garbage collection in large object areas. In this case, the service provider can create, maintain, support, etc. a computer infrastructure that performs the process steps of the invention for one or more customers. In return, the service provider can receive payment from the customer(s) under a subscription and/or fee agreement, and/or the service provider can receive payment from the sale of advertising content to one or more third parties.
The flowchart in
While embodiments of the present invention have been described herein for purposes of illustration, many modifications and changes will become apparent to those skilled in the art. Accordingly, the appended claims are intended to encompass all such modifications and changes as fall within the true spirit and scope of this invention.