This disclosure relates in general to the field of communications and, more particularly, to remote memory access with memory mapped addressing among multiple compute nodes.
Compute nodes such as microservers and hypervisor-based virtual machines executing in a single chassis can provide scaled out workloads in hyper-scale data centers. Microservers are an emerging trend of servers for processing lightweight workloads with large numbers (e.g., tens or even hundreds) of relatively lightweight server nodes bundled together in a shared chassis infrastructure, for example, sharing power, cooling fans, and input/output components, eliminating space and power consumption demands of duplicate infrastructure components. The microserver topology facilitates density, lower power per node, reduced costs, and increased operational efficiency. Microservers are generally based on small form-factor, system-on-a-chip (SoC) boards, which pack processing capability, memory, and system input/output onto a single integrated circuit. Unlike the relatively newer microservers, hypervisor-based virtual machines have been in use for several years. Yet, sharing data across the compute nodes with more effective and efficient inter-process communication has always been a challenge.
To provide a more complete understanding of the present disclosure and features and advantages thereof, reference is made to the following description, taken in conjunction with the accompanying figures, wherein like reference numerals represent like parts, in which:
An example method for facilitating remote memory access with memory mapped addressing among multiple compute nodes is executed at an input/output (IO) adapter in communication with the compute nodes over a Peripheral Component Interconnect Express (PCIE) bus, the method including: receiving a memory request from a first compute node to permit access by a second compute node to a local memory region of the first compute node; generating a remap window region in a memory element of the IO adapter, the remap window region corresponding to a base address register (BAR) of the second compute node in the 10 adapter; and configuring the remap window region to point to the local memory region of the first compute node, wherein access by the second compute node to the BAR corresponding with the remap window region results in direct access of the local memory region of the first compute node by the second compute node. As used herein, the term “compute node” refers to a hardware processing apparatus, in which user applications (e.g., software programs) are executed.
Turning to
In a general sense, compute nodes 14 include capabilities for processing, memory, network and storage resources. For example, as shown in greater detail in the figure, compute node Host1 runs (e.g., executes) an operating system 24 and various applications 26. A device driver (also referred to herein as a driver) 28 operates or controls a particular type of device that is attached to compute node 14. For example, each PCIE device visible to (e.g., accessible by) Host1 may be associated with a separate device driver in some embodiments. In another example, all PCIE endpoints visible to Host1 may be associated with a single PCIE device driver. In a general sense device driver 28 provides a software interface to hardware devices, enabling operating system 24 and applications 26 to access hardware functions (e.g., memory access) without needing to know precise details of the hardware being used.
In many embodiments, substantially all PCIE endpoints appear as hardware device to the accessing compute node, irrespective of its actual form. For example, in some embodiments, compute nodes 14 may comprise virtual machines; however, because one compute node is visible as a PCIE device to another compute node, they appear as hardware devices to each other and are associated with corresponding device drivers. Driver 28 communicates with the hardware device through PCIE bus 22. When one of applications 26 invokes a routine in driver 28, driver 28 issues commands to the hardware device it is associated with. Thus, driver 28 facilitates communication (e.g., acts as a translator) between its associated hardware device and applications 26. Driver 28 is hardware dependent and operating-system-specific.
In various embodiments, each of compute nodes 14, as shown using example Host1, includes various hardware components, such as one or more sockets 30 (e.g., socket refers to a hardware receptacle that enables a collection of central processing unit (CPU) cores with a direct pipe to memory); each socket holds one processor 32; each processor comprises one or more CPU cores 34; each CPU core 34 executes instructions (e.g., computations, such as Floating-point Operations Per Second (FLOPS)); a memory element 36 may facilitate operations of CPU cores 34.
Common IO adapter 18 facilitates communication to and from each of compute nodes 14. In various embodiments, IO adapter 18 services both network and storage access requests from compute nodes 14 in chassis 12, facilitating a cost efficient architecture. In various embodiments, a memory element 38 may be associated with (e.g., accessed by) IP adapter 18. Memory element 38 includes various base address registers (BARs) 40 and remap windows 42 for various operations as described herein. A remap window helper register 44 and firmware 46 are also included (among other components) in IO adapter 18. As used herein the term “firmware” comprises machine-readable and executable instructions and associated data that are stored in (e.g., embedded in, forming an integral part of, etc.) hardware, such as a read-only memory, or flash memory, or an ASIC, or a field programmable gate array (FPGA) and executed by one or more processors (not shown) in IO adapter 18 to control the operations of IO adapter 18. In a general sense, firmware 46 comprises a combination of software and hardware used exclusively to control operations of IO adapter 18.
In a general sense, network traffic between compute nodes 14 and network 16 may be termed as “North-South Traffic”; network traffic among compute nodes 14 may be termed as “East-West Traffic”. Note that compute nodes 14 are unaware of the physical location of other compute nodes, for example, whether they exist in same chassis 12, or are located remotely, over network 16. Thus, compute nodes 14 are agnostic to the direction of network traffic they originate or terminate, such as whether the traffic is North-South, or East-West, and thereby use the same addressing mechanism (e.g., L2 Ethernet MAC address/IP address) for addressing nodes located in same chassis 12 or located in a remote node in same L2/L3 domain.
According to various embodiments of communication system 10, a memory access scheme using low latency and low overhead protocols implemented in IO adapter 18 allows any one (or more) compute nodes 14, for example, Host1, to share and access remote memory of another compute node (e.g., across different servers; across a hypervisor; across different operating systems), for example, Host2. Host2 may include an operating system different from that of Host1 without departing from the scope of the embodiments. The protocols as described herein do not require any particularized (e.g., custom) support from the operating systems or networking stack of Host1 or Host2. The scheme is completely transparent to the operating systems of Host1 and Host2, allowing suitable throughput while communicating in different memory domains.
For purposes of illustrating the techniques of communication system 10, it is important to understand the communications that may be traversing the system shown in
In any server ecosystem, typical challenges to achieving better inter-process communication or sharing the data across the servers include reliable tunnels for the data sharing, low latency for the communication, low overhead while working with remote server etc. There are several solutions available in the market that predominantly use network tunnels to communicate in two distinct physical servers. Typical examples would be Remote Direct Memory Access (RDMA), RDMA over Converged Ethernet (RoCE) and InfiniBand. Although proven and in use for several years, such network based communication can be limited by various parameters, such as latency, networking stack dependency (e.g., network stack awareness), OS interference (e.g., OS dependency, OS awareness, OS configuration), IO semantics and key exchanges for security, necessity for protocol awareness, complex channel semantics and tedious channel setup procedures. Moreover, not all operating systems support RDMA (and its variants).
For example, RDMA communication is based on a set of three queues: (i) a send queue and (ii) a receive queue, comprising a Queue Pair (QP) and (iii) a Completion Queue (CQ). Posts in the QP are used to initiate the sending or receiving of data. A sending application (e.g., driver) places instructions, called Work Queue Elements (WQE), on its work queues that generate buffers in the sender's adapter to send data. The WQE placed on the send queue contains a pointer to the message to be sent; a pointer in the WQE on the receive queue contains a pointer to a buffer where an incoming message can be placed. The sender's adapter consumes WQE from the send queue at the egress side and streams the data from the memory region to the remote receiver. When data arrives at the remote receiver, the receiver's adapter consumes the WQEs at the receive queue at the ingress side and places the received data in appropriate memory regions of the receiving application. Any memory sharing or access between a sending compute node and the receiving compute node thus requires tedious channel setup, RDMA protocols, etc.
Moreover, in a chassis where several compute nodes share a common IO adapter, such remote memory access sharing protocols can have unnecessary overhead. For example, every packet from any compute node, say Host1, has to hit a port of upstream switch 20 and then return on the same pipe back to IO adapter 18, which then redirects it to the destination compute node, say Host2. Such east-west data sharing can cause inefficient utilization of bandwidth in the common pipe, which is potentially used by various other compute nodes performing extensive north-south traffic with network 16. The east-west traffic pattern also increases application response latency, for example, due to longer path to be traversed by network packets.
Communication system 10 is configured to address these issues (among others) to offer a system and method for facilitating remote memory access with memory mapped addressing among multiple compute nodes 14 sharing IO adapter 18. In various embodiments, PCIE, which is typically supported by almost all operating systems, is used to share data from a memory region on one compute node, say Host1, with a different memory region of another compute node, say Host2. As used herein, the term “memory region” comprises a block (e.g., section, portion, slice, chunk, piece, space, etc.) of memory that can be accessed through a contiguous range of memory addresses (e.g., a memory address is a unique identifier (e.g., binary identifier) used by a processor for tracking a location of each memory byte stored in the memory). As used herein, the term “window” in the context of memory regions refers to a memory region comprising a contiguous range of memory addresses, either virtual or physical.
In various embodiments, IO adapter 18 is connected to compute nodes 14 by means of PCIE bus 22. IO adapter 18 includes an embedded operating system hosting multiple VNICs configured with memory resources of memory element 38. Each VNIC accesses a separate, exclusive region of memory element 38. Each PCIE endpoint, namely VNICs is typically associated with a host software driver, namely device driver 28. In an example embodiment, each VNIC that requires a separate driver is considered a separate PCIE device.
For ease of explanation of various embodiments, a brief overview of PCIE protocol is provided herein. A PCIe data transfer subsystem in a computing system (such as that of an IO adapter) includes a PCIe root complex comprising a computer hardware chipset that handles communications between the PCIE endpoints. The root complex enables PCIe endpoints to be discovered, enumerated and worked upon by the host operating system. The base PCIe switching structure of a single root complex has a tree topology, which addresses PCIe endpoints through a bus numbering scheme. Configuration software on the root complex detects every bus, device and function (e.g., storage adapter, networking adapter, graphics adapter, hard drive interface, device controller, Ethernet controller, etc.) within a given PCIe topology.
The IO adapter's operating system assigns address space in the IO adapter memory element 38 to each PCIe endpoint (e.g., VNIC) so that the PCIe endpoint can understand at what address space it is identified by the IO adapter and map the corresponding interrupts accordingly. After the configuration of the PCIe endpoint device is complete, the PCIe's device driver 28 compatible with the host operating system 24 can work efficiently with the PCIe endpoint and facilitate appropriate device specific functionality.
Each PCIE endpoint is enabled on IO adapter 18 by being mapped into a memory-mapped address space in memory element 18 referred to as configuration space (e.g., register, typically consisting of 256 bytes). The configuration space contains a number of base address registers (BARs) 40, comprising the starting address of a contiguous mapped address in IO adapter memory element 38. For example, a 32-bit BAR0 is offset 10 h in PCI Compatible Configuration Space—and post enumeration would contain the start address of BAR. Any other PCIE endpoint, to access (e.g., read data from or write data to) the PCIE endpoint associated with a specific BAR, would submit a request with the address of that BAR. An enumeration software allocates memory for the PCIE endpoints and writes to corresponding BARs. Firmware 46 programs the PCIe endpoint's BARs to inform the PCIe endpoints of its address mapping. When the BAR for a particular PCIe endpoint is written, all memory transactions generated to that bus address range are claimed by the particular PCIe endpoint.
Typically, when a PCIE endpoint, say a flash memory device, is discovered on one of the compute nodes, say, Host1, OS 24 provides a physical address to BAR 40 and allocates the address space for device driver 28 to interact with the flash memory device. When device driver 28 is loaded, it requests the memory mapped address from OS 24 corresponding to the physical address so that it can work with the flash memory device using the address handle. Subsequent accesses to BAR 40 from device driver 28 are completely transparent to OS 24 as it has already carved out the address space sufficient to work with the flash memory device. Thus, typical PCIE data access is between application 26 and the PCIE endpoint, such as the flash memory device. PCIE data access is not typically used across two different compute nodes 14. In other words, one compute node typically cannot share its memory space with another compute node using native PCIE protocols.
Nevertheless, according to various embodiments, appropriate configuration of IO adapter 18 with multiple ports and remap window feature can support memory sharing between compute nodes 14 using PCIE. Remap window feature includes a remap window base and remap window region for memory mapping for the purpose of remapping root complex IO and memory BARs to address ranges that are directly addressable by the processor. Remap window base is used to configure a start address of a memory region which can be mapped to any other memory region. The remap window region refers to the mapped region in memory element 38 differentiated according to a virtual network interface card (VNIC) identifier (ID) configured in remap window helper register 44 in IO adapter 18.
The VNIC ID could map to any host-based VNIC or root complex VNIC. In some embodiments, four remap window regions, each capable of addressing 4 MB may be allocated for the remap window feature, permitting easy access of up to 16 MB of memory either in host memory or Root Complex endpoint device memory. Moreover, multiple PCIE ports on PCIE bus 22 distinguish different PCIE lanes associated with distinct compute nodes 14. Each memory region in Root Complex endpoint device memory is associated with a distinct PCIE lane that is completely independent of each other such that no two memory regions share any PCIE activity with each other.
In an example embodiment, an administrator configures the VNIC ID of computing nodes 14 through respective service profiles. A unified computing system manager (e.g., network management application such as Cisco® UCSM) programs the VNIC ID in IO adapter 18 through appropriate control management protocol. Upon reception, firmware 46 populates the VNIC ID information in remap window helper register 44 and also makes the VNICs ready and discoverable from corresponding computing nodes 14. Firmware 46 adds the BAR size of a specific BAR, for example, BAR3, to the memory region allocated with each VNIC ID. In an example embodiment, 16 MB may be added to accommodate all four remap window regions.
After one of computing nodes 14, say Host1 is powered up, the enumeration software (BIOS) of IO adapter 18 discovers the new PCIE device. Through PCI enumeration protocol, the BIOS identifies BAR size requirements, and associates a physical address to corresponding Host1 in BAR 40. In various embodiments, three separate BARs are provided for each VNIC, namely, BAR0, BAR1 and BAR2. Device driver 28 upon loading in Host1 requests OS 24 to provide memory mapped equivalent of the physical address for each BAR. It identifies that BAR2 is the remap window region according to a preconfigured protocol between firmware 46 and driver 28. The memory mapped 10 address comprises an address handle given by OS 24 to access the BAR2 region of memory element 38 in IO adapter 18. Applications 26 using device driver 28 understands the capability of remap window exposed by device driver 28. Similar sequence of events occurs in another compute node, say Host2, when it powers up and its device driver is loaded in its OS. Through a pre-determined protocol, applications 26 in compute nodes 14, say Host1 and Host2, exchange their respective address handles through firmware 46 and request corresponding memory access.
The memory access mechanisms described herein can present one of the lowest latency protocols to communicate with different servers, virtual machines, or other such compute nodes 14. In some embodiments, the memory access mechanisms described herein can also be used as IPC between two compute nodes 14. Note that the operating system or network stacks do not need any separate, or distinct configuration to enable such remote memory access. In some embodiments, IO adapter 18 servicing a hypervisor can use the described mechanisms to allow various applications executing in separate virtual machines (e.g., guest domains) to communicate with each other without having to go through specially installed IPC software (e.g., VMWARE ESX/ESXi) or other external memory management/sharing applications.
In an example embodiment wherein compute nodes 14 comprise microservers the storage and network is shared across multiple servers (e.g., in some cases sixteen servers). The network ecosystem (e.g., of network 16) may support different classes and QoS policies for network traffic, which can result in different priority flows. However, storage traffic does not typically have any associated QoS. Such differentiated traffic types (e.g., some traffic having QoS, other traffic not having QoS) can create imbalance of traffic performance across different servers causing some servers using (or allocated) larger bandwidths and other servers using (or allocated) poor bandwidth. With large amounts of input/output among (or from/to) servers, the condition can become worse with performance drops becoming noticeable in some servers. In other words, performance of some servers drops when other unrelated servers are experiencing heavy network traffic.
To have balanced throughput across the servers, a cooperative I/O scheduling across the servers may be implemented. For example, every server monitors and records a number of IO requests issued to IO adapter 18. Such IO statistics are shared with other servers through the local memory mapped scheme in BAR3 as described herein. Such data sharing can facilitate decisions at the individual servers regarding whether to send a SCSI BUSY message to its OS storage stack. Thus, even though the associated storage VNIC has bandwidth to push the IOs to IO adapter 18, it will not schedule the IO requests, voluntarily relinquishing claim on storage for some time, until the network traffic bottleneck clears up. Such actions can lead to other VNICs balancing out storage traffic pattern in chassis 12, maintaining the IO equilibrium therein.
In various embodiments, IO adapter 18 receives a memory request from one of compute nodes 14, say Host1, to permit access by another of compute nodes 14, say Host2, to a local memory region of Host1 (assume the local memory region is in memory element 36). The memory request comprises a host identifier of Host2 and address of the local memory region of Host1 in some embodiments. The host identifier can be obtained from a resource map providing identifying information of compute nodes 14 in communication with IO adapter 18 over PCIE bus 22.
Firmware 46 in IO adapter 18 generates remap window region 42 in memory element 38 of IO adapter 18, remap window region 42 corresponding to BAR 40 (e.g., BAR2) of Host2 in IO adapter 18. Firmware 46 configures remap window region 42 to point to the local memory region of Host1, access by Host2 to BAR2 corresponding with remap window region 42 resulting in direct access of the local memory region of Host1 by Host2. Note that compute nodes 14 are associated with unique PCIE endpoints on PCIE bus 22; therefore, each has distinct BARs 40 associated therewith. Moreover, the direct access of the local memory region of Host1 by Host2 does not involve operating systems of Host1 and/or Host2. BAR2 associated with remap window region 42 can comprise one of a plurality of BARs associated with Host2.
In various embodiments, device driver 28 of Host2 associates BAR2 with remap window region, such that application 26 executing in Host2 can access the local memory region of Host1 through appropriate access requests to BAR2 using device driver 28. In various embodiments, configuring remap window region 42 comprises configuring a remap window base in a BAR Resource Table (BRT) to be a start address of the local memory region.
Turning to the infrastructure of communication system 10, network topology of the network including chassis 12 can include any number of compute nodes, servers, hardware accelerators, virtual machines, switches (including distributed virtual switches), routers, and other nodes inter-connected to form a large and complex network. A node may be any electronic device, client, server, peer, service, application, or other object capable of sending, receiving, or forwarding information over communications channels in a network. Elements of
Communication system 10 may include a configuration capable of TCP/IP communications for the electronic transmission or reception of data packets in a network. Communication system 10 may also operate in conjunction with a User Datagram Protocol/Internet Protocol (UDP/IP) or any other suitable protocol, where appropriate and based on particular needs. In addition, gateways, routers, switches, and any other suitable nodes (physical or virtual) may be used to facilitate electronic communication between various nodes in the network.
Note that the numerical and letter designations assigned to the elements of
The example network environment may be configured over a physical infrastructure that may include one or more networks and, further, may be configured in any form including, but not limited to, local area networks (LANs), wireless local area networks (WLANs), VLANs, metropolitan area networks (MANs), VPNs, Intranet, Extranet, any other appropriate architecture or system, or any combination thereof that facilitates communications in a network.
In some embodiments, a communication link may represent any electronic link supporting a LAN environment such as, for example, cable, PCIE, Ethernet, wireless technologies (e.g., IEEE 802.11x), ATM, fiber optics, etc. or any suitable combination thereof. In other embodiments, communication links may represent a remote connection through any appropriate medium (e.g., digital subscriber lines (DSL), telephone lines, T1 lines, T3 lines, wireless, satellite, fiber optics, cable, Ethernet, etc. or any combination thereof) and/or through any additional networks such as a wide area networks (e.g., the Internet).
In various embodiments, chassis 12 may comprise a rack-mounted enclosure, blade enclosure, or a rack computer that accepts plug-in compute nodes 14. Note that chassis 12 can include, in a general sense, any suitable network element, which encompasses computers, network appliances, servers, routers, switches, gateways, bridges, load-balancers, firewalls, processors, modules, or any other suitable device, component, element, or object operable to exchange information in a network environment. Moreover, the network elements may include any suitably configured hardware provisioned with suitable software, components, modules, interfaces, or objects that facilitate the operations thereof. This may be inclusive of appropriate algorithms and communication protocols that allow for the effective exchange of data or information.
Compute nodes 14 may comprise printed circuit boards, for example, manufactured with empty sockets. Each printed circuit board may hold more than one processor (e.g., within the same processor family, differing core counts, with a wide range of frequencies and vastly differing memory cache structures may be included in a single processor/socket combination). In some embodiments, compute nodes 14 may include hypervisors and virtual machines. IO adapter 18 may include an electronic circuit, expansion card or plug-in module that accepts input and generates output in a particular format. IO adapter 18 facilitates conversion of data format and electronic timing between input/output streams and internal computer circuits of chassis 12. In some embodiments, IO adapter 18 may comprise a hypervisor, and compute nodes 14 may comprise separate virtual machines.
Turning to
To explain further, consider Host1. Note that the descriptions herein for Host1 apply equally for Host2. Operating system 24 in Host1 enumerates BARs 40(1) associated with Host1 and maps IO address space in host memory 36 to each BAR such that any access to the corresponding mapped addresses in the mapped IO address space in Host1 will point to (e.g., correspond with, associate with) the appropriate one of BARs 40(1): BAR0, BAR1 and BAR2 in IO adapter 18. Note that whereas mapped addresses in Host1 may be virtual, they point to the physical memory region in IO adapter 18. Device driver 28 accesses BARs 40(1) using the memory mapped addresses returned by OS 24.
In various embodiments, BAR2 is reserved for remap window 42, which is identified by the device driver in respective compute nodes 14. For example, BAR2 of BAR 40(1) is reserved for remap window region 42(1) and BAR2 of BAR 40(2) is reserved for remap window region 42(2). In other words, device driver 28 in Host1 understands BAR2 of 40(1) to be associated with remap window 42(1). When device driver 28 (or application 26) in Host1 wants to allow another compute node, such as Host2, to access its local memory 56(1), firmware 46 configures remap window 42(2) of Host2 to point to memory addressed space 56(1) of Host1. Similarly, when device driver 28 (or application 26) in Host2 wants to allow Host1 to access its local memory 56(2), firmware 46 configures remap window 42(1) to point to memory addressed space 56(2) of Host2.
In other words, BAR2 of BAR 40(1) associated with Host1 refers to memory space 56(2) of Host2; likewise, BAR2 of BAR 40(2) associated with Host2 refers to memory space 56(1) of Host1. Anything written to BAR2 of BAR 40(1) by Host1 will be as if written directly into memory space 56(2) of Host2, without any intervening protocols or communication. Thus applications in separate compute nodes can easily access the memory present in their peer's memory domain.
Turning to
Firmware 46 of IO adapter 18 presents a virtualized view of PCIE endpoints' configuration space to compute nodes 14. When Host1 configures memory/IO bar window(s) in the VNIC's configuration space, Host1's BAR address windows are translated by remap window helper register 44 to map them to the local root complex endpoint's BAR windows in IO adapter's local address space. For example, memory region 56(2) of Host1 is mapped to remap window region 42(2) of Host2 in memory element 38. After enumeration and virtualization of the configuration space of the PCIE endpoints, the device drivers running on compute nodes 14 may post work requests using their assigned memory bar windows.
During operation, a memory request from Host1 to allow access to a specific memory region 56(1) by a remote PCIE endpoint, say Host2 may proceed as follows. Host1 sends a memory request to firmware 46, including HostID=identifier of remote peer, say Host2; type=remote_memory_access; address=address of local memory 56(1). The memory request is converted into a search key to BMT 44(1), triggering a lookup (e.g., ternary content-addressable memory (TCAM)) of BMT 44(1), which outputs a hit index to RAM 44(2) that activates a read of appropriate entry in BRT 44(3). In some embodiments, the memory request from Host1 may reference a VNIC number, which may be converted into the corresponding host identifier by suitable modules. Firmware 46 programs the appropriate entry in BRT 44(3) to point to the provided address 56(1) of Host1. The specific memory region of the appropriate entry in BRT 44(3) is already pre-mapped to BAR2 of Host2 as remap window 42(2). In other words, the entry in BRT 44(3) references remap window region 42(2), which now directly points to memory space 56(1) of Host1 after configuration by firmware 46. Any memory requests going through remap window region 42(1) will be tagged with the VNIC of the destination compute node 14. Any writes by Host1 into local memory region 56(1) can be directly accessed by Host2 through its mapped remap window region 42(2) without any intervention by operating systems or CPUs.
Turning to
Firmware 46 configures remap window base 58 with the given address and sets up application specific integrated circuit (ASIC) data structures to be ready for remap window region access. Firmware 46 discovers the destination VNIC of Host1 that wants to access the memory region as given by driver 28. Firmware 46 configures BRT0, corresponding to BAR2 of the destination VNIC Host1, with the remap window region address and offset that would correspond to the remap window base 58. Configured BRT0 corresponds to remap window 42(1) and points to memory region 56(2) of Host2.
After remap window region 42(1) is configured on behalf of the destination VNIC, firmware 46 sends notification to driver 28 running in Host1 that its BAR2 is ready to access the Host2 memory. Upon receiving the notification from firmware 46, driver 28 running in Host1 passes the notification to application A. Application A already has memory mapped the BAR2 region with appropriate IOMMU configuration (e.g., addresses). Subsequently application A's read/write access to BAR2 of Host1 maps to remote memory region 56(2) present in Host2's memory domain. Likewise, application B's read/write access to memory region 56(2) maps to BAR2 of Host1. Thus both application A and application B running in different compute nodes 14 can communicate with each other without any OS intervention.
Turning to
Turning to
IO adapter 18 includes a resource map providing resource information, for example, its memory offset and length, associated with the corresponding VNIC. In some embodiments, the resource map associates memory address offsets (also referred to herein as “memory offsets,” or simply “offsets”) with the BAR of one or more I/O resources (the I/O resource corresponding to a PCIE device, such as VNIC). For example, the resource map may include information identifying each PCIE device on PCIE bus 22 and its corresponding BARs. In many embodiments, the resource map may be comprised in remap window help register 44. In various embodiments, BAR0 of each PCIE endpoint may point to the resource map stored in IO adapter 18. The PCIE endpoints may be identified using host indices, or other suitable identifiers. On parsing the resource map, device driver 28 in producer 72 identifies other compute nodes 14 present in chassis 12. Consumer 74 notifies firmware 46 of its intent to read the contents of the memory of Host1 through a resource update. Firmware 46 decodes the request, identifies the source and destination VNICs and sends notification to the VNIC whose associated memory is to be read.
At 76, producer 72 creates data at memory offset with dirty bit reset. (Note that the dirty bit is well known in the art to be associated with a block of memory and indicates whether or not the corresponding block of memory has been modified; if the bit is set (or reset), the data has been modified since the last time it was read). At 78, device driver 28 in Host1 notifies firmware 46 about the data availability. The notification's meta-data includes the address to be read from at its local memory space 56(1) including: address, length, destination host index of consumer and a key. At 80, firmware 46 configures remap window base 58 with the memory offset. At 82, firmware 46 configures remap window region 42(2) of BAR2 associated with Host2 to point to memory space 56(1) of Host1. At 84, firmware 46 configures remap window 42(2) only once (e.g., for all transactions between producer 72 and same consumer 74).
At 86, firmware 46 notifies consumer 74 that the data is ready. At 88, consumer 74 reads BAR2 at the memory offset specified by firmware 46. Reading BAR2 at the memory offset is identical to accessing memory region 56(1) of producer 72. At 90, consumer 74 marks the dirty bit in the meta data, indicating that the data has been read. At 92, consumer 74 may notify firmware 46 that consumer 74 has completed reading the data. At 94, firmware 46 may notify producer 72 that data has been consumed by consumer 74. At 96, producer 72 writes the next set of data to the memory offset, and the operations resume from 76 and continue thereafter.
Turning to
At 108, consumer application 74 notifies firmware 46 of intent to read contents of memory 56(1) of Host1 through a resource update message (or other suitable mechanism). At 110, firmware 46 decodes the request, identifies the source and destination VNICs and sends notification to Host1 VNIC from which the memory is to be read. At 112, producer application 72 provides to firmware 46 the address of memory region 56(1) to be read from, through an appropriate memory request. At 114, firmware 46 sends the remap window region information (e.g., remap window base 58, remap window region 42(2)) to consumer application 74 (e.g., through associated VNIC) BAR2 at known offset and notifies the consumer VNIC.
At 116, firmware 46 maps the address belonging to the producer application VNIC, namely, address of memory region 56(1) into the consumer VNIC's BAR2 remap window region 42(2) at known offset. At 118, consumer VNIC passes the remap window information to consumer application 74. At 120, consumer application 74 reads the data from its memory mapped BAR2 which corresponds to memory 56(1) of producer application 72.
Turning to
At 140, Host1 is powered up. At 142, BIOS enumeration software discovers PCIE endpoint and through PCIE enumeration protocol, identifies BAR size requirements and associates physical address to each BAR. At 144, device driver 28 of Host1 upon loading, requests OS 24 to provide memory mapped equivalent of physical address for each BAR. At 146, device driver 28 identifies that BAR2 is remap window region according to preconfigured protocol with firmware 46. At 148, application 26 running in Host1's operating system 24 making use of device driver 28 understands capability of remap window exposed by device driver 28. At 150, application 26 exchanges handles with peer host (e.g., Host2) through firmware 46 and permits memory to be accessed by the peer host.
Turning to
Note that in this Specification, references to various features (e.g., elements, structures, modules, components, steps, operations, characteristics, etc.) included in “one embodiment”, “example embodiment”, “an embodiment”, “another embodiment”, “some embodiments”, “various embodiments”, “other embodiments”, “alternative embodiment”, and the like are intended to mean that any such features are included in one or more embodiments of the present disclosure, but may or may not necessarily be combined in the same embodiments. Furthermore, the words “optimize,” “optimization,” and related terms are terms of art that refer to improvements in speed and/or efficiency of a specified outcome and do not purport to indicate that a process for achieving the specified outcome has achieved, or is capable of achieving, an “optimal” or perfectly speedy/perfectly efficient state.
Embodiments described herein may be used as or to support firmware instructions executed upon some form of processing core (such as the processor of IO adapter 18) or otherwise implemented or realized upon or within a machine-readable medium. A machine-readable medium includes any mechanism for storing or transmitting information in a form readable by a machine (e.g., a computer). For example, a machine-readable medium can include such as a read only memory (ROM); a random access memory (RAM); a magnetic disk storage media; an optical storage media; and a flash memory device, etc. In addition, a machine-readable medium can include propagated signals such as electrical, optical, acoustical or other form of propagated signals (e.g., carrier waves, infrared signals, digital signals, etc.).
In example implementations, at least some portions of the activities outlined herein may be implemented in software in, for example, IO adapter 18. In some embodiments, one or more of these features may be implemented in hardware, provided external to these elements, or consolidated in any appropriate manner to achieve the intended functionality. The various network elements may include software (or reciprocating software) that can coordinate in order to achieve the operations as outlined herein. In still other embodiments, these elements may include any suitable algorithms, hardware, software, components, modules, interfaces, or objects that facilitate the operations thereof.
Furthermore, IO adapter 18 described and shown herein (and/or their associated structures) may also include suitable interfaces for receiving, transmitting, and/or otherwise communicating data or information in a network environment. Additionally, some of the processors and memory elements associated with the various nodes may be removed, or otherwise consolidated such that a single processor and a single memory element are responsible for certain activities. In a general sense, the arrangements depicted in the FIGURES may be more logical in their representations, whereas a physical architecture may include various permutations, combinations, and/or hybrids of these elements. It is imperative to note that countless possible design configurations can be used to achieve the operational objectives outlined here. Accordingly, the associated infrastructure has a myriad of substitute arrangements, design choices, device possibilities, hardware configurations, software implementations, equipment options, etc.
In some of example embodiments, one or more memory elements (e.g., memory element 38, memory element 36) can store data used for the operations described herein. This includes the memory element being able to store instructions (e.g., software, logic, code, etc.) in non-transitory media, such that the instructions are executed to carry out the activities described in this Specification. A processor can execute any type of instructions associated with the data to achieve the operations detailed herein in this Specification. In one example, processors could transform an element or an article (e.g., data) from one state or thing to another state or thing. In another example, the activities outlined herein may be implemented with fixed logic or programmable logic (e.g., software/computer instructions executed by a processor) and the elements identified herein could be some type of a programmable processor, programmable digital logic (e.g., a field programmable gate array (FPGA), an erasable programmable read only memory (EPROM), an electrically erasable programmable read only memory (EEPROM)), an ASIC that includes digital logic, software, code, electronic instructions, flash memory, optical disks, CD-ROMs, DVD ROMs, magnetic or optical cards, other types of machine-readable mediums suitable for storing electronic instructions, or any suitable combination thereof.
These devices may further keep information in any suitable type of non-transitory storage medium (e.g., random access memory (RAM), read only memory (ROM), field programmable gate array (FPGA), erasable programmable read only memory (EPROM), electrically erasable programmable ROM (EEPROM), etc.), software, hardware, or in any other suitable component, device, element, or object where appropriate and based on particular needs. The information being tracked, sent, received, or stored in communication system 10 could be provided in any database, register, table, cache, queue, control list, or storage structure, based on particular needs and implementations, all of which could be referenced in any suitable timeframe. Any of the memory items discussed herein should be construed as being encompassed within the broad term ‘memory element.’ Similarly, any of the potential processing elements, modules, and machines described in this Specification should be construed as being encompassed within the broad term ‘processor.’
It is also important to note that the operations and steps described with reference to the preceding FIGURES illustrate only some of the possible scenarios that may be executed by, or within, the system. Some of these operations may be deleted or removed where appropriate, or these steps may be modified or changed considerably without departing from the scope of the discussed concepts. In addition, the timing of these operations may be altered considerably and still achieve the results taught in this disclosure. The preceding operational flows have been offered for purposes of example and discussion. Substantial flexibility is provided by the system in that any suitable arrangements, chronologies, configurations, and timing mechanisms may be provided without departing from the teachings of the discussed concepts.
Although the present disclosure has been described in detail with reference to particular arrangements and configurations, these example configurations and arrangements may be changed significantly without departing from the scope of the present disclosure. For example, although the present disclosure has been described with reference to particular communication exchanges involving certain network access and protocols, communication system 10 may be applicable to other exchanges or routing protocols. Moreover, although communication system 10 has been illustrated with reference to particular elements and operations that facilitate the communication process, these elements, and operations may be replaced by any suitable architecture or process that achieves the intended functionality of communication system 10.
Numerous other changes, substitutions, variations, alterations, and modifications may be ascertained to one skilled in the art and it is intended that the present disclosure encompass all such changes, substitutions, variations, alterations, and modifications as falling within the scope of the appended claims. In order to assist the United States Patent and Trademark Office (USPTO) and, additionally, any readers of any patent issued on this application in interpreting the claims appended hereto, Applicant wishes to note that the Applicant: (a) does not intend any of the appended claims to invoke paragraph six (6) of 35 U.S.C. section 112 as it exists on the date of the filing hereof unless the words “means for” or “step for” are specifically used in the particular claims; and (b) does not intend, by any statement in the specification, to limit this disclosure in any way that is not otherwise reflected in the appended claims.
The instant application is a Divisional of and claims priority to U.S. patent application Ser. No. 15/174,718 entitled REMOTE MEMORY ACCESS USING MEMORY MAPPED ADDRESSING AMONG MULTIPLE COMPUTE NODES filed Jun. 6, 2016, the contents of which are expressly incorporated by reference herein in its entirety.
Number | Name | Date | Kind |
---|---|---|---|
4688695 | Hirohata | Aug 1987 | A |
5263003 | Cowles et al. | Nov 1993 | A |
5339445 | Gasztonyi | Aug 1994 | A |
5430859 | Norman et al. | Jul 1995 | A |
5457746 | Dolphin | Oct 1995 | A |
5535336 | Smith et al. | Jul 1996 | A |
5588012 | Oizumi | Dec 1996 | A |
5617421 | Chin et al. | Apr 1997 | A |
5680579 | Young et al. | Oct 1997 | A |
5690194 | Parker et al. | Nov 1997 | A |
5740171 | Mazzola et al. | Apr 1998 | A |
5742604 | Edsall et al. | Apr 1998 | A |
5764636 | Edsall | Jun 1998 | A |
5809285 | Hilland | Sep 1998 | A |
5812814 | Sukegawa | Sep 1998 | A |
5812950 | Tom | Sep 1998 | A |
5838970 | Thomas | Nov 1998 | A |
5999930 | Wolff | Dec 1999 | A |
6035105 | McCloghrie et al. | Mar 2000 | A |
6043777 | Bergman et al. | Mar 2000 | A |
6101497 | Ofek | Aug 2000 | A |
6148414 | Brown et al. | Nov 2000 | A |
6185203 | Berman | Feb 2001 | B1 |
6188694 | Fine et al. | Feb 2001 | B1 |
6202135 | Kedem et al. | Mar 2001 | B1 |
6208649 | Kloth | Mar 2001 | B1 |
6209059 | Ofer et al. | Mar 2001 | B1 |
6219699 | McCloghrie et al. | Apr 2001 | B1 |
6219753 | Richardson | Apr 2001 | B1 |
6223250 | Yokono | Apr 2001 | B1 |
6226771 | Hilla et al. | May 2001 | B1 |
6260120 | Blumenau et al. | Jul 2001 | B1 |
6266705 | Ullum et al. | Jul 2001 | B1 |
6269381 | St. Pierre et al. | Jul 2001 | B1 |
6269431 | Dunham | Jul 2001 | B1 |
6295575 | Blumenau et al. | Sep 2001 | B1 |
6400730 | Latif et al. | Jun 2002 | B1 |
6408406 | Parris | Jun 2002 | B1 |
6542909 | Tamer et al. | Apr 2003 | B1 |
6542961 | Matsunami et al. | Apr 2003 | B1 |
6553390 | Gross et al. | Apr 2003 | B1 |
6564252 | Hickman et al. | May 2003 | B1 |
6647474 | Yanai et al. | Nov 2003 | B2 |
6675258 | Bramhall et al. | Jan 2004 | B1 |
6683883 | Czeiger et al. | Jan 2004 | B1 |
6694413 | Mimatsu et al. | Feb 2004 | B1 |
6708227 | Cabrera et al. | Mar 2004 | B1 |
6715007 | Williams et al. | Mar 2004 | B1 |
6728791 | Young | Apr 2004 | B1 |
6772231 | Reuter et al. | Aug 2004 | B2 |
6820099 | Huber et al. | Nov 2004 | B1 |
6847647 | Wrenn | Jan 2005 | B1 |
6848759 | Doornbos et al. | Feb 2005 | B2 |
6850955 | Sonoda et al. | Feb 2005 | B2 |
6876656 | Brewer et al. | Apr 2005 | B2 |
6880062 | Ibrahim et al. | Apr 2005 | B1 |
6898670 | Nahum | May 2005 | B2 |
6907419 | Pesola et al. | Jun 2005 | B1 |
6912668 | Brown et al. | Jun 2005 | B1 |
6952734 | Gunlock et al. | Oct 2005 | B1 |
6976090 | Ben-Shaul et al. | Dec 2005 | B2 |
6978300 | Beukema et al. | Dec 2005 | B1 |
6983303 | Pellegrino et al. | Jan 2006 | B2 |
6986015 | Testardi | Jan 2006 | B2 |
6986069 | Oehler et al. | Jan 2006 | B2 |
7051056 | Rodriguez-Rivera et al. | May 2006 | B2 |
7069465 | Chu et al. | Jun 2006 | B2 |
7073017 | Yamamoto | Jul 2006 | B2 |
7108339 | Berger | Sep 2006 | B2 |
7149858 | Kiselev | Dec 2006 | B1 |
7171514 | Coronado et al. | Jan 2007 | B2 |
7171668 | Molloy et al. | Jan 2007 | B2 |
7174354 | Andreasson | Feb 2007 | B2 |
7200144 | Terrell et al. | Apr 2007 | B2 |
7222255 | Claessens et al. | May 2007 | B1 |
7237045 | Beckmann et al. | Jun 2007 | B2 |
7240188 | Takata et al. | Jul 2007 | B2 |
7246260 | Brown et al. | Jul 2007 | B2 |
7266718 | Idei et al. | Sep 2007 | B2 |
7269168 | Roy et al. | Sep 2007 | B2 |
7277431 | Walter et al. | Oct 2007 | B2 |
7277948 | Igarashi et al. | Oct 2007 | B2 |
7305658 | Hamilton et al. | Dec 2007 | B1 |
7328434 | Swanson et al. | Feb 2008 | B2 |
7340555 | Ashmore et al. | Mar 2008 | B2 |
7346751 | Prahlad et al. | Mar 2008 | B2 |
7352706 | Klotz et al. | Apr 2008 | B2 |
7353305 | Pangal et al. | Apr 2008 | B2 |
7359321 | Sindhu et al. | Apr 2008 | B1 |
7383381 | Faulkner et al. | Jun 2008 | B1 |
7403987 | Marinelli et al. | Jul 2008 | B1 |
7433326 | Desai et al. | Oct 2008 | B2 |
7433948 | Edsall | Oct 2008 | B2 |
7434105 | Rodriguez-Rivera et al. | Oct 2008 | B1 |
7441154 | Klotz et al. | Oct 2008 | B2 |
7447839 | Uppala | Nov 2008 | B2 |
7487321 | Muthiah et al. | Feb 2009 | B2 |
7500053 | Kavuri et al. | Mar 2009 | B1 |
7512744 | Banga et al. | Mar 2009 | B2 |
7542681 | Cornell et al. | Jun 2009 | B2 |
7558872 | Senevirathne et al. | Jul 2009 | B1 |
7587570 | Sarkar et al. | Sep 2009 | B2 |
7631023 | Kaiser et al. | Dec 2009 | B1 |
7643505 | Colloff | Jan 2010 | B1 |
7654625 | Amann et al. | Feb 2010 | B2 |
7657796 | Kaiser et al. | Feb 2010 | B1 |
7668981 | Nagineni et al. | Feb 2010 | B1 |
7669071 | Cochran et al. | Feb 2010 | B2 |
7689384 | Becker | Mar 2010 | B1 |
7694092 | Mizuno | Apr 2010 | B2 |
7697554 | Ofer et al. | Apr 2010 | B1 |
7706303 | Bose et al. | Apr 2010 | B2 |
7707481 | Kirschner et al. | Apr 2010 | B2 |
7716648 | Vaidyanathan et al. | May 2010 | B2 |
7752360 | Galles | Jul 2010 | B2 |
7757059 | Ofer et al. | Jul 2010 | B1 |
7774329 | Peddy et al. | Aug 2010 | B1 |
7774839 | Nazzal | Aug 2010 | B2 |
7793138 | Rastogi et al. | Sep 2010 | B2 |
7840730 | D'Amato et al. | Nov 2010 | B2 |
7843906 | Chidambaram et al. | Nov 2010 | B1 |
7895428 | Boland, IV et al. | Feb 2011 | B2 |
7904599 | Bennett | Mar 2011 | B1 |
7930494 | Goheer et al. | Apr 2011 | B1 |
7975175 | Votta et al. | Jul 2011 | B2 |
7979670 | Saliba et al. | Jul 2011 | B2 |
7984259 | English | Jul 2011 | B1 |
8031703 | Gottumukkula et al. | Oct 2011 | B2 |
8032621 | Upalekar et al. | Oct 2011 | B1 |
8051197 | Mullendore et al. | Nov 2011 | B2 |
8086755 | Duffy, IV et al. | Dec 2011 | B2 |
8161134 | Mishra et al. | Apr 2012 | B2 |
8196018 | Forhan et al. | Jun 2012 | B2 |
8205951 | Boks | Jun 2012 | B2 |
8218538 | Chidambaram et al. | Jul 2012 | B1 |
8230066 | Heil | Jul 2012 | B2 |
8234377 | Cohn | Jul 2012 | B2 |
8266238 | Zimmer et al. | Sep 2012 | B2 |
8272104 | Chen et al. | Sep 2012 | B2 |
8274993 | Sharma et al. | Sep 2012 | B2 |
8290919 | Kelly et al. | Oct 2012 | B1 |
8297722 | Chambers et al. | Oct 2012 | B2 |
8301746 | Head et al. | Oct 2012 | B2 |
8335231 | Kloth et al. | Dec 2012 | B2 |
8341121 | Claudatos et al. | Dec 2012 | B1 |
8345692 | Smith | Jan 2013 | B2 |
8352941 | Protopopov et al. | Jan 2013 | B1 |
8392760 | Kandula et al. | Mar 2013 | B2 |
8442059 | de la Iglesia et al. | May 2013 | B1 |
8479211 | Marshall et al. | Jul 2013 | B1 |
8495356 | Ashok et al. | Jul 2013 | B2 |
8514868 | Hill | Aug 2013 | B2 |
8532108 | Li et al. | Sep 2013 | B2 |
8560663 | Baucke et al. | Oct 2013 | B2 |
8619599 | Even | Dec 2013 | B1 |
8626891 | Guru et al. | Jan 2014 | B2 |
8630983 | Sengupta et al. | Jan 2014 | B2 |
8660129 | Brendel et al. | Feb 2014 | B1 |
8661299 | Ip | Feb 2014 | B1 |
8677485 | Sharma et al. | Mar 2014 | B2 |
8683296 | Anderson et al. | Mar 2014 | B2 |
8706772 | Hartig et al. | Apr 2014 | B2 |
8719804 | Jain | May 2014 | B2 |
8725854 | Edsall | May 2014 | B2 |
8768981 | Milne et al. | Jul 2014 | B1 |
8775773 | Acharya et al. | Jul 2014 | B2 |
8793372 | Ashok et al. | Jul 2014 | B2 |
8805918 | Chandrasekaran et al. | Aug 2014 | B1 |
8805951 | Faibish et al. | Aug 2014 | B1 |
8832330 | Lancaster | Sep 2014 | B1 |
8855116 | Rosset et al. | Oct 2014 | B2 |
8856339 | Mestery et al. | Oct 2014 | B2 |
8868474 | Leung et al. | Oct 2014 | B2 |
8887286 | Dupont et al. | Nov 2014 | B2 |
8898385 | Jayaraman et al. | Nov 2014 | B2 |
8909928 | Ahmad et al. | Dec 2014 | B2 |
8918510 | Gmach et al. | Dec 2014 | B2 |
8918586 | Todd et al. | Dec 2014 | B1 |
8924720 | Raghuram et al. | Dec 2014 | B2 |
8930747 | Levijarvi et al. | Jan 2015 | B2 |
8935500 | Gulati et al. | Jan 2015 | B1 |
8949677 | Brundage et al. | Feb 2015 | B1 |
8996837 | Bono et al. | Mar 2015 | B1 |
9003086 | Schuller et al. | Apr 2015 | B1 |
9007922 | Mittal et al. | Apr 2015 | B1 |
9009427 | Sharma et al. | Apr 2015 | B2 |
9009704 | McGrath et al. | Apr 2015 | B2 |
9075638 | Barnett et al. | Jul 2015 | B2 |
9141554 | Candelaria | Sep 2015 | B1 |
9141785 | Mukkara et al. | Sep 2015 | B2 |
9164795 | Vincent | Oct 2015 | B1 |
9176677 | Fradkin et al. | Nov 2015 | B1 |
9201704 | Chang et al. | Dec 2015 | B2 |
9203784 | Chang et al. | Dec 2015 | B2 |
9207882 | Rosset et al. | Dec 2015 | B2 |
9207929 | Katsura | Dec 2015 | B2 |
9213612 | Candelaria | Dec 2015 | B2 |
9223564 | Munireddy et al. | Dec 2015 | B2 |
9223634 | Chang et al. | Dec 2015 | B2 |
9244761 | Yekhanin et al. | Jan 2016 | B2 |
9250969 | Lager-Cavilla et al. | Feb 2016 | B2 |
9264494 | Factor et al. | Feb 2016 | B2 |
9270754 | Iyengar et al. | Feb 2016 | B2 |
9280487 | Candelaria | Mar 2016 | B2 |
9304815 | Vasanth et al. | Apr 2016 | B1 |
9313048 | Chang et al. | Apr 2016 | B2 |
9374270 | Nakil et al. | Jun 2016 | B2 |
9378060 | Jansson et al. | Jun 2016 | B2 |
9396251 | Boudreau et al. | Jul 2016 | B1 |
9448877 | Candelaria | Sep 2016 | B2 |
9471348 | Zuo et al. | Oct 2016 | B2 |
9501473 | Kong et al. | Nov 2016 | B1 |
9503523 | Rosset et al. | Nov 2016 | B2 |
9565110 | Mullendore et al. | Feb 2017 | B2 |
9575828 | Agarwal et al. | Feb 2017 | B2 |
9582377 | Dhoolam et al. | Feb 2017 | B1 |
9614763 | Dong et al. | Apr 2017 | B2 |
9658868 | Hill | May 2017 | B2 |
9658876 | Chang et al. | May 2017 | B2 |
9733868 | Chandrasekaran et al. | Aug 2017 | B2 |
9763518 | Charest et al. | Sep 2017 | B2 |
9830240 | George et al. | Nov 2017 | B2 |
9853873 | Dasu et al. | Dec 2017 | B2 |
20020049980 | Hoang | Apr 2002 | A1 |
20020053009 | Selkirk et al. | May 2002 | A1 |
20020073276 | Howard et al. | Jun 2002 | A1 |
20020083120 | Soltis | Jun 2002 | A1 |
20020095547 | Watanabe et al. | Jul 2002 | A1 |
20020103889 | Markson et al. | Aug 2002 | A1 |
20020103943 | Lo et al. | Aug 2002 | A1 |
20020112113 | Karpoff et al. | Aug 2002 | A1 |
20020120741 | Webb et al. | Aug 2002 | A1 |
20020138675 | Mann | Sep 2002 | A1 |
20020156971 | Jones et al. | Oct 2002 | A1 |
20030023885 | Potter et al. | Jan 2003 | A1 |
20030026267 | Oberman et al. | Feb 2003 | A1 |
20030055933 | Ishizaki et al. | Mar 2003 | A1 |
20030056126 | O'Connor et al. | Mar 2003 | A1 |
20030065986 | Fraenkel et al. | Apr 2003 | A1 |
20030084359 | Bresniker et al. | May 2003 | A1 |
20030118053 | Edsall et al. | Jun 2003 | A1 |
20030131105 | Czeiger et al. | Jul 2003 | A1 |
20030131165 | Asano et al. | Jul 2003 | A1 |
20030131182 | Kumar et al. | Jul 2003 | A1 |
20030140134 | Swanson et al. | Jul 2003 | A1 |
20030140210 | Testardi | Jul 2003 | A1 |
20030149763 | Heitman et al. | Aug 2003 | A1 |
20030154271 | Baldwin et al. | Aug 2003 | A1 |
20030159058 | Eguchi et al. | Aug 2003 | A1 |
20030174725 | Shankar | Sep 2003 | A1 |
20030189395 | Doornbos et al. | Oct 2003 | A1 |
20030210686 | Terrell et al. | Nov 2003 | A1 |
20040024961 | Cochran et al. | Feb 2004 | A1 |
20040030857 | Krakirian et al. | Feb 2004 | A1 |
20040039939 | Cox et al. | Feb 2004 | A1 |
20040054776 | Klotz et al. | Mar 2004 | A1 |
20040057389 | Klotz et al. | Mar 2004 | A1 |
20040059807 | Klotz et al. | Mar 2004 | A1 |
20040088574 | Walter et al. | May 2004 | A1 |
20040117438 | Considine et al. | Jun 2004 | A1 |
20040123029 | Dalai et al. | Jun 2004 | A1 |
20040128470 | Hetzler et al. | Jul 2004 | A1 |
20040128540 | Roskind | Jul 2004 | A1 |
20040153863 | Klotz et al. | Aug 2004 | A1 |
20040190901 | Fang | Sep 2004 | A1 |
20040215749 | Tsao | Oct 2004 | A1 |
20040230848 | Mayo et al. | Nov 2004 | A1 |
20040250034 | Yagawa et al. | Dec 2004 | A1 |
20050033936 | Nakano et al. | Feb 2005 | A1 |
20050036499 | Dutt et al. | Feb 2005 | A1 |
20050050211 | Kaul et al. | Mar 2005 | A1 |
20050050270 | Horn et al. | Mar 2005 | A1 |
20050053073 | Kloth et al. | Mar 2005 | A1 |
20050055428 | Terai et al. | Mar 2005 | A1 |
20050060574 | Klotz et al. | Mar 2005 | A1 |
20050060598 | Klotz et al. | Mar 2005 | A1 |
20050071851 | Opheim | Mar 2005 | A1 |
20050076113 | Klotz et al. | Apr 2005 | A1 |
20050091426 | Horn et al. | Apr 2005 | A1 |
20050114611 | Durham et al. | May 2005 | A1 |
20050114615 | Ogasawara et al. | May 2005 | A1 |
20050117522 | Basavaiah et al. | Jun 2005 | A1 |
20050117562 | Wrenn | Jun 2005 | A1 |
20050138287 | Ogasawara et al. | Jun 2005 | A1 |
20050169188 | Cometto et al. | Aug 2005 | A1 |
20050185597 | Le et al. | Aug 2005 | A1 |
20050188170 | Yamamoto | Aug 2005 | A1 |
20050198523 | Shanbhag et al. | Sep 2005 | A1 |
20050235072 | Smith et al. | Oct 2005 | A1 |
20050283658 | Clark et al. | Dec 2005 | A1 |
20060015861 | Takata et al. | Jan 2006 | A1 |
20060015928 | Setty et al. | Jan 2006 | A1 |
20060034302 | Peterson | Feb 2006 | A1 |
20060045021 | Deragon et al. | Mar 2006 | A1 |
20060075191 | Lolayekar et al. | Apr 2006 | A1 |
20060098672 | Schzukin et al. | May 2006 | A1 |
20060117099 | Mogul | Jun 2006 | A1 |
20060136684 | Le et al. | Jun 2006 | A1 |
20060184287 | Belady et al. | Aug 2006 | A1 |
20060198319 | Schondelmayer et al. | Sep 2006 | A1 |
20060215297 | Kikuchi | Sep 2006 | A1 |
20060230227 | Ogasawara et al. | Oct 2006 | A1 |
20060242332 | Johnsen et al. | Oct 2006 | A1 |
20060251111 | Kloth et al. | Nov 2006 | A1 |
20070005297 | Beresniewicz et al. | Jan 2007 | A1 |
20070067593 | Satoyama et al. | Mar 2007 | A1 |
20070079068 | Draggon | Apr 2007 | A1 |
20070091903 | Atkinson | Apr 2007 | A1 |
20070094465 | Sharma et al. | Apr 2007 | A1 |
20070101202 | Garbow | May 2007 | A1 |
20070121519 | Cuni et al. | May 2007 | A1 |
20070136541 | Herz et al. | Jun 2007 | A1 |
20070162969 | Becker | Jul 2007 | A1 |
20070211640 | Palacharla et al. | Sep 2007 | A1 |
20070214316 | Kim | Sep 2007 | A1 |
20070250838 | Belady et al. | Oct 2007 | A1 |
20070258380 | Chamdani et al. | Nov 2007 | A1 |
20070263545 | Foster et al. | Nov 2007 | A1 |
20070276884 | Hara et al. | Nov 2007 | A1 |
20070283059 | Ho et al. | Dec 2007 | A1 |
20080016412 | White et al. | Jan 2008 | A1 |
20080034149 | Sheen | Feb 2008 | A1 |
20080052459 | Chang et al. | Feb 2008 | A1 |
20080059698 | Kabir et al. | Mar 2008 | A1 |
20080114933 | Ogasawara et al. | May 2008 | A1 |
20080126509 | Subrannanian et al. | May 2008 | A1 |
20080126734 | Murase | May 2008 | A1 |
20080168304 | Flynn et al. | Jul 2008 | A1 |
20080201616 | Ashmore | Aug 2008 | A1 |
20080244184 | Lewis et al. | Oct 2008 | A1 |
20080256082 | Davies et al. | Oct 2008 | A1 |
20080267217 | Colville et al. | Oct 2008 | A1 |
20080288661 | Galles | Nov 2008 | A1 |
20080294888 | Ando et al. | Nov 2008 | A1 |
20090063766 | Matsumura et al. | Mar 2009 | A1 |
20090083484 | Basham et al. | Mar 2009 | A1 |
20090089567 | Boland, IV et al. | Apr 2009 | A1 |
20090094380 | Qiu et al. | Apr 2009 | A1 |
20090094664 | Butler et al. | Apr 2009 | A1 |
20090125694 | Innan et al. | May 2009 | A1 |
20090193223 | Saliba et al. | Jul 2009 | A1 |
20090201926 | Kagan et al. | Aug 2009 | A1 |
20090222733 | Basham et al. | Sep 2009 | A1 |
20090240873 | Yu et al. | Sep 2009 | A1 |
20090282471 | Green et al. | Nov 2009 | A1 |
20090323706 | Germain et al. | Dec 2009 | A1 |
20100011365 | Gerovac et al. | Jan 2010 | A1 |
20100030995 | Wang et al. | Feb 2010 | A1 |
20100046378 | Knapp et al. | Feb 2010 | A1 |
20100083055 | Ozonat | Apr 2010 | A1 |
20100174968 | Charles et al. | Jul 2010 | A1 |
20100318609 | Lahiri et al. | Dec 2010 | A1 |
20100318837 | Murphy et al. | Dec 2010 | A1 |
20110010394 | Carew et al. | Jan 2011 | A1 |
20110022691 | Banerjee et al. | Jan 2011 | A1 |
20110029824 | Schöler et al. | Feb 2011 | A1 |
20110035494 | Pandey et al. | Feb 2011 | A1 |
20110075667 | Li et al. | Mar 2011 | A1 |
20110087848 | Trent | Apr 2011 | A1 |
20110119556 | de Buen | May 2011 | A1 |
20110142053 | Van Der Merwe et al. | Jun 2011 | A1 |
20110161496 | Nicklin | Jun 2011 | A1 |
20110173303 | Rider | Jul 2011 | A1 |
20110228679 | Varma et al. | Sep 2011 | A1 |
20110231899 | Pulier et al. | Sep 2011 | A1 |
20110239039 | Dieffenbach et al. | Sep 2011 | A1 |
20110252274 | Kawaguchi et al. | Oct 2011 | A1 |
20110255540 | Mizrahi et al. | Oct 2011 | A1 |
20110276584 | Cotner et al. | Nov 2011 | A1 |
20110276675 | Singh et al. | Nov 2011 | A1 |
20110276951 | Jain | Nov 2011 | A1 |
20110299539 | Rajagopal et al. | Dec 2011 | A1 |
20110307450 | Hahn et al. | Dec 2011 | A1 |
20110313973 | Srivas et al. | Dec 2011 | A1 |
20120023319 | Chin et al. | Jan 2012 | A1 |
20120030401 | Cowan et al. | Feb 2012 | A1 |
20120054367 | Ramakrishnan et al. | Mar 2012 | A1 |
20120072578 | Alam | Mar 2012 | A1 |
20120072985 | Davne et al. | Mar 2012 | A1 |
20120075999 | Ko et al. | Mar 2012 | A1 |
20120084445 | Brock et al. | Apr 2012 | A1 |
20120084782 | Chou et al. | Apr 2012 | A1 |
20120096134 | Suit | Apr 2012 | A1 |
20120130874 | Mane et al. | May 2012 | A1 |
20120131174 | Ferris et al. | May 2012 | A1 |
20120134672 | Banerjee | May 2012 | A1 |
20120144014 | Natham et al. | Jun 2012 | A1 |
20120159112 | Tokusho et al. | Jun 2012 | A1 |
20120167094 | Suit | Jun 2012 | A1 |
20120173581 | Hartig et al. | Jul 2012 | A1 |
20120173589 | Kwon et al. | Jul 2012 | A1 |
20120177039 | Berman | Jul 2012 | A1 |
20120177041 | Berman | Jul 2012 | A1 |
20120177042 | Berman | Jul 2012 | A1 |
20120177043 | Berman | Jul 2012 | A1 |
20120177044 | Berman | Jul 2012 | A1 |
20120177045 | Berman | Jul 2012 | A1 |
20120177370 | Berman | Jul 2012 | A1 |
20120179909 | Sagi et al. | Jul 2012 | A1 |
20120201138 | Yu et al. | Aug 2012 | A1 |
20120210041 | Flynn et al. | Aug 2012 | A1 |
20120254440 | Wang | Oct 2012 | A1 |
20120257501 | Kucharczyk | Oct 2012 | A1 |
20120265976 | Spiers et al. | Oct 2012 | A1 |
20120281706 | Agarwal et al. | Nov 2012 | A1 |
20120297088 | Wang et al. | Nov 2012 | A1 |
20120303618 | Dutta et al. | Nov 2012 | A1 |
20120311106 | Morgan | Dec 2012 | A1 |
20120311568 | Jansen | Dec 2012 | A1 |
20120320788 | Venkataramanan et al. | Dec 2012 | A1 |
20120324114 | Dutta et al. | Dec 2012 | A1 |
20120331119 | Bose et al. | Dec 2012 | A1 |
20130003737 | Sinicrope | Jan 2013 | A1 |
20130013664 | Baird et al. | Jan 2013 | A1 |
20130028135 | Berman | Jan 2013 | A1 |
20130036212 | Jibbe et al. | Feb 2013 | A1 |
20130036213 | Hasan et al. | Feb 2013 | A1 |
20130036449 | Mukkara et al. | Feb 2013 | A1 |
20130054888 | Bhat et al. | Feb 2013 | A1 |
20130061089 | Valiyaparambil et al. | Mar 2013 | A1 |
20130067162 | Jayaraman et al. | Mar 2013 | A1 |
20130080823 | Roth et al. | Mar 2013 | A1 |
20130086340 | Fleming et al. | Apr 2013 | A1 |
20130100858 | Kamath et al. | Apr 2013 | A1 |
20130111540 | Sabin | May 2013 | A1 |
20130138816 | Kuo et al. | May 2013 | A1 |
20130138836 | Cohen et al. | May 2013 | A1 |
20130139138 | Kakos | May 2013 | A1 |
20130144933 | Hinni et al. | Jun 2013 | A1 |
20130152076 | Patel | Jun 2013 | A1 |
20130152175 | Hromoko et al. | Jun 2013 | A1 |
20130163426 | Beliveau et al. | Jun 2013 | A1 |
20130163606 | Bagepalli et al. | Jun 2013 | A1 |
20130179941 | McGloin et al. | Jul 2013 | A1 |
20130182712 | Aguayo et al. | Jul 2013 | A1 |
20130185433 | Zhu et al. | Jul 2013 | A1 |
20130191106 | Kephart et al. | Jul 2013 | A1 |
20130198730 | Munireddy et al. | Aug 2013 | A1 |
20130208888 | Agrawal et al. | Aug 2013 | A1 |
20130212130 | Rahnama | Aug 2013 | A1 |
20130223236 | Dickey | Aug 2013 | A1 |
20130238641 | Mandelstein et al. | Sep 2013 | A1 |
20130266307 | Garg et al. | Oct 2013 | A1 |
20130268922 | Tiwari et al. | Oct 2013 | A1 |
20130275470 | Cao et al. | Oct 2013 | A1 |
20130297769 | Chang et al. | Nov 2013 | A1 |
20130318134 | Bolik et al. | Nov 2013 | A1 |
20130318288 | Khan et al. | Nov 2013 | A1 |
20140006708 | Huynh et al. | Jan 2014 | A1 |
20140016493 | Johnsson et al. | Jan 2014 | A1 |
20140019684 | Wei et al. | Jan 2014 | A1 |
20140025770 | Warfield et al. | Jan 2014 | A1 |
20140029441 | Nydell | Jan 2014 | A1 |
20140029442 | Wallman | Jan 2014 | A1 |
20140039683 | Zimmermann et al. | Feb 2014 | A1 |
20140040473 | Ho et al. | Feb 2014 | A1 |
20140040883 | Tompkins | Feb 2014 | A1 |
20140047201 | Mehta | Feb 2014 | A1 |
20140053264 | Dubrovsky et al. | Feb 2014 | A1 |
20140059187 | Rosset et al. | Feb 2014 | A1 |
20140059266 | Ben-Michael et al. | Feb 2014 | A1 |
20140086253 | Yong | Mar 2014 | A1 |
20140089273 | Borshack et al. | Mar 2014 | A1 |
20140095556 | Lee et al. | Apr 2014 | A1 |
20140096249 | Dupont et al. | Apr 2014 | A1 |
20140105009 | Vos et al. | Apr 2014 | A1 |
20140108474 | David et al. | Apr 2014 | A1 |
20140109071 | Ding et al. | Apr 2014 | A1 |
20140112122 | Kapadia et al. | Apr 2014 | A1 |
20140123207 | Agarwal et al. | May 2014 | A1 |
20140156557 | Zeng et al. | Jun 2014 | A1 |
20140164666 | Yang | Jun 2014 | A1 |
20140164866 | Bolotov et al. | Jun 2014 | A1 |
20140172371 | Zhu et al. | Jun 2014 | A1 |
20140173060 | Jubran et al. | Jun 2014 | A1 |
20140173195 | Rosset et al. | Jun 2014 | A1 |
20140173579 | McDonald et al. | Jun 2014 | A1 |
20140189278 | Peng | Jul 2014 | A1 |
20140198794 | Mehta et al. | Jul 2014 | A1 |
20140211661 | Gorkemli et al. | Jul 2014 | A1 |
20140215265 | Mohanta et al. | Jul 2014 | A1 |
20140215590 | Brand | Jul 2014 | A1 |
20140219086 | Cantu′ et al. | Aug 2014 | A1 |
20140222953 | Karve et al. | Aug 2014 | A1 |
20140229790 | Goss et al. | Aug 2014 | A1 |
20140244585 | Sivasubramanian et al. | Aug 2014 | A1 |
20140244897 | Goss et al. | Aug 2014 | A1 |
20140245435 | Belenky | Aug 2014 | A1 |
20140269390 | Ciodaru et al. | Sep 2014 | A1 |
20140281700 | Nagesharao et al. | Sep 2014 | A1 |
20140297941 | Rajani et al. | Oct 2014 | A1 |
20140307578 | DeSanti | Oct 2014 | A1 |
20140317206 | Lomelino et al. | Oct 2014 | A1 |
20140324862 | Bingham et al. | Oct 2014 | A1 |
20140325208 | Resch et al. | Oct 2014 | A1 |
20140331276 | Frascadore et al. | Nov 2014 | A1 |
20140348166 | Yang et al. | Nov 2014 | A1 |
20140355450 | Bhikkaji et al. | Dec 2014 | A1 |
20140366155 | Chang et al. | Dec 2014 | A1 |
20140376550 | Khan et al. | Dec 2014 | A1 |
20150003450 | Salam et al. | Jan 2015 | A1 |
20150003458 | Li et al. | Jan 2015 | A1 |
20150003463 | Li et al. | Jan 2015 | A1 |
20150010001 | Duda et al. | Jan 2015 | A1 |
20150016461 | Qiang | Jan 2015 | A1 |
20150030024 | Venkataswami et al. | Jan 2015 | A1 |
20150046123 | Kato | Feb 2015 | A1 |
20150063353 | Kapadia et al. | Mar 2015 | A1 |
20150067001 | Koltsidas | Mar 2015 | A1 |
20150082432 | Eaton et al. | Mar 2015 | A1 |
20150092824 | Wicker, Jr. et al. | Apr 2015 | A1 |
20150120907 | Niestemski et al. | Apr 2015 | A1 |
20150120969 | He | Apr 2015 | A1 |
20150121131 | Kiselev et al. | Apr 2015 | A1 |
20150127979 | Doppalapudi | May 2015 | A1 |
20150142840 | Baldwin et al. | May 2015 | A1 |
20150169313 | Katsura | Jun 2015 | A1 |
20150180672 | Kuwata | Jun 2015 | A1 |
20150207763 | Bertran Ortiz et al. | Jun 2015 | A1 |
20150205974 | Talley et al. | Jul 2015 | A1 |
20150222444 | Sarkar | Aug 2015 | A1 |
20150229546 | Somaiya et al. | Aug 2015 | A1 |
20150248366 | Bergsten et al. | Sep 2015 | A1 |
20150248418 | Bhardwaj et al. | Sep 2015 | A1 |
20150254003 | Lee et al. | Sep 2015 | A1 |
20150254088 | Chou et al. | Sep 2015 | A1 |
20150261446 | Lee | Sep 2015 | A1 |
20150263887 | Sajeepa | Sep 2015 | A1 |
20150263993 | Kuch et al. | Sep 2015 | A1 |
20150269048 | Marr et al. | Sep 2015 | A1 |
20150277804 | Arnold et al. | Oct 2015 | A1 |
20150281067 | Wu | Oct 2015 | A1 |
20150303949 | Jafarkhani et al. | Oct 2015 | A1 |
20150341237 | Cuni et al. | Nov 2015 | A1 |
20150341239 | Bertran Ortiz et al. | Nov 2015 | A1 |
20150358136 | Medard | Dec 2015 | A1 |
20150379150 | Duda | Dec 2015 | A1 |
20160004611 | Lakshman et al. | Jan 2016 | A1 |
20160011936 | Luby | Jan 2016 | A1 |
20160011942 | Golbourn et al. | Jan 2016 | A1 |
20160054922 | Awasthi et al. | Feb 2016 | A1 |
20160062820 | Jones et al. | Mar 2016 | A1 |
20160070652 | Sundararaman et al. | Mar 2016 | A1 |
20160087885 | Tripathi et al. | Mar 2016 | A1 |
20160088083 | Bharadwaj et al. | Mar 2016 | A1 |
20160098358 | Cha | Apr 2016 | A1 |
20160119159 | Zhao et al. | Apr 2016 | A1 |
20160119421 | Semke et al. | Apr 2016 | A1 |
20160139820 | Fluman et al. | May 2016 | A1 |
20160149639 | Pham et al. | May 2016 | A1 |
20160205189 | Mopur et al. | Jul 2016 | A1 |
20160210161 | Rosset et al. | Jul 2016 | A1 |
20160231928 | Lewis et al. | Aug 2016 | A1 |
20160274926 | Narasimhamurthy et al. | Sep 2016 | A1 |
20160285760 | Dong | Sep 2016 | A1 |
20160292359 | Tellis et al. | Oct 2016 | A1 |
20160294983 | Kliteynik et al. | Oct 2016 | A1 |
20160334998 | George et al. | Nov 2016 | A1 |
20160366094 | Mason et al. | Dec 2016 | A1 |
20160378624 | Jenkins, Jr. et al. | Dec 2016 | A1 |
20160380694 | Guduru | Dec 2016 | A1 |
20170010874 | Rosset | Jan 2017 | A1 |
20170010930 | Dutta et al. | Jan 2017 | A1 |
20170019475 | Metz et al. | Jan 2017 | A1 |
20170068630 | Iskandar et al. | Mar 2017 | A1 |
20170139799 | Puri | May 2017 | A1 |
20170168970 | Sajeepa et al. | Jun 2017 | A1 |
20170177541 | Berman | Jun 2017 | A1 |
20170177860 | Suarez et al. | Jun 2017 | A1 |
20170199768 | Arroyo | Jul 2017 | A1 |
20170212858 | Chu et al. | Jul 2017 | A1 |
20170273019 | Park et al. | Sep 2017 | A1 |
20170277655 | Das | Sep 2017 | A1 |
20170337097 | Sipos et al. | Nov 2017 | A1 |
20170340113 | Charest et al. | Nov 2017 | A1 |
20170371558 | George et al. | Dec 2017 | A1 |
20180097707 | Wright et al. | Apr 2018 | A1 |
20180225254 | Hu | Aug 2018 | A1 |
Number | Date | Country |
---|---|---|
2228719 | Sep 2010 | EP |
2439637 | Apr 2012 | EP |
2680155 | Jan 2014 | EP |
2350028 | May 2001 | GB |
2000-242434 | Sep 2000 | JP |
1566104 | Jan 2017 | TW |
WO 2004077214 | Sep 2004 | WO |
WO 2016003408 | Jan 2016 | WO |
WO 2016003489 | Jan 2016 | WO |
Entry |
---|
Author Unknown, “5 Benefits of a Storage Gateway in the Cloud,” Blog, TwinStrata, Inc., posted Jul. 10, 2012, 4 pages, https://web.archive.org/web/20120725092619/http://blog.twinstrata.com/2012/07/10//5-benefits-of-a-storage-gateway-in-the-cloud. |
Author Unknown, “Configuration Interface for IBM System Storage DS5000, IBM DS4000, and IBM DS3000 Systems,” IBM SAN Volume Controller Version 7.1, IBM® System Storage® SAN Volume Controller Information Center, Jun. 16, 2013, 3 pages. |
Author Unknown, “Coraid EtherCloud, Software-Defined Storage with Scale-Out Infrastructure,” Solution Brief, 2013, 2 pages, Coraid, Redwood City, California, U.S.A. |
Author Unknown, “Coraid Virtual DAS (VDAS) Technology: Eliminate Tradeoffs between DAS and Networked Storage,” Coraid Technology Brief, © 2013 Cora id, Inc., Published on or about Mar. 20, 2013, 2 pages. |
Author Unknown, “Creating Performance-based SAN SLAs Using Finisar's NetWisdom” May 2006, 7 pages, Finisar Corporation, Sunnyvale, California, U.S.A. |
Author Unknown, “Data Center, Metro Cloud Connectivity: Integrated Metro SAN Connectivity in 16 Gbps Switches,” Brocade Communication Systems, Inc., Apr. 2011, 14 pages. |
Author Unknown, “Data Center, SAN Fabric Administration Best Practices Guide, Support Perspective,” Brocade Communication Systems, Inc., May 2013, 21 pages. |
Author Unknown, “delphi—Save a CRC value in a file, without altering the actual CRC Checksum?” Stack Overflow, stackoverflow.com, Dec. 23, 2011, XP055130879, 3 pages http://stackoverflow.com/questions/8608219/save-a-crc-value-in-a-file-wihout-altering-the-actual-crc-checksum. |
Author Unknown, “EMC Unisphere: Innovative Approach to Managing Low-End and Midrange Storage; Redefining Simplicity in the Entry-Level and Midrange Storage Markets,” Data Sheet, EMC Corporation; published on or about Jan. 4, 2013 [Retrieved and printed Sep. 12, 2013] 6 pages http://www.emc.com/storage/vnx/unisphere.htm. |
Author Unknown, “HP XP Array Manager Software—Overview & Features,” Storage Device Management Software; Hewlett-Packard Development Company, 3 pages; © 2013 Hewlett-Packard Development Company, L.P. |
Author Unknown, “Joint Cisco and VMWare Solution for Optimizing Virtual Desktop Delivery: Data Center 3.0: Solutions to Accelerate Data Center Virtualization,” Cisco Systems, Inc. and VMware, Inc., Sep. 2008, 10 pages. |
Author Unknown, “Network Transformation with Software-Defined Networking and Ethernet Fabrics,” Positioning Paper, 2012, 6 pages, Brocade Communications Systems. |
Author Unknown, “Recreating Real Application Traffic in Junosphere Lab,” Solution Brief, Juniper Networks, Dec. 2011, 3 pages. |
Author Unknown, “Shunra for HP Softwarer, Enabiling Confidence in Application Performance Before Deployment,” 2010, 2 pages. |
Author Unknown, “Software Defined Networking: The New Norm for Networks,” White Paper, Open Networking Foundation, Apr. 13, 2012, 12 pages. |
Author Unknown, “Software Defined Storage Networks An Introduction,” White Paper, Doc # 01-000030-001 Rev. A, Dec. 12, 2012, 8 pages; Jeda Networks, Newport Beach, California, U.S.A. |
Author Unknown, “Standard RAID Levels,” Wikipedia, the Free Encyclopedia, last updated Jul. 18, 2014, 7 pages; http://en.wikipedia.org/wiki/Standard_RAID_levels. |
Author Unknown, “Storage Infrastructure for the Cloud,” Solution Brief, © 2012, 3 pages; coraid, Redwood City, California, U.S.A. |
Author Unknown, “Storage Area Network—NPIV: Emulex Virtual HBA and Brocade, Proven Interoperability and Proven Solution,” Technical Brief, Apr. 2008, 4 pages, Emulex and Brocade Communications Systems. |
Author Unknown, “The Fundamentals of Software-Defined Storage, Simplicity at Scale for Cloud-Architectures” Solution Brief, 2013, 3 pages; Coraid, Redwood City, California, U.S.A. |
Author Unknown, “VirtualWisdom® SAN Performance Probe Family Models: Probe FC8, HD, and HD48,” Virtual Instruments Data Sheet, Apr. 2014 Virtual Instruments. All Rights Reserved; 4 pages. |
Author Unknown, “Xgig Analyzer: Quick Start Feature Guide 4.0,” Feb. 2008, 24 pages, Finisar Corporation, Sunnyvale, California, U.S.A. |
Author Unknown, “Sun Storage Common Array Manager Installation and Setup Guide,” Software Installation and Setup Guide Version 6.7.x 821-1362-10, Appendix D: Configuring In-Band Management, Sun Oracle; retrieved and printed Sep. 12, 2013, 15 pages. |
Author Unknown, “Vblock Solution for SAP: Simplified Provisioning for Operation Efficiency,” VCE White Paper, VCE—The Virtual Computing Environment Company, Aug. 2011, 11 pages. |
Aweya, James, et al., “Multi-level active queue management with dynamic thresholds,” Elsevier, Computer Communications 25 (2002) pp. 756-771. |
Berman, Stuart, et al., “Start-Up Jeda Networks in Software Defined Storage Network Technology,” Press Release, Feb. 25, 2013, 2 pages, http://www.storagenewsletter.com/news/startups/jeda-networks. |
Borovick, Lucinda, et al., “White Paper, Architecting the Network for the Cloud,” IDC Analyze the Future, Jan. 2011, pp. 1-8. |
Chakrabarti, Kaushik, et al., “Locally Adaptive Dimensionality Reduction for Indexing Large Time Series Databases,” ACM Transactions on Database Systems, vol. 27, No. 2, Jun. 2009, pp. 188-228. |
Chandola, Varun, et al., “A Gaussian Process Based Online Change Detection Algorithm for Monitoring Periodic Time Series,” Proceedings of the Eleventh SIAM International Conference on Data Mining, SDM 2011, Apr. 28-30, 2011, 12 pages. |
Cisco Systems, Inc. “N-Port Virtualization in the Data Center,” Cisco White Paper, Cisco Systems, Inc., Mar. 2008, 7 pages. |
Cisco Systems, Inc., “Best Practices in Deploying Cisco Nexus 1000V Series Switches on Cisco UCS B and C Series Cisco UCS Manager Servers,” White Paper, Cisco Systems, Inc., Apr. 2011, 36 pages. |
Cisco Systems, Inc., “Cisco Prime Data Center Network Manager 6.1,” At-A-Glance, © 2012, 3 pages. |
Cisco Systems, Inc., “Cisco Prime Data Center Network Manager,” Release 6.1 Data Sheet, © 2012, 10 pages. |
Cisco Systems, Inc., “Cisco Unified Network Services: Overcome Obstacles to Cloud-Ready Deployments,” White Paper, Cisco Systems, Inc., Jan. 2011, 6 pages. |
Clarke, Alan, et al., “Open Data Center Alliance Usage: Virtual Machine (VM) Interoperability in a Hybrid Cloud Environment Rev. 1.2,” Open Data Center Alliance, Inc., 2013, pp. 1-18. |
Cummings, Roger, et al., Fibre Channel-Fabric Generic Requirements (FC-FG), Dec. 4, 1996, 33 pages, American National Standards Institute, Inc., New York, New York, U.S.A. |
Farber, Franz, et al. “An In-Memory Database System for Multi-Tenant Applications,” Proceedings of 14th Business, Technology and Web (BTW) Conference on Database Systems for Business, Technology, and Web, Feb. 28 to Mar. 4, 2011, 17 pages, University of Kaiserslautern, Germany. |
Guo, Chang Jie, et al., “IBM Resarch Report: Data Integration and Composite Business Services, Part 3, Building a Multi-Tenant Data Tier with with [sic] Access Control and Security,” RC24426 (C0711-037), Nov. 19, 2007, 20 pages, IBM. |
Hatzieleftheriou, Andromachi, et al., “Host-side Filesystem Journaling for Durable Shared Storage,” 13th USENIX Conference on File and Storage Technologies (FAST '15), Feb. 16-19, 2015, 9 pages; https://www.usenix.org/system/files/conference/fast15/fast15-paper-hatzieleftheriou.pdf. |
Hedayat, K., et al., “A Two-Way Active Measurement Protocol (TWAMP),” Network Working Group, RFC 5357, Oct. 2008, 26 pages. |
Horn, C., et al., “Online anomaly detection with expert system feedback in social networks,” 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), May 22-27, 2011, 2 pages, Prague; [Abstract only]. |
Hosterman, Cody, et al., “Using EMC Symmetrix Storage inVMware vSph ere Environments,” Version 8.0, EMC2Techbooks, EMC Corporation; published on or about Jul. 8, 2008, 314 pages; [Retrieved and printed Sep. 12, 2013]. |
Hu, Yuchong, et al., “Cooperative Recovery of Distributed Storage Systems from Multiple Losses with Network Coding,” University of Science & Technology of China, Feb. 2010, 9 pages. |
Keogh, Eamonn, et al., “Dimensionality Reduction for Fast Similarity Search in Large Time Series Databases,” KAIS Long Paper submitted May 16, 2000; 19 pages. |
Kolyshkin, Kirill, “Virtualization in Linux,” Sep. 1, 2006, pp. 1-5. |
Kovar, Joseph F., “Startup Jeda Networks Takes SDN Approach to Storage Networks,” CRN Press Release, Feb. 22, 2013, 1 page, http://www.crn.com/240149244/printablearticle.htm. |
Lampson, Butler, W., et al., “Crash Recovery in a Distributed Data Storage System,” Jun. 1, 1979, 28 pages. |
Lewis, Michael E., et al., “Design of an Advanced Development Model Optical Disk-Based Redundant Array of Independent Disks (RAID) High Speed Mass Storage Subsystem,” Final Technical Report, Oct. 1997, pp. 1-211. |
Lin, Jessica, “Finding Motifs in Time Series,” SIGKDD'02 Jul. 23-26, 2002, 11 pages, Edmonton, Alberta, Canada. |
Linthicum, David, “VM Import could be a game changer for hybrid clouds”, InfoWorld, Dec. 23, 2010, 4 pages. |
Long, Abraham Jr., “Modeling the Reliability of RAID Sets,” Dell Power Solutions, May 2008, 4 pages. |
Ma, Ao, et al., “RAIDShield: Characterizing, Monitoring, and Proactively Protecting Against Disk Failures,” FAST '15, 13th USENIX Conference on File and Storage Technologies, Feb. 16-19, 2015, 17 pages, Santa Clara, California, U.S.A. |
Mahalingam, M., et al., “Virtual eXtensible Local Area Network (VXLAN): A Framework for Overlaying Virtualized Layer 2 Networks over Layer 3 Networks,” Independent Submission, RFC 7348, Aug. 2014, 22 pages; http://www.hjp.at/doc/rfc/rfc7348.html. |
McQuerry, Steve, “Cisco UCS M-Series Modular Servers for Cloud-Scale Workloads,” White Paper, Cisco Systems, Inc., Sep. 2014, 11 pages. |
Monia, Charles, et al., IFCP—A Protocol for Internet Fibre Channel Networking, draft-monia-ips-ifcp-00.txt, Dec. 12, 2000, 6 pages. |
Mueen, Abdullah, et al., “Online Discovery and Maintenance of Time Series Motifs,” KDD'10 The 16th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Jul. 25-28, 2010, 10 pages, Washington, DC, U.S.A. |
Muglia, Bob, “Decoding SDN,” Jan. 14, 2013, Juniper Networks, pp. 1-7, http://forums.juniper.net/t5/The-New-Network/Decoding-SDN/ba-p/174651. |
Murray, Joseph F., et al., “Machine Learning Methods for Predicting Failures in Hard Drives: A Multiple-Instance Application,” Journal of Machine Learning Research 6 (2005), pp. 783-816; May 2005, 34 pages. |
Nelson, Mark, “File Verification Using CRC,” Dr. Dobb's Journal, May 1, 1992, pp. 1-18, XP055130883. |
Pace, Alberto, “Technologies for Large Data Management in Scientific Computing,” International Journal of Modern Physics C., vol. 25, No. 2, Feb. 2014, 72 pages. |
Petersen, Chris, “Introducing Lightning: A flexible NVMe JBOF,” Mar. 9, 2016, 6 pages. |
Pinheiro, Eduardo, et al., “Failure Trends in a Large Disk Drive Population,” FAST '07, 5th USENIX Conference on File and Storage Technologies, Feb. 13-16, 2007, 13 pages, San Jose, California, U.S.A. |
Raginsky, Maxim, et al., “Sequential Anomaly Detection in the Presence of Noise and Limited Feedback,” arXiv:0911.2904v4 [cs.LG] Mar. 13, 2012, 19 pages. |
Saidi, Ali G., et al., “Performance Validation of Network-Intensive Workloads on a Full-System Simulator,” Interaction between Operating System and Computer Architecture Workshop, (IOSCA 2005), Austin, Texas, Oct. 2005, 10 pages. |
Sajassi, A., et al., “BGP MPLS Based Ethernet VPN,” Network Working Group, Oct. 18, 2014, 52 pages. |
Sajassi, Ali, et al., “A Network Virtualization Overlay Solution using EVPN,” L2VPN Workgroup, Nov. 10, 2014, 24 pages; http://tools.ietf.org/pdf/draft-ietf-bess-evpn-overlay-00.pdf. |
Sajassi, Ali, et al., “Integrated Routing and Bridging in EVPN,” L2VPN Workgroup, Nov. 11, 2014, 26 pages; http://tools.ietf.org/pdf/draft-ietf-bess-evpn-inter-subnet-forwarding-00.pdf. |
Schroeder, Bianca, et al., “Disk failures in the real world: What does an MTTF of 1,000,000 hours mean to you?” FAST '07: 5th USENIX Conference on File and Storage Technologies, Feb. 13-16, 2007, 16 pages, San Jose, California, U.S.A. |
Shue, David, et al., “Performance Isolation and Fairness for Multi-Tenant Cloud Storage,” USENIX Association, 10th USENIX Symposium on Operating Systems Design Implementation (OSDI '12), 2012, 14 pages; https://www.usenix.org/system/files/conference/osdi12/osdi12-final-215.pdf. |
Staimer, Marc, “Inside Cisco Systems' Unified Computing System,” Dragon Slayer Consulting, Jul. 2009, 5 pages. |
Stamey, John, et al., “Client-Side Dynamic Metadata in Web 2.0,” SIGDOC '07, Oct. 22-24, 2007, pp. 155-161. |
Swami, Vijay, “Simplifying SAN Management for VMWare Boot from SAN, Utilizing Cisco UCS and Palo,” posted May 31, 2011, 6 pages. |
Tate, Jon, et al., “Introduction to Storage Area Networks and System Networking,” Dec. 2017, 302 pages, ibm.com/redbooks. |
Vuppala, Vibhavasu, et al., “Layer-3 Switching Using Virtual Network Ports,” Computer Communications and Networks, 1999, Proceedings, Eight International Conference in Boston, MA, USA, Oct. 11-13, 1999, Piscataway, NJ, USA, IEEE, ISBN: 0-7803-5794-9, pp. 642-648. |
Wang, Feng, et al. “OBFS: A File System for Object-Based Storage Devices,” Storage System Research Center, MSST. vol. 4., Apr. 2004, 18 pages. |
Weil, Sage A., “Ceph: Reliable, Scalable, and High-Performance Distributed Storage,” Dec. 2007, 239 pages, University of California, Santa Cruz. |
Weil, Sage A., et al. “CRUSH: Controlled, Scalable, Decentralized Placement of Replicated Data.” Proceedings of the 2006 ACM/IEEE conference on Supercomputing. ACM, Nov. 11, 2006, 12 pages. |
Weil, Sage A., et al. “Ceph: A Scalable, High-performance Distributed File System,” Proceedings of the 7th symposium on Operating systems design and implementation. USENIX Association, Nov. 6, 2006, 14 pages. |
Wu, Joel, et al., “The Design, and Implementation of AQuA: An Adaptive Quality of Service Aware Object-Based Storage Device,” Department of Computer Science, MSST, May 17, 2006, 25 pages; http://storageconference.us/2006/Presentations/30Wu.pdf. |
Xue, Chendi, et al. “A Standard framework for Ceph performance profiling with latency breakdown,” Ceph, Jun. 30, 2015, 3 pages. |
Zhou, Zihan, et al., “Stable Principal Component Pursuit,” arXiv:1001.2363v1 [cs.IT], Jan. 14, 2010, 5 pages. |
Zhu, Yunfeng, et al., “A Cost-based Heterogeneous Recovery Scheme for Distributed Storage Systems with RAID-6 Codes,” University of Science & Technology Of China, 2012, 12 pages. |
Number | Date | Country | |
---|---|---|---|
20190370216 A1 | Dec 2019 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 15174718 | Jun 2016 | US |
Child | 16542952 | US |