None.
Not applicable.
Not applicable.
Electronic communications may carry a wide variety of content, for example electronic mail, medical records, financial transactions, and other confidential information. A wireless communication network may store this information in databases on the network. These databases may expend a significant amount of network resources, in turn placing a strain on the network infrastructure. The confidential information may also be at risk to be fraudulently modified or tampered with by a third party.
In an embodiment, a method of combining chains of blocks in a network is described. The method comprises creating a plurality of birth blocks of a plurality of chains of blocks by a block foundry application in a network, where each birth block is associated with a chain of blocks that records events of a network entity and comprises a nonce value, a hash value, and a transaction and for each chain of blocks, creating blocks by a plurality of nodes in the network, where each block comprises a nonce value, a transaction, a hash value, and the hash value of the previous block, wherein the hash value of the previous block links the current block and the previous block, terminating the chain of blocks by the network entity, wherein the entity sends a termination request to create an end block to the plurality of nodes, and creating the end block based on the termination request by the plurality of nodes, wherein the end block is the final block of the chain of blocks, the end block comprising a final nonce value, a final transaction, a final hash value, and the hash value of the previous block, wherein the hash value of the previous block links the end block and previous block. The method further comprises, in response to the creation of the end block, sending a request by the network entity to the plurality of nodes to create a block of a meta-chain of blocks and creating based on the block request the block of the meta-chain of blocks by the plurality of nodes, wherein the block comprises a nonce value, a transaction, a hash value, the hash value of the previous block of the meta-chain of blocks, and the hash value of the end block of the chain of blocks. In yet another embodiment, a network system is described. The network system comprises a block foundry server, a plurality of network entities, and a plurality of nodes in a network, comprising a non-transitory memory. The system further comprises a plurality of chains of blocks stored in the non-transitory memory, wherein each chain of blocks comprises a birth block, a plurality of blocks, and an end block, each block comprising a block number, a nonce value, a transaction, a hash value, and the hash value of the previous block, wherein the hash value of the previous block links the block and the previous block and a meta-chain of blocks stored in the non-transitory memory, wherein the meta-chain of blocks comprises at least one block that comprises a block number, a nonce value, a transaction, a hash value, a hash value of the previous block of the meta-chain of blocks, and a hash value of an end block of a chain of blocks.
In yet another embodiment, a method of tracking, monitoring, and preserving temporal network function virtualization (NFV) events on a distributed ledger technology (DLT) computer system is described. The method comprises, initiating by a hypervisor executing on a computer system a plurality of virtual servers providing a plurality of virtualized network functions (VNFs) that comprises a tracking application on a network function virtualization (NFV) core network, and for each virtual server, sending a request to create a birth block based on the initiation of the virtual server by the tracking application to a block foundry application in the network, creating based on the block request by the block foundry application, the birth block of a chain of blocks comprising a first nonce value, a first transaction, and a first hash value, and detecting a plurality of changes of state of the virtual server by the tracking application. The method further comprises for each change of state of the virtual server, sending a request to create a new block based on the change of state of the virtual server by the hypervisor to a plurality of nodes in the network and creating based on the new block request by the plurality of nodes, a new block of the chain of blocks, wherein the new block follows the previous block, comprising a new nonce value, a new transaction, a new hash value, and the hash value of the previous block, wherein the hash value of the previous block links the new block and the previous block. The method further comprises terminating the virtual server on the NFV core network by the hypervisor, sending a termination request to create an end block based on the termination of the virtual server by the hypervisor to the plurality of nodes in the network, creating based on the termination block request by the plurality of nodes, an end block of the chain of blocks, wherein the end block is the final block of the chain of blocks, the end block comprising a final nonce value, a final transaction, a final hash value, and the hash value of the previous block, wherein the hash value of the previous block links the end block and the previous block, sending a request by the hypervisor to the plurality of nodes to create a block of a meta-chain of blocks, wherein the meta-chain of blocks tracks related chains of blocks in the NFV core network, and creating by the plurality of nodes the block of the meta-chain of blocks, wherein the block comprises a nonce value, a transaction, a hash value, the hash value of the previous block of the meta-chain of blocks, and the hash value of the end block of the chain of blocks.
These and other features will be more clearly understood from the following detailed description taken in conjunction with the accompanying drawings and claims.
For a more complete understanding of the present disclosure, reference is now made to the following brief description, taken in connection with the accompanying drawings and detailed description, wherein like reference numerals represent like parts.
It should be understood at the outset that although illustrative implementations of one or more embodiments are illustrated below, the disclosed systems and methods may be implemented using any number of techniques, whether currently known or not yet in existence. The disclosure should in no way be limited to the illustrative implementations, drawings, and techniques illustrated below, but may be modified within the scope of the appended claims along with their full scope of equivalents.
Network function virtualization (NFV) systems may be implemented by communication service providers to increase efficiency and improve service flexibility. Virtualized network functions (VNF) may reduce the service providers' hardware dependencies and allow for more scalable distribution of virtual resources, which may reduce capital and operational expenditures. However, tracking usage for these virtual resources may be difficult since they do not have persistent physical identities. Logging the usage of the virtual machines and storing the logs in data stores distributed on the network may place a strain on network infrastructure and use up valuable network resources. Additionally, these logs may be modified or corrupted by fraudulent revisions by parties on the network. The unauthorized alterations of the logs may lead to inaccurate information across the network, and the service provider may erroneously bill a user. This may cause a loss of confidence amongst parties in the network without an efficient method to verify the usage logs. In an embodiment, a service provider may want to monitor NFV events in order to determine when and where to increase or decrease hardware assets for NFV support in a timely manner. For example, the service provider may decide to expand hardware assets when 80% of virtual resources of a VNF are used.
The present disclosure teaches several related methods for tracking, monitoring, and preserving network function virtualization (NFV) events based on a distributed ledger technology (DLT) computer system. Distributed ledger technology (DLT) utilizes a distributed database without a central administrator managed on a peer-to-peer (P2P) network in order to maintain records in a secure manner by a consensus mechanism amongst nodes in the network to prevent modification of the records. The distributed database may be spread across multiple sites, regions, and/or parties. DLT may be used to validate and authenticate data exchanges. Records may be stored in the network once a majority of the nodes in the network achieve consensus. Distributed ledger technology may be applied in many aspects that involve secure data storage. For example, distributed ledgers may be used to record contracts (e.g., financial contracts, smart contracts), financial transactions, and other tangible and intangible goods.
In some forms of DLT, blocks are used to store information and then connected in a chronological technique. These blocks are inherently resistant to modification of the data. One skilled in the art understands the process of how DLT is used to securely store records, but a basic description is provided here. The first block may be created by a block foundry in the network through an algorithm. A request for a new block to be created may be sent to a plurality of nodes in the network. A block typically comprises a block number, a nonce value, a timestamp, a plurality of transactions, a hash value, and the hash value of the previous block, which links the block with the previous block. A hash is a string of data with a given size (e.g., 64-bits, 128-bits, 256-bits) that is based on the information in the block (e.g., the block number, the nonce value, the transactions, and the previous hash). The nonce value is varied by the nodes in the network in order to solve the hash value based on a hashing algorithm. No meaningful data can be derived from the hash of the block. In other words, the transaction data cannot be back-solved based on the hash. The nodes in the network collectively validate new blocks when a request is sent, and once a block is validated, it cannot be retroactively altered without invalidating the subsequent blocks. Typically, blocks are validated on a majority rule (e.g., 50%+1 vote) in the network, wherein the majority of nodes in the network are in consensus, or agreement. Each block comprises the hash and the hash of the previous block, which links the two blocks.
In an embodiment, some of a communication network's functionality may be provided with a network function virtualization (NFV) communication paradigm. Network node functions, for example, the functionality of a home subscriber server, a mobility management entity, a policy and charging rules function node, and a serving gateway—rather than each being provided by separate computers or computer systems—may be provided by virtual servers executing in a virtual computing environment, for example executing in a cloud computing environment. The network functions may be composed of common functions. In an embodiment, common functions are executed in one or more virtual servers executing in a virtual computing environment.
The factorization of common functions in combination with virtualized execution can provide a variety of advantages to a wireless communication service provider. This paradigm provides the ability to conveniently and quickly expand or contract compute resources as needed. This paradigm allows a wireless communication service provider to pay for compute resources as they are used, on an as needed basis, versus investing significant capital in purchasing and installing computing equipment in anticipation of future needs, needs which may not materialize or may not be as large (or as small!) as projected. Virtualization of common network functions and network function virtualization is described further hereinafter.
A virtual computing environment may support ease of maintenance, ease of upgrading, ease of expansion and contraction of computer resources. The virtual computing environment may be a private cloud computing environment owned, operated, and managed by a wireless communication service provider. Alternatively, the virtual computing environment may be a public cloud computing deployment owned, operated, and managed by a cloud computing service provider and supporting not only the core network functionality of a wireless communication service provider but also supporting computing needs of web-based enterprises, larger on-line retail sales enterprises, governmental entities, and the like.
Virtual servers execute on compute resources. The physical details of the compute resources may be abstracted so a virtual server only interacts with a logical view of the underlying compute resources, thereby decoupling the virtual servers from the physical details of the compute resources. Decoupling the virtual servers from the physical details of the underlying compute resources may make instantiating the virtual server on different physical compute resources more convenient and may entail less system administrator intervention. A virtual server may execute in the context of or under the management of a hypervisor or may execute in the context of a logical operating system that abstracts away the details of the underlying compute resources. Compute resources comprising the customary computer system components: processor, main memory, secondary memory or mass memory, and network interfaces. Services rendered in the NFV core network may be monitored and billed. Additional virtual servers may be dynamically started to execute additional instances of a common function and stopped as needed to handle dynamically changing communication loads in the NFV core network. There may be tens of thousands of virtual machines running on a network at a given time.
A distributed ledger used to record changes of state of a virtual server in a network function virtualization (NFV) core network is taught herein. When a virtual server providing a virtualized network function (VNF) is initiated on the network, a request to create a first block may be sent to a block foundry application in the network. The virtual server may comprise a tracking application that monitors the state of the virtual server. The tracking application may be configured to send a request to create a block upon initiation of a virtual server. The block foundry application may then create the first block in the chain of blocks that comprises a block number, a nonce, a timestamp, and a hash. In some embodiments, the first block may be referred to as the birth block. Since there is no block preceding the birth block, there may not be a previous hash or the previous hash may be assigned value of 0. As the virtual server operates on the network, the tracking application may detect a change of state on the virtual server. A request may be sent to the plurality of nodes in the network to create another block that records the change of state of the virtual server. The plurality of nodes may operate as a plurality of consensus servers. A change of state of the virtual server may occur when the virtual server processes a predefined number of wireless communication services from a plurality of UEs on the network. For example, a change of state may occur when the virtual server conducts one thousand authentication functions for UEs on the network.
This process may repeat until the virtual server is terminated. Upon termination of the virtual server, the tracking application may be triggered to send a request for an end block, and the nodes in the network may create an end block for the chain of blocks. The blocks link together to form a chain since each block after the birth block comprises a hash and the hash of the previous block that links the blocks and allows the order to be followed back to the birth block. Any modification to a block in the chain of blocks invalidates the hash of that particular block which in turn invalidates the blocks that come after the invalidated block. This allows for a trustworthy method of storing transaction data since the chain of blocks is inherently resistant to alterations in the data.
In an embodiment, each virtual server creates its own chain of blocks to record events (e.g., transactions and/or changes of state). A plurality of chains of blocks may be concurrently generated and stored on the nodes in the network. There may be upwards of tens of thousands of virtual servers operating in a hypervised state at a given time. Each chain of blocks may be in different stages of its life cycle. For example, one chain of blocks may have been created more recently relative to another chain of blocks. Due to the multitude of blocks being created in the network at a given time, the network may assign a priority to a new block request, wherein the priority indicates the distribution of network resources to create the new block by the plurality of nodes in the network. Since the plurality of nodes in the network solve for the hash value of each block and desire consensus, a larger number of block requests may cause the nodes to take a longer amount of time to create new blocks. A higher priority of a new block request would indicate a larger proportion of network resources should be allotted to creating the new block.
It may be desired in some embodiments to group transactions according to a class of user such as by client, function, or some other category. For example, a class of user may be a category of business (e.g., healthcare business, insurance business, technology business) or a specific company. This may further safeguard the data records and prevent clients from accessing other potentially proprietary information within the same network. In some instances, a plurality of parties may desire to access and view the one another's data stored on their chains of blocks. With permission from the network and/or the parties associated with the blocks, the blocks may be released in a read-only state. The virtual computing environment may be executing in a public network, a private network, or a combination thereof.
In an embodiment, it may be desired to combine two or more classes of a chain of blocks or more than one chain of blocks in general. A meta-chain of blocks may provide a secure method for combining multiple chains of blocks into a single chain of blocks in order for the relevant parties to view information of the virtual servers. For example, if a first virtual server is associated with a Company A, a second virtual server is associated with a Company B, and Company A and Company B request to view one another's virtual servers, a meta-chain of blocks may be created. A block of the meta-chain of blocks may comprise a block number, a nonce, a transaction, a hash value, the hash value of the previous block of the meta-chain of blocks, and the hash value of an end block of a chain of blocks.
A new block in the meta-chain of blocks may be requested when a first virtual server is terminated, which thereby terminates a first chain of blocks with an end block. The new block in the meta-chain of blocks may comprise a nonce value, a transaction, a hash value, the hash value of the previous block in the meta-chain of blocks, and the hash value of the end block of the terminated first chain of blocks. In an embodiment, the hash value of the end block of the terminated chain of blocks may be stored in the transaction of the block. When a second virtual server is terminated, another block of the meta-chain of blocks may be created comprising a nonce, a transaction, a hash value, the hash value of the previous block, and the hash value of the end block of the second chain of blocks. As more virtual servers are terminated on the network, more blocks may be created on the meta-chain of blocks, wherein each block comprises the hash value of the end block of the chain of blocks. In an embodiment, a block of the meta-chain of blocks may comprise more than one hash of an end block.
The meta-chain of blocks allows parties on the network a more efficient way of viewing activity of virtual servers on the network. Since a block of the meta-chain of blocks comprises the hash value of the end block of a particular chain of blocks, the hash value of the end block may be used to find the particular chain of blocks and track the history of the virtual server back to the birth block of the particular chain of blocks. The meta-chain of blocks allows parties on the network to view historical information in a trustworthy way since the blocks may not be modified after they are established on the chain. The meta-chain of blocks is described in further detail hereinafter.
In an embodiment, a service provider may desire to bill a user or a plurality of users for network services rendered in the NFV core network. A service provider may determine the network services used by the users from the meta-chain of blocks. A service provider may track the virtual resources used efficiently through the blocks on the meta-chain of blocks associated with a plurality of virtual servers. In another embodiment, a licensor may request an audit that challenges the reported usage of a license associated with the licensor. For example, the license may be related to a licensed entity such as licensed software, licensed logo, licensed design, or licensed intellectual property in which a fee is paid on a per-use basis to the licensor by a licensee. The service provider may comply with the audit request and provide the meta-chain of blocks related to the license to the licensor. Since the meta-chain of blocks and thereby chains of blocks that comprise it are immutable, both the licensor and licensee may trust the information in the blocks as an accurate report regarding the use of the license. Reviewing the appropriate block chain, for example, may promote establishing time durations and numbers of instances of use of the licensed entity.
In an embodiment, a service provider of the network may monitor the usage of the virtual resources in the virtual computing environment. Virtual resources may comprise compute resources, memory resources, I/O resources, and other types of resources. It may be useful for service providers to track virtual resource usage in order to determine when to add or remove hardware assets that support the virtual computing environment. The chain of blocks may track the succession of the changes of states of a virtual server, and service providers may use this information to forecast at some time in the future whether or not hardware assets that support the virtual computing environment should be added or removed. For example, a service provider may determine to increase hardware assets at 5 P.M. on weekdays since mobile communication services increase while people are leaving work. Since the chains of blocks may not be modified without detection by parties in the network, the service provider can be confident of the information it receives about the network.
Turning now to
The system 100 further comprises a block server 114, a plurality of consensus servers 116, and a virtual computing environment 118 comprising a physical host 120. The physical host 120 may further comprise a plurality of virtual servers 122 each comprising a virtualized network function (VNF) 124 that comprises a tracking application 126. A VNF 124 may be a common function executing in a virtual server 122 in the virtual computing environment 118. For example, a VNF 124 may be an attach function, an authentication function, a mobility function, a bearer function, a context function, a policy function, or a data function. Other VNFs 124 may be factorized from traditional network functions. A single virtual server 122 may concurrently execute a plurality of instances of a VNF 124. A plurality of virtual servers 122, each executing instances of the same VNF 120 may be deployed to carry a communication load. In an embodiment, a virtual server 122 executes instances of only one VNF 124. For example, no virtual server 122 would concurrently execute instances of different VNFs 124. A hypervisor 128 may be deployed on the physical host 120 to manage the virtual servers 122. Although not shown, the physical host 120 further comprises processors (e.g., microprocessors, digital signal processors, graphics processors), main memory (e.g., disk drives), mass storage, and network interfaces. A single virtual server 122 may execute on one physical host 120 or a plurality of virtual servers 122 may execute on one physical host 120. Any number of physical hosts 120 may exist on the virtual computing environment 118.
The plurality of consensus servers 116 may store a plurality of blocks 140. The plurality of blocks 140 may be linked in a way that forms a chain of blocks 202. The block server 114 comprises a block foundry application 130. The tracking application 126 of the VNF application 124 may be configured to actively monitor the events of the virtual server 122. When the tracking application 126 detects the initiation of a virtual server 122 by the hypervisor 128 on a physical host 120, it may be triggered to send a request to the block foundry application 130 to create a block 140. The block 140 created by the block foundry application 130 may be referred to as a birth block 140 in some contexts since it is the first block of a chain of blocks. The birth block 140 may be transmitted and stored in the plurality of consensus servers 116. Each consensus server 116 stores replicates of the same birth block 140. Consensus servers 116 may be distributed in different geographical locations.
Turning now to
Each consensus server 116 receives a copy of the transaction 146. Upon receiving the transaction 146, the consensus servers 116 independently create a block by calculating a satisfactory hash 152 based on the block number 142, the nonce 144, data from the transaction 146, the timestamp 148, and the previous hash 150. A hash 152 is calculated by a one-way hashing function from an input data, wherein the same input data returns the same hash 152. Changing as much as a single character of the input data may result in an entirely different hash 152. The hash 152 may be a string comprising a fixed size (e.g., 32-bits, 64-bits, 256-bits). No meaningful data can be derived from the hash 152 about the block 140. In other words, data from the block 140 such as data from the transaction 146 cannot be recovered from the hash 152. There may be a certain condition to fulfill before a hash 152 is deemed satisfactory, such as the hash 152 containing a predetermined number of leading zeros. The nonce 144 is an arbitrary numerical value that may be incrementally varied until the hash 152 satisfies the condition. A consensus server 116 may vary the nonce 144 in order solve the hash 152.
In an embodiment, three or more consensus servers 116 may exist on the network 106. Each consensus server 116 receives a copy of the same transactions 146 and creates a block 140 by calculating a satisfactory hash 152 from varying the nonce 144. The first consensus server 116 to generate a hash 152 that satisfies the condition transmits its block 140 to the other consensus servers 116 for validation. A majority (50%+1 vote) of the consensus servers 116 may confirm the block 140 in order to validate it, where each consensus server 116 receives 1 vote. A copy of the block 140 may be stored in each of the plurality of consensus servers 116. The block number 142 and timestamp 148 may be assigned based on when the block 140 is created relative to other blocks in the chain of blocks 202 that are already stored in the consensus servers 116. The previous hash 150 allows the block 140 to be linked to the most recent block in the consensus server 116. The links between blocks 140 generate a chain of blocks 202, wherein the blocks 140 are in a fixed order. The birth block may not comprise a previous hash 150 or the previous hash 150 may consist of a string of zeros since no block precedes the birth block. Since the hash 152 is based on the block 140, the data of the block 140 may not be modified or altered once the block 140 is validated and stored without invalidating the block 140. The chain of blocks may assure a secure storage of records such as transaction history that may not be altered by a party.
A consensus server 116 may store a plurality of chains of blocks 202. A group of three or more consensus servers 116 each store a copy of the same chains of blocks 202 in order to provide validation for the next block 140. In an embodiment, a greater number of consensus servers 116 may indicate a greater confidence in the accuracy of a validated block 140 since more votes are accounted for. One chain of blocks 202 may record the transactions 146 that occur on one virtual server 122. In an embodiment, a first plurality of consensus servers 116 may store different chains of blocks than a second plurality of consensus servers 116 on the network 106. For example, if ten chains of blocks 202a to 202j record transactions of virtual servers 122a-122j wherein chain of blocks 202a records virtual server 122a, chain of blocks 202b records virtual server 122b and so on and so forth, the first plurality of consensus servers 116 may store chains of blocks 202a-202f, and the second plurality of consensus servers 116 may store chains of blocks 202g-202j. The tracking application 126 of the virtual server 122 may determine to which plurality of consensus servers 116 to distribute the transactions 146. The virtual servers 122a-122j may perform plurality of VNF applications 124. For example, virtual servers 122a-122c may perform an authentication function, virtual servers 122d-122e may perform an attach function, virtual servers 122f-122i may perform a mobility function, and virtual server 122j may perform a context function.
The tracking application 126 may continue to monitor the virtual server 122. When the virtual server 122 is destroyed, the tracking application 126 may send a transaction 146 of the termination of the virtual server 122 to the plurality of consensus servers 116 associated with the chain of blocks 202 of the virtual server 122. The consensus servers 116 may create an end block 140 for the chain of blocks 202, where the end block 140 comprises a block number 142, a nonce 144, the transaction 146, a timestamp 148, a previous hash 150 of the previous block, and a hash 152 based on the transaction 146. After the end block 140 is validated by the plurality of consensus servers, no additional blocks may be inserted or added to the chain of blocks 202 without invalidating the record of the virtual server 122. A party of the network 106 (e.g., a wireless communication service provider) may use the previous hash 150 and hash 152 of the blocks 140 to follow the sequence of changes of state of the virtual server 122 back to the birth block.
In an embodiment, a chain of blocks 202 may comprise confidential or proprietary information (e.g., financial transactions, contracts) or generally information that the parties involved do not wish to be publically available. In the network 106, virtual servers 122 may carry out VNFs 124 for a multitude of separate parties such as different clients, companies, and/or entities. A chain of blocks 202 may be associated with a class. A class may be a client, entity, business category, or some other way of categorizing the chain of blocks 202. These parties may not want other parties in the network 106 to access their transaction record history. To view a chain of blocks 202, a party may use the previous hash 150 and block number 142 where both the previous hash 150 and block number 142 reference the same block 140. As an additional security measure, a party may also assign a chain of blocks 202 with an access credential grants access to a block 140. This way, it would be more difficult for unapproved or nefarious parties to view the transaction history.
In an embodiment, it may be desired to view the chain of blocks 202 as a method for a network service provider to accurately bill a client for services. The service provider may charge clients for network resources used such as the storage and transfer of data. Invoices may be sent on a periodic basis (e.g., every week, every month, every six months, or another period of time). Sometimes, clients may disagree with the amount of network resources used that the service provider presents. At other times, the service provider may incorrectly determine the amount of network resources used. Both the service provider and the client may inspect the chains of blocks 202 to determine the true amount of network resources used. Since the transactions 146 of the chains of blocks 202 may not be modified or deleted without alerting either party, they offer a trustworthy means to verify information.
The chains of blocks 202 may also be used in a multitude of other modes where tracking transaction history is important. For example, a licensing term may exist for the usage of an entity. A client may pay a fee for using the licensed entity (e.g., software, logo, design, intellectual property). The client may be charged the fee on a per-use basis to be paid to the licensor. On a virtual server 122, it may be difficult to accurately track the use of the licensed entity, so a client may under-report the usage which violates the terms of the licensing agreement. On the other hand, a client may also over-report the usage to avoid breaching the licensing agreement, which causes a waste of money for the client. The chains of blocks 202 allow for both the client and licensor to track the usage of the licensed entity as a measure to verify accurate reporting.
Turning now to
With reference to
At time instance A, the first virtual server 122a is initiated on the network 106, and the tracking application 126 of the first virtual server 122a may request a birth block 204a from the block foundry application 130. The birth block 204a of the first chain of blocks 202a may be created by the block foundry application 130, where the birth block 204a comprises a hash 152a. The birth block 204a may comprise a previous hash 150a (not shown) that comprises a value of zero or a string of zeros. The birth block 204a may be transmitted to the plurality of consensus servers 116 to be stored. At the same time instance A, the second virtual server 122b may be initiated on the network 106, and the tracking application 126 of the second virtual server 122b may request a birth block 204b from the block foundry application 130. The birth block 204b of the second chain of blocks 202b may be created by the block foundry application 130, where the birth block 202b comprises a hash 152d. The birth block 204b may be transmitted to the plurality of consensus servers 116.
At time instance B, the X-th virtual server 122X may be initiated on the network 106, and the tracking application 126 of the X-th virtual server 122X may request a birth block 204X from the block foundry application 130. The birth block 204X of the X-th chain of blocks 202X may be created by the block foundry application 130, where the birth block 204X comprises a hash 152g. The birth block 204X may be transmitted to the plurality of consensus servers 116. Furthermore at time instance B, the tracking application 126 of each of the virtual servers 122a, 122b may request more blocks 206a, 206b to be created by the plurality of consensus servers 116 for the chains of blocks 202a, 202b. Each of the blocks 206a, 206b may comprise a previous hash 150b, 150e and hash 152b, 152e as the virtual servers 122a, 122b undergo changes of state. The tracking application 126 may determine when to request new blocks 206 from the plurality of consensus servers 116.
At time instance C, chain of blocks 202X creates a new block 206X comprising previous hash 150h and hash 152h in response to a change of state of the virtual server 122X. The tracking application 126 of virtual servers 122a, 122b, 122X continue to request blocks 206a, 206b, 206X to be created by the consensus servers 116 for the chains of blocks 202a, 202b, 202X. The chains of blocks 202a, 202b, 202X may comprise any number of blocks 206, and there may be any number of active chains of blocks 202a, 202b, 202X at a given time.
At time instance D, the first virtual server 122a is destroyed, thereby terminating chain of blocks 202a. The tracking application 126 may request an end block from the plurality of consensus servers 116. The plurality of consensus servers 116 may create an end block 208a that comprises a previous hash 150c and hash 152c. In response to the end block 208a of the first chain of blocks 202a, the tracking application 126 may also request a birth block 170a of a meta-chain of blocks 160 from the block foundry application 130. The birth block 170a may comprise a hash 174a and the hash 152c of the end block 208a of the first chain of blocks 202a. The birth block 170a may be transmitted and stored in the plurality of consensus servers 116. Although not shown, the birth block 170a may also comprise a block number 142, a nonce 144, a transaction 146, and a timestamp 148. In some embodiments, the hash 152c may be stored as a transaction 146 of the birth block 170a.
At time instance E, the X-th virtual server 122X is destroyed, thereby terminating chain of blocks 202X, and an end block 208X that comprises a previous hash 150i and hash 152i may be created by the consensus servers 116. The tracking application 126 of the X-th virtual server 122X may request the plurality of consensus servers 116 to create a block 170b of the meta-chain of blocks 160. The block 170b comprising a hash 174b, a hash 172b of the previous birth block 170a, and the hash 152i of the block 208X may also be created at time instance E. The block 170b may follow the birth block 170a and be stored in the consensus servers 116. The previous hash 172b links the block 170b and the birth block 170a. At time instance F, the second virtual server 122b is destroyed, thereby terminating chain of blocks 202b, and an end block 208b that comprises a previous hash 150c and hash 152c may be created and stored in the plurality of consensus servers 116. An end block 170c of the meta-chain of blocks 160 comprising a hash 174c, a hash 172c of the previous block 170b, and the hash 152f of the block 208b may also be created and stored in the plurality of consensus servers 116 at time instance F.
In an embodiment, the meta-chain of blocks 160 may comprise any number of blocks 170. Blocks 170 of the meta-chain of blocks 160 may continue to be requested and created on the network 106 for a plurality of virtual servers 122. While three virtual servers 122a, 122b, 122X were used in this embodiment, any number of virtual servers 122 comprising any number of blocks may be used to create a meta-chain of blocks 160. The meta-chain of blocks 160 may be created when a first virtual server 122a is destroyed on the virtual computing environment 118 and the meta-chain of blocks 160 may be terminated with an end block 170c when the last virtual server 122X of a plurality of virtual servers 122 of the virtual computing environment 118 is destroyed. The plurality of virtual servers 122 may be related to one another. For example, the plurality of virtual servers 122 may be related by function, by client, or by class. Blocks 170 may be created as virtual servers 122 continue to be destroyed. Virtual servers 122 and thereby chains of blocks 202 may be initiated at any time. The meta-chain of blocks 160 may be stored in the plurality of consensus servers 116 on the network 106. Any number of meta-chains of blocks 160 may be concurrently active and stored in the consensus servers 116 at a given time. It is understood that the blocks 170 of the meta-chain of blocks may also comprise a block number 142, a nonce 144, a transaction 146, and a timestamp 148. In some embodiments, the hash 152 of each of the end blocks 208a, 208b, 208X may be stored as a transaction 146 of the blocks 170.
Like the independent chains of blocks 202a, 202b, 202X, the blocks 170 of the meta-chain of blocks 160 comprise previous hashes 172 and hashes 174. The previous hashes 172 allow the blocks 170 to be linked to create a chain that comprises blocks 170 in a chronological order. The blocks 170 also comprise the hash 152 of the end block 208 of a chain of blocks 202. Due to the linkage between the blocks of a chain, the events of the virtual server 122 may be followed back stage-by-stage until the birth block 204 is reached. The meta-chain of blocks 160 may reside in a plurality of consensus servers 116 and may be viewed by approved parties of the network 106. The meta-chain of blocks 160 allows the parties to view information about the relevant virtual servers 122 in a much more efficient and straightforward way than if the chains of blocks were viewed independently. Since the meta-chain of blocks 160 comprises references to hashes 152 of the chains of blocks 202, a party can use the hashes 152 to access more detailed data about a virtual server 122, if desired.
Turning now to
At block 310, for each change of state of the virtual server, the hypervisor sends a request to create a new block based on the change of state of the virtual server to a plurality of nodes in the network. At block 312, the plurality of nodes creates based on the new block request, a new block of the chain of blocks, wherein the new block follows the previous block, comprising a new nonce value, a new transaction, a new hash value, and the hash value of the previous block, wherein the hash value of the previous block links the new block and the previous block. At block 314, the hypervisor terminates the virtual server on the NFV core network. At block 316, the hypervisor sends a termination request to the plurality of nodes in the network to create an end block based on the termination of the virtual server. At block 318, the plurality of nodes creates based on the termination block request, an end block of the chain of blocks, wherein the end block is the final data block of the chain of blocks, the end block comprising a final nonce value, a final transaction, a final hash value, and the hash value of the previous block, wherein the hash value of the previous block links the end block and the previous block.
At block 320, the hypervisor sends a termination request to the plurality of nodes in the network to create a block of a meta-chain of blocks, wherein the meta-chain of blocks tracks related chains of blocks in the NFV core network. At block 322, the plurality of nodes created the block of the meta-chain of blocks, wherein the block comprises a nonce, a transaction, a hash value, the hash value of the previous block of the meta-chain of blocks, and the hash value of the end block of the chain of blocks.
In an embodiment, method 200 may further comprise billing a user of the virtual server based on usage of the network resources determined from the meta-chain of blocks. Method 200 may further comprise scaling hardware assets used by the NFV core network in response to a change in the frequency of NFV events by a service provider.
Turning now to
At block 338, the plurality of nodes create the end block based on the termination request, wherein the end block is the final block of the chain of blocks, the end block comprising a final nonce value, a final transaction, a final hash value, and the hash value of the previous block, wherein the hash value of the previous blocks links the end block and previous block. At block 340, the network entity sends a request to the plurality of nodes to create a block of a meta-chain of blocks in response to the creation of the end block. At block 342, the plurality of nodes create based on the block request, a block of the meta-chain of blocks, wherein the block comprises a nonce value, a transaction, a hash value, the hash value of the previous block, and the hash value of the end block of the chain of blocks.
Method 330 may further comprise receiving an audit request from a licensor by a wireless communication service provider, wherein the audit request challenges the usage amount of at least one license associated with the licensor on the wireless communication network; complying with the audit request by the wireless communication service provider; and verifying, by using the meta-chain of blocks (i.e., scanning, traversing, reading the content and/or data fields thereof), the usage of the at least one license, wherein the usages of the at least one license on a network entity are stored in a block associated with the meta-chain of blocks. Method 330 may further comprise assigning a priority by the network in the block request, wherein the priority is associated with the proportion of computing power distributed by the plurality of nodes to creating the block.
The DSP 502 or some other form of controller or central processing unit operates to control the various components of the UE 400 in accordance with embedded software or firmware stored in memory 504 or stored in memory contained within the DSP 502 itself. In addition to the embedded software or firmware, the DSP 502 may execute other applications stored in the memory 504 or made available via information carrier media such as portable data storage media like the removable memory card 520 or via wired or wireless network communications. The application software may comprise a compiled set of machine-readable instructions that configure the DSP 502 to provide the desired functionality, or the application software may be high-level software instructions to be processed by an interpreter or compiler to indirectly configure the DSP 502.
The DSP 502 may communicate with a wireless network via the analog baseband processing unit 510. In some embodiments, the communication may provide Internet connectivity, enabling a user to gain access to content on the Internet and to send and receive e-mail or text messages. The input/output interface 518 interconnects the DSP 502 and various memories and interfaces. The memory 504 and the removable memory card 520 may provide software and data to configure the operation of the DSP 502. Among the interfaces may be the USB port 522 and the infrared port 524. The USB port 522 may enable the UE 400 to function as a peripheral device to exchange information with a personal computer or other computer system. The infrared port 524 and other optional ports such as a Bluetooth® interface or an IEEE 802.11 compliant wireless interface may enable the UE 400 to communicate wirelessly with other nearby handsets and/or wireless base stations. In an embodiment, the UE 400 may comprise a near field communication (NFC) transceiver. The NFC transceiver may be used to complete payment transactions with point-of-sale terminals or other communications exchanges. In an embodiment, the UE 400 may comprise a radio frequency identify (RFID) reader and/or writer device.
The switches 528 may couple to the DSP 502 via the input/output interface 518 to provide one mechanism for the user to provide input to the UE 400. Alternatively, one or more of the switches 528 may be coupled to a motherboard of the UE 400 and/or to components of the UE 400 via a different path (e.g., not via the input/output interface 518), for example coupled to a power control circuit (power button) of the UE 400. The touch screen display 530 is another input mechanism, which further displays text and/or graphics to the user. The touch screen LCD controller 532 couples the DSP 502 to the touch screen display 530. The GPS receiver 538 is coupled to the DSP 502 to decode global positioning system signals, thereby enabling the UE 400 to determine its position.
It is understood that by programming and/or loading executable instructions onto the computer system 380, at least one of the CPU 382, the RAM 388, and the ROM 386 are changed, transforming the computer system 380 in part into a particular machine or apparatus having the novel functionality taught by the present disclosure. It is fundamental to the electrical engineering and software engineering arts that functionality that can be implemented by loading executable software into a computer can be converted to a hardware implementation by well-known design rules. Decisions between implementing a concept in software versus hardware typically hinge on considerations of stability of the design and numbers of units to be produced rather than any issues involved in translating from the software domain to the hardware domain. Generally, a design that is still subject to frequent change may be preferred to be implemented in software, because re-spinning a hardware implementation is more expensive than re-spinning a software design. Generally, a design that is stable that will be produced in large volume may be preferred to be implemented in hardware, for example in an application specific integrated circuit (ASIC), because for large production runs the hardware implementation may be less expensive than the software implementation. Often a design may be developed and tested in a software form and later transformed, by well-known design rules, to an equivalent hardware implementation in an application specific integrated circuit that hardwires the instructions of the software. In the same manner as a machine controlled by a new ASIC is a particular machine or apparatus, likewise a computer that has been programmed and/or loaded with executable instructions may be viewed as a particular machine or apparatus.
Additionally, after the system 380 is turned on or booted, the CPU 382 may execute a computer program or application. For example, the CPU 382 may execute software or firmware stored in the ROM 386 or stored in the RAM 388. In some cases, on boot and/or when the application is initiated, the CPU 382 may copy the application or portions of the application from the secondary storage 384 to the RAM 388 or to memory space within the CPU 382 itself, and the CPU 382 may then execute instructions that the application is comprised of. In some cases, the CPU 382 may copy the application or portions of the application from memory accessed via the network connectivity devices 392 or via the I/O devices 390 to the RAM 388 or to memory space within the CPU 382, and the CPU 382 may then execute instructions that the application is comprised of. During execution, an application may load instructions into the CPU 382, for example load some of the instructions of the application into a cache of the CPU 382. In some contexts, an application that is executed may be said to configure the CPU 382 to do something, e.g., to configure the CPU 382 to perform the function or functions promoted by the subject application. When the CPU 382 is configured in this way by the application, the CPU 382 becomes a specific purpose computer or a specific purpose machine.
The secondary storage 384 is typically comprised of one or more disk drives or tape drives and is used for non-volatile storage of data and as an over-flow data storage device if RAM 388 is not large enough to hold all working data. Secondary storage 384 may be used to store programs which are loaded into RAM 388 when such programs are selected for execution. The ROM 386 is used to store instructions and perhaps data which are read during program execution. ROM 386 is a non-volatile memory device which typically has a small memory capacity relative to the larger memory capacity of secondary storage 384. The RAM 388 is used to store volatile data and perhaps to store instructions. Access to both ROM 386 and RAM 388 is typically faster than to secondary storage 384. The secondary storage 384, the RAM 388, and/or the ROM 386 may be referred to in some contexts as computer readable storage media and/or non-transitory computer readable media.
I/O devices 390 may include printers, video monitors, liquid crystal displays (LCDs), touch screen displays, keyboards, keypads, switches, dials, mice, track balls, voice recognizers, card readers, paper tape readers, or other well-known input devices.
The network connectivity devices 392 may take the form of modems, modem banks, Ethernet cards, universal serial bus (USB) interface cards, serial interfaces, token ring cards, fiber distributed data interface (FDDI) cards, wireless local area network (WLAN) cards, radio transceiver cards that promote radio communications using protocols such as code division multiple access (CDMA), global system for mobile communications (GSM), long-term evolution (LTE), worldwide interoperability for microwave access (WiMAX), near field communications (NFC), radio frequency identity (RFID), and/or other air interface protocol radio transceiver cards, and other well-known network devices. These network connectivity devices 392 may enable the processor 382 to communicate with the Internet or one or more intranets. With such a network connection, it is contemplated that the processor 382 might receive information from the network, or might output information to the network in the course of performing the above-described method steps. Such information, which is often represented as a sequence of instructions to be executed using processor 382, may be received from and outputted to the network, for example, in the form of a computer data signal embodied in a carrier wave.
Such information, which may include data or instructions to be executed using processor 382 for example, may be received from and outputted to the network, for example, in the form of a computer data baseband signal or signal embodied in a carrier wave. The baseband signal or signal embedded in the carrier wave, or other types of signals currently used or hereafter developed, may be generated according to several methods well-known to one skilled in the art. The baseband signal and/or signal embedded in the carrier wave may be referred to in some contexts as a transitory signal.
The processor 382 executes instructions, codes, computer programs, scripts which it accesses from hard disk, floppy disk, optical disk (these various disk based systems may all be considered secondary storage 384), flash drive, ROM 386, RAM 388, or the network connectivity devices 392. While only one processor 382 is shown, multiple processors may be present. Thus, while instructions may be discussed as executed by a processor, the instructions may be executed simultaneously, serially, or otherwise executed by one or multiple processors. Instructions, codes, computer programs, scripts, and/or data that may be accessed from the secondary storage 384, for example, hard drives, floppy disks, optical disks, and/or other device, the ROM 386, and/or the RAM 388 may be referred to in some contexts as non-transitory instructions and/or non-transitory information.
In an embodiment, the computer system 380 may comprise two or more computers in communication with each other that collaborate to perform a task. For example, but not by way of limitation, an application may be partitioned in such a way as to permit concurrent and/or parallel processing of the instructions of the application. Alternatively, the data processed by the application may be partitioned in such a way as to permit concurrent and/or parallel processing of different portions of a data set by the two or more computers. In an embodiment, virtualization software may be employed by the computer system 380 to provide the functionality of a number of servers that is not directly bound to the number of computers in the computer system 380. For example, virtualization software may provide twenty virtual servers on four physical computers. In an embodiment, the functionality disclosed above may be provided by executing the application and/or applications in a cloud computing environment. Cloud computing may comprise providing computing services via a network connection using dynamically scalable computing resources. Cloud computing may be supported, at least in part, by virtualization software. A cloud computing environment may be established by an enterprise and/or may be hired on an as-needed basis from a third party provider. Some cloud computing environments may comprise cloud computing resources owned and operated by the enterprise as well as cloud computing resources hired and/or leased from a third party provider.
In an embodiment, some or all of the functionality disclosed above may be provided as a computer program product. The computer program product may comprise one or more computer readable storage medium having computer usable program code embodied therein to implement the functionality disclosed above. The computer program product may comprise data structures, executable instructions, and other computer usable program code. The computer program product may be embodied in removable computer storage media and/or non-removable computer storage media. The removable computer readable storage medium may comprise, without limitation, a paper tape, a magnetic tape, magnetic disk, an optical disk, a solid state memory chip, for example analog magnetic tape, compact disk read only memory (CD-ROM) disks, floppy disks, jump drives, digital cards, multimedia cards, and others. The computer program product may be suitable for loading, by the computer system 380, at least portions of the contents of the computer program product to the secondary storage 384, to the ROM 386, to the RAM 388, and/or to other non-volatile memory and volatile memory of the computer system 380. The processor 382 may process the executable instructions and/or data structures in part by directly accessing the computer program product, for example by reading from a CD-ROM disk inserted into a disk drive peripheral of the computer system 380. Alternatively, the processor 382 may process the executable instructions and/or data structures by remotely accessing the computer program product, for example by downloading the executable instructions and/or data structures from a remote server through the network connectivity devices 392. The computer program product may comprise instructions that promote the loading and/or copying of data, data structures, files, and/or executable instructions to the secondary storage 384, to the ROM 386, to the RAM 388, and/or to other non-volatile memory and volatile memory of the computer system 380.
In some contexts, the secondary storage 384, the ROM 386, and the RAM 388 may be referred to as a non-transitory computer readable medium or a computer readable storage media. A dynamic RAM embodiment of the RAM 388, likewise, may be referred to as a non-transitory computer readable medium in that while the dynamic RAM receives electrical power and is operated in accordance with its design, for example during a period of time during which the computer system 380 is turned on and operational, the dynamic RAM stores information that is written to it. Similarly, the processor 382 may comprise an internal RAM, an internal ROM, a cache memory, and/or other internal non-transitory storage blocks, sections, or components that may be referred to in some contexts as non-transitory computer readable media or computer readable storage media.
While several embodiments have been provided in the present disclosure, it should be understood that the disclosed systems and methods may be embodied in many other specific forms without departing from the spirit or scope of the present disclosure. The present examples are to be considered as illustrative and not restrictive, and the intention is not to be limited to the details given herein. For example, the various elements or components may be combined or integrated in another system or certain features may be omitted or not implemented.
Also, techniques, systems, subsystems, and methods described and illustrated in the various embodiments as discrete or separate may be combined or integrated with other systems, modules, techniques, or methods without departing from the scope of the present disclosure. Other items shown or discussed as directly coupled or communicating with each other may be indirectly coupled or communicating through some interface, device, or intermediate component, whether electrically, mechanically, or otherwise. Other examples of changes, substitutions, and alterations are ascertainable by one skilled in the art and could be made without departing from the spirit and scope disclosed herein.
Number | Name | Date | Kind |
---|---|---|---|
5928363 | Ruvolo | Jul 1999 | A |
5983350 | Minear et al. | Nov 1999 | A |
6775772 | Binding et al. | Aug 2004 | B1 |
7305712 | Watt et al. | Dec 2007 | B2 |
7325083 | Watt et al. | Jan 2008 | B2 |
7340573 | Watt | Mar 2008 | B2 |
7370210 | Symes | May 2008 | B2 |
7661104 | Watt et al. | Feb 2010 | B2 |
7702951 | Yamamoto et al. | Apr 2010 | B2 |
7849296 | Watt et al. | Dec 2010 | B2 |
8005000 | Srinivasan | Aug 2011 | B1 |
8090797 | Chinta et al. | Jan 2012 | B2 |
8171483 | Nord et al. | May 2012 | B2 |
8271976 | Vega et al. | Sep 2012 | B2 |
8280944 | Laadan et al. | Oct 2012 | B2 |
8443230 | James-Roxby et al. | May 2013 | B1 |
8649770 | Cope et al. | Feb 2014 | B1 |
8712407 | Cope et al. | Apr 2014 | B1 |
8797875 | Garcia Martin et al. | Aug 2014 | B2 |
8842578 | Zisapel et al. | Sep 2014 | B1 |
8863252 | Katzer et al. | Oct 2014 | B1 |
8918860 | Enderwick | Dec 2014 | B1 |
8935318 | Konerding et al. | Jan 2015 | B1 |
8984110 | Asveren | Mar 2015 | B1 |
8996644 | Pope | Mar 2015 | B2 |
9047441 | Xie et al. | Jun 2015 | B2 |
9077709 | Dall et al. | Jul 2015 | B1 |
9104836 | Burstein et al. | Aug 2015 | B2 |
9118655 | Paczkowski et al. | Aug 2015 | B1 |
9161227 | Bye et al. | Oct 2015 | B1 |
9183606 | Paczkowski et al. | Nov 2015 | B1 |
9191865 | Paczkowski et al. | Nov 2015 | B1 |
9203842 | Venkataramu et al. | Dec 2015 | B2 |
9230085 | Paczkowski et al. | Jan 2016 | B1 |
9237084 | Chapman | Jan 2016 | B2 |
9251384 | Potlapally et al. | Feb 2016 | B1 |
9274974 | Chen et al. | Mar 2016 | B1 |
9282898 | McRoberts et al. | Mar 2016 | B2 |
9288148 | Krishnaswamy et al. | Mar 2016 | B1 |
9298515 | McMurry et al. | Mar 2016 | B2 |
9317689 | Aissi | Apr 2016 | B2 |
9319220 | Grewal et al. | Apr 2016 | B2 |
9324016 | Cordes et al. | Apr 2016 | B1 |
9374363 | Paczkowski et al. | Jun 2016 | B1 |
9384028 | Felstaine et al. | Jul 2016 | B1 |
9392446 | Paczkowski et al. | Jul 2016 | B1 |
9396016 | Marquardt et al. | Jul 2016 | B1 |
9407612 | Sood et al. | Aug 2016 | B2 |
9413634 | Nadeau et al. | Aug 2016 | B2 |
9444886 | Rahman et al. | Sep 2016 | B2 |
9450866 | He et al. | Sep 2016 | B2 |
9460286 | Felstaine et al. | Oct 2016 | B1 |
9462084 | Connor et al. | Oct 2016 | B2 |
9503344 | Chakrabarti et al. | Nov 2016 | B2 |
9503363 | Sivabalan et al. | Nov 2016 | B2 |
9509587 | Marquardt et al. | Nov 2016 | B1 |
9519563 | Manghirmalani et al. | Dec 2016 | B2 |
9537741 | Chakrabarti et al. | Jan 2017 | B2 |
9549321 | Slavov et al. | Jan 2017 | B2 |
9560078 | Sood et al. | Jan 2017 | B2 |
9565168 | Marquardt et al. | Feb 2017 | B1 |
9578008 | Sood et al. | Feb 2017 | B2 |
9578664 | Paczkowski et al. | Feb 2017 | B1 |
9613190 | Ford et al. | Apr 2017 | B2 |
9628445 | Antonakakis | Apr 2017 | B2 |
9648044 | Islam et al. | May 2017 | B2 |
9667665 | Paczkowski et al. | May 2017 | B1 |
9686237 | DeCusatis et al. | Jun 2017 | B2 |
9686240 | Ray et al. | Jun 2017 | B1 |
9692743 | Islam et al. | Jun 2017 | B2 |
9729515 | Anantharaju | Aug 2017 | B1 |
9742790 | Sood et al. | Aug 2017 | B2 |
9749294 | Marquardt et al. | Aug 2017 | B1 |
9769854 | Paczkowski et al. | Sep 2017 | B1 |
9781016 | Marquardt et al. | Oct 2017 | B1 |
9811686 | Marquardt et al. | Nov 2017 | B1 |
9819661 | Stern | Nov 2017 | B2 |
9871768 | Ray et al. | Jan 2018 | B1 |
9979699 | Marquardt et al. | May 2018 | B1 |
10044572 | Marquardt et al. | Aug 2018 | B1 |
10250498 | Bales et al. | Apr 2019 | B1 |
20020091650 | Ellis | Jul 2002 | A1 |
20020107958 | Faraldo | Aug 2002 | A1 |
20030177387 | Osterwalder et al. | Sep 2003 | A1 |
20040139352 | Shewchuk et al. | Jul 2004 | A1 |
20040177269 | Belnet et al. | Sep 2004 | A1 |
20040181682 | Orino et al. | Sep 2004 | A1 |
20040187117 | Orion et al. | Sep 2004 | A1 |
20040221163 | Jorgensen | Nov 2004 | A1 |
20040240468 | Chin et al. | Dec 2004 | A1 |
20040260910 | Watt et al. | Dec 2004 | A1 |
20050068981 | Park et al. | Mar 2005 | A1 |
20050102603 | Tapper et al. | May 2005 | A1 |
20050114616 | Tune et al. | May 2005 | A1 |
20050124382 | Britt et al. | Jun 2005 | A1 |
20050138421 | Fedronic et al. | Jun 2005 | A1 |
20050160210 | Watt et al. | Jul 2005 | A1 |
20050185672 | Endo et al. | Aug 2005 | A1 |
20050261985 | Miller et al. | Nov 2005 | A1 |
20060146767 | Moganti | Jul 2006 | A1 |
20060190614 | Altman et al. | Aug 2006 | A1 |
20060253701 | Kim et al. | Nov 2006 | A1 |
20060259641 | Kim et al. | Nov 2006 | A1 |
20070174253 | Hodnett et al. | Jul 2007 | A1 |
20070282572 | Larus | Dec 2007 | A1 |
20080020745 | Bae et al. | Jan 2008 | A1 |
20080049640 | Heinz et al. | Feb 2008 | A1 |
20080162877 | Altman et al. | Jul 2008 | A1 |
20080165259 | Nobels | Jul 2008 | A1 |
20080189468 | Schmidt et al. | Aug 2008 | A1 |
20080301779 | Garg et al. | Dec 2008 | A1 |
20090007100 | Field et al. | Jan 2009 | A1 |
20090199177 | Edwards et al. | Aug 2009 | A1 |
20090210526 | Howell et al. | Aug 2009 | A1 |
20090241108 | Edwards et al. | Sep 2009 | A1 |
20090254984 | Nice et al. | Oct 2009 | A1 |
20090300605 | Edwards et al. | Dec 2009 | A1 |
20090320048 | Watt et al. | Dec 2009 | A1 |
20100103837 | Jungck et al. | Apr 2010 | A1 |
20100106568 | Grimes | Apr 2010 | A1 |
20100192230 | Steeves et al. | Jul 2010 | A1 |
20100306361 | Boberg et al. | Dec 2010 | A1 |
20110119748 | Edwards et al. | May 2011 | A1 |
20110173443 | Osterwalder et al. | Jul 2011 | A1 |
20120040662 | Rahman et al. | Feb 2012 | A1 |
20120117379 | Thornewell et al. | May 2012 | A1 |
20120304244 | Xie et al. | Nov 2012 | A1 |
20120331550 | Raj et al. | Dec 2012 | A1 |
20130055256 | Banga et al. | Feb 2013 | A1 |
20130070745 | Nixon et al. | Mar 2013 | A1 |
20130091568 | Sharif et al. | Apr 2013 | A1 |
20130160139 | Goel et al. | Jun 2013 | A1 |
20130219010 | Mahendran et al. | Aug 2013 | A1 |
20130290563 | Fleischman et al. | Oct 2013 | A1 |
20130305333 | Katzer et al. | Nov 2013 | A1 |
20130333008 | Tapling et al. | Dec 2013 | A1 |
20130345530 | McRoberts et al. | Dec 2013 | A1 |
20140013327 | Sherwood et al. | Jan 2014 | A1 |
20140033316 | Paczkowski et al. | Jan 2014 | A1 |
20140047548 | Bye et al. | Feb 2014 | A1 |
20140052922 | Moyer et al. | Feb 2014 | A1 |
20140053003 | Moyer et al. | Feb 2014 | A1 |
20140075567 | Raleigh et al. | Mar 2014 | A1 |
20140092753 | Vasseur et al. | Apr 2014 | A1 |
20140201374 | Ashwood-Smith et al. | Jul 2014 | A1 |
20140229945 | Barkai et al. | Aug 2014 | A1 |
20140241247 | Kempf et al. | Aug 2014 | A1 |
20140259115 | Bakshi et al. | Sep 2014 | A1 |
20140281529 | Epp et al. | Sep 2014 | A1 |
20140281544 | Paczkowski et al. | Sep 2014 | A1 |
20140289826 | Croome | Sep 2014 | A1 |
20140298477 | Castro et al. | Oct 2014 | A1 |
20140304803 | Pope et al. | Oct 2014 | A1 |
20140325681 | Kleidermacher et al. | Oct 2014 | A1 |
20140331297 | Innes et al. | Nov 2014 | A1 |
20140337940 | Slavov et al. | Nov 2014 | A1 |
20140344912 | Chapman et al. | Nov 2014 | A1 |
20140373012 | Ylitalo et al. | Dec 2014 | A1 |
20150046676 | Archibald et al. | Feb 2015 | A1 |
20150072726 | Stern | Mar 2015 | A1 |
20150074745 | Stern et al. | Mar 2015 | A1 |
20150074764 | Stern | Mar 2015 | A1 |
20150089246 | Kanai et al. | Mar 2015 | A1 |
20150117409 | Ghai | Apr 2015 | A1 |
20150117455 | Umesh et al. | Apr 2015 | A1 |
20150169885 | Paczkowski et al. | Jun 2015 | A1 |
20150172928 | Katzer et al. | Jun 2015 | A1 |
20150180730 | Felstaine et al. | Jun 2015 | A1 |
20150195149 | Vasseur et al. | Jul 2015 | A1 |
20150195281 | Venkataramu et al. | Jul 2015 | A1 |
20150215308 | Manolov et al. | Jul 2015 | A1 |
20150220937 | Iannace et al. | Aug 2015 | A1 |
20150237035 | Islam et al. | Aug 2015 | A1 |
20150244717 | Jin et al. | Aug 2015 | A1 |
20150248283 | Gschwind et al. | Sep 2015 | A1 |
20150248554 | Dumitru et al. | Sep 2015 | A1 |
20150358248 | Saha et al. | Dec 2015 | A1 |
20150365352 | Xiang | Dec 2015 | A1 |
20150370704 | Kato | Dec 2015 | A1 |
20150373050 | Dayan et al. | Dec 2015 | A1 |
20150381423 | Xiang | Dec 2015 | A1 |
20160004876 | Bye et al. | Jan 2016 | A1 |
20160007190 | Wane | Jan 2016 | A1 |
20160043944 | Felstaine et al. | Feb 2016 | A1 |
20160057102 | Wei et al. | Feb 2016 | A1 |
20160057171 | DeCusatis et al. | Feb 2016 | A1 |
20160057788 | Sharma et al. | Feb 2016 | A1 |
20160073283 | Grayson et al. | Mar 2016 | A1 |
20160080323 | Mackay et al. | Mar 2016 | A1 |
20160086172 | Kamal et al. | Mar 2016 | A1 |
20160094573 | Sood et al. | Mar 2016 | A1 |
20160119141 | Jing et al. | Apr 2016 | A1 |
20160119374 | Williams et al. | Apr 2016 | A1 |
20160127323 | Antonakakis | May 2016 | A1 |
20160127333 | Sood et al. | May 2016 | A1 |
20160142396 | McRoberts et al. | May 2016 | A1 |
20160149748 | Pan | May 2016 | A1 |
20160149921 | Potlapally et al. | May 2016 | A1 |
20160157084 | Tsubouchi | Jun 2016 | A1 |
20160170848 | Yang et al. | Jun 2016 | A1 |
20160180089 | Dalcher | Jun 2016 | A1 |
20160182499 | Sharaga et al. | Jun 2016 | A1 |
20160182567 | Sood et al. | Jun 2016 | A1 |
20160205004 | Chou et al. | Jul 2016 | A1 |
20160212016 | Vrzic et al. | Jul 2016 | A1 |
20160212620 | Paczkowski et al. | Jul 2016 | A1 |
20160219076 | Paczkowski et al. | Jul 2016 | A1 |
20160226663 | Jones et al. | Aug 2016 | A1 |
20160226907 | Weiss et al. | Aug 2016 | A1 |
20160226912 | Clark et al. | Aug 2016 | A1 |
20160226913 | Sood et al. | Aug 2016 | A1 |
20160234725 | Paczkowski et al. | Aug 2016 | A1 |
20160253664 | Yuan et al. | Sep 2016 | A1 |
20160323200 | Xiang et al. | Nov 2016 | A1 |
20160337329 | Sood et al. | Nov 2016 | A1 |
20160344560 | Caceres et al. | Nov 2016 | A1 |
20160350150 | Marquardt et al. | Dec 2016 | A1 |
20160352537 | Marquardt et al. | Dec 2016 | A1 |
20160366105 | Smith et al. | Dec 2016 | A1 |
20160366123 | Smith et al. | Dec 2016 | A1 |
20160373474 | Sood et al. | Dec 2016 | A1 |
20160378685 | Spurlock et al. | Dec 2016 | A1 |
20160379003 | Kapoor et al. | Dec 2016 | A1 |
20170005990 | Birger et al. | Jan 2017 | A1 |
20170012968 | Feng et al. | Jan 2017 | A1 |
20170012975 | Ilyadis et al. | Jan 2017 | A1 |
20170034284 | Smith et al. | Feb 2017 | A1 |
20170068817 | Ali et al. | Mar 2017 | A1 |
20170093806 | Phegade et al. | Mar 2017 | A1 |
20170102957 | Marquardt et al. | Apr 2017 | A1 |
20170126413 | Grobman et al. | May 2017 | A1 |
20170142024 | Fromentoux et al. | May 2017 | A1 |
20170142163 | Sood et al. | May 2017 | A1 |
20170149798 | Antonakakis | May 2017 | A1 |
20170161501 | Sood et al. | Jun 2017 | A1 |
20170187723 | Islam et al. | Jun 2017 | A1 |
20170208038 | Hinaman et al. | Jul 2017 | A1 |
20170214694 | Yan | Jul 2017 | A1 |
20170230428 | Paczkowski et al. | Aug 2017 | A1 |
20170308872 | Uhr | Oct 2017 | A1 |
20170310647 | Hu et al. | Oct 2017 | A1 |
20170347308 | Chou et al. | Nov 2017 | A1 |
20180082390 | Leidner | Mar 2018 | A1 |
20180218342 | Lim | Aug 2018 | A1 |
20190019188 | Leidner | Jan 2019 | A1 |
Number | Date | Country |
---|---|---|
WO2015039699 | Mar 2015 | WO |
WO2016102297 | Jun 2016 | WO |
WO2017062101 | Apr 2017 | WO |
Entry |
---|
FAIPP Pre-Interview Communication dated Oct. 9, 2018, U.S. Appl. No. 151284,506, filed Oct. 3, 2016. |
Office Action dated Oct. 30, 2018, U.S. Appl. No. 14/872,936, filed Oct. 1, 2015. |
Notice of Allowance dated Nov. 28, 2018, U.S. Appl. No. 15/284,506, filed Oct. 3, 2016. |
Mahy, et al., entitled, “Traversal Using Relays around NAT (TURN): Relay Extensions to Session Traversal Utilities for NAT (STUN),” Apr. 2010, Internet Engineering Task Force (IETF), Request for Comments: 5766. (Year: 2010). |
Notice of Allowance Dated dated Apr. 16, 2018, U.S. Appl. No. 15/686,324, filed Aug. 25, 2017. |
Final Office Action dated Apr. 17, 2018, U.S. Appl. No. 14/872,936, filed Oct. 1, 2015. |
Foreign Communication from a Related Counterpart , Preliminary Report on Patentability, dated Apr. 19, 2018, Application No. PCT/US2016/046648, filed on Aug. 11, 2016. |
Notice of Allowance dated Oct. 5, 2016, U.S. Appl. No. 14/746,615, filed Jun. 22, 2015. |
Notice of Allowance dated May 12, 2017, U.S. Appl. No. 15/403,166, filed Jan. 10, 2017. |
FAIPP Pre-Interview Communication dated Jul. 25, 2016, U.S. Appl. No. 14/703,885, filed May 5, 2015. |
Notice of Allowance dated Sep. 20, 2016, U.S. Appl. No. 14/703,885, filed May 5, 2015. |
FAIPP Pre-Interview Communication dated Feb. 7, 2017, U.S. Appl. No. 14/847,992, filed Sep. 8, 2015. |
Notice of Allowance dated Apr. 25, 2017, U.S. Appl. No. 14/847,992, filed Sep. 8, 2015. |
FAIPP Pre-Interview Communication dated Dec. 15, 2017, U.S. Appl. No. 15/655,080, filed Jul. 20, 2017. |
Notice of Allowance dated Jan. 30, 2018, U.S. Appl. No. 15/655,080, filed Jul. 20, 2017. |
FAIPP Pre-Interview Communication dated Mar. 9, 2017, U.S. Appl. No. 14/930,146, filed Nov. 2, 2015. |
Notice of Allowance dated May 26, 2017, U.S. Appl. No. 14/930,146, filed Nov. 2, 2015. |
FAIPP Pre-Interview Communication dated Jan. 12, 2018, U.S. Appl. No. 15/686,324, filed Aug. 25, 2017. |
FAIPP Pre-Interview Communication dated Apr. 20, 2017, U.S. Appl. No. 14/879,324, filed Oct. 9, 2015. |
Notice of Allowance dated Jul. 3, 2017, U.S. Appl. No. 14/879,324, filed Oct. 9, 2015. |
FAIPP Pre-Interview Communication dated Oct. 19, 2017, U.S. Appl. No. 14/872,936, filed Oct. 1, 2015. |
Office Action dated Jul. 24, 2017, U.S. Appl. No. 14/879,327, filed Oct. 9, 2015. |
Final Office Action dated Feb. 7, 2018, U.S. Appl. No. 14/879,327, filed Oct. 9, 2015. |
Notice of Allowance dated Feb. 21, 2017, U.S. Appl. No. 14/793,344, filed Jul. 7, 2015. |
Notice of Allowance dated Sep. 11, 2017, U.S. Appl. No. 15/602,057, filed May 22, 2017. |
Network Functions Virtualisation, “An Introduction, Benefits, Enablers, Challenges Call for Action,” Oct. 22-24, 2012, “SDN an OpenFlow World Congress,” Darmstadt-Germany. |
Network Functions Virtualisation, “Network Operator Perspectives on Industry Progress,” Oct. 14-17, 2012, “SDN an OpenFlow World Congress,” Darmstadt-Germany. |
Foreign Communication from a Related Counterpart—International Search Report and Written Opinion dated Oct. 25, 2016, filed on Aug. 11, 2016, International Application No. PCT/US2016/046648. |
Hwang, Jinho, et al., entitled, “NetVM: High Performance and Flexible Networking Using Virtualization on Commodity Platforms,” 11th USENIX Symposium on Networked Systems Design and Implementation (NSDI 14); Apr. 2-5, 2014, Seattle WA, US, ISBN 978-1-931971-09-06. |
Marquardt, Ronald R., et al., “System and Method of Establishing Trusted Operability Between Networks in a Network Functions Virtualization Environment,” filed Jul. 20, 2017, U.S. Appl. No. 15/655,080. |
Marquardt, Ronald R., et al., “Dynamic Addition of Network Function Services,” filed on Aug. 25, 2017, U.S. Appl. No. 15/686,324. |
Marquardt, Ronald R., et al., “Securing Communication in a Network Function Virtualization (NFV) Core Network,” filed on Oct. 1, 2015, U.S. Appl. No. 14/872,936. |
Bales, Mark R., et al., “Session Aggregator Brokering of Data Stream Communication,” filed on Oct. 3, 2016, U.S. Appl. No. 15/284,506. |
Final Office Action dated Apr. 19, 2019, U.S. Appl. No. 14/872,936, filed Oct. 1, 2015. |