HTTP-AWARE CONTENT CACHING

Information

  • Patent Application
  • 20180234518
  • Publication Number
    20180234518
  • Date Filed
    August 11, 2015
    9 years ago
  • Date Published
    August 16, 2018
    6 years ago
Abstract
Certain aspects of the present disclosure provide methods and apparatus for caching content at a network component and providing the cached content to requesting UEs. An example method generally includes receiving a request for content from a user equipment, inspecting the request for information about the requested content, determining, based on information about the requested content obtained by inspecting the request, if the requested content is on a service list, and providing the requested content from a cache to the UE if it is determined the requested content is on the service list and cached at the base station or retrieving the requested content from a remote source, storing the requested content in the cache, and providing the requested content to the UE if it is determined the requested content is on the service list but not cached at the base station.
Description
TECHNICAL FIELD

Certain embodiments of the present disclosure generally relate to caching content at network components.


BACKGROUND

Wireless communication systems are widely deployed to provide various types of communication content such as voice, data, and so on. These systems may be multiple-access systems capable of supporting communication with multiple users by sharing the available system resources (e.g., bandwidth and transmit power). Examples of such multiple-access systems include code division multiple access (CDMA) systems, time division multiple access (TDMA) systems, frequency division multiple access (FDMA) systems, 3GPP Long Term Evolution (LTE) systems, and orthogonal frequency division multiple access (OFDMA) systems.


Generally, a wireless multiple-access communication system can simultaneously support communication for multiple wireless terminals. Each terminal communicates with one or more base stations via transmissions on the forward and reverse links. The forward link (or downlink) refers to the communication link from the base stations to the terminals, and the reverse link (or uplink) refers to the communication link from the terminals to the base stations. This communication link may be established via a single-in-single-out, multiple-in-single-out or a multiple-in-multiple-out (MIMO) system.


Wireless devices comprise user equipments (UEs) and remote devices. A UE is a device that operates under direct control by humans. Some examples of UEs include cellular phones (e.g., smart phones), personal digital assistants (PDAs), wireless modems, handheld devices, laptop computers, tablets, netbooks, smartbooks, ultrabooks, robots, drones, wearable devices (e.g., smart watch, smart bracelet, smart clothing, smart glasses), etc. A remote device is a device that operates without being directly controlled by humans. Some examples of remote devices include autonomous robots, autonomous drones, sensors, meters, location tags, monitoring devices, etc. A remote device may communicate with a base station, another remote device, or some other entity. Machine type communication (MTC) refers to communication involving at least one remote device on at least one end of the communication.


SUMMARY

Certain aspects of the present disclosure provide a method for communicating by a base station. The method generally includes receiving a request for content from a user equipment (UE), inspecting the request for information about the requested content, determining, based on information about the requested content obtained by inspecting the request, if the requested content is on a service list, and providing the requested content from a cache to the UE if it is determined the requested content is on the service list and stored in the cache or retrieving the requested content from a remote source, storing the requested content in the cache, and providing the requested content to the UE if it is determined the requested content is on the service list but not cached at the base station.


Certain aspects of the present disclosure provide a method for communicating by a user equipment (UE). The method generally includes transmitting a request for content to a base station, and receiving the requested content from the base station if the requested content is cached at the base station or receiving the requested content from a remote source if the requested content is not cached at the base station.


Certain aspects of the present disclosure also include various apparatuses and computer program products capable of performing the operations described above.





BRIEF DESCRIPTION OF THE DRAWINGS

The features, nature, and advantages of the present disclosure will become more apparent from the detailed description set forth below when taken in conjunction with the drawings in which like reference characters identify correspondingly throughout and wherein:



FIG. 1 illustrates a multiple access wireless communication system, according to aspects of the present disclosure.



FIG. 2 is a block diagram of a communication system, according to aspects of the present disclosure.



FIG. 3 illustrates an example frame structure, according to aspects of the present disclosure.



FIG. 4 illustrates an example subframe resource element mapping, according to aspects of the present disclosure.



FIG. 5 illustrates example operations that may be performed by a base station to cache content for transmission to a plurality of user equipments (UEs), in accordance with an aspect of the present disclosure.



FIG. 6 illustrates an example network architecture with caching performed at a plurality of eNodeBs, in accordance with an aspect of the present disclosure.



FIG. 7 illustrates an example of synchronization between base stations, in accordance with an aspect of the present disclosure.



FIG. 8 illustrates an example flow diagram of operations for caching content at a network edge and transmitting cached content to a UE, in accordance with an aspect of the present disclosure.



FIG. 9 illustrates an example flow diagram of operations for caching content at a network edge and transmitting cached content to a UE, in accordance with an aspect of the present disclosure.



FIG. 10 illustrates an example flow diagram of operations for caching content and transmitting cached content to a UE, in accordance with an aspect of the present disclosure.



FIG. 11 illustrates example operations that may be performed by a base station to provide content to a user equipment (UE) from a cache or a remote source, in accordance with an aspect of the present disclosure.



FIG. 12 illustrates example operations that may be performed by a user equipment (UE) to obtain content from a base station or a remote source, in accordance with an aspect of the present disclosure.



FIG. 13 illustrates an example network architecture for HTTP-aware caching, in accordance with an aspect of the present disclosure.



FIG. 14 illustrates an example network protocol stack for HTTP-aware caching, in accordance with an aspect of the present disclosure.





DETAILED DESCRIPTION

Aspects of the present disclosure provide techniques for caching content at a network edge. Caching content at a network edge may allow for reduced latency in the provisioning of content to requesting UEs.


The detailed description set forth below, in connection with the appended drawings, is intended as a description of various configurations and is not intended to represent the only configurations in which the concepts described herein may be practiced. The detailed description includes specific details for the purpose of providing a thorough understanding of the various concepts. However, it will be apparent to those skilled in the art that these concepts may be practiced without these specific details. In some instances, well-known structures and components are shown in block diagram form in order to avoid obscuring such concepts.


The techniques described herein may be used for various wireless communication networks such as Code Division Multiple Access (CDMA) networks, Time Division Multiple Access (TDMA) networks, Frequency Division Multiple Access (FDMA) networks, Orthogonal FDMA (OFDMA) networks, Single-Carrier FDMA (SC-FDMA) networks, etc. The terms “networks” and “systems” are often used interchangeably. A CDMA network may implement a radio technology such as Universal Terrestrial Radio Access (UTRA), cdma2000, etc. UTRA includes Wideband-CDMA (W-CDMA) and Low Chip Rate (LCR). cdma2000 covers IS-2000, IS-95 and IS-856 standards. A TDMA network may implement a radio technology such as Global System for Mobile Communications (GSM). An OFDMA network may implement a radio technology such as Evolved UTRA (E-UTRA), IEEE 802.11, IEEE 802.16, IEEE 802.20, Flash-OFDM®, etc. UTRA, E-UTRA, and GSM are part of Universal Mobile Telecommunication System (UMTS). Long Term Evolution (LTE) is an upcoming release of UMTS that uses E-UTRA. UTRA, E-UTRA, GSM, UMTS and LTE are described in documents from an organization named “3rd Generation Partnership Project” (3GPP). cdma2000 is described in documents from an organization named “3rd Generation Partnership Project 2” (3GPP2). These various radio technologies and standards are known in the art. For clarity, certain aspects of the techniques are described below for LTE, and LTE terminology is used in much of the description below.


Single carrier frequency division multiple access (SC-FDMA), which utilizes single carrier modulation and frequency domain equalization is a technique. SC-FDMA has similar performance and essentially the same overall complexity as those of OFDMA system. SC-FDMA signal has lower peak-to-average power ratio (PAPR) because of its inherent single carrier structure. SC-FDMA has drawn great attention, especially in the uplink communications where lower PAPR greatly benefits the mobile terminal in terms of transmit power efficiency. It is currently a working assumption for uplink multiple access scheme in 3GPP Long Term Evolution (LTE), or Evolved UTRA.



FIG. 1 shows a wireless communication network 100 in which aspects of the present disclosure may be practiced. For example, evolved Node Bs 110 may cache content and transmit the cached content to user equipments (UEs) 120 as described herein.


Wireless communication network 100 may be an LTE network. The wireless network 100 may include a number of evolved Node Bs (eNBs) 110 and other network entities. An eNB may be a station that communicates with the UEs and may also be referred to as a base station, an access point, etc. A Node B is another example of a station that communicates with the UEs.


Each eNB 110 may provide communication coverage for a particular geographic area. In 3GPP, the term “cell” can refer to a coverage area of an eNB and/or an eNB subsystem serving this coverage area, depending on the context in which the term is used.


An eNB may provide communication coverage for a macro cell, a pico cell, a femto cell, and/or other types of cell. A macro cell may cover a relatively large geographic area (e.g., several kilometers in radius) and may allow unrestricted access by UEs with service subscription. A pico cell may cover a relatively small geographic area and may allow unrestricted access by UEs with service subscription. A femto cell may cover a relatively small geographic area (e.g., a home) and may allow restricted access by UEs having association with the femto cell (e.g., UEs in a Closed Subscriber Group (CSG), UEs for users in the home, etc.). An eNB for a macro cell may be referred to as a macro eNB. An eNB for a pico cell may be referred to as a pico eNB. An eNB for a femto cell may be referred to as a femto eNB or a home eNB. In the example shown in FIG. 1, the eNBs 110a, 110b and 110c may be macro eNBs for the macro cells 102a, 102b and 102c, respectively. The eNB 110x may be a pico eNB for a pico cell 102x. The eNBs 110y and 110z may be femto eNBs for the femto cells 102y and 102z, respectively. An eNB may support one or multiple (e.g., three) cells.


The wireless network 100 may also include relay stations. A relay station is a station that receives a transmission of data and/or other information from an upstream station (e.g., an eNB or a UE) and sends a transmission of the data and/or other information to a downstream station (e.g., a UE or an eNB). A relay station may also be a UE that relays transmissions for other UEs. In the example shown in FIG. 1, a relay station 110r may communicate with the eNB 110a and a UE 120r in order to facilitate communication between the eNB 110a and the UE 120r. A relay station may also be referred to as a relay eNB, a relay, etc.


The wireless network 100 may be a heterogeneous network that includes eNBs of different types, e.g., macro eNBs, pico eNBs, femto eNBs, relays, etc. These different types of eNBs may have different transmit power levels, different coverage areas, and different impact on interference in the wireless network 100. For example, macro eNBs may have a high transmit power level (e.g., 20 Watts) whereas pico eNBs, femto eNBs and relays may have a lower transmit power level (e.g., 1 Watt).


The wireless network 100 may support synchronous or asynchronous operation. For synchronous operation, the eNBs may have similar frame timing, and transmissions from different eNBs may be approximately aligned in time. For asynchronous operation, the eNBs may have different frame timing, and transmissions from different eNBs may not be aligned in time. The techniques described herein may be used for both synchronous and asynchronous operation.


A network controller 130 may couple to a set of eNBs and provide coordination and control for these eNBs. The network controller 130 may communicate with the eNBs 110 via a backhaul. The eNBs 110 may also communicate with one another, e.g., directly or indirectly via wireless or wireline backhaul.


The UEs 120 may be dispersed throughout the wireless network 100, and each UE may be stationary or mobile. A UE may also be referred to as a terminal, a mobile station, a subscriber unit, a station, etc. A UE may be a cellular phone, a personal digital assistant (PDA), a wireless modem, a wireless communication device, a handheld device, a laptop computer, a cordless phone, a wireless local loop (WLL) station, etc. A UE may be able to communicate with macro eNBs, pico eNBs, femto eNBs, relays, etc. In FIG. 1, a solid line with double arrows indicates desired transmissions between a UE and a serving eNB, which is an eNB designated to serve the UE on the downlink and/or uplink. A dashed line with double arrows indicates interfering transmissions between a UE and an eNB.


LTE utilizes orthogonal frequency division multiplexing (OFDM) on the downlink and single-carrier frequency division multiplexing (SC-FDM) on the uplink. OFDM and SC-FDM partition the system bandwidth into multiple (K) orthogonal subcarriers, which are also commonly referred to as tones, bins, etc. Each subcarrier may be modulated with data. In general, modulation symbols are sent in the frequency domain with OFDM and in the time domain with SC-FDM. The spacing between adjacent subcarriers may be fixed, and the total number of subcarriers (K) may be dependent on the system bandwidth. For example, the spacing of the subcarriers may be 15 kHz and the minimum resource allocation (called a ‘resource block’) may be 12 subcarriers (or 180 kHz). Consequently, the nominal FFT size may be equal to 128, 256, 512, 1024 or 2048 for system bandwidth of 1.25, 2.5, 5, 10 or 20 megahertz (MHz), respectively. The system bandwidth may also be partitioned into subbands. For example, a subband may cover 1.08 MHz (i.e., 6 resource blocks), and there may be 1, 2, 4, 8 or 16 subbands for system bandwidth of 1.25, 2.5, 5, 10 or 20 MHz, respectively.


The wireless network 100 may also include UEs 120 capable of communicating with a core network via one or more radio access networks (RANs) that implement one or more radio access technologies (RATs). For example, according to certain aspects provided herein, the wireless network 100 may include co-located access points (APs) and/or base stations that provide communication through a first RAN implementing a first RAT and a second RAN implementing a second RAT. According to certain aspects, the first RAN may be a wide area wireless access network (WWAN) and the second RAN may be a wireless local area network (WLAN). Examples of WWAN may include, but not be limited to, for example, radio access technologies (RATs) such as LTE, UMTS, cdma2000, GSM, and the like. Examples of WLAN may include, but not be limited to, for example, RATs such as Wi-Fi or IEEE 802.11 based technologies, and the like.


According to certain aspects provided herein, the wireless network 100 may include co-located Wi-Fi access points (APs) and femto eNBs that provide communication through Wi-Fi and cellular radio links. As used herein, the term “co-located” generally means “in close proximity to,” and applies to Wi-Fi APs or femto eNBs within the same device enclosure or within separate devices that are in close proximity to each other. According to certain aspects of the present disclosure, as used herein, the term “femtoAP” may refer to a co-located Wi-Fi AP and femto eNB.



FIG. 2 is a block diagram of an embodiment of a transmitter system 210 (also known as an access point (AP)) and a receiver system 250 (also known as a user equipment (UE)) in a system, such as a MIMO system 200. Aspects of the present disclosure may practiced in the transmitter system (AP) 210 and the receiver system (UE) 250. Although referred to as transmitter system 210 and receiver system 250, these systems can transmit as well as receive, depending on the application. For example, transmitter system 210 may be configured to cache data requested by a receiver system 250 and provide the data to other receiver systems from the cache, as described below with reference to FIG. 5.


At the transmitter system 210, traffic data for a number of data streams is provided from a data source 212 to a transmit (TX) data processor 214. In an aspect, each data stream is transmitted over a respective transmit antenna. TX data processor 214 formats, codes, and interleaves the traffic data for each data stream based on a particular coding scheme selected for that data stream to provide coded data.


The coded data for each data stream may be multiplexed with pilot data using OFDM techniques. The pilot data is typically a known data pattern that is processed in a known manner and may be used at the receiver system to estimate the channel response. The multiplexed pilot and coded data for each data stream is then modulated (i.e., symbol mapped) based on a particular modulation scheme (e.g., BPSK, QSPK, M-PSK, or M-QAM) selected for that data stream to provide modulation symbols. The data rate, coding, and modulation for each data stream may be determined by instructions performed by processor 230.


The modulation symbols for all data streams are then provided to a TX MIMO processor 220, which may further process the modulation symbols (e.g., for OFDM). TX MIMO processor 220 then provides NT modulation symbol streams to NT transmitters (TMTR) 222a through 222t. In certain embodiments, TX MIMO processor 220 applies beamforming weights to the symbols of the data streams and to the antenna from which the symbol is being transmitted.


Each transmitter 222 receives and processes a respective symbol stream to provide one or more analog signals, and further conditions (e.g., amplifies, filters, and upconverts) the analog signals to provide a modulated signal suitable for transmission over the MIMO channel. NT modulated signals from transmitters 222a through 222t are then transmitted from NT antennas 224a through 224t, respectively.


At receiver system 250, the transmitted modulated signals are received by NR antennas 252a through 252r, and the received signal from each antenna 252 is provided to a respective receiver (RCVR) 254a through 254r. Each receiver 254 conditions (e.g., filters, amplifies, and downconverts) a respective received signal, digitizes the conditioned signal to provide samples, and further processes the samples to provide a corresponding “received” symbol stream.


An RX data processor 260 then receives and processes the NR received symbol streams from NR receivers 254 based on a particular receiver processing technique to provide NT “detected” symbol streams. The RX data processor 260 then demodulates, deinterleaves, and decodes each detected symbol stream to recover the traffic data for the data stream. The processing by RX data processor 260 is complementary to that performed by TX MIMO processor 220 and TX data processor 214 at transmitter system 210.


A processor 270 periodically determines which pre-coding matrix to use. Processor 270 formulates a reverse link message comprising a matrix index portion and a rank value portion.


The reverse link message may comprise various types of information regarding the communication link and/or the received data stream. The reverse link message is then processed by a TX data processor 238, which also receives traffic data for a number of data streams from a data source 236, modulated by a modulator 280, conditioned by transmitters 254a through 254r, and transmitted back to transmitter system 210.


At transmitter system 210, the modulated signals from receiver system 250 are received by antennas 224, conditioned by receivers 222, demodulated by a demodulator 240, and processed by a RX data processor 242 to extract the reserve link message transmitted by the receiver system 250. Processor 230 then determines which pre-coding matrix to use for determining the beamforming weights and then processes the extracted message.


According to certain aspects, the controllers/processors 230 and 270 may direct the operation at the transmitter system 210 and the receiver system 250, respectively. According to an aspect, the processor 230, TX data processor 214, and/or other processors and modules at the transmitter system 210 may perform or direct processes for the techniques described herein. According to another aspect, the processor 270, RX data processor 260, and/or other processors and modules at the receiver system 250 may perform or direct processes for the techniques described herein. For example, the processor 230, TX data processor 214, and/or other processors and modules at the transmitter system 210 may perform or direct operations 500 in FIG. 5. For example, the processor 270, RX data processor 260, and/or other processors and modules at the receiver system 250 may perform or direct operations at receiver system 250.


In an aspect, logical channels are classified into Control Channels and Traffic Channels. Logical Control Channels comprise Broadcast Control Channel (BCCH), which is a DL channel for broadcasting system control information. Paging Control Channel (PCCH) is a DL channel that transfers paging information. Multicast Control Channel (MCCH) is a point-to-multipoint DL channel used for transmitting Multimedia Broadcast and Multicast Service (MBMS) scheduling and control information for one or several MTCHs. Generally, after establishing an RRC connection, this channel is only used by UEs that receive MBMS (Note: old MCCH+MSCH). Dedicated Control Channel (DCCH) is a point-to-point bi-directional channel that transmits dedicated control information used by UEs having an RRC connection. In an aspect, Logical Traffic Channels comprise a Dedicated Traffic Channel (DTCH), which is a point-to-point bi-directional channel, dedicated to one UE, for the transfer of user information. Also, a Multicast Traffic Channel (MTCH) is a point-to-multipoint DL channel for transmitting traffic data.


In an aspect, Transport Channels are classified into DL and UL. DL Transport Channels comprise a Broadcast Channel (BCH), Downlink Shared Data Channel (DL-SDCH), and a Paging Channel (PCH), the PCH for support of UE power saving (DRX cycle is indicated by the network to the UE), broadcasted over entire cell and mapped to PHY resources which can be used for other control/traffic channels. The UL Transport Channels comprise a Random Access Channel (RACH), a Request Channel (REQCH), an Uplink Shared Data Channel (UL-SDCH), and a plurality of PHY channels. The PHY channels comprise a set of DL channels and UL channels.


In an aspect, a channel structure is provided that preserves low PAPR (at any given time, the channel is contiguous or uniformly spaced in frequency) properties of a single carrier waveform.



FIG. 3 shows an exemplary frame structure 300 for FDD in LTE. The transmission timeline for each of the downlink and uplink may be partitioned into units of radio frames. Each radio frame may have a predetermined duration (e.g., 10 milliseconds (ms)) and may be partitioned into 10 subframes with indices of 0 through 9. Each subframe may include two slots. Each radio frame may thus include 20 slots with indices of 0 through 19. Each slot may include L symbol periods, e.g., seven symbol periods for a normal cyclic prefix (as shown in FIG. 2) or six symbol periods for an extended cyclic prefix. The 2L symbol periods in each subframe may be assigned indices of 0 through 2L−1.


In LTE, an eNB may transmit a primary synchronization signal (PSS) and a secondary synchronization signal (SSS) on the downlink in the center 1.08 MHz of the system bandwidth for each cell supported by the eNB. The PSS and SSS may be transmitted in symbol periods 6 and 5, respectively, in subframes 0 and 5 of each radio frame with the normal cyclic prefix, as shown in FIG. 3. The PSS and SSS may be used by UEs for cell search and acquisition. During cell search and acquisition the terminal detects the cell frame timing and the physical-layer identity of the cell from which the terminal learns the start of the references-signal sequence (given by the frame timing) and the reference-signal sequence of the cell (given by the physical layer cell identity). The eNB may transmit a cell-specific reference signal (CRS) across the system bandwidth for each cell supported by the eNB. The CRS may be transmitted in certain symbol periods of each subframe and may be used by the UEs to perform channel estimation, channel quality measurement, and/or other functions. In aspects, different and/or additional reference signals may be employed. The eNB may also transmit a Physical Broadcast Channel (PBCH) in symbol periods 0 to 3 in slot 1 of certain radio frames. The PBCH may carry some system information. The eNB may transmit other system information such as System Information Blocks (SIBs) on a Physical Downlink Shared Channel (PDSCH) in certain subframes. The eNB may transmit control information/data on a Physical Downlink Control Channel (PDCCH) in the first B symbol periods of a subframe, where B may be configurable for each subframe. The eNB may transmit traffic data and/or other data on the PDSCH in the remaining symbol periods of each subframe.



FIG. 4 shows two exemplary subframe formats 410 and 420 for the downlink with the normal cyclic prefix. The available time frequency resources for the downlink may be partitioned into resource blocks. Each resource block may cover 12 subcarriers in one slot and may include a number of resource elements. Each resource element may cover one subcarrier in one symbol period and may be used to send one modulation symbol, which may be a real or complex value.


Subframe format 410 may be used for an eNB equipped with two antennas. A CRS may be transmitted from antennas 0 and 1 in symbol periods 0, 4, 7 and 11. A reference signal is a signal that is known a priori by a transmitter and a receiver and may also be referred to as a pilot. A CRS is a reference signal that is specific for a cell, e.g., generated based on a cell identity (ID). In FIG. 4, for a given resource element with label Ra, a modulation symbol may be transmitted on that resource element from antenna a, and no modulation symbols may be transmitted on that resource element from other antennas. Subframe format 420 may be used for an eNB equipped with four antennas. A CRS may be transmitted from antennas 0 and 1 in symbol periods 0, 4, 7 and 11 and from antennas 2 and 3 in symbol periods 1 and 8. For both subframe formats 410 and 420, a CRS may be transmitted on evenly spaced subcarriers, which may be determined based on cell ID. Different eNBs may transmit their CRSs on the same or different subcarriers, depending on their cell IDs. For both subframe formats 410 and 420, resource elements not used for the CRS may be used to transmit data (e.g., traffic data, control data, and/or other data).


The PSS, SSS, CRS and PBCH in LTE are described in 3GPP TS 36.211, entitled “Evolved Universal Terrestrial Radio Access (E-UTRA); Physical Channels and Modulation,” which is publicly available.


An interlace structure may be used for each of the downlink and uplink for FDD in LTE. For example, Q interlaces with indices of 0 through Q−1 may be defined, where Q may be equal to 4, 6, 8, 10, or some other value. Each interlace may include subframes that are spaced apart by Q frames. In particular, interlace q may include subframes q, q+Q, q+2Q, etc., where q∈{0, . . . , Q−1}.


The wireless network may support hybrid automatic retransmission (HARQ) for data transmission on the downlink and uplink. For HARQ, a transmitter (e.g., an eNB) may send one or more transmissions of a packet until the packet is decoded correctly by a receiver (e.g., a UE) or some other termination condition is encountered. For synchronous HARQ, all transmissions of the packet may be sent in subframes of a single interlace. For asynchronous HARQ, each transmission of the packet may be sent in any subframe.


A UE may be located within the coverage area of multiple eNBs. One of these eNBs may be selected to serve the UE. The serving eNB may be selected based on various criteria such as received signal strength, received signal quality, pathloss, etc. Received signal quality may be quantified by a signal-to-noise-and-interference ratio (SINR), or a reference signal received quality (RSRQ), or some other metric. The UE may operate in a dominant interference scenario in which the UE may observe high interference from one or more interfering eNBs.


Caching Content at Network Edges

Certain aspects of the present disclosure provide mechanisms for caching requested content at a network edge. Caching content at network edges may allow for reduced latency in provisioning content to a plurality of UEs requesting the same content.


Caching content generally allows for reduced latency in delivery of content. Reductions in latency may be useful in a variety of scenarios, such as industrial automation, delivery of content in real-time applications (e.g., online video games), to provide for increased TCP throughput, and so on. Caching at a network edge may reduce an amount of backhaul transmissions and reduce processing delays in provisioning the same content to a plurality of UEs. Requested data may be served, thus, from a device on the network edge rather than from a remote source, which may reduce the amount of traffic on a core network.



FIG. 5 illustrates example operations 500 that may be performed to cache content at a network edge and provide the content to a plurality of UEs from the cache, according to an aspect of the present disclosure.


Operations 500 begin at 502, where a base station receives, from a first user equipment (UE), a request for content from a remote source. At 504, the base station retrieves the content from the remote source and provides the content to the first UE. At 506, the base station stores at least a portion of the content in a local cache at the base station. At 508, the base station receives a request for the content from at least a second UE. At 510, the base station retrieves the content from the local cache and provides the content to the second UE.


In an aspect, content may be cached at a local gateway. For example, the content may be cached in a local gateway (LGW) collocated with the eNB or a standalone LGW. If a LGW is not supported by the network, the content may be cached in a Packet Data Network (PDN) gateway (P-GW). In a network using a broadcast-multicast architecture, content may be cached at a broadcast/multicast service center (BM-SC) located in a local network. Caching content at a local gateway or BM-SC may allow for a global view of content consumption; that is, an operator may be able to determine, based on content requests and the provisioning of content, what content users on the network are interested in.


In an example, the LGW, P-GW, or BM-SC may transmit a Dynamic Adaptive Streaming over HTTP (DASH) or HTTP request (e.g., an HTTP GET request) to a remote server to request content. The requested content may be pushed from the remote server to the LGW, P-GW, or BM-SC. For example, if content is requested using DASH, the remote server may transmit a plurality of DASH segments to the LGW, P-GW, or BM-SC. Similarly, in a Content Delivery Network (CDN), when a UE performs a DNS look up for content (for example www.video.example.com), the DNS look up can direct the UE to access the local server or cache located in the L-GW or P-GW. When the UE requests the content, the LGW or the P-GW may send the requested segments to the UE through the unicast channel. The BM-SC may transmit DASH segments to the UEs, e.g., via an enhanced multimedia broadcast/multicast service (eMBMS) channel. UEs may fetch missing content from the BM-SC, for example, using file repair procedures; however, instead of retrieving the missing content from a remote server, the UEs may retrieve the missing content from cached content at the local server (e.g., located in the BM-SC).


In some cases, a network may include two gateways. Each gateway may have a distinct access point name (APN). A UE may communicate through one gateway for low latency traffic and through another gateway for regular traffic. The gateway used for low latency traffic may include a local cache at which content may be cached for transmission to other UEs, as described herein. If requested content is not cached at the gateway used for low latency traffic, the eNB may request content from the remote server via the gateway used for regular traffic, which may route communications to the remote server through the core network.


In an aspect, content may be cached at an eNodeB (eNB). In this case, the eNB is served as local proxy or local server. The eNB can cache content requested by a first UE and transmit the same content to other UEs that request the same content. An eNB may determine that other UEs are requesting the same content, for example, by performing Deep Packet Inspection (DPI). In streaming applications (e.g., DASH), the eNB may act as a client to fetch content for the plurality of requesting UEs. In some cases, the eNB may predict that certain content will be requested by a UE (or a plurality of UEs) and begin to fetch and catch that content preemptively for retransmission to requesting UEs in the future.


In some cases, the eNB may act as a proxy server. Resources requested by one UE may be made available to other UEs, whether connected to the same eNB or to a neighbor eNB. In some cases, a remote server can predict that a large number of UEs may request particular content from the remote server and push the predicted content to the eNB, which then caches the content and provides the content to UEs from the cache. The eNB may receive the content from the remote server through a P-GW or directly from the remote server via an internet connection or a low latency network connection. The UE is assigned an IP address from the P-GW. When a UE moves from the eNB to a neighbor eNB, the UE's IP address need not change.


In some cases, whether the eNB acts as a proxy server may be transparent to the UE. The eNB may implement a full network protocol stack. A TCP session may be established between a UE and the eNB, and between the eNB and the remote server. The eNB performs DPI. If the eNB detects it does not have the requested content in cache, the eNB may request the data from the remote server directly through internet or CDN. Alternatively, the eNB may request content from the remote server, for example by transmitting the request through a GPRS Transport Protocol (GTP) tunnel (e.g., a GTP-User Plane (GTP-U) tunnel), or by forwarding a request for content to the packet data network gateway (P-GW) using an IP interface. The GTP-U tunnel may be per eNB instead of per UE tunnel.



FIG. 6 illustrates an example network architecture in which multiple eNodeBs cache content, in accordance with aspects of the present disclosure. In some aspects, for UE mobility and load balancing, the same content or segments of content may be duplicated and cached in neighboring eNBs. Alternatively, different content or segments of content can be disturbed and cached in the neighboring eNBs. When the UE moves from one eNB to another eNB, the UE may continue to receive data from the previous proxy/server which is located in the source eNB over an X2 interface. In some aspects, a session with the previous proxy located in the source eNB may be discontinued, and the UE may re-establish a session (for example, TCP or UDP) between the UE and the new proxy in the target eNB. The UE may continue to retrieve the data from the previous proxy until the UE establishes the connection with the new proxy. In some aspects, HTTP redirection from the previous proxy may be used to redirect the UE to the target proxy. A TCP session may be transferred from the source proxy to the target proxy. When a TCP session transfer is triggered by an X2 context transfer, the source proxy can move a socket to the target proxy. The TCP session may be transferred from a source proxy/local server having a first IP address to a target proxy/local server having a second IP address. The transferred session may include TCP sequences and TCP segments that have not been acknowledged. Each server may have a virtual interface with an IP address (IPx). The UE may use the virtual IPx for access to a local proxy/server, and the eNB may route content to local proxy/server's IP address.


In some cases, the eNB may switch from transmitting the requested content to the requesting UEs via unicast to transmitting via broadcast or multicast (e.g., using eMBMS or Single-Cell Point to Multipoint (SC-PTM) transmission). If the eNB determines that a number of UEs requesting the same content exceeds a threshold value, which may be a configurable value, the eNB can switch to eMBMS or SC-PTM transmission for traffic uploading. If the eNB switches to SC-PTM transmission, the eNB may hand over requesting UEs or redirect requesting UEs to a shared physical downlink shared channel (PDSCH).


In some cases, the eNB may transmit data using different radio access technologies (RATs). For example, if the eNB determines that a number of UEs requesting the same content exceeds a threshold value, which may be a configurable value, the eNB can switch data transmission of cached content from one RAT to another RAT.


In some cases, an eNB can enable eMBMS transmission if content consumption is detected in neighboring eNBs. Detecting that the same content is being requested by UEs served by multiple eNBs may be performed, for example, through a protocol across eNBs (e.g., using an X2 interface or an enhanced X2 interface), or through an interface between eNBs and higher layer components (e.g., a mobility management entity (MME) or multi-cell/multicast coordination entity (MCE)).


An eNB may determine that a neighbor eNB has cached content in a variety of ways. In an aspect, the first eNB that retrieves content from a remote server may push the content to neighbor eNBs (e.g., via an X2 interface). In an aspect, the first eNB that retrieves content may inform other eNBs that the first eNB has retrieved content associated with a particular address (e.g., a particular URL). Other eNBs may retrieve the content from the first eNB when a UE served by one of the other eNBs transmits a request for content associated with the same address. In an aspect, the first eNB may announce the particular address associated with requested content to an MME. eNBs within a network may query the MME for content associated with a particular address. If the MME indicates that the first eNB has previously retrieved the content, other eNBs may request the content from the first eNB rather than the remote server. If the MME indicates that the content has not previously been retrieved by an eNB, the eNB may forward a request for the content to a remote server (and cache the content, as described above).


The eNB or MCE (e.g., an “anchor” eNB) may coordinate multicast/broadcast single frequency network operations with other eNBs by synchronizing timestamps across the eNBs. For example, the anchor eNB may set the time stamp based on the furthermost eNB. An eNB may indicate to an MCE that an MBMS session should be initiated. If the MCE is centralized, an indication that an MBMS session should be initiated may be made, for example, using an enhanced M1 interface (e.g., an eNB-MCE interface). If the MCE is co-located with an eNB, an indication that an MBMS session should be initiated may be made, for example, using an enhanced X2 interface (e.g., between eNBs), or by using an MCE-MCE interface. Once the eMBMS channel is set up, the eNB may redirect UEs requesting the content to tune to the eMBMS channel.



FIG. 7 illustrates synchronization among eNBs for multicast/broadcast services, according to an aspect of the present disclosure. As illustrated, an “anchor eNB” may cache content and may transmit MBMS packets (including the cached content) to the UEs (e.g., via a broadcast or multicast channel). The “anchor eNB” may synchronize with other eNBs to create a single frequency network to serve the cached data to a plurality of UEs. “SYNC” may be a protocol to synchronize data used to generate a certain radio frame. The eNBs may be part of an MBSFN.


In some cases, user services description (USD) information may be transmitted to UEs from a network edge device (e.g., an eNB or BM-SC). USD may be transmitted, for example, through dedicated signaling (e.g., via radio resource control (RRC) or network access stratum (NAS) signaling), on an SC-PTM channel, or on an eMBMS channel. In some cases, if USD is transmitted on an eMBMS channel, a separate USD channel may be implemented for communication of locally generated USD information.



FIG. 8 illustrates an example of content caching at an eNodeB, according to aspects of the present disclosure. As illustrated, a first UE requests content from a remote server. In a streaming video example, the first UE may request a media presentation description (MPD) file that describes the requested content and provides information for a client to stream content by requesting media segments from a server. The request is transmitted from a UE through an eNodeB to an eventual endpoint of a server hosting the file.


On receiving the request for the MPD file, the remote server may transmit the MPD file to the UE through the eNodeB, and the eNodeB may cache this file. As illustrated, a second UE may request the same MPD file from the same remote server. Because the MPD file has been cached at the eNodeB, the request for the MPD file may be satisfied by the eNodeB transmitting the cached MPD file to the second UE, thus eliminating the additional time needed to request the MPD file from the remote server.


The first UE may, after reading the MPD file, request segments of a video file from the remote server. The remote server may respond with a requested segment of video, which the eNodeB may store in a cache. The second UE, after reading the MPD file, may also request segments of a video file from the remote server. Since these segments have already been cached at the eNodeB, the eNodeB, rather than the remote server, may provide the requested segments to the second UE.


As illustrated, an eNodeB may detect that many UEs are requesting the same content. In response, the eNodeB may enable broadcast or multicast service (e.g., eMBMS or SC-PTM) for traffic offload, and UEs requesting content may be redirected to the broadcast or multicast service to receive the content.



FIG. 9 illustrates an example of content caching at an eNodeB, according to an aspect of the present disclosure. Similarly to the example illustrated in FIG. 8, the first UE may request content (e.g., an MPD file) from a remote server. The eNodeB may cache the MPD file for provisioning to other UEs that request the same content (e.g., the second UE illustrated in this example). The eNodeB may also act as a client device requesting the content described in the MPD file from the remote server. This content (e.g., segments of a video file) may also be cached at the eNodeB.


As illustrated, both the first and second UEs may request a segment of a video file described by the MPD file. Since the eNodeB has already retrieved the requested segment from the remote server and cached the segment, the eNodeB, rather than the remote server, may provide the requested segment to the first and second UEs. As with the example illustrated in FIG. 8, when an eNodeB detects that many UEs are requesting the same content, the eNodeB may enable broadcast or multicast service and redirect the UEs requesting the content to the broadcast or multicast service to receive the content.



FIG. 10 illustrates an example of content caching at a plurality of eNodeBs, according to an aspect of the present disclosure. As illustrated herein, the first UE communicates with the first eNodeB, and the second UE communicates with the second eNodeB. Similarly to the examples illustrated in FIGS. 8 and 9, a first UE may request content (e.g., an MPD file) from a remote server. As illustrated, a first eNodeB may cache the MPD file for provisioning to other UEs that request the same content (e.g., the second UE illustrated in this example). The first eNodeB may further announce to neighboring eNodeBs (e.g., the second eNodeB illustrated in this example) that the first eNodeB has cached the content and an indication of the associated URL.


As illustrated, the second UE may request the same MPD file from a second eNodeB. Since the first eNodeB has announced that the MPD file has been previously requested and cached at the first eNodeB, the second eNodeB may retrieve the cached MPD file from the first eNodeB, and then transmit the MPD file to the second UE.


The first UE may request a segment of a video file described by the MPD file from the remote server. Similarly to the request of the MPD file, the eNodeB may cache the requested segment and announce to neighboring eNodeBs that the segment has been cached at the first eNodeB, along with the URL of the cached segment. When a second UE requests the segment, via the second eNodeB, the second eNodeB may retrieve the cached segment from the first eNodeB and transmit the segment to the second UE without requesting the segment from the remote server.


Http-Aware Content Caching

Certain aspects of the present disclosure provide for HTTP-aware content caching at network devices. By using the native HTTP caching mechanisms at network components (e.g., eNodeBs and UEs), a mobile content delivery network (CDN) may provide content to requesting UEs and reduce backhaul communications between eNodeBs acting as a mobile CDN.


Content delivery networks may be used to accelerate transmission of large amounts of data, such as video content. While selected IP traffic offloading (SIPTO) allows caching of content close to a proxy gateway (PGW), such caching may expose eNodeBs to the internet, which may be a security risk. Additionally, when UEs move from one eNodeB to another, service may be interrupted (if the UE changes IP addresses), or content may not be offloaded from a backhaul between an anchor eNodeB and a serving eNodeB. By caching content at eNodeBs (e.g., using HTTP-aware techniques), eNodeBs may provide for offloading of internet traffic from the backhaul of an eNodeB.



FIG. 11 illustrates example operations 1100 that may be performed to provide content to a UE from a cache at an eNodeB or a remote source based on information in a request, according to an aspect of the present disclosure.


Operations 1100 begin at 1102, where a base station receives a request for content from a user equipment (UE). At 1104, the base station inspects the request for information about the requested content. At 1106, the base station determines, based on information about the requested content obtained by inspecting the request, if the requested content is on a service list. At 1108, the base station provides the requested content to the UE from a cache if it is determined the requested content is on a service list and is cached at the base station or retrieves the requested content from a remote source, stores the requested content in the cache, and provides the requested content to the UE if it is determined the requested content is on a service list but not cached at the base station.



FIG. 12 illustrates example operations 1200 that may be performed by a user equipment to retrieve content from a cache, according to an aspect of the present disclosure.


Operations 1200 begin at 1202, where a user equipment transmits a request for content to a base station. At 1204, the base station receives the requested content from the base station if the requested content is cached at the base station or receives the requested content from a remote source if the requested content is not cached at the base station.


In an aspect, an eNodeB may inspect requests for information about requested content using deep packet inspection (DPI). An eNodeB may decode hypertext transport protocol (HTTP), transmission control protocol (TCP), and internet protocol (IP) messages above packet data convergence protocol (PDCP) packets to identify the content being requested via a UE. In some cases, an identification of the content may be a uniform resource locator (URL) identifying the address of the content to be retrieved. Because DPI may be computationally expensive, DPI may be performed only for certain packets. In an aspect, a flag may be set in a PDCP header to identify packets that are subject to DPI.



FIG. 13 illustrates an example network architecture for HTTP-aware content caching, in accordance with an aspect of the present disclosure. At a UE, an HTTP client application may be running on an application processor, and a modem may transmit and receive PDCP packets. To allow for a flag to be set in a PDCP header for packets subject to DPI, an HTTP proxy may be implemented inside a UE's modem. The HTTP proxy may intercept an HTTP request generated by a client application and inspect the HTTP request to determine whether or not the request is for content from a cacheable source (e.g., a multimedia file from a particular website). If the HTTP proxy at the UE's modem determines that the request is for content from a cacheable source, the modem can set a flag in the PDCP header indicating that the packet is subject to DPI and encapsulate the HTTP request in a PDCP packet.


At the eNodeB, a modem receives the PDCP packet and determines, based on the DPI flag, whether a packet is subject to DPI or not. If the packet is subject to DPI, the eNodeB can extract the HTTP request from the PDCP packet and use an HTTP proxy at the modem to inspect the HTTP request for information about the requested content. Based on the information about the requested content, the eNodeB can search for the requested content in a cache at the eNodeB and provide the requested content from the cache, or, if the content is not stored in the cache at the eNodeB, retrieve the content from a remote source for transmission to the requesting UE.


In an aspect, the modem at a UE and the eNodeB may be logically considered to be an HTTP proxy. In this case, the air interface communication may be considered internal communications within an HTTP proxy, and thus may not need to strictly follow the HTTP/TCP protocol. Thus, HTTP messages may be delivered over PDCP directly and need not be transmitted using TCP/IP.



FIG. 14 illustrates an example protocol stack for transmitting HTTP messages over PDCP, in accordance with an aspect of the present disclosure. At the UE, HTTP messages may be transmitted using the PDCP stack. Likewise, at the eNodeB, HTTP messages may be received over PDCP, and an eNodeB may access a proxy gateway or serving gateway using the IP stack, and an eNodeB may access remote content sources (e.g., an internet CDN or a content server) via a proxy gateway or serving gateway.


In an aspect, an HTTP message may be routed, at the UE, to the HTTP proxy by configuring client applications to communicate via the HTTP proxy. In some aspects, a transparent proxy mechanism may be used to determine if the HTTP request is to be transmitted via an HTTP proxy server at the UE's modem to an HTTP proxy server at the eNodeB. When an IP packet is received, the modem may check the contents of the IP packet to determine whether the address of the requested content matches a list of addresses that may be accelerated (e.g., cached at a network component). If the address matches an address in a list of addresses that may be accelerated, the HTTP message may be transmitted to the HTTP proxy at the eNodeB. Otherwise, the packet is transmitted to the eNodeB using TCP/IP over PDCP.


For example, the service list may be a list of URLs, uniform resource indicators (URIs) or TCP/IP address information corresponding to remote content sources that may be cached at a network component. In some cases, the list may be an Access Control List (ACL) including filtering parameters for IP packets. The parameters may include, for example, IP addresses and TCP or user datagram protocol (UDP) ports. The list may configured and/or updated by a network operator to include, for example, service providers that are under contract with the network operator. To update a list, an operator may signal the UE via Open Mobile Alliance (OMA) Device Management (DM), system information, Radio Resource Control, or Non-Access Stratum messages.


In some aspects, the HTTP proxy in the UE's modem may enable local caching and pre-caching. UE-accessed content may be cached using the native HTTP caching mechanism. In some cases, the HTTP proxy in the UE's modem and/or the network operator may predict what content a user may be interested in and pre-cache the predicted content at the UE. Predictions may be based, for example, on historical content access information, user configurations, and/or social networking information.


If requested content is not locally cached at a serving eNodeB, the eNodeB retrieves the requested content for the URL from the remote source. In some cases, an eNodeB may directly access a content server or internet CDN to retrieve the requested content; however, direct access may expose the eNodeB to the internet.


In some cases, an eNodeB can send the HTTP request message to a concent server or internet CDN via a proxy gateway or serving gateway. To do so, the eNodeB may transmit a TCP/IP packet with the HTTP request over a GPRS Tunneling Protocol User Plane (GTP-U) tunnel to the proxy gateway or serving gateway. The GTP-U tunnel may be the tunnel used by the requesting UE or a tunnel set up for a virtual UE. If a tunnel is set up for a virtual UE, the virtual UE may be assigned with a high quality of service so that the virtual UE can be used to retrieve content for multiple real UEs. With a virtual server, HTTP requests and responses are not affected by UE mobility. The requesting UE may move to another eNodeB between transmitting the HTTP request and receiving the HTTP response from the remote source (e.g., content server or internet CDN), and the context and pending packets may be forwarded from the original serving eNodeB to the target eNodeB using context transfer and a data forwarding tunnel. If the target eNodeB does not support HTTP caching, communications between the UE and eNodeB may fall back to standard communications between the UE and eNodeB (i.e., without the use of HTTP proxies at the UE and eNodeB).


In some cases, UEs may maintain independent caches and lists of cacheable content (or cacheable content sources). The lists of cacheable content or content sources may be configured by each user according to user preferences. In some cases, the UEs may cooperatively cache content. For example, UEs in a network may be aware of the content cached by each UE. When a UE requests content, and if the requested content is cached by a UE in the network, the requesting UE may retrieve the content from the UE at which the content is cached. The requesting UE may retrieve the content from the UE at which the content is cached, for example, via device-to-device communications, over a Wi-Fi connection, or using a short-range wireless protocol such as Bluetooth.


In aspects, the present disclosure provides methods and apparatus for wireless communications over a network by a base station configured to cache requested content and provide the cached content to requesting UEs. According to aspects, the processor 230, TX data processor 214, and/or other processors and modules at the transmitter system 210 may perform or direct processes for such methods.


It is understood that the specific order or hierarchy of steps in the processes disclosed is an example of exemplary approaches. Based upon design preferences, it is understood that the specific order or hierarchy of steps in the processes may be rearranged while remaining within the scope of the present disclosure. The accompanying method claims present elements of the various steps in a sample order, and are not meant to be limited to the specific order or hierarchy presented.


Those of skill in the art would understand that information and signals may be represented using any of a variety of different technologies and techniques. For example, data, instructions, commands, information, signals, bits, symbols and chips that may be referenced throughout the above description may be represented by voltages, currents, electromagnetic waves, magnetic fields or particles, optical fields or particles, or any combination thereof.


Those of skill would further appreciate that the various illustrative logical blocks, modules, circuits, and algorithm steps described in connection with the embodiments disclosed herein may be implemented as hardware or software/firmware, or combinations thereof. To clearly illustrate this interchangeability of hardware and software/firmware, various illustrative components, blocks, modules, circuits, and steps have been described above generally in terms of their functionality. Whether such functionality is implemented as hardware or software/firmware depends upon the particular application and design constraints imposed on the overall system. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present disclosure.


The various illustrative logical blocks, modules, and circuits described in connection with the embodiments disclosed herein may be implemented or performed with a general purpose processor, a digital signal processor (DSP), an application specific integrated circuit (ASIC), a field programmable gate array (FPGA) or other programmable logic device, discrete gate or transistor logic, discrete hardware components, or any combination thereof designed to perform the functions described herein. A general purpose processor may be a microprocessor, but in the alternative, the processor may be any conventional processor, controller, microcontroller, or state machine. A processor may also be implemented as a combination of computing devices, e.g., a combination of a DSP and a microprocessor, a plurality of microprocessors, one or more microprocessors in conjunction with a DSP core, or any other such configuration.


The steps of a method or algorithm described in connection with the embodiments disclosed herein may be embodied directly in hardware, in a software/firmware module executed by a processor, or in a combination of the two. A software/firmware module may reside in RAM memory, flash memory, ROM memory, EPROM memory, EEPROM memory, phase change memory (PCM), registers, hard disk, a removable disk, a CD-ROM, or any other form of storage medium known in the art. An exemplary storage medium is coupled to the processor such the processor can read information from, and write information to, the storage medium. In the alternative, the storage medium may be integral to the processor. The processor and the storage medium may reside in an ASIC. The ASIC may reside in a user terminal. In the alternative, the processor and the storage medium may reside as discrete components in a user terminal. As used herein, a phrase referring to “at least one of” a list of items refers to any combination of those items, including single members. As an example, “at least one of: a, b, or c” is intended to cover a, b, c, a-b, a-c, b-c, and a-b-c., as well as any combination with multiples of the same element (e.g., a-a, a-a-a, a-a-b, a-a-c, a-b-b, a-c-c, b-b, b-b-b, b-b-c, c-c, and c-c-c or any other ordering of a, b, and c).


The previous description of the disclosed embodiments is provided to enable any person skilled in the art to make or use the present disclosure. Various modifications to these embodiments will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other embodiments without departing from the spirit or scope of the disclosure. Thus, the present disclosure is not intended to be limited to the embodiments shown herein but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.

Claims
  • 1. A method for wireless communications by a base station, comprising: receiving a request for content from a user equipment (UE);inspecting the request for information about the requested content;determining, based on information about the requested content obtained by inspecting the request, if the requested content is on a service list; andproviding the requested content from a cache to the UE if it is determined the requested content is on the service list and stored in the cache or retrieving the requested content from a remote source, storing the requested content in the cache, and providing the requested content to the UE if it is determined the requested content is on the service list but not cached at the base station.
  • 2. The method of claim 1, wherein the request comprises a hypertext transport protocol (HTTP) packet.
  • 3. The method of claim 2, wherein the inspecting the request for information about the requested content comprises using deep packet inspection (DPI) to obtain the information about the requested content.
  • 4. The method of claim 3, wherein the request comprises an flag indicating whether or not the request is subject to DPI, and wherein the inspecting the request occurs if the flag indicates that the request is subject to DPI.
  • 5. The method of claim 2, wherein the request is received via an HTTP proxy server at the base station.
  • 6. The method of claim 2, wherein the HTTP packet is received in a packet data convergence protocol (PDCP) packet.
  • 7. The method of claim 1, wherein the information comprises a uniform resource identifier (URI) identifying the requested content.
  • 8. The method of claim 1, wherein the service list comprises an access control list, wherein the access control list comprises a list of transmission control protocol/internet protocol (TCP/IP) addresses corresponding to remote sources hosting cacheable content.
  • 9. The method of claim 8, wherein the information comprises a TCP/IP address of a remote source hosting the requested content.
  • 10. The method of claim 1, wherein the service list is maintained on a per-UE basis.
  • 11. The method of claim 1, wherein retrieving the requested content from the remote source comprises retrieving the requested content via a GPRS Tunneling Protocol User Plane (GTP-U) tunnel used by the UE.
  • 12. The method of claim 1, further comprising: initializing a virtual UE, wherein the virtual UE is assigned a GTP-U tunnel; andwherein the retrieving the requested content from a remote source comprises retrieving the requested content via the GTP-U tunnel assigned to the virtual UE.
  • 13. The method of claim 1, further comprising: detecting that the UE has moved to a second base station; andforwarding one or more pending packets containing requested content from the base station to the second base station.
  • 14. A method for wireless communications by a user equipment (UE), comprising: transmitting a request for content to a base station; andreceiving the requested content from the base station if the requested content is cached at the base station, or receiving the content from a remote source if the requested content is not cached at the base station.
  • 15. The method of claim 14, wherein the request comprises a hypertext transport protocol (HTTP) packet.
  • 16. The method of claim 15, wherein the HTTP packet is transmitted in a packet data convergence protocol (PDCP) packet.
  • 17. The method of claim 15, wherein the request comprises an indication that the base station is to perform deep packet inspection (DPI) to obtain information about the requested content.
  • 18. The method of claim 15, wherein the request is transmitted via an HTTP proxy server at the UE to an HTTP proxy server at the base station.
  • 19. The method of claim 18, wherein the transmitting a request for content to a base station comprises: prior to transmitting the request for content to the base station, determining if a remote source associated with the requested content matches one or more sources on a list of sources for caching; andif a match is found, transmitting the request via the HTTP proxy server at the UE to the HTTP proxy server at the base station.
  • 20. The method of claim 14, further comprising: storing content in a local cache at the UE.
  • 21. The method of claim 20, further comprising: prior to transmitting the request for content to the base station, determining if the requested content is stored in the local cache; andif the requested content is stored in the local cache, providing the requested content from the local cache.
  • 22. The method of claim 20, wherein the storing content in the local cache comprises predictively storing content in a local cache at the UE.
  • 23. The method of claim 22, wherein predictively storing content comprises storing content based on at least historical content access.
  • 24. The method of claim 22, wherein predictively storing content comprises storing content based on at least a user configuration.
  • 25. The method of claim 22, wherein predictively storing content comprises storing content based on at least information about social groups a user belongs to.
  • 26. The method of claim 20, wherein the UE and one or more second UEs cooperatively cache content from one or more remote sources.
  • 27. The method of claim 26, wherein cooperatively caching content comprises transmitting, to one or more second UEs, information about the content stored in the local cache.
  • 28. The method of claim 27, further comprising: receiving, from one of the second UEs, a request for content stored in the local cache; andtransmitting the requested content to the one of the second UEs.
  • 29. The method of claim 28, wherein the transmitting the requested content to the one of the second UEs is performed using a device-to-device protocol.
  • 30. The method of claim 28, wherein the transmitting the requested content to the one of the second UEs is performed using a Wi-Fi connection.
  • 31. The method of claim 28, wherein the transmitting the requested content to the one of the second UEs is performed using a Bluetooth connection between the UE and the one of the second UEs.
PCT Information
Filing Document Filing Date Country Kind
PCT/CN2015/086657 8/11/2015 WO 00