The present invention is generally directed to reassembly free scanning of files in a peer to peer network. More specifically, the present invention scans file data without reassembling a file even when parts of the file are received out of order.
Data communicated over computer networks today pass through various layers in a computer system architecture. Typically data is received at a network interface of a computer at a link layer. The link layer is a layer in the architecture of a computer that includes physical hardware. The link layer connects the computer to other computers in a computer network. Link layers also are used to transmit data from one computer to another over a computer network.
Other layers above the link layer in computer system architectures commonly include a network layer, a transport layer, and an application layer. The network layer receives data packets from and provides data packets to the link layer. The network layer may also receive data in segments from the transport layer and send data in segments to the transport layer. Commonly when the network layer receives a segment of data from the transport layer it will generate a packet or an internet protocol (IP) datagram for transmission to another computer. This process may include encapsulating the segmented data received from the transport layer and adding a header that includes a destination IP address when generating an IP packet. In certain instances more than one IP packet may be associated with a data segment. The network layer may also receive IP packets from the link layer and may pass segmented data to the transport layer.
When a series of IP packets are used to transport data to a computer, those packets may be received out of order at the network layer. When this occurs, the transport layer may re-order the data segments from a plurality of packets before sending the re-ordered data to the application layer. Conventionally data received at an application layer must be received in-order (i.e. sequentially). For example, in a client-server environment file data received at the application layer of a client or a server must be in-order before it can be processed. This is because the client-server environment expects received data to be in order. While communication transferred over a computer network according to the Transmission Control Protocol (TCP) will re-order packets, communications over other transport layer protocols, such as the User Datagram Protocol (UDP) do not.
Typically in a client-server environment a server will send a data set or a file sequentially from the application layer to the transport layer, the transport layer may then send that data to the network layer. The network layer then packetizes the data and sends a plurality of packets to a client. Even though the packetized data may be sent out of order, data contained in the packets will be re-ordered before that data is received at the application layer at the client. Because of this, application layers at a client or a server in a client-server environment may never receive file data that is out of order.
Peer to peer (P2P) networks, however, do not operate in the same way as a client-server environment. For example P2P networks may receive data at an application layer that is out of order. This is because P2P networks fundamentally have a different type of architecture as compared to a client-server environment.
In a P2P network a computer accessing file data may receive parts of data from a file from a plurality of computers. A P2P network is capable of transmitting file data in pieces where each piece of data may be transmitted from a different computer. Because of this a first piece of data received from a first computer may be out of order as compared to a second piece of data received from a second computer. When this occurs the network layer and the transport layer at a receiving computer will not be aware that the first data piece and the second data piece have been received out of order. This is true even when packetized data sent from the first computer (or the second computer) to the receiving computer have been re-ordered. This is because the network layer and the transport layer at the receiving computer do not check whether application data received from different peer computers are received in order. Conventionally, the network layer and the transport layer are only capable of re-ordering packetized data that has been transmitted from a single source computer to a destination computer.
P2P networks may also break a file into a number of pieces where each piece may include a pre-determined or specific number of blocks. Information relating to a number of pieces that a data file is broken into may be included in metadata (or a metadata file) that is associated with the data file. Once a number of pieces are identified, a file size divided by the number of pieces will correspond to a number of blocks that the file may be broken into in the P2P network.
Limitations included in the network layers and in the transport layers of computers today mean that file data received at an application layer of a computer cannot easily scan the received data for malicious content (such as computer worms, viruses, or other attacking software). Conventionally the scanning of data for malicious content at the application layer in a P2P network either cannot be done reliably or must be done in an inefficient manner. For example, if data from a file is scanned out of order, the scan can miss a virus contained within the data, because malicious content are characterized by a sequential ordered series of characters, not an out of order series of characters. In another example, when the application layer re-orders received data before scanning it, data from the out of order pieces must be stored until the data pieces can be re-ordered and scanned. Thus, the first example is unreliable and the second example is inefficient.
Application data that includes interleaved out of order data received at an application layer of a computer system if scanned in the order received may result in missing malicious content contained within the received interleaved data. Furthermore, data received out of order may also result in scanning software falsely detecting malicious content. For example, when the character sequence of “car” is associated a virus and two pieces of data that were received out of order where a later piece of data ending with the character “c” is scanned before an earlier piece of data that begins with “ar,” malicious scanning software will falsely identify that these pieces of data include the virus, when they do not.
What is needed to increase the reliability and efficiency of P2P networks are systems and methods that scan pieces of data received out of order at an application layer without storing and re-ordering data pieces that have been received out of order. What is also needed are systems and methods that scan interleaved data reliably at an application layer. The reliable scanning of received data at an application level increases the reliability of detecting malicious content while reducing the likelihood that malicious content scanning software will falsely associate received data with malicious content.
The presently claimed invention relates to an apparatus, a method, and a non-transitory computer readable storage medium for deep packet inspection scanning at an application layer of a computer. A method of the presently claimed invention may receive a portion of data at an application layer of a computer system after which the received portion of data may be scanned from a first input state at the application layer. When the first input state includes a portion of information known to be included in malicious content, the method then identifies that the data set can include malicious content after the received portion of data has been scanned from the first input state.
The presently claimed invention may also be implemented as a non-transitory computer readable storage medium where a processor executing instructions out of a memory receives a portion of data at an application layer of a computer system after which the received portion of data may be scanned from a first input state at the application layer. When the first input state includes a portion of information known to be included in malicious content, the processor executing instructions out of the memory then identifies that the data set can include malicious content after the received portion of data has been scanned from the first input state.
An apparatus of the presently claimed invention may include a network interface that receives information, a memory, and a processor. The processor executing instructions out of the memory then receives a portion of a data set at an application layer. The received portion of data is then scanned at the application layer from a first input state. When the first input state includes a portion of information known to be included in a piece of malicious content, the processor executing instructions out of the memory then identifies that the data set can include malicious content after the received portion of data has been scanned from the first input state.
The present disclosure relates to an apparatus, a method, and a non-transitory computer readable storage medium for deep packet inspection scanning at an application layer of a computer. A method of the presently claimed invention may scan data received at an application layer from a first input state. The first input state including information, such as, one or more characters that can be associated with malicious content. The method may then identify that the data set may include malicious content when the first input state combined with the scan of the received data matches a known piece of malicious content.
The peer to peer (P2P) network illustrated 150 in
For example, when malicious content is identified as receiving the character “c” followed by character “a” that is, in turn, followed by character “r,” then the identified input states associated with malicious content include state “c” and the state “ca.” In the instance where a preceding data block ends with characters “ca” and a following data block begins with the character “r,” then malicious code “car” is present in these data blocks. Similarly in the instance where the preceding data block ends with the character “c” and the following data block begins with characters “ar,” then the malicious code “car” is also present in these data blocks. Malicious content “car” may correspond to a rule that identifies “car” as being malicious content.
In the instance where characters “car” are associated with malicious code by a rule when data block B 2-2 is received and when data block B 2-1 has not yet been received (as in
In an apparatus consistent with the presently disclosed invention, malicious code can be associated with one or more different sequences of characters. For example, the character sequence of “Apple” may also be associated with malicious content via a rule (R2) that identifies “Apple” as being a virus. In the instance when “Apple” is associated with malicious content, identified input states (sub-states) may include an empty string “ ”, “A,” “Ap,” “App,” and “Appl.”.
After data block B 2-2 is scanned using each of the identified input states in
Finally after data block B 1-1 is received, data block B 1-1 may be scanned from an initial state (such as a null state) and output states 1-1 may be output. The dotted line 320 indicates that output states 1-1 may be compared with the identified input states when identifying whether the combined data blocks B 1-2 and B 2-1 include malicious content. Note that this process scans data blocks received out of order for malicious content without reassembling the data blocks. Instead a series of identified input states may be used when scanning an out of order packet for malicious content. The presently disclosed invention, thus, identifies malicious content by comparing output states with identified input states that may be included in a data block that has not yet been received. Later when the out of order data block is received, the out of order data block may be scanned generating one or more output states. When an output state of the out of order packet includes an identified input state of a subsequently ordered data block, the two different data blocks may include malicious content.
In certain instances one or more output states associated with different pieces of a data set may be stored in memory where each of these output states may be associated with a possible identified input state associated with yet other pieces of the data set. When one or more output states and one or more possible identified input states are stored in memory and an outstanding piece of the data set is received, the outstanding piece of the data set may be scanned generating an output state associated with the outstanding piece of the data set. In such an instance, each of the output states and possible identified input states may be assembled in a chain when identifying that the data set includes malicious content.
For example, when data blocks are received in the order illustrated in
Next data block B 1-2 is scanned using the identified input states generating output states 1-2. When rules that identify malicious content as being “car” and “Apple” the character sequences an empty string (i.e. an initial state), “c,” “ca”, “A,” Ap,” “App,” and “Appl” each are identified input states that are associated with malicious content. Since data block B 1-2 consists of “aaaa,” the only output state that corresponds to an identified input state that may be associated with malicious content is the empty string (i.e. an initial state). This is because the character sequences of “a,” “aa,” “aaa,” and “aaaa” are not associated with malicious content according to rules that identify “Apple” and “car” as being malicious content. This means that data block B 2-1 may be scanned from just the empty string. Thus, in this example, the process of reduction identifies that the only identified input state of all of the identified input states that data block B 2-1 should be scanned from is the empty string.
When data block B 2-1 is received, it is scanned from only the empty string. After data block B 2-1 is scanned, output states 2-1 will be generated. Since data block B 2-1 consists of “araA,” the only output state that corresponds to an identified input state is the identified output state of “A.” Since data block B 2-2 has already been scanned and identified as including “pple,” malicious content of “Apple” will be detected in the data set when preceding data block B 2-1 ends with the character “A.” Since, in this example, the identified input state of “A” precedes data block B 2-2, the malicious content of “Apple” is detected in the data set. Once malicious content has been detected in the data set, the receipt of additional data blocks, such as data block B 1-1, may be blocked. Note also that a chain of only one possible identified input state of “A” of data block B 2-2 and the output state “pple” of output states 2-2 are used to identify malicious content in this example. Note also that only a reduced number of input states coupled with a number of output states 2-2 requires limited storage as compared to storing the data blocks received.
The present disclosure is not limited to malicious content spanning one or two data blocks, as methods consistent with the present disclosure may detect malicious content that spans any number of data blocks in a data set, including all of the data blocks.
Similarly state flow may move from state S0 to state S6 when the character “c” is in the data set, then moves to state S7 when a subsequent character “a” is encountered in sequence in the data set, and then moves from state S7 to state S8 when a subsequent character “r” is in the data set. Note that state S8 is identified as item 420 in
Dashed lines in
The components shown in
Storage device 530, which may include mass storage implemented with a magnetic disk drive or an optical disk drive, may be a non-volatile storage device for storing data and instructions for use by processor unit 510. Storage device 530 can store the system software for implementing embodiments of the present invention for purposes of loading that software into main memory 510.
Portable storage device of storage 530 operates in conjunction with a portable non-volatile storage medium, such as a floppy disk, compact disk or Digital video disc, to input and output data and code to and from the computer system 500 of
Antenna 540 may include one or more antennas for communicating wirelessly with another device. Antenna 540 may be used, for example, to communicate wirelessly via Wi-Fi, Bluetooth, with a cellular network, or with other wireless protocols and systems. The one or more antennas may be controlled by a processor 510, which may include a controller, to transmit and receive wireless signals. For example, processor 510 execute programs stored in memory 520 to control antenna 540 transmit a wireless signal to a cellular network and receive a wireless signal from a cellular network.
The system 500 as shown in
Display system 570 may include a liquid crystal display (LCD), LED display, or other suitable display device. Display system 570 receives textual and graphical information, and processes the information for output to the display device.
Peripherals 580 may include any type of computer support device to add additional functionality to the computer system. For example, peripheral device(s) 580 may include a modem or a router.
The components contained in the computer system 500 of
Actions taken when the content included in a data set or file received at an application level at a peer device may vary and may depend on one or more actions identified by a user of the peer computer. In certain instances, user preferred actions may be selected in a user interface displayed on a display at the peer computer. In other instances actions taken after detecting malicious content may be according to a set of pre-defined or default actions set in an application program. Actions that may be taken after detecting malicious content in a data set include, yet are not limited to marking the data set or file as including malicious content, blocking reception of data associated with the data set or file, and resetting one or more TCP connections associated with the data set or file.
When the data set or file is marked as including malicious content, that marking may be stored in a table or database at the peer computer that received and detected the malicious data. The data set or file may be identified by a name (i.e. by a filename) or may be identified using a Hash function or checksum of information that identifies the data set. Once a data file has been identified subsequent attempts to download the file may be blocked.
Hash functions identifying a data set or file may be generated from metadata downloaded from a peer when downloading a portion of a data set. In certain instances the downloaded metadata may include a peer identifier, an internet protocol (IP) address, a domain name, or a port number.
As soon as a file is identified as being associated with malicious content, one or more communication sessions associated with the file may be reset. This may include resetting communications sessions between a plurality of peer computers that are providing parts of the file.
The presently disclosure is not limited to files received over a peer to peer network as file data received in an interleaved (out of order) sequence at the application level may also be scanned according to the present disclosure. For example, interleaved data received using the server message block (SMB) 2.0 standard may be scanned in order without reassembly at the application layer of a computer for malicious content.
Embodiments of the present disclosure may be implemented by a non-transitory computer readable storage medium by a processor executing instructions out of a memory, by a DPI scanner implemented in a field programmable gate array (FPGA).
The presently disclosed invention may be implemented in software (i.e. as a non-transitory computer readable storage medium executable by a processor), may be implemented in whole or in part in a field programmable gate array, may be implemented in whole or in part in a hardware state machine, or may be implemented in a combination of hardware and software.
The various methods may be performed by software operating in conjunction with hardware. For example, instructions executed by a processor, the instructions otherwise stored in a non-transitory computer readable medium such as memory. Various interfaces may be implemented—both communications and interface. One skilled in the art will appreciate the various requisite components of a mobile device and integration of the same with one or more of the foregoing figures and/or descriptions.
The foregoing detailed description of the technology has been presented for purposes of illustration and description. It is not intended to be exhaustive or to limit the technology to the precise form disclosed. Many modifications and variations are possible in light of the above teaching. The described embodiments were chosen in order to best explain the principles of the technology, its practical application, and to enable others skilled in the art to utilize the technology in various embodiments and with various modifications as are suited to the particular use contemplated. It is intended that the scope of the technology be defined by the claim.
This application is a continuation and claims the priority benefit of U.S. patent application Ser. No. 17/174,182 filed Feb. 11, 2021, now U.S. Pat. No. 11,695,784, which is a continuation and claims the priority benefit of U.S. patent application Ser. No. 16/853,360 filed Apr. 20, 2020, now U.S. Pat. No. 11,005,858, which is a continuation and claims the priority benefit of U.S. patent application Ser. No. 15/860,623 filed Jan. 2, 2018, now U.S. Pat. No. 10,630,697, which is a continuation and claims the priority benefit of U.S. patent application Ser. No. 14/965,866 filed Dec. 10, 2015, now U.S. Pat. No. 9,860,259, the disclosures of which are incorporated herein by reference.
Number | Name | Date | Kind |
---|---|---|---|
6961783 | Cook et al. | Nov 2005 | B1 |
7206765 | Gilliam et al. | Apr 2007 | B2 |
7849502 | Bloch et al. | Dec 2010 | B1 |
7849507 | Bloch et al. | Dec 2010 | B1 |
7984149 | Grayson | Jul 2011 | B1 |
8316446 | Campbell et al. | Nov 2012 | B1 |
8320372 | Menten | Nov 2012 | B2 |
8352998 | Kougiouris et al. | Jan 2013 | B1 |
8578489 | Dubrovsky et al. | Nov 2013 | B1 |
8782771 | Chen et al. | Jul 2014 | B2 |
8813221 | Dubrovsky et al. | Aug 2014 | B1 |
8826443 | Raman et al. | Sep 2014 | B1 |
8850567 | Hsieh et al. | Sep 2014 | B1 |
8856869 | Brinskelle | Oct 2014 | B1 |
9031937 | Guha | May 2015 | B2 |
9350750 | Aval et al. | May 2016 | B1 |
9438699 | Shetty et al. | Sep 2016 | B1 |
9557889 | Raleigh et al. | Jan 2017 | B2 |
9723027 | Vazquez Carames | Aug 2017 | B2 |
9860259 | Ling et al. | Jan 2018 | B2 |
9923870 | Dusi et al. | Mar 2018 | B2 |
10277610 | Dubrovsky et al. | Apr 2019 | B2 |
10491566 | Vazquez Carames | Nov 2019 | B2 |
10630697 | Ling et al. | Apr 2020 | B2 |
11695784 | Ling et al. | Jul 2023 | B2 |
20040179477 | Lincoln et al. | Sep 2004 | A1 |
20050108518 | Pandya | May 2005 | A1 |
20060233101 | Luft et al. | Oct 2006 | A1 |
20070041402 | Sekaran et al. | Feb 2007 | A1 |
20070064702 | Bates et al. | Mar 2007 | A1 |
20070192861 | Varghese | Aug 2007 | A1 |
20070226362 | Johnson | Sep 2007 | A1 |
20080127349 | Ormazabal et al. | May 2008 | A1 |
20080219169 | Sargor | Sep 2008 | A1 |
20080235755 | Blaisdell et al. | Sep 2008 | A1 |
20080262991 | Kapoor et al. | Oct 2008 | A1 |
20080320582 | Chen et al. | Dec 2008 | A1 |
20090164560 | Fiatal | Jun 2009 | A1 |
20090260087 | Ishida et al. | Oct 2009 | A1 |
20090316698 | Menten | Dec 2009 | A1 |
20100005118 | Sezer | Jan 2010 | A1 |
20100172257 | Yu | Jul 2010 | A1 |
20110013527 | Varadarajan | Jan 2011 | A1 |
20110035469 | Smith et al. | Feb 2011 | A1 |
20110125748 | Wood et al. | May 2011 | A1 |
20110153802 | Steiner | Jun 2011 | A1 |
20110211586 | Zhu | Sep 2011 | A1 |
20110219426 | Kim et al. | Sep 2011 | A1 |
20110231924 | Devdhar et al. | Sep 2011 | A1 |
20120144061 | Song | Jun 2012 | A1 |
20120230200 | Wentink | Sep 2012 | A1 |
20120240185 | Kapoor et al. | Sep 2012 | A1 |
20120291087 | Agrawal | Nov 2012 | A1 |
20120324099 | Perez Martinez et al. | Dec 2012 | A1 |
20130074177 | Varadhan et al. | Mar 2013 | A1 |
20130128742 | Yu | May 2013 | A1 |
20130167192 | Hickman et al. | Jun 2013 | A1 |
20130286860 | Dorenbosch et al. | Oct 2013 | A1 |
20130291107 | Marck et al. | Oct 2013 | A1 |
20140053239 | Narayanswamy | Feb 2014 | A1 |
20140157405 | Joll | Jun 2014 | A1 |
20140181972 | Karta et al. | Jun 2014 | A1 |
20140258456 | Lee et al. | Sep 2014 | A1 |
20140304766 | Livne | Oct 2014 | A1 |
20140359764 | Dubrovsky | Dec 2014 | A1 |
20150058488 | Backholm | Feb 2015 | A1 |
20150058916 | Rostami-Hesarsorkh et al. | Feb 2015 | A1 |
20150088897 | Sherman et al. | Mar 2015 | A1 |
20150278798 | Lerch et al. | Oct 2015 | A1 |
20150312220 | Crawford | Oct 2015 | A1 |
20150326613 | Devarajan et al. | Nov 2015 | A1 |
20150373167 | Murashov et al. | Dec 2015 | A1 |
20160036833 | Ardeli | Feb 2016 | A1 |
20160056927 | Liu et al. | Feb 2016 | A1 |
20160057185 | Zhang | Feb 2016 | A1 |
20160119198 | Kfir et al. | Apr 2016 | A1 |
20160127305 | Droms et al. | May 2016 | A1 |
20160164825 | Riedel et al. | Jun 2016 | A1 |
20160182537 | Tatourian et al. | Jun 2016 | A1 |
20160205072 | Dusi et al. | Jul 2016 | A1 |
20170048260 | Peddemors et al. | Feb 2017 | A1 |
20170099310 | Di Pietro et al. | Apr 2017 | A1 |
20170134428 | Vazquez Carames | May 2017 | A1 |
20170171222 | Ling et al. | Jun 2017 | A1 |
20170302628 | Vazquez Carames | Oct 2017 | A1 |
20180198804 | Ling et al. | Jul 2018 | A1 |
20200213278 | Vazquez Carames | Jul 2020 | A1 |
20200351280 | Ling et al. | Nov 2020 | A1 |
20210234873 | Ling et al. | Jul 2021 | A1 |
Number | Date | Country |
---|---|---|
2901391 | Aug 2018 | EP |
2006052714 | May 2006 | WO |
Entry |
---|
U.S. Appl. No. 14/965,866 Office Action May 18, 2017. |
U.S. Appl. No. 15/860,623 Office Action Jun. 13, 2019. |
U.S. Appl. No. 15/860,623 Final Office Action Nov. 27, 2018. |
U.S. Appl. No. 15/860,623 Office Action May 3, 2018. |
U.S. Appl. No. 17/174,182 Office Action Aug. 15, 2022. |
U.S. Appl. No. 15/636,148 Office Action Feb. 19, 2019. |
U.S. Appl. No. 15/636,148 Final Office Action Jul. 26, 2018. |
U.S. Appl. No. 15/636,148 Office Action Feb. 22, 2018. |
U.S. Appl. No. 16/697,082 Final Office Action Jan. 19, 2021. |
U.S. Appl. No. 16/697,082 Office Action Oct. 6, 2020. |
Number | Date | Country | |
---|---|---|---|
20230336570 A1 | Oct 2023 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 17174182 | Feb 2021 | US |
Child | 18215669 | US | |
Parent | 16853360 | Apr 2020 | US |
Child | 17174182 | US | |
Parent | 15860623 | Jan 2018 | US |
Child | 16853360 | US | |
Parent | 14965866 | Dec 2015 | US |
Child | 15860623 | US |