The present invention generally relates to cryptology, and more particularly, to utilizing primitives for secure hash functions and/or stream ciphers.
As digital communication becomes more commonplace, the need for securing the communication channels becomes increasingly more important. For example, current technologies allow a user to remotely access bank accounts, medical information, and other private and sensitive data.
Cryptology has been widely used to provide secure digital communication. Cryptology relates to the enciphering (or encrypting) and deciphering (decrypting) of messages. The encryption and decryption generally uses some secret information (such as a key) for processing the data to be secured. In different encryption methods, a single key or multiple keys may be used for encryption and/or decryption.
Currently, two types of symmetric cipher are in common use. Generally, the encryption is said to be symmetric-key if for each associated encryption/decryption key pair, it is computationally “easy” to determine each of the keys in the key pair from the other key in the key pair. The first type of symmetric cipher is a block cipher which operates on a large block of data. The second type is a stream cipher which, by contrast, operates on relatively smaller units of text (such as bits). Depending on its implementation, stream cipher is believed to be much faster than block cipher and, hence, of special interest.
Additionally, stream ciphers can generate streams (also known as a “keystreams”) while approaching the high security of one-time pad. Generally, one-time pad cipher generates a keystream with the same length as the text message being encrypted. The one-time pad keystream is believed to be completely random, yielding very high security levels.
One of the most widely used stream cipher techniques is the alleged RC4 (Ron's Code 4—variable-key-size encryption algorithm by Ron Rivest of RSA (Rivest, Shamir, and Adleman public key encryption technology)). For example, the alleged RC4 is used in SSL (secure socket layer) which is implemented in many Internet-based communications (e.g., through HTTP (hyper-text transfer protocol) and FTP (file transfer protocol)).
The alleged RC4, however, can require eight to sixteen machine cycles to output a single byte. While this can be run in software, it requires a substantial amount of hardware resources. Also, implementing the alleged RC4 in hardware limits the flexibility of running the alleged RC4 on different computing devices (i.e., because it would require the presence of alleged RC4-specific hardware in each device).
Another method used for securing data is a hash function. A hash function transforms an input string into a fixed-size output string (also known as a “hash value”). The size of the output string is referred to as a message “digest.” It is generally desirable to provide a hash function that is one-way (i.e., hard to invert) and collision-free (i.e., different hash values are generated for different messages). One standard hash function is SHA-1 (secure hash algorithm as revised in 1994) which generates a 160-bit digest from an input stream of less than 264 bits.
Cryptographic primitives such as stream ciphers and secure hash functions are an essential part of many cryptosystems, making their efficient computation an important problem. Most practical implementations, however, use custom designs (message digests (e.g., MD5 and MD6), SHA-1 or alleged RC4, Seal, and Scream) that depend on empirical analysis, as opposed to formal complexity theoretic or information theoretic analysis.
Accordingly, the current solutions fail to provide an efficient methodology for quick and secure data encryption/decryption, while maintaining sufficient efficiency and/or speed for implementation in software.
Techniques are disclosed to enable efficient implementation of secure hash functions and/or stream ciphers. More specifically, a family of graphs is described that has relatively large girth, large daw, and/or rapid mixing properties. The graphs are suitable for construction of cryptographic primitives such as collision resistant hash functions and stream ciphers, which allow efficient software implementation.
The detailed description is described with reference to the accompanying figures. In the figures, the left-most digit(s) of a reference number identifies the figure in which the reference number first appears. The use of the same reference numbers in different figures indicates similar or identical items.
The following discussion assumes that the reader is familiar with cryptography techniques. For a basic introduction of cryptography, the reader is directed to a text written by A. Menezes, R van Oorschot, and S. Vanstone entitled, “Handbook of Applied Cryptography,” fifth printing (August 2001), published by CRC Press.
Overview
The following disclosure describes a class of graphs that may be employed in software to implement secure hash functions and/or stream ciphers. More specifically, a collection of matrices are described that can be used in the construction of graphs suitable for provision of cryptographic primitives. The speed of the methodologies described here may be further improved by utilizing hardware such as an application specific integrated circuit (ASIC). Also, each stage discussed with reference to methodologies herein may be performed by software, hardware, firmware, or combinations thereof.
Secure Hash Function
Accordingly, instead of a simplified approach that may add the hash values of each input block yielding an implementation prone to security risks (e.g., because changing the order of input blocks will still result in the same final hash value), the method 100 ensures a more secure implementation by generating a final hash value recursively based on previously calculated hash values. Other implementations will be further discussed herein that utilize graphical and matrix-based methodologies to provide a final hash value.
Graph-Based Hash Compression Function
A hash function may be constructed by starting at a node z0 and tracing a path which is described by the sequence x1, . . . ,xd (208). The final point of the trace indicates the value of the compression function (210).
In an implementation, the compression function determination requires two properties:
Various types of intersecting paths may be considered for the graph girth. For example, a directed parallel cycle may include two directed paths starting at a same node and ending at a same node. A directed claw may be considered when the start nodes are different with a same end node. And, a directed cycle would start and end at a same node.
Property (1) is important so that one does not find collisions on the first block. If a randomizing function (such as a block cipher done by the data encryption standard (DES), MD5, and the like) is applied at the end of the first block, then on two different inputs property (2) avoids a collision in the next block and so on. Specific groups of matrices will be introduced under the section entitled “matrix graphs” below that will possess these properties. Moreover, the matrices are carefully chosen to contain small entries and admit efficient implementation (e.g., when implemented in software).
Matrix Graphs and Block Functions
In one implementation, the block function is based on a walk on a Cayley graph (referred to herein as a “matrix graph”) defined by the matrices. Let some m=2l (e.g., m=232), where m is the number of nodes in the matrix graph. Let A={σ1, . . . ,σk} be a set of generators (such as those shown in
Given such a graph (G), the block function B may be defined as follows. Let Σ be an alphabet of size k (the number of generators of G). Then B will map V×Σn→V, so that given an initial location M and input X=x1, . . . ,xt with each xi a character over Σ, we have that B(M,X) is a matrix in V. By taking any correspondence between Σ and A, each xi may be viewed as a matrix BiεA. We can now define:
As discussed above with respect to properties (1) and (2), the desired characteristics of the block function may be summarized as:
Let g be a random one-way function mapping some domain D to D, with a bijection φ:D→V. Let I be the identity matrix in V. Let the input be X1, . . . ,Xb with each XiεΣt as above. We assume that the input is an integral number of blocks; padding issues for inputs of different sizes can be dealt with using standard methods. Define Y0=I and for 1≦i≦b1, Yi=φ(g(φ−1(B(Yi-1,Xi)))). We may define the hash function as H(X1, . . . ,Xb)=Yb.
In one implementation, if such a bijection φ is difficult to find, a relaxed bijection ψ can be used instead. More specifically, let f(Xi) be a secure hash evaluated at Xi that outputs w bits, which are interpreted by any convenient technique as numbers a1, a2, and a3 and defining Yiψ (Xi) by:
Accordingly, it can be seen that given two inputs X and X′, if they collide on an interblock step, the girth arguments show that if they differ on the following block, the outputs at the end of that block will differ also.
If d is chosen sufficiently small, property (2) can be shown for all graphs. In particular, let n be the number of nodes in a graph, and let d=(log n)/(2+ε). Let B(u) be the number of nodes within distance d of a node u; then |B(u)| ε o(√n). Now if a node v is chosen at random, we have as well |B(v)| ε o(√n). Accordingly, the probability that u and v form a claw (e.g., reaching a same node) is bounded by |B(u)∩B(v)|/∩n ε o(1). Thus, if d is logarithmic in the number of nodes in the graph, the probability of picking a claw is negligible.
Hash Implementation
In a stage 404, the initial matrix M is set to be the identity matrix (such as discussed with reference to the matrix I in the vertex set V). A stage 406 processes the data input blocks (e.g., as 9-bit blocks in the example discussed with reference to
Accordingly, instead of a simplified approach that may add the hash values of each input block yielding an implementation prone to security risks (e.g., because changing the order of input blocks will still result in the same final hash value), the method 400 ensures a more secure implementation by multiplying the hash values of each input block by a set of generator matrices (such as those discussed with reference to
In one implementation, using a table of 256 entries based on a set of four generator matrices is not as efficient as the example discussed with reference to
Stream Cipher Amplification
The matrix graphs discussed herein also have expansion properties that make them suitable for constructing stream ciphers. For example, the alleged RC4 can be thought of as a random walk on a graph with a set of permutations on {1, . . . ,256}. The matrix graphs have the additional property that their directed girth is relatively large. So, a random walk can not return to a given node in a relatively short amount of time. An immediate way to use this graph is to stretch the outputs of the alleged RC4, by using the alleged RC4's output as a way to perform a random walk on the graph. For example, after being initialized with a key k which sets the internal state, the stream cipher outputs a sequence of bytes, x1,x2, . . . . Each xi output is a function of the current internal state of the cipher, which is changed with each output. In an implementation, the output of this stream cipher is free of the correlations that have been found in the alleged RC4.
As discussed with reference to the block functions, by a suitable choice of generators A, each xi output may be associated by the block cipher with a matrix B(xi), and form a walk on the matrix graph generated by A. For a matrix
let L(B)=(top(c),top(f) be the label output for each matrix at each step in the random walk. The empirical expansion properties enjoyed by G(A) indicate that the sequence L(B(x1)), L(B(x2)), . . . should be a pseudorandom sequence, if the xi are.
A remaining portion of this description (immediately following this section) deals with proof that the generator matrices used herein (such as those discussed with reference to
Free Monoids
The monoid M generated by the matrices S and T below is free, that is, isomorphic to the monoid of binary strings.
One way to prove this is to show that given AεM, only one of A′·S=A or A′·T=A holds with A′ ε M. In fact, the maximum component of A′ is shown to be strictly less than that of A, so that an easy induction accomplishes the proof.
Thus, an easy way to show that a monoid M generated by some set G of matrices is free, is to show that for any AεM, there is a unique U AεG with A′·U=A, where A′ ε M and smaller in some sense than A.
Considering the finite monoid generated by S and T (where operations are reduced modulo m=2l), the matrix graph so defined has relatively high girth.
Lemma. The matrix graph generated by S and T with operations modulo m=2l has girth l−1.
Proof. As the graph is a Cayley graph, it suffices to show that the identity I is not involved in a short cycle. Consider a matrix A on a path from the identity, and let amax be the maximum coefficient in A. If amax<m/2, then the maximum coefficient of either AS or AT is at most 2amax. Hence, if A is arrived at from a path of length t from I, and A′ is the matrix in the infinite monoid over □, if t<l, then amax<m, and A=A′ component-wise. In particular, as the infinite is free, A≠I. Thus I is not involved in a cycle of length less than l, proving the lemma.
Accordingly, this shows that it may generally be enough to find free monoids in one implementation. Moreover, the construction of the free monoid above suggests that to show that a monoid M generated by some set G of matrices is free, one should show that for any AεM, there is a unique U AεG with A′·U=A, where A′ ε M and smaller in some sense than A.
A Free Monoid
For a simple construction, let G1={T1, . . . ,T4} and M1 be the monoid generated by G1, with:
Note that if σi, for i=1 . . . 4, are the nontrivial degree-5 Gabber-Galil generators, then:
T1=σ2, T2=σ1σ4, T3=σ3σ2, T4=σ32
Lemma. The monoid M1 is free.
Proof. Given a matrix A, let A′ be the upper-left 2×2 submatrix of A. We note that σi0 is either S for i=1,2 and T for s=3,4. Also, as the first two columns of the last row of any element of G1 are zero, if A=Π Ui for some UiεG1, then A°=ΠUi0. Thus, given any AεM1, we have that A° is equivalent to a string sA over {S,T}, by the free-ness of the 2×2 monoid. Examining the generators, one can see that:
τ10=S·S, τ20=S·T, τ30=T·S, τ40=T·T
Hence, given the last two characters of sA, one can see that there is a unique Ti such that A′·Ti=A, with A′εM1. Furthermore, as |s′A|=|sA|−2, there is a norm on M1 that has been reduced, and so one can induct that M1 is indeed free.
General 3×3 Monoids
Any element of SL2(□+) can be identified with a binary string by forming the unique product equaling that element. The monoid of the previous section is then constructed from a set of SL2(□+) matrices whose strings are suffix-free. The entries in the third column are in this sense spurious. At the very least, aesthetic considerations may compel discovering matrices where all columns are involved in forming unique products.
A general class of free monoids are illustrate, of which the matrices of
Definition. Let the components of a matrix be:
A matrix is called feasible if the following holds:
a+b>c, and d+e>f
Lemma. Any product of feasible generators has the sum of the first two columns strictly greater than the third, for the upper two components.
Proof. The proof is by induction on the length of the product. The base case is clear by the definition of feasible. Now suppose:
Assume that the sum of the first two columns is strictly greater than the sum of the third. Then, if one takes a step by a feasible generator, the first row is [Aa+Bd, Ab+Be, Ac+Bf+C]. Summing the first two entries results in:
A(a+b)+B(d+e)≧Ac+Bf+A+B>Ac+Bf+C
Based on the property assumed of the generator matrix, the extra A and B result from the inequalities (and the strict inequality from the assumption that at least one of the generator inequalities is strict). Finally, the assumption on M gives A+B>C, proving the lemma.
Next, a general construction of generators whose monoid is free is described. Note this is one of many obvious ways to generalize the construction. A set of matrices C1, . . . ,Ck is said to be free if the monoid generated by them is free. This is equivalent to their expressions as products of SL2(□+) being prefix-free.
Lemma. Let C1, . . . ,Ck be any free matrices in SL2(□+). Fix a column cj of Cj, and let gj(i) for i=1,2 be elements of SL3(□+) defined from Cj by:
where the uj(i) and vj(i) are any integers satisfying:
as well as the conditions of previous Lemma. Then, the monoid generated by G={gj(i)}, with j=1, . . . ,k and i=1,2, is free.
Proof. Let M be the monoid generated by the G of the lemma. The same technique as before may be used, namely, given any XεM, it can be shown that either XεG, or there is a unique TεM with X′r=X for some X′ εM. Given Xε (M/G), as before one can see that
for some i1, . . . ij, so a j can be found so that either X=X′gj(1) or X=X′gj(2). The question now is to disambiguate between these two choices. Let xj be the column vector of the first two coordinates of the column of X corresponding to cj (i.e., the first or second column, depending on what was used to choose the third column of gj(i). Similarly, let x3 be the first two coordinates of the third column of X. Let ui=uj(i) and vi=vj(i). The disambiguation is performed by picking gj(i) iff xj≧x3. To show this works, suppose that X=X′gj(2) but xj≧x3. If j=1:
then xj≧x3 implies:
as u2>w and v2>y. But the first and final inequalities are impossible, as all numbers are ≧0. For the other direction, assume that X=X′gj(1) but xj is not greater than or equal to x3. Then, either
wa+yb<u1a+v1b+c≦wa+yb+c−a−b,
or wd+ye<u1d+v1e+f≦wd+ye+f−d−e,
by the choice of u1 and v1. But this implies either a+b≦c or d+e≦f, and as the gk(i) are feasible generators, neither situation is possible by the previous Lemma. The case when j=2 is analogous.
Accordingly, the vectors {u(i), v(i)}i=1,2 are a disambiguating pair with respect to a matrix C ε SL2(+) if they satisfy the conditions of the lemma above.
Remark Accordingly, this construction can be easily extended to generators of larger dimension.
Hardware Implementation
Computer environment 600 includes a general-purpose computing device in the form of a computer 602. The components of computer 602 can include, but are not limited to, one or more processors or processing units 604 (optionally including a cryptographic processor or co-processor), a system memory 606, and a system bus 608 that couples various system components including the processor 604 to the system memory 606.
The system bus 608 represents one or more of any of several types of bus structures, including a memory bus or memory controller, a peripheral bus, an accelerated graphics port, and a processor or local bus using any of a variety of bus architectures. By way of example, such architectures can include an Industry Standard Architecture (ISA) bus, a Micro Channel Architecture (MCA) bus, an Enhanced ISA (EISA) bus, a Video Electronics Standards Association (VESA) local bus, and a Peripheral Component Interconnects (PCI) bus also known as a Mezzanine bus.
Computer 602 typically includes a variety of computer-readable media. Such media can be any available media that is accessible by computer 602 and includes both volatile and non-volatile media, removable and non-removable media.
The system memory 606 includes computer-readable media in the form of volatile memory, such as random access memory (RAM) 610, and/or non-volatile memory, such as read only memory (ROM) 612. A basic input/output system (BIOS) 614, containing the basic routines that help to transfer information between elements within computer 602, such as during start-up, is stored in ROM 612. RAM 610 typically contains data and/or program modules that are immediately accessible to and/or presently operated on by the processing unit 604.
Computer 602 may also include other removable/non-removable, volatile/non-volatile computer storage media. By way of example,
The disk drives and their associated computer-readable media provide non-volatile storage of computer-readable instructions, data structures, program modules, and other data for computer 602. Although the example illustrates a hard disk 616, a removable magnetic disk 620, and a removable optical disk 624, it is to be appreciated that other types of computer-readable media which can store data that is accessible by a computer, such as magnetic cassettes or other magnetic storage devices, flash memory cards, CD-ROM, digital versatile disks (DVD) or other optical storage, random access memories (RAM), read only memories (ROM), electrically erasable programmable read-only memory (EEPROM), and the like, can also be utilized to implement the exemplary computing system and environment.
Any number of program modules can be stored on the hard disk 616, magnetic disk 620, optical disk 624, ROM 612, and/or RAM 610, including by way of example, an operating system 626, one or more application programs 628, other program modules 630, and program data 632. Each of such operating system 626, one or more application programs 628, other program modules 630, and program data 632 (or some combination thereof) may implement all or part of the resident components that support the distributed file system. Example of program modules and data is discussed below with reference to
A user can enter commands and information into computer 602 via input devices such as a keyboard 634 and a pointing device 636 (e.g., a “mouse”). Other input devices 638 (not shown specifically) may include a microphone, joystick, game pad, satellite dish, serial port, scanner, and/or the like. These and other input devices are connected to the processing unit 604 via input/output interfaces 640 that are coupled to the system bus 608, but may be connected by other interface and bus structures, such as a parallel port, game port, or a universal serial bus (USB).
A monitor 642 or other type of display device can also be connected to the system bus 608 via an interface, such as a video adapter 644. In addition to the monitor 642, other output peripheral devices can include components such as speakers (not shown) and a printer 646 which can be connected to computer 602 via the input/output interfaces 640.
Computer 602 can operate in a networked environment using logical connections to one or more remote computers, such as a remote computing device 648. By way of example, the remote computing device 648 can be a personal computer, portable computer, a server, a router, a network computer, a peer device or other common network node, game console, and the like. The remote computing device 648 is illustrated as a portable computer that can include many or all of the elements and features described herein relative to computer 602.
Logical connections between computer 602 and the remote computer 648 are depicted as a local area network (LAN) 650 and a general wide area network (WAN) 652. Such networking environments are commonplace in offices, enterprise-wide computer networks, intranets, and the Internet.
When implemented in a LAN networking environment, the computer 602 is connected to a local network 650 via a network interface or adapter 654. When implemented in a WAN networking environment, the computer 602 typically includes a modem 656 or other means for establishing communications over the wide network 652. The modem 656, which can be internal or external to computer 602, can be connected to the system bus 608 via the input/output interfaces 640 or other appropriate mechanisms. It is to be appreciated that the illustrated network connections are exemplary and that other means of establishing communication link(s) between the computers 602 and 648 can be employed.
In a networked environment, such as that illustrated with computing environment 600, program modules depicted relative to the computer 602, or portions thereof, may be stored in a remote memory storage device. By way of example, remote application programs 658 reside on a memory device of remote computer 648. For purposes of illustration, application programs and other executable program components such as the operating system are illustrated herein as discrete blocks, although it is recognized that such programs and components reside at various times in different storage components of the computing device 602, and are executed by the data processor(s) of the computer.
Various modules and techniques may be described herein in the general context of computer-executable instructions, such as program modules, executed by one or more computers or other devices. Generally, program modules include routines, programs, objects, components, data structures, etc. that perform particular tasks or implement particular abstract data types. Typically, the functionality of the program modules may be combined or distributed as desired in various implementations.
An implementation of these modules and techniques may be stored on or transmitted across some form of computer-readable media. Computer-readable media can be any available media that can be accessed by a computer. By way of example, and not limitation, computer-readable media may include “computer storage media” and “communications media.”
“Computer storage media” includes volatile and non-volatile, removable and non-removable media implemented in any method or technology for storage of information such as computer-readable instructions, data structures, program modules, or other data. Computer storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can be accessed by a computer.
“Communication media” typically includes computer-readable instructions, data structures, program modules, or other data in a modulated data signal, such as carrier wave or other transport mechanism Communication media also includes any information delivery media. The term “modulated data signal” means a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal. By way of example, and not limitation, communication media includes wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, RF, infrared, and other wireless media. Combinations of any of the above are also included within the scope of computer-readable media.
This application is a divisional of and claims priority to U.S. patent application Ser. No. 10/775,485, filed Feb. 9, 2004 now U.S. Pat. No. 7,289,629.
Number | Name | Date | Kind |
---|---|---|---|
4316055 | Feistel | Feb 1982 | A |
5058137 | Shah | Oct 1991 | A |
6069954 | Moreau | May 2000 | A |
6928602 | Yamagishi et al. | Aug 2005 | B2 |
7421076 | Stein et al. | Sep 2008 | B2 |
20020101986 | Roelse | Aug 2002 | A1 |
20030152219 | Coppersmith et al. | Aug 2003 | A1 |
Number | Date | Country | |
---|---|---|---|
20090022308 A1 | Jan 2009 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 10775485 | Feb 2004 | US |
Child | 11873339 | US |