The present invention relates to the technical field of cryptography, cryptographic processing attacks, and side channel attack (SCA) mitigations. More particularly, embodiments relate to micro-electronics, circuits and configurations thereof for disrupting side channel attack learning and SCA profiling of deep-learning machines.
Broadly, it relates to a new and developed way to cryptographically protect electronics using cryptographic algorithms and Public Key Infrastructure (PKI) systems for signature/verification, encryption/decryption, key exchange, and key encapsulation, which are sensitive to side-channel attack.
More specifically, it envisages a means of desynchronizing timing of register transfers used in cryptographic processing at the hardware bit-level to protect profiling of secure intense computations.
Advantageously, the invention provides a technical mechanism for protecting hardware against power analysis profiling by supervised attacks such as deep-learning side-channel attacks.
Cyber-attacks are unwelcome attempts to steal, expose, alter, disable or destroy information through unauthorized access to computer systems. Cyberattacks are now a common and repeating occurrence. One of these attacks is the Side Channel Attack (SCA), which monitors security critical operations and exploits physical leakages from a cryptographic device rather than the weakness in the cryptographic algorithm itself. It exploits system behaviors in an attempt to extract secrets from a semiconductor chip or electronics by measuring or analyzing various physical parameters, for example, timing information, power consumption, heat, acoustics and electromagnetic leaks. Side-channel leakages from a physical electronic device can be exploited to recover a secret such as keys used in encryption mechanisms, or digital signatures, from block and stream ciphers, to public key algorithms post-quantum cryptography. Today, side-channel attacks are considered as the most practical attacks, and countering them is currently considered mandatory for implementing physically secure cryptographic algorithms.
A deep learning side-channel attack (DL-SCA) is a new form of a profiling side-channel attack method that uses a machine-intelligence system's statistical learning abilities to reveal a cryptographic module's secret key 102 from such side-channel information. A deep learning algorithm (executing on machine 107) can learn a mapping between leaked information and the individual bits of the private key, in what is called “labeling” 111. As an example of a DL-SCA, a neural network can be trained from the output of power and timing signals captured from a target cryptographic module, or cryptographic processor 101. The trained neural network extracts relevant hardware processing information (e.g. transistor state changes, gate switching, logic operations, etc.) and external information (e.g. electrical signals, EM signals/waveforms, etc.) from the leaky part that depends on the private key 102.
DL-SCA's can also exploit algorithm specific implementation details (whether in software, hardware, or combination thereof) to gain knowledge about the private key 102, such as the time the algorithm (at the high-level software operating-system level, or the low-level gate logic) takes to execute the computation. For instance, the time duration taken by the known square-and-multiply algorithm used in classical implementation of the modular exponentiation depends linearly on the number of bits in the key. This linear relation can be exploited by an attacker, to infer the number of bits in the key by timing how long it takes to perform the computation for diverse RSA keys. That is, it can link an operation sequence of machine instructions, for instance, add and multiply operations of an SCA algorithm, to bits in the private key 102 during its access and use by the cryptographic processor 101, to label operations sequences to bit values of the private key, thereby compromising the security of the system.
Simple power analysis is a method of side-channel attack that examines a chip's current consumption over time. Because different electronic computation operations will exhibit different power profiles during security relevant tasks, an attacker can determine what type of operation is being performed at a given time. As each operation has its individual power signature, it is possible to determine which operations are performed within the chip by solely looking at its power consumption between clock cycles. Moreover, the power trace of the execution of the same operation looks slightly different if the data supplied to that operation is different as well because more internal states need to be switched and this involves higher power consumption, which can be analyzed over time. For systems which are not especially hardened against these kinds of attacks, an attacker might thus be able to determine which operations are performed within a microchip by solely looking at the power traces.
Herein provided is an electronic device for securing and desynchronizing register transfers as a mitigation strategy to side channel attacks that employ power analysis profiling, whereby leakage information produced as a result of register switching at the bit-level, by way of random delay cell insertions, is decorrelated from actual power consumption. The technical effect of the solution provided by way of the electronic device constructively produces overlaps of current power profiles/signatures, maximally spanning across bit value profiles of a register (or across data paths of sequential logic elements), which when analyzed from a Hamming Weight or Distance model introduces a wider time-spread of resulting current power profiles thereby impeding deep learning of the transistor switching/flipping state behavior, and consequently disrupting efficacy of profile/signature matching/mapping of transistor state changes characteristic of a traditional power analysis side channel attack.
In another embodiment, an electronic circuit configuration incorporates programmable random delay cells to desynchronize transitions across sequential circuit elements (e.g. transistors, CMOS gates, flip-flops, sequential elements, etc.) representing the bit-level values, for example, a register used in a cryptographic operation. The programmable random delay cells receive as input a random state to introduce randomness in the transistor state changes, yet are constricted to a predetermined time/delay period, such that the representation of the state (e.g. 0 or 1) is maximally stretched or compressed within a permitted state boundary. The random state can be provided by way of a random masking bit, random number generator (RNG), a random selector or combination thereof. The predetermined time/delay period that defines the permitted state boundary can itself be programmed, or learned, from analysis of multiple power consumption waveforms for same operation sequences when executed several times.
In yet another embodiment, a method for decorrelating information leaked by an electronic device from its power consumption is provided. The method includes securing and desynchronizing register transfers by way of a realized circuit design (e.g. VHDL/Verilog for FPGA, RTL for IC/SoC design, etc.) and/or method of manufacture thereof, to incorporate programable delay cells and bit-level randomizations that desynchronize transitions among sequential circuit elements (e.g. transistors, CMOS gates, flip-flops, etc.) representing bit-level values is provided, for example, a register design used for cryptographic operation. The method includes adding a programmable delay cell between a start point and an end point of a data path traversing combinational logic. The programmable delay can be sourced via a random state element, such as a random masking bit, a random selector or combination thereof. In one arrangement, the programmable delay cell is inserted along a gated clock path (synchronous) data path where the start point and end point are clocked. In another arrangement, the programmable delay cell is inserted at along a non-gated clock path (asynchronous), for example, from an input port to a set/clear pin of a sequential element.
In another embodiment a temporaly-skewed latch is provided. By way of random level triggering (e.g. a random masking bit, a random selector or combination thereof), the latch injects randomness on the bus lines between sequential logic circuits (spanning combinational logic) by adding information coming from a random number generator (RNG). The randomness maximizes a current (electromagnetic) waveform spread of each bit-flip such that the representation of each bit value overlaps as many neighboring bit values (e.g., waveform tails and heads of adjacent current envelopes/signals) as permitted between sequential circuits to fully diversify a signature profile for a same value of the bit. That is, the spread of each flip is such that each bit overlaps as many neighborhoods as possible to have many different signatures for the same value. This spreading of the bit-value waveforms from the viewpoint of a current profile analysis renders any labeling or linking activity of power consumption activity to bit value less effective.
Advantageously, these embodiments complement the traditional/regular protection measures (scrambling, hiding, masking, etc.) on sequential logic registers to mitigate/cancel correlations between the data and the power consumption. The addition of a random programmable delay ensures current profiles/signatures targeted by power analysis are varied and different when an operation is executed several times, which further conceals power leakage on the combinational logic thereby reducing the strength of a power analysis profiling by a side channel attack. The inventive design can be used to mitigate a supervised attack because it further granulizes the modeling of a whole register to the bit thereby decreasing and thwarting the possibility of supervised attacks.
The features of the present invention, which are believed to be novel, are set forth with particularity in the appended claims. The invention, together with further objects and advantages thereof, may best be understood by reference to the following description, taken in conjunction with the accompanying drawings, in the several figures of which like reference numerals identify like elements, and in which:
While the specification concludes with claims defining the features of the invention that are regarded as novel, it is believed that the invention will be better understood from a consideration of the following description in conjunction with the drawing figures.
Over the years, research on Side Channel Attacks has confirmed that power leakage is equivalent to an information leakage, and thereby allows an attacker to recover secret data of electronics devices by monitoring the power consumption, and/or the resulting leakage consequent to power consumption activities (e.g., EM, heat, photonic emission, etc.). Because side-channel attacks rely on the relationship between information emitted (leaked) through a side channel and the secret data, countermeasures generally fall into two main categories: a) eliminate or reduce the release of such information, and b) eliminate the relationship between the leaked information and the secret data, that is, make the leaked information unrelated, or rather uncorrelated, to the secret data, typically through some form of randomization of the internal variables that transforms the data in a way that can be undone after the cryptographic operation is completed. The use of random delays in embedded software has been proposed as a generic countermeasure against side channel analysis, such as simple power analysis and statistical analysis of the power consumption or electromagnetic emanations. And, hardware based random process interrupts have been considered as candidates for adding clock cycles into a “process” at random points to create desynchronization. In most of these state of the art cases however, this is accomplished at a macro-level, for example, by adding clock jitter, or random controls above the machine word level, or at random points in hardware.
Briefly, a register is a fast accessible memory available to a CPU. A register may hold an instruction, a storage address, or any kind of data (such as a bit sequence or individual characters). Some instructions specify registers as part of the instruction. For instance, an instruction may specify that the contents of two defined registers be added together and then placed in a specified register. A processor typically contains multiple index registers, also known as address registers or registers of modification. The effective address of any entity in a computer includes the base, index, and relative addresses, all of which are stored in the index register. A shift register is another type. Bits enter the shift register at one end and emerge from the other end. Flip flops, also known as bistable gates, store and process the data.
Here, the register 200 includes elements that each represent an electronic state. A register is made of several gates, each of them storing an atomic information 0 or 1. A CMOS gate is made of two transistors at minimum. More complex gates with dozens of transistors are also contemplated and applicable. As one example, the CMOS gate (208) may be in a 0 (low) or 1 (high) state. Complementary metal-oxide-semiconductor, is a type of metal-oxide-semiconductor field-effect transistor (MOSFET) fabrication process that uses complementary and symmetrical pairs of p-type and n-type MOSFETs for logic functions. CMOS technology is used for constructing integrated circuit (IC) chips, including microprocessors, microcontrollers, memory chips (including CMOS BIOS), and other digital logic circuits. Although the wire connections between the gates induce a parasitic capacitance, CMOS has high noise immunity and uses little power, except when it is switching.
Electronic computers (microcontrollers, microprocessors, FPGAs, etc.) have two components to their power consumption: static and dynamic. The static power consumption is the power required to keep the device running, and depends on things like the number of transistors inside the device. Secondly, and more importantly for side channel attack mitigation analysis, dynamic power consumption depends on the data moving around inside the device. Every time a bit is changed from a 0 to a 1 (like those in register 200), or vice versa, some electronic current is required to charge or discharge the data lines. This dynamic power is the point of exposure interest in simple power analysis of side channel attacks. Transistor power consumption is asymmetric in the sense that when the transistor state changes, the stream of electrons yield a dynamic power consumption, while otherwise only static power consumption occurs. Thus, when one of the CMOS gates in register 200 switches from one state to another (e.g. 0 or 1), it does not consume the same amount of power as when is does not switch. Accordingly, the dynamic power consumption can be perceived as having a wider power consumption variance that the static power consumption. And further it doesn't consume the same amount of power when is switches from one state to another. Accordingly, the Electro Magnetic (EM) field shown by 202 and 203, evolves as a function of voltage variation.
When the register 200 is loaded with a value v (e.g., bits 0 to 7: 10010001), the induced power consumption is dependent on the static and dynamic states of the transistors representing the value. Usually the induced power consumption is assumed to be linked to the Hamming Weight of v (i.e. the number of bits equal to ‘1’ in the binary representation of v). Hence we can write power(v)=HW(v)+N where the noise is assumed to some electronics perturbations induced by surrounding elements. N is usually assumed to be Gaussian.
Referring to
Other leakages model exist where the most information known about the state of the transistor changes is the Hamming Distance (i.e. the number of bits that have switched between the latest value of the register and the current one). Hence we write power(v)n=HDn-1->n(v)+N which means that the power of v depends on the distance between the value of the register at state n−1 and the value of the register at state n. N remains the same as the previous Gaussian noise. The Hamming distance model provides a convenient method for determining the expected power consumption of a circuit over a given time interval. Where a change in data is observable, the hamming Distance is preferred over the Hamming Weight model.
In the register 200 model of
On this point, Template Attacks are a special type of side channel attack that exploit knowledge of this flipping behavior at the transistor level by building a dictionary of power signatures in an identification step (that is, a learning step or templating step) and by further confronting a power signature candidate to this dictionary to recover the corresponding value, which occurs in a matching step (also known as an exploitation step or recovery step). One contemplated means for mitigating template attacks would is to force the bits to flip at the exact same moment. However, from a semiconductor design standpoint, this is a huge constraint that demands significant increases in silicon area and/or high resolution clocking rates (frequencies), and/or power consumption increase and likely a complex and costly balancing between all these factors.
Thus there is a need to obfuscate a template attack's ability, as well as other power analysis based side channel attack approaches, that attempt to link power consumption (and observable behaviors related to processor instruction executions and resulting current profiles/signatures) with transistor state change behaviors and render such side channel attack strategies less effective.
In one embodiment, an electronic device is provided to mitigate power analysis profiling of a side channel attack. The device comprises a first sequential circuit, a second sequential circuit, a combinational logic between said first and second sequential circuits, and a clock. The sequential circuits are clocked along a data path whereby said electronic device produces power traces with a time delay specific to said electronic device. A random programmable delay line inserted along said data path from said first sequential circuit to said combinational logic desynchronizes transitions within said first and second sequential circuit. The delay line, constructively produces overlaps of said power traces to maximally span bit value profiles of said first and second sequential circuit.
Advantageously, the delay line (401) produces a more numerous and wider time-spread of resulting power trace profiles than without said delay line, thereby impeding deep learning (107) of transistor switching and flipping state behavior when said power traces are analyzed using a Hamming (250,260) Weight or Distance model. Advantageously, the randomizations provided by the random programmable delay line (401) selectively and dynamically modify bus lines of said data path (307), thereby spreading a timing race of each bus line. Advantageously, each random delay cell (420) can be selectively turned off or on for performance tuning at the macro-level; and each individual cell in said series of delay cells (425) can be selectively turned off or on for performance tuning at the micro-level.
It aims at protecting register transfers and can be used not only for cryptographic purposes before operational execution of commands on a processor, but also during those operations. It serves to obfuscate detection of exchanges between low-level device, for example, from memory or register to another register, including keys that will further be used in cryptographic algorithms, including, but not limited to Public Key Infrastructure activities, such as asymmetric (e.g. ECC and RSA) or symmetric key (e.g. AES). For instance, the inventive approach/method can be incorporated in a smart card or a secure element implementing RSA signature/verification using modular exponential algorithms and ECC signature/verification using point addition algorithms to protect profiling of the highly intense math operations (e.g., point multiplication, point addition, etc.).
Briefly, the term bit “flipping” means the state of changing a bit value from a binary 0 to 1 or vice versa. The term bit “switching” means the temporal manner by which a transistor charges or discharges in changing from one state to another. The term “state”, when referring to a transistor, can mean either on or off, or, 0 or 1.
In this illustration, logic paths flow can be visualized from left to right, and up to down; namely, starting from the launch points (310-319), across the combinational logic, and ending at the capture points (330-339). Down the column of launch synchronous gates are a series of sequential circuits (here, flip-flops: FF 310). Down the column of capture synchronous gates are also a series of sequential circuits (here, flip-flops: FF 330). Between the two columns is the combinational logic 320, whose output is a function of only the present input. Along each data path, between the respective launch (start) 310 and capture (end) 330 points, sits logic 320 (e.g. AND, OR, etc.). The logic 320 is individually specific to the data path, and its synchronous gates before (e.g. 310) and after (e.g. 330). All the sequential circuits are synchronized by way of a clock 360; the circuit changes from the present state to the next state on the clock control input. The combinational logic 320 does not use memory and does not use the clock. The change of its internal state occurs when there is a change in the input variable. Combinational logic is characterized by its propagation delay and contamination delay. The propagation delay is the maximum time from when any input changes until the output or outputs reach their final value. The contamination delay is the minimum time from when any input changes until any output starts to change its value.
The flip flop (310 or 330) is a sequential circuit which samples its inputs and changes its outputs only at particular instants of time and not continuously. A flip-flop is used to store a single binary bit and has two states; one of its two states represents “1”, the other represents “0”. A flip flop is said to be edge sensitive or edge triggered rather than being level triggered like latches. The flip-flop is however considered “latched” when the Q terminal output holds the last input condition. Flip-flops are characterized by their propagation delay time. For flip-flops, data must arrive before the rising edge of the clock phase, rather than the falling edge. Data always departs the flop at the rising edge. One must therefore separately track arrival and departure times and introduce a set of departure constraints that relate arrival and departure times.
Briefly, a register is a group of flip-flops used to store a binary word, where one flip-flop is needed for each bit in the data word Accordingly, the group of launch synchronous gates can also represent the bits of a first register (e.g. 200 of
During this register transfer/transition time, the transistors in between switch so that single signals or a whole bus get switched from one state into the other. The switching and the capacitances of the various structures within the chip takes time and draws current as a result. If the power consumption is measured over time, for a given data path with same inputs, a unique power trace will results. Understandable, the resulting power trace along each path will be different because it is characteristic to each logic block 320. Although power trace waveforms A, B, C in
As noted in the background, there are two types of timing analysis: static and dynamic. Static timing analysis is the approach for validating performance of a design by checking all possible paths for timing violations under worst case conditions. It considers the worst possible delay through each logic element, but not the logical operation of the circuit. Delays such as gate delay and net delays are considered in each path and these delays are compared against their required maximum and minimum values. In contrast, dynamic timing analysis verifies functionality of the design by applying input vectors and checking for correct output vectors whereas Static Timing Analysis checks static delay requirements of the circuit without any input or output vectors. Dynamic timing analysis can be used for synchronous as well as asynchronous designs and is preferable for designs having clocks crossing multiple domains.
Data travel is visualized for our timing analysis purposes here as flowing sequentially from left to right across a row. As can be seen (e.g. Tf 340, Ts 350), the transient current activity timing is unique and stable for a given logic path. Transient means when the transistors in the circuits are changing states (e.g. flipping bits). For instance, along the upper-most row, the fastest bus line with a capture time, Tf 340, is observed. Along the lower-most row, the slowest bus line with a capture time, Ts 350, is observed. Allow this appears in the illustration to result from propogation delays associated with the circuitry, the fastest and slowest times can differ from that shown according to the internal delay of the combinational logic 320; that is, the fastest bus line may be the lower one, and the slowest bus line could be along the data path in the middle of the diagram.
The current activity profile of each bus line is depending on multiple factors such as clock skew, logic Boolean equation, power supply variation. And, all those factors are stable over time. As a consequence, a unique current profile exists for a given vector v. That is a key point for the learning step. In order to make the learning phase more complex, and make a side channel attack based on power analysis less effective, as will be seen ahead in
The circuit 400 (whether configured as specialized register transfer logic, or, configured as general sequential transfer logic) has setup and hold time constraints that dictate the maximum and minimum delays of the combinational logic 320 due to the flip-flops 310 on the left side comprising the launch synchronous gates, and the flip-fops 330 on the right side comprising the capture synchronous gates; namely, endpoints of the sequential logic circuit. Within these time constraints, in accordance with the inventive aspects herein, the circuit 400 is re-configured/re-designed from that of circuit 300 to maximize neighboring overlaps of the bus lines relevant to the data paths; for instance, the data signals of the fastest bus line, (see Tf 340 in
In the illustration of
Here, in the inventive embodiment, the desynchronizing is at a finer grain level; namely, at the individual bit-level between clocking times of sequential circuits along data paths. This is different from, and an advantage over, usual state of the art jitter mechanisms where all bits are shifted at the same time, or where noise is intentionally introduced to confound side channel attack success. The random delays are added on the data path of combinatorial logic, between any sequential elements, such as registers, memories and so on. Although the random delay cell 420 is added on the link (e.g. data path) between sequential elements, the bus is not mandatory per se. Here, each random delay cell 420 is performed on a per bit basis and not a nibble/byte/word basis. This allows for overlap between bits, and accordingly produces a wider range of resulting signatures (power traces). Advantageously, it provides for temporal misalignment of current signals (see A→C→B (451,452,453) in
The random programmable delay line (401) is suitable for insertion amongst multiple data paths (307) connecting a first sequential circuit (301) to a combinational logic (320) to a second sequential circuit (302) to desynchronize transitions there between. The delay line (401) is programmable and introduces a randomness of data transfer at a bit-level along said data path by way of one or more random delay cells (420) therein, and with said sequential circuit (301) together produce power traces (451-453) with a randomly characteristic time delay (A,B,C). The one or more random delay cells (420) constructively produce overlaps (A,B,C) of said power traces (451-453) to maximally span bit-value profiles of said first and second sequential circuit. The randomness selectively and dynamically modifies bus lines of said multiple data paths (307) to spread a timing race of each bus line, thereby imparting said randomly characteristic time delay. The random selector (427) and random masking bit (402) by way of said series of delay cells (425) and said multiplexer (426) introduce said randomness into the data path (307) thereby desynchronizing transitions of said first and second sequential logic circuits. Each random delay cell (420) can be programmatically turned off or on for performance tuning at the macro-level. Similarly, each individual cell in said series of delay cells (425) can be programmatically turned off or on for performance tuning at the micro-level, wherein said performance tuning imparts said randomly characteristic time delay to said power traces (451-453). The performance tuning includes stacking delays in the series of delay cells (425) to maximize or minimize an amount of bit-level overlaps, wherein this stacking is mapped to the range of the random number generator. The performance tuning can be learned via training or hard-coded.
The random masking bit 402 and a random selector 427 can each, alone or in combination, be sourced by a random number generator. A random number generator is a hardware device or software algorithm that generates a number that is taken from a limited or unlimited distribution and outputs it. The two main types of random number generators are pseudo random number generators and true random number (TRNG) generators. A hardware random number generator (HRNG) is a true random number generator, and is a device that generates random numbers from a physical process, rather than by means of an algorithm. Such devices are often based on microscopic phenomena that generate low-level, statistically random “noise” signals, such as thermal noise, the photoelectric effect, involving a beam splitter, and other quantum phenomena. It is emphasized however, that the exemplary schematic 450 is not the only configuration for introducing randomness into the data path of sequential logic circuits contemplated above. It merely serves as one example on how to introduce delays for purposes of desynchronizing transitions across sequential circuit elements.
It should also be noted that the one or more random delay cells 420 constituting the random programmable delay line 401, are programmable. That is, for example, each random delay cell 420 at the macro-level (and individual delay cells 425 at the micro-level) can be selectively turned off or on for performance tuning purposes. For the purposes of performance tuning, the number of delays in the delay line 425 can be stacked to maximize or curtail (minimize) the amount of bit-level overlaps mentioned above, and this stacking can be mapped to the range of the random number generator as an advantage of the technical effect of the solution. The performance tuning can be learned via training (see
The random delay cell (420) desynchronizes transitions within an electronic device and aid in mitigating supervised power analysis profiling of a side channel attack. The cell (420) comprises a series of delay cells (425), and a multiplexer (426). Each random delay cell (420) receives as input: a random masking bit (402) to said series of delay cells (425); and a random selector (427) to said multiplexer (426). The random selector (427) and said random masking bit (402) by way of said series of delay cells (425) and said multiplexer (426) introduce a randomness into the data path (307). One embodiment realized is a configuration wherein said random masking bit (402) is or'd with an output of a first sequential circuit (301). The output of the first sequential circuit (301) is then staged through said series of delay cells (425) as staged inputs into said multiplexer (426). One of said staged inputs is chosen according to said random selector (427), and an output of the mux (426) receiving a chosen staged input is then or'd with the random masking bit (402). The chosen staged input then connects to said combinational logic (329). The random masking bit (402) and random selector (427), alone or in combination, is sourced by a hardware random number generator (HRNG), though other sources can be used.
The point of
The temporally-skewed register (510) desynchronizes transitions within an electronic device and aid in mitigating supervised power analysis profiling of a side channel attack. The register (501) comprises a clock (518); one or more flip-flops (521) serially coupled to produce parallel data that is shifted into combinational logic (320) by way of said clock (518), and a random delay cell (420) inserted at an output of each said flip-flop to desynchronize transitions between said flip-flop and said combinational logic. The random delay cell (420) comprises a series of delay cells (425), and a multiplexer (426). The random delay cell receives as input: a random masking bit (402) to said series of delay cells (425); and a random selector (427) to said multiplexer (426). The random selector (427) and said random masking bit (402) by way of said series of delay cells (425) and said multiplexer (426) introduce a randomness into the data path (307) to produces overlaps (A,B,C) of resulting power traces (451-453) to maximally span across bit value profiles of said flip-flop. The random masking bit (402) and random selector (427), alone or in combination, is sourced by a hardware random number generator (HRNG), though other sources can be used.
Various embodiments for securing and desynchronizing register transfers as a mitigation strategy to side channel attacks that employ power analysis profiling, whereby leakage information produced as a result of register switching at the bit-level, by way of random delay cell insertions, is decorrelated from actual power consumption have been provided. delay element represents a latch. The examples so far have includes sequential/synchronous circuits. However other embodiments are contemplated for extending the inventive aspect to latches. A latch is a circuit that can hold 1 bit of data and is asynchronous (i.e. without a clock). In contrast, a register (as seen in
The temporally-skewed latch (501) desynchronizes transitions within an electronic device and aids in mitigating supervised power analysis profiling of a side channel attack. The latch (501) comprises a latch (601); and thereto communicatively coupled, a random delay cell (420) inserted along a data path (307) from said latch to a combinational logic (320) to desynchronize transitions between said latch and said combinational logic. The random delay cell (420) comprises a series of delay cells (425), and a multiplexer (426). The random delay cell receives as input: a random masking bit (402) to said series of delay cells (425); and a random selector (427) to said multiplexer (426). The random selector (427) and said random masking bit (402) by way of said series of delay cells (425) and said multiplexer (426) introduce a randomness into the data path (307) to produces overlaps (A,B,C) of resulting power traces (451-453) to maximally span across bit value profiles of said latch. The random masking bit (402) and random selector (427), alone or in combination, is sourced by a hardware random number generator (HRNG), though other sources can be used.
These embodiments can be implemented on a computer, hardware, or other device such as a machine.
The machine may comprise a server computer, a client user computer, a personal computer (PC), a tablet PC, a laptop computer, a desktop computer, a mobile device, a cell phone, a control system, a network router, switch or bridge, or any machine capable of executing a set of instructions (sequential or otherwise) that specify actions to be taken by that machine. It will be understood that a device of the present disclosure includes broadly any electronic device that provides voice, video or data communication. Further, while a single machine is illustrated, the term “machine” shall also be taken to include any collection of machines that individually or jointly execute a set (or multiple sets) of instructions to perform any one or more of the methodologies discussed herein.
The computer system 700 may include a processor 702 (e.g., a central processing unit (CPU), a graphics processing unit (GPU, or both), a main memory 704 and a static memory 706, which communicate with each other via a bus 708. The computer system 700 may further include a video display unit 710 (e.g., a liquid crystal display or LCD), a flat panel, a solid state display, or a cathode ray tube (CRT)). The computer system 700 may include an input device 712 (e.g., a keyboard, screen), a cursor control device 714 (e.g., a mouse, screen), a disk drive unit 716, a signal generation device 718 (e.g., a speaker or remote control) and a network interface device 720.
The disk drive unit 716 may include a machine-readable medium 722 on which is stored one or more sets of instructions (e.g., software 724) embodying any one or more of the methodologies or functions described herein, including those methods illustrated above. The instructions 724 may also reside, completely or at least partially, within the main memory 704, the static memory 706, and/or within the processor 702 during execution thereof by the computer system 700. The main memory 704 and the processor 702 also may constitute machine-readable media.
Dedicated hardware implementations including, but not limited to, application specific integrated circuits, programmable logic arrays and other hardware devices can likewise be constructed to implement the methods described herein. Applications that may include the apparatus and systems of various embodiments broadly include a variety of electronic and computer systems. Some embodiments implement functions in two or more specific interconnected hardware modules or devices with related control and data signals communicated between and through the modules, or as portions of an application-specific integrated circuit. Thus, the example system is applicable to software, firmware, and hardware implementations.
In accordance with various embodiments of the present disclosure, the methods described herein are intended for operation as software programs running on a computer processor. Furthermore, software implementations can include, but not limited to, distributed processing or component/object distributed processing, parallel processing, or virtual machine processing can also be constructed to implement the methods described herein.
While the machine-readable medium 722 is shown in an example embodiment to be a single medium, the term “machine-readable medium” should be taken to include a single medium or multiple media (e.g., a centralized or distributed database, and/or associated caches and servers) that store the one or more sets of instructions. The term “machine-readable medium” shall also be taken to include any medium that is capable of storing, encoding or carrying a set of instructions for execution by the machine and that cause the machine to perform any one or more of the methodologies of the present disclosure.
The term “machine-readable medium” shall accordingly be taken to include, but not be limited to: solid-state memories such as a memory card or other package that houses one or more read-only (non-volatile) memories, random access memories, or other re-writable (volatile) memories; magneto-optical or optical medium such as a disk or tape; and carrier wave signals such as a signal embodying computer instructions in a transmission medium; and/or a digital file attachment to e-mail or other self-contained information archive or set of archives is considered a distribution medium equivalent to a tangible storage medium. Accordingly, the disclosure is considered to include any one or more of a machine-readable medium or a distribution medium, as listed herein and including art-recognized equivalents and successor media, in which the software implementations herein are stored.
The illustrations of embodiments described herein are intended to provide a general understanding of the structure of various embodiments, and they are not intended to serve as a complete description of all the elements and features of apparatus and systems that might make use of the structures described herein. Many other embodiments will be apparent to those of skill in the art upon reviewing the above description. Other embodiments may be utilized and derived therefrom, such that structural and logical substitutions and changes may be made without departing from the scope of this disclosure. Figures are also merely representational and may not be drawn to scale. Certain proportions thereof may be exaggerated, while others may be minimized. Accordingly, the specification and drawings are to be regarded in an illustrative rather than a restrictive sense.
Such embodiments of the inventive subject matter may be referred to herein, individually and/or collectively, by the term “invention” merely for convenience and without intending to voluntarily limit the scope of this application to any single invention or inventive concept if more than one is in fact disclosed. Thus, although specific embodiments have been illustrated and described herein, it should be appreciated that any arrangement calculated to achieve the same purpose may be substituted for the specific embodiments shown. This disclosure is intended to cover any and all adaptations or variations of various embodiments. Combinations of the above embodiments, and other embodiments not specifically described herein, will be apparent to those of skill in the art upon reviewing the above description.
In the above-description of various embodiments of the present disclosure, aspects of the present disclosure may be illustrated and described herein in any of a number of patentable classes or contexts including any new and useful process, machine, manufacture, or composition of matter, or any new and useful improvement thereof. Accordingly, aspects of the present disclosure may be implemented in entirely hardware, entirely software (including firmware, resident software, micro-code, etc.) or combining software and hardware implementation that may all generally be referred to herein as a “circuit,” “module,” “component,” or “system.” Furthermore, aspects of the present disclosure may take the form of a computer program product comprising one or more computer readable media having computer readable program code embodied thereon.
Any combination of one or more computer readable media may be used. The computer readable media may be a computer readable signal medium or a computer readable storage medium. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. In the context of this document, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
A computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device. Program code embodied on a computer readable signal medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
Computer program code for carrying out operations for aspects of the present disclosure may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Scala, Smalltalk, Scheme, Go, C++, C #, VB.NET, Python or the like, conventional procedural programming languages, such as the “C” programming language, Perl, PHP, dynamic programming languages such as Python, Ruby and Groovy, or other programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer, entirely on the remote computer or server, or within the Cloud or other computer network. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider) or in a cloud computing environment or offered as a service such as a Software as a Service (SaaS), Backend as a Service (BaaS) for connecting mobile apps to cloud based services, and Security as a Service (SECaas).
Aspects of the present disclosure are described herein with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the disclosure. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable instruction execution apparatus, create a mechanism for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer readable medium that when executed can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions when stored in the computer readable medium produce an article of manufacture including instructions which when executed, cause a computer to implement the function/act specified in the flowchart and/or block diagram block or blocks. The computer program instructions may also be loaded onto a computer, other programmable instruction execution apparatus, or other devices to cause a series of operational steps to be performed on the computer, other programmable apparatuses or other devices to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
It is to be understood that the terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. Unless otherwise defined, all terms (including technical and scientific terms) used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this disclosure belongs. It will be further understood that terms, such as those defined in commonly used dictionaries, should be interpreted as having a meaning that is consistent with their meaning in the context of this specification and the relevant art and will not be interpreted in an idealized or overly formal sense unless expressly so defined herein.
The flowchart and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods, and computer program products according to various aspects of the present disclosure. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems that perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
The terminology used herein is for the purpose of describing particular aspects only and is not intended to be limiting of the disclosure. As used herein, the singular forms “a”, “an” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms “comprises” and/or “comprising,” when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof. As used herein, the term “and/or” includes any and all combinations of one or more of the associated listed items. Like reference numbers signify like elements throughout the description of the figures.
The corresponding structures, materials, acts, and equivalents of any means or step plus function elements in the claims below are intended to include any disclosed structure, material, or act for performing the function in combination with other claimed elements as specifically claimed. The description of the present disclosure has been presented for purposes of illustration and description, but is not intended to be exhaustive or limited to the disclosure in the form disclosed. Many modifications and variations will be apparent to those of ordinary skill in the art without departing from the scope and spirit of the disclosure. The aspects of the disclosure herein were chosen and described in order to best explain the principles of the disclosure and the practical application, and to enable others of ordinary skill in the art to understand the disclosure with various modifications as are suited to the particular use contemplated.
| Number | Date | Country | Kind |
|---|---|---|---|
| 22305273.9 | Mar 2022 | EP | regional |
| Filing Document | Filing Date | Country | Kind |
|---|---|---|---|
| PCT/EP2023/056253 | 3/10/2023 | WO |