Baby-step giant-step
In group theory, a branch of mathematics, the baby-step giant-step is a meet-in-the-middle algorithm for computing the discrete logarithm. The discrete log problem is of fundamental importance to the area of public key cryptography. Many of the most commonly used cryptography systems are based on the assumption that the discrete log is extremely difficult to compute; the more difficult it is, the more security it provides a data transfer. One way to increase the difficulty of the discrete log problem is to base the cryptosystem on a larger group.
Contents
Theory
The algorithm is based on a space–time tradeoff. It is a fairly simple modification of trial multiplication, the naive method of finding discrete logarithms.
Given a cyclic group of order , a generator of the group and a group element , the problem is to find an integer such that
The baby-step giant-step algorithm is based on rewriting as , with and and . Therefore, we have:
The algorithm precomputes for several values of . Then it fixes an and tries values of in the left-hand side of the congruence above, in the manner of trial multiplication. It tests to see if the congruence is satisfied for any value of , using the precomputed values of .
The algorithm
Input: A cyclic group G of order n, having a generator α and an element β.
Output: A value x satisfying .
- m ← Ceiling(√n)
- For all j where 0 ≤ j < m:
- Compute α^{j} and store the pair (j, α^{j}) in a table. (See section "In practice")
- Compute α^{−m}.
- γ ← β. (set γ = β)
- For all i where 0 ≤ i < m:
- Check to see if γ is the second component (α^{j}) of any pair in the table.
- If so, return im + j.
- If not, γ ← γ • α^{−m}.
C algorithm with the GNU MP lib
Contrary to the recommendation below to use a hash table, this implementation uses binary search in a sorted array as table; this incurs an extra factor in the lookup complexity. Indeed, since the bit-size of the group elements involved is also , this is arguably enough to make the table lookups dominate the total time complexity.
void baby_step_giant_step (mpz_t g, mpz_t h, mpz_t p, mpz_t n, mpz_t x ){
unsigned long int i;
long int j = 0;
mpz_t N;
mpz_t* gr ; /* list g^r */
unsigned long int* indices; /* indices[ i ] = k <=> gr[ i ] = g^k */
mpz_t hgNq ; /* hg^(Nq) */
mpz_t inv ; /* inverse of g^(N) */
mpz_init (N) ;
mpz_sqrt (N, n ) ;
mpz_add ui (N, N, 1 ) ;
gr = malloc (mpz_get_ui (N) * sizeof (mpz_t) ) ;
indices = malloc ( mpz_get_ui (N) * sizeof (long int ) ) ;
mpz_init_set_ui (gr[ 0 ], 1);
/* find the sequence {g^r} r = 1 ,.. ,N (Baby step ) */
for ( i = 1 ; i <= mpz_get_ui (N) ; i++) {
indices[i - 1] = i - 1 ;
mpz_init (gr[ i ]) ;
mpz_mul (gr[ i ], gr[ i - 1 ], g ); /* multiply gr[i - 1] for g */
mpz_mod (gr[ i ], gr[ i ], p );
}
/* sort the values (k , g^k) with respect to g^k */
qsort ( gr, indices, mpz_get_ui (N), mpz_cmp ) ;
/* compute g^(-Nq) (Giant step) */
mpz_init_set (inv, g);
mpz_powm (inv, inv, N, p); /* inv <- inv ^ N (mod p) */
mpz_invert (inv, inv, p) ;
mpz_init_set (hgNq, h);
/* find the elements in the two sequences */
for ( i = 0 ; i <= mpz_get_ui (N) ; i++){
/* find hgNq in the sequence gr ) */
j = bsearch (gr, hgNq, 0, mpz_get_ui (N), mpz_cmp ) ;
if ( j >= 0 ){
mpz_mul_ui (N, N, i);
mpz_add_ui (N, N, indices [j]);
mpz_set (x, N) ;
break;
}
/* if j < 0, find the next value of g^(Nq) */
mpz_mul (hgNq, hgNq, inv);
mpz_mod (hgNq, hgNq, p);
}
}
In practice
The best way to speed up the baby-step giant-step algorithm is to use an efficient table lookup scheme. The best in this case is a hash table. The hashing is done on the second component, and to perform the check in step 1 of the main loop, γ is hashed and the resulting memory address checked. Since hash tables can retrieve and add elements in O(1) time (constant time), this does not slow down the overall baby-step giant-step algorithm.
The running time of the algorithm and the space complexity is , much better than the O(n) running time of the naive brute force calculation.
The Baby-step giant-step algorithm is often used to solve for the shared key in the Diffie Hellman key exchange, when the modulus is a prime number. If the modulus is not prime, the Pohlig–Hellman algorithm has a smaller algorithmic complexity, and solves the same problem.
Notes
- The baby-step giant-step algorithm is a generic algorithm. It works for every finite cyclic group.
- It is not necessary to know the order of the group G in advance. The algorithm still works if n is merely an upper bound on the group order.
- Usually the baby-step giant-step algorithm is used for groups whose order is prime. If the order of the group is composite then the Pohlig–Hellman algorithm is more efficient.
- The algorithm requires O(m) memory. It is possible to use less memory by choosing a smaller m in the first step of the algorithm. Doing so increases the running time, which then is O(n/m). Alternatively one can use Pollard's rho algorithm for logarithms, which has about the same running time as the baby-step giant-step algorithm, but only a small memory requirement.
- The algorithm is usually credited to Daniel Shanks, but a 1994 paper by Nechaev^{[1]} states it was known to Gel'fond in 1962.
References
- ^ V. I. Nechaev, Complexity of a determinate algorithm for the discrete logarithm, Mathematical Notes, vol. 55, no. 2 1994 (165-172)
- H. Cohen, A course in computational algebraic number theory, Springer, 1996.
- D. Shanks. Class number, a theory of factorization and genera. In Proc. Symp. Pure Math. 20, pages 415—440. AMS, Providence, R.I., 1971.
- A. Stein and E. Teske, Optimized baby step-giant step methods, Journal of the Ramanujan Mathematical Society 20 (2005), no. 1, 1–32.
- A. V. Sutherland, Order computations in generic groups, PhD thesis, M.I.T., 2007.
- D. C. Terr, A modification of Shanks’ baby-step giant-step algorithm, Mathematics of Computation 69 (2000), 767–773. doi:10.1090/S0025-5718-99-01141-2