7 #ifndef _SECP256K1_FIELD_REPR_IMPL_H_ 8 #define _SECP256K1_FIELD_REPR_IMPL_H_ 10 #if defined HAVE_CONFIG_H 19 #if defined(USE_ASM_X86_64) 35 const uint64_t *d = a->
n;
36 int m = a->normalized ? 1 : 2 * a->magnitude, r = 1;
38 r &= (d[0] <= 0xFFFFFFFFFFFFFULL *
m);
39 r &= (d[1] <= 0xFFFFFFFFFFFFFULL *
m);
40 r &= (d[2] <= 0xFFFFFFFFFFFFFULL *
m);
41 r &= (d[3] <= 0xFFFFFFFFFFFFFULL *
m);
42 r &= (d[4] <= 0x0FFFFFFFFFFFFULL *
m);
43 r &= (a->magnitude >= 0);
44 r &= (a->magnitude <= 2048);
46 r &= (a->magnitude <= 1);
47 if (r && (d[4] == 0x0FFFFFFFFFFFFULL) && ((d[3] & d[2] & d[1]) == 0xFFFFFFFFFFFFFULL)) {
48 r &= (d[0] < 0xFFFFEFFFFFC2FULL);
60 uint64_t t0 = r->
n[0], t1 = r->
n[1], t2 = r->
n[2], t3 = r->
n[3], t4 = r->
n[4];
64 uint64_t x = t4 >> 48; t4 &= 0x0FFFFFFFFFFFFULL;
67 t0 += x * 0x1000003D1ULL;
68 t1 += (t0 >> 52); t0 &= 0xFFFFFFFFFFFFFULL;
69 t2 += (t1 >> 52); t1 &= 0xFFFFFFFFFFFFFULL;
m = t1;
70 t3 += (t2 >> 52); t2 &= 0xFFFFFFFFFFFFFULL;
m &= t2;
71 t4 += (t3 >> 52); t3 &= 0xFFFFFFFFFFFFFULL;
m &= t3;
77 x = (t4 >> 48) | ((t4 == 0x0FFFFFFFFFFFFULL) & (
m == 0xFFFFFFFFFFFFFULL)
78 & (t0 >= 0xFFFFEFFFFFC2FULL));
81 t0 += x * 0x1000003D1ULL;
82 t1 += (t0 >> 52); t0 &= 0xFFFFFFFFFFFFFULL;
83 t2 += (t1 >> 52); t1 &= 0xFFFFFFFFFFFFFULL;
84 t3 += (t2 >> 52); t2 &= 0xFFFFFFFFFFFFFULL;
85 t4 += (t3 >> 52); t3 &= 0xFFFFFFFFFFFFFULL;
91 t4 &= 0x0FFFFFFFFFFFFULL;
93 r->
n[0] = t0; r->
n[1] = t1; r->
n[2] = t2; r->
n[3] = t3; r->
n[4] = t4;
103 uint64_t t0 = r->
n[0], t1 = r->
n[1], t2 = r->
n[2], t3 = r->
n[3], t4 = r->
n[4];
106 uint64_t x = t4 >> 48; t4 &= 0x0FFFFFFFFFFFFULL;
109 t0 += x * 0x1000003D1ULL;
110 t1 += (t0 >> 52); t0 &= 0xFFFFFFFFFFFFFULL;
111 t2 += (t1 >> 52); t1 &= 0xFFFFFFFFFFFFFULL;
112 t3 += (t2 >> 52); t2 &= 0xFFFFFFFFFFFFFULL;
113 t4 += (t3 >> 52); t3 &= 0xFFFFFFFFFFFFFULL;
118 r->
n[0] = t0; r->
n[1] = t1; r->
n[2] = t2; r->
n[3] = t3; r->
n[4] = t4;
127 uint64_t t0 = r->
n[0], t1 = r->
n[1], t2 = r->
n[2], t3 = r->
n[3], t4 = r->
n[4];
131 uint64_t x = t4 >> 48; t4 &= 0x0FFFFFFFFFFFFULL;
134 t0 += x * 0x1000003D1ULL;
135 t1 += (t0 >> 52); t0 &= 0xFFFFFFFFFFFFFULL;
136 t2 += (t1 >> 52); t1 &= 0xFFFFFFFFFFFFFULL;
m = t1;
137 t3 += (t2 >> 52); t2 &= 0xFFFFFFFFFFFFFULL;
m &= t2;
138 t4 += (t3 >> 52); t3 &= 0xFFFFFFFFFFFFFULL;
m &= t3;
144 x = (t4 >> 48) | ((t4 == 0x0FFFFFFFFFFFFULL) & (
m == 0xFFFFFFFFFFFFFULL)
145 & (t0 >= 0xFFFFEFFFFFC2FULL));
148 t0 += 0x1000003D1ULL;
149 t1 += (t0 >> 52); t0 &= 0xFFFFFFFFFFFFFULL;
150 t2 += (t1 >> 52); t1 &= 0xFFFFFFFFFFFFFULL;
151 t3 += (t2 >> 52); t2 &= 0xFFFFFFFFFFFFFULL;
152 t4 += (t3 >> 52); t3 &= 0xFFFFFFFFFFFFFULL;
158 t4 &= 0x0FFFFFFFFFFFFULL;
161 r->
n[0] = t0; r->
n[1] = t1; r->
n[2] = t2; r->
n[3] = t3; r->
n[4] = t4;
171 uint64_t t0 = r->
n[0], t1 = r->
n[1], t2 = r->
n[2], t3 = r->
n[3], t4 = r->
n[4];
177 uint64_t x = t4 >> 48; t4 &= 0x0FFFFFFFFFFFFULL;
180 t0 += x * 0x1000003D1ULL;
181 t1 += (t0 >> 52); t0 &= 0xFFFFFFFFFFFFFULL; z0 = t0; z1 = t0 ^ 0x1000003D0ULL;
182 t2 += (t1 >> 52); t1 &= 0xFFFFFFFFFFFFFULL; z0 |= t1; z1 &= t1;
183 t3 += (t2 >> 52); t2 &= 0xFFFFFFFFFFFFFULL; z0 |= t2; z1 &= t2;
184 t4 += (t3 >> 52); t3 &= 0xFFFFFFFFFFFFFULL; z0 |= t3; z1 &= t3;
185 z0 |= t4; z1 &= t4 ^ 0xF000000000000ULL;
190 return (z0 == 0) | (z1 == 0xFFFFFFFFFFFFFULL);
194 uint64_t t0, t1, t2, t3, t4;
205 t0 += x * 0x1000003D1ULL;
208 z0 = t0 & 0xFFFFFFFFFFFFFULL;
209 z1 = z0 ^ 0x1000003D0ULL;
212 if ((z0 != 0ULL) & (z1 != 0xFFFFFFFFFFFFFULL)) {
220 t4 &= 0x0FFFFFFFFFFFFULL;
223 t2 += (t1 >> 52); t1 &= 0xFFFFFFFFFFFFFULL; z0 |= t1; z1 &= t1;
224 t3 += (t2 >> 52); t2 &= 0xFFFFFFFFFFFFFULL; z0 |= t2; z1 &= t2;
225 t4 += (t3 >> 52); t3 &= 0xFFFFFFFFFFFFFULL; z0 |= t3; z1 &= t3;
226 z0 |= t4; z1 &= t4 ^ 0xF000000000000ULL;
231 return (z0 == 0) | (z1 == 0xFFFFFFFFFFFFFULL);
236 r->
n[1] = r->
n[2] = r->
n[3] = r->
n[4] = 0;
245 const uint64_t *t = a->
n;
250 return (t[0] | t[1] | t[2] | t[3] | t[4]) == 0;
267 for (i=0; i<5; i++) {
280 for (i = 4; i >= 0; i--) {
281 if (a->
n[i] > b->
n[i]) {
284 if (a->
n[i] < b->
n[i]) {
293 r->
n[0] = r->
n[1] = r->
n[2] = r->
n[3] = r->
n[4] = 0;
294 for (i=0; i<32; i++) {
296 for (j=0; j<2; j++) {
297 int limb = (8*i+4*j)/52;
298 int shift = (8*i+4*j)%52;
299 r->
n[limb] |= (uint64_t)((a[31-i] >> (4*j)) & 0xF) << shift;
302 if (r->
n[4] == 0x0FFFFFFFFFFFFULL && (r->
n[3] & r->
n[2] & r->
n[1]) == 0xFFFFFFFFFFFFFULL && r->
n[0] >= 0xFFFFEFFFFFC2FULL) {
320 for (i=0; i<32; i++) {
323 for (j=0; j<2; j++) {
324 int limb = (8*i+4*j)/52;
325 int shift = (8*i+4*j)%52;
326 c |= ((a->
n[limb] >> shift) & 0xF) << (4 * j);
337 r->
n[0] = 0xFFFFEFFFFFC2FULL * 2 * (
m + 1) - a->
n[0];
338 r->
n[1] = 0xFFFFFFFFFFFFFULL * 2 * (
m + 1) - a->
n[1];
339 r->
n[2] = 0xFFFFFFFFFFFFFULL * 2 * (
m + 1) - a->
n[2];
340 r->
n[3] = 0xFFFFFFFFFFFFFULL * 2 * (
m + 1) - a->
n[3];
341 r->
n[4] = 0x0FFFFFFFFFFFFULL * 2 * (
m + 1) - a->
n[4];
343 r->magnitude =
m + 1;
372 r->magnitude += a->magnitude;
408 uint64_t mask0, mask1;
409 mask0 = flag + ~((uint64_t)0);
411 r->
n[0] = (r->
n[0] & mask0) | (a->
n[0] & mask1);
412 r->
n[1] = (r->
n[1] & mask0) | (a->
n[1] & mask1);
413 r->
n[2] = (r->
n[2] & mask0) | (a->
n[2] & mask1);
414 r->
n[3] = (r->
n[3] & mask0) | (a->
n[3] & mask1);
415 r->
n[4] = (r->
n[4] & mask0) | (a->
n[4] & mask1);
417 if (a->magnitude > r->magnitude) {
418 r->magnitude = a->magnitude;
420 r->normalized &= a->normalized;
425 uint64_t mask0, mask1;
426 mask0 = flag + ~((uint64_t)0);
428 r->
n[0] = (r->
n[0] & mask0) | (a->
n[0] & mask1);
429 r->
n[1] = (r->
n[1] & mask0) | (a->
n[1] & mask1);
430 r->
n[2] = (r->
n[2] & mask0) | (a->
n[2] & mask1);
431 r->
n[3] = (r->
n[3] & mask0) | (a->
n[3] & mask1);
438 r->
n[0] = a->
n[0] | a->
n[1] << 52;
439 r->
n[1] = a->
n[1] >> 12 | a->
n[2] << 40;
440 r->
n[2] = a->
n[2] >> 24 | a->
n[3] << 28;
441 r->
n[3] = a->
n[3] >> 36 | a->
n[4] << 16;
445 r->
n[0] = a->
n[0] & 0xFFFFFFFFFFFFFULL;
446 r->
n[1] = a->
n[0] >> 52 | ((a->
n[1] << 12) & 0xFFFFFFFFFFFFFULL);
447 r->
n[2] = a->
n[1] >> 40 | ((a->
n[2] << 24) & 0xFFFFFFFFFFFFFULL);
448 r->
n[3] = a->
n[2] >> 28 | ((a->
n[3] << 36) & 0xFFFFFFFFFFFFFULL);
449 r->
n[4] = a->
n[3] >> 16;
#define VERIFY_CHECK(cond)
static SECP256K1_INLINE void secp256k1_fe_sqr_inner(uint32_t *r, const uint32_t *a)
static void secp256k1_fe_get_b32(unsigned char *r, const secp256k1_fe *a)
static int secp256k1_fe_set_b32(secp256k1_fe *r, const unsigned char *a)
static SECP256K1_INLINE void secp256k1_fe_cmov(secp256k1_fe *r, const secp256k1_fe *a, int flag)
static void secp256k1_fe_sqr(secp256k1_fe *r, const secp256k1_fe *a)
static void secp256k1_fe_normalize(secp256k1_fe *r)
static SECP256K1_INLINE void secp256k1_fe_mul_inner(uint32_t *r, const uint32_t *a, const uint32_t *SECP256K1_RESTRICT b)
static int secp256k1_fe_normalizes_to_zero(secp256k1_fe *r)
static void secp256k1_fe_verify(const secp256k1_fe *a)
#define SECP256K1_RESTRICT
static void secp256k1_fe_to_storage(secp256k1_fe_storage *r, const secp256k1_fe *a)
static SECP256K1_INLINE void secp256k1_fe_storage_cmov(secp256k1_fe_storage *r, const secp256k1_fe_storage *a, int flag)
static SECP256K1_INLINE void secp256k1_fe_mul_int(secp256k1_fe *r, int a)
static int secp256k1_fe_normalizes_to_zero_var(secp256k1_fe *r)
static SECP256K1_INLINE void secp256k1_fe_negate(secp256k1_fe *r, const secp256k1_fe *a, int m)
static void secp256k1_fe_normalize_weak(secp256k1_fe *r)
static SECP256K1_INLINE void secp256k1_fe_clear(secp256k1_fe *a)
static SECP256K1_INLINE int secp256k1_fe_is_zero(const secp256k1_fe *a)
static int secp256k1_fe_cmp_var(const secp256k1_fe *a, const secp256k1_fe *b)
static void secp256k1_fe_mul(secp256k1_fe *r, const secp256k1_fe *a, const secp256k1_fe *SECP256K1_RESTRICT b)
static void secp256k1_fe_normalize_var(secp256k1_fe *r)
static SECP256K1_INLINE void secp256k1_fe_add(secp256k1_fe *r, const secp256k1_fe *a)
static SECP256K1_INLINE void secp256k1_fe_from_storage(secp256k1_fe *r, const secp256k1_fe_storage *a)
static SECP256K1_INLINE int secp256k1_fe_is_odd(const secp256k1_fe *a)
static SECP256K1_INLINE void secp256k1_fe_set_int(secp256k1_fe *r, int a)