7 #ifndef SECP256K1_FIELD_REPR_IMPL_H
8 #define SECP256K1_FIELD_REPR_IMPL_H
15 #if defined(USE_ASM_X86_64)
22 static void secp256k1_fe_impl_verify(
const secp256k1_fe *a) {
23 const uint64_t *d = a->
n;
24 int m = a->normalized ? 1 : 2 * a->magnitude;
32 if ((d[4] == 0x0FFFFFFFFFFFFULL) && ((d[3] & d[2] & d[1]) == 0xFFFFFFFFFFFFFULL)) {
40 r->
n[0] = 0xFFFFFFFFFFFFFULL * 2 *
m;
41 r->
n[1] = 0xFFFFFFFFFFFFFULL * 2 *
m;
42 r->
n[2] = 0xFFFFFFFFFFFFFULL * 2 *
m;
43 r->
n[3] = 0xFFFFFFFFFFFFFULL * 2 *
m;
44 r->
n[4] = 0x0FFFFFFFFFFFFULL * 2 *
m;
48 uint64_t t0 = r->
n[0], t1 = r->
n[1], t2 = r->
n[2], t3 = r->
n[3], t4 = r->
n[4];
52 uint64_t x = t4 >> 48; t4 &= 0x0FFFFFFFFFFFFULL;
55 t0 += x * 0x1000003D1ULL;
56 t1 += (t0 >> 52); t0 &= 0xFFFFFFFFFFFFFULL;
57 t2 += (t1 >> 52); t1 &= 0xFFFFFFFFFFFFFULL;
m = t1;
58 t3 += (t2 >> 52); t2 &= 0xFFFFFFFFFFFFFULL;
m &= t2;
59 t4 += (t3 >> 52); t3 &= 0xFFFFFFFFFFFFFULL;
m &= t3;
65 x = (t4 >> 48) | ((t4 == 0x0FFFFFFFFFFFFULL) & (
m == 0xFFFFFFFFFFFFFULL)
66 & (t0 >= 0xFFFFEFFFFFC2FULL));
69 t0 += x * 0x1000003D1ULL;
70 t1 += (t0 >> 52); t0 &= 0xFFFFFFFFFFFFFULL;
71 t2 += (t1 >> 52); t1 &= 0xFFFFFFFFFFFFFULL;
72 t3 += (t2 >> 52); t2 &= 0xFFFFFFFFFFFFFULL;
73 t4 += (t3 >> 52); t3 &= 0xFFFFFFFFFFFFFULL;
79 t4 &= 0x0FFFFFFFFFFFFULL;
81 r->
n[0] = t0; r->
n[1] = t1; r->
n[2] = t2; r->
n[3] = t3; r->
n[4] = t4;
85 uint64_t t0 = r->
n[0], t1 = r->
n[1], t2 = r->
n[2], t3 = r->
n[3], t4 = r->
n[4];
88 uint64_t x = t4 >> 48; t4 &= 0x0FFFFFFFFFFFFULL;
91 t0 += x * 0x1000003D1ULL;
92 t1 += (t0 >> 52); t0 &= 0xFFFFFFFFFFFFFULL;
93 t2 += (t1 >> 52); t1 &= 0xFFFFFFFFFFFFFULL;
94 t3 += (t2 >> 52); t2 &= 0xFFFFFFFFFFFFFULL;
95 t4 += (t3 >> 52); t3 &= 0xFFFFFFFFFFFFFULL;
100 r->
n[0] = t0; r->
n[1] = t1; r->
n[2] = t2; r->
n[3] = t3; r->
n[4] = t4;
104 uint64_t t0 = r->
n[0], t1 = r->
n[1], t2 = r->
n[2], t3 = r->
n[3], t4 = r->
n[4];
108 uint64_t x = t4 >> 48; t4 &= 0x0FFFFFFFFFFFFULL;
111 t0 += x * 0x1000003D1ULL;
112 t1 += (t0 >> 52); t0 &= 0xFFFFFFFFFFFFFULL;
113 t2 += (t1 >> 52); t1 &= 0xFFFFFFFFFFFFFULL;
m = t1;
114 t3 += (t2 >> 52); t2 &= 0xFFFFFFFFFFFFFULL;
m &= t2;
115 t4 += (t3 >> 52); t3 &= 0xFFFFFFFFFFFFFULL;
m &= t3;
121 x = (t4 >> 48) | ((t4 == 0x0FFFFFFFFFFFFULL) & (
m == 0xFFFFFFFFFFFFFULL)
122 & (t0 >= 0xFFFFEFFFFFC2FULL));
125 t0 += 0x1000003D1ULL;
126 t1 += (t0 >> 52); t0 &= 0xFFFFFFFFFFFFFULL;
127 t2 += (t1 >> 52); t1 &= 0xFFFFFFFFFFFFFULL;
128 t3 += (t2 >> 52); t2 &= 0xFFFFFFFFFFFFFULL;
129 t4 += (t3 >> 52); t3 &= 0xFFFFFFFFFFFFFULL;
135 t4 &= 0x0FFFFFFFFFFFFULL;
138 r->
n[0] = t0; r->
n[1] = t1; r->
n[2] = t2; r->
n[3] = t3; r->
n[4] = t4;
142 uint64_t t0 = r->
n[0], t1 = r->
n[1], t2 = r->
n[2], t3 = r->
n[3], t4 = r->
n[4];
148 uint64_t x = t4 >> 48; t4 &= 0x0FFFFFFFFFFFFULL;
151 t0 += x * 0x1000003D1ULL;
152 t1 += (t0 >> 52); t0 &= 0xFFFFFFFFFFFFFULL; z0 = t0; z1 = t0 ^ 0x1000003D0ULL;
153 t2 += (t1 >> 52); t1 &= 0xFFFFFFFFFFFFFULL; z0 |= t1; z1 &= t1;
154 t3 += (t2 >> 52); t2 &= 0xFFFFFFFFFFFFFULL; z0 |= t2; z1 &= t2;
155 t4 += (t3 >> 52); t3 &= 0xFFFFFFFFFFFFFULL; z0 |= t3; z1 &= t3;
156 z0 |= t4; z1 &= t4 ^ 0xF000000000000ULL;
161 return (z0 == 0) | (z1 == 0xFFFFFFFFFFFFFULL);
165 uint64_t t0, t1, t2, t3, t4;
176 t0 += x * 0x1000003D1ULL;
179 z0 = t0 & 0xFFFFFFFFFFFFFULL;
180 z1 = z0 ^ 0x1000003D0ULL;
183 if ((z0 != 0ULL) & (z1 != 0xFFFFFFFFFFFFFULL)) {
191 t4 &= 0x0FFFFFFFFFFFFULL;
194 t2 += (t1 >> 52); t1 &= 0xFFFFFFFFFFFFFULL; z0 |= t1; z1 &= t1;
195 t3 += (t2 >> 52); t2 &= 0xFFFFFFFFFFFFFULL; z0 |= t2; z1 &= t2;
196 t4 += (t3 >> 52); t3 &= 0xFFFFFFFFFFFFFULL; z0 |= t3; z1 &= t3;
197 z0 |= t4; z1 &= t4 ^ 0xF000000000000ULL;
202 return (z0 == 0) | (z1 == 0xFFFFFFFFFFFFFULL);
207 r->
n[1] = r->
n[2] = r->
n[3] = r->
n[4] = 0;
211 const uint64_t *
t = a->
n;
212 return (
t[0] |
t[1] |
t[2] |
t[3] |
t[4]) == 0;
221 for (i=0; i<5; i++) {
228 for (i = 4; i >= 0; i--) {
229 if (a->
n[i] > b->
n[i]) {
232 if (a->
n[i] < b->
n[i]) {
240 r->
n[0] = (uint64_t)a[31]
241 | ((uint64_t)a[30] << 8)
242 | ((uint64_t)a[29] << 16)
243 | ((uint64_t)a[28] << 24)
244 | ((uint64_t)a[27] << 32)
245 | ((uint64_t)a[26] << 40)
246 | ((uint64_t)(a[25] & 0xF) << 48);
247 r->
n[1] = (uint64_t)((a[25] >> 4) & 0xF)
248 | ((uint64_t)a[24] << 4)
249 | ((uint64_t)a[23] << 12)
250 | ((uint64_t)a[22] << 20)
251 | ((uint64_t)a[21] << 28)
252 | ((uint64_t)a[20] << 36)
253 | ((uint64_t)a[19] << 44);
254 r->
n[2] = (uint64_t)a[18]
255 | ((uint64_t)a[17] << 8)
256 | ((uint64_t)a[16] << 16)
257 | ((uint64_t)a[15] << 24)
258 | ((uint64_t)a[14] << 32)
259 | ((uint64_t)a[13] << 40)
260 | ((uint64_t)(a[12] & 0xF) << 48);
261 r->
n[3] = (uint64_t)((a[12] >> 4) & 0xF)
262 | ((uint64_t)a[11] << 4)
263 | ((uint64_t)a[10] << 12)
264 | ((uint64_t)a[9] << 20)
265 | ((uint64_t)a[8] << 28)
266 | ((uint64_t)a[7] << 36)
267 | ((uint64_t)a[6] << 44);
268 r->
n[4] = (uint64_t)a[5]
269 | ((uint64_t)a[4] << 8)
270 | ((uint64_t)a[3] << 16)
271 | ((uint64_t)a[2] << 24)
272 | ((uint64_t)a[1] << 32)
273 | ((uint64_t)a[0] << 40);
278 return !((r->
n[4] == 0x0FFFFFFFFFFFFULL) & ((r->
n[3] & r->
n[2] & r->
n[1]) == 0xFFFFFFFFFFFFFULL) & (r->
n[0] >= 0xFFFFEFFFFFC2FULL));
283 r[0] = (a->
n[4] >> 40) & 0xFF;
284 r[1] = (a->
n[4] >> 32) & 0xFF;
285 r[2] = (a->
n[4] >> 24) & 0xFF;
286 r[3] = (a->
n[4] >> 16) & 0xFF;
287 r[4] = (a->
n[4] >> 8) & 0xFF;
288 r[5] = a->
n[4] & 0xFF;
289 r[6] = (a->
n[3] >> 44) & 0xFF;
290 r[7] = (a->
n[3] >> 36) & 0xFF;
291 r[8] = (a->
n[3] >> 28) & 0xFF;
292 r[9] = (a->
n[3] >> 20) & 0xFF;
293 r[10] = (a->
n[3] >> 12) & 0xFF;
294 r[11] = (a->
n[3] >> 4) & 0xFF;
295 r[12] = ((a->
n[2] >> 48) & 0xF) | ((a->
n[3] & 0xF) << 4);
296 r[13] = (a->
n[2] >> 40) & 0xFF;
297 r[14] = (a->
n[2] >> 32) & 0xFF;
298 r[15] = (a->
n[2] >> 24) & 0xFF;
299 r[16] = (a->
n[2] >> 16) & 0xFF;
300 r[17] = (a->
n[2] >> 8) & 0xFF;
301 r[18] = a->
n[2] & 0xFF;
302 r[19] = (a->
n[1] >> 44) & 0xFF;
303 r[20] = (a->
n[1] >> 36) & 0xFF;
304 r[21] = (a->
n[1] >> 28) & 0xFF;
305 r[22] = (a->
n[1] >> 20) & 0xFF;
306 r[23] = (a->
n[1] >> 12) & 0xFF;
307 r[24] = (a->
n[1] >> 4) & 0xFF;
308 r[25] = ((a->
n[0] >> 48) & 0xF) | ((a->
n[1] & 0xF) << 4);
309 r[26] = (a->
n[0] >> 40) & 0xFF;
310 r[27] = (a->
n[0] >> 32) & 0xFF;
311 r[28] = (a->
n[0] >> 24) & 0xFF;
312 r[29] = (a->
n[0] >> 16) & 0xFF;
313 r[30] = (a->
n[0] >> 8) & 0xFF;
314 r[31] = a->
n[0] & 0xFF;
319 VERIFY_CHECK(0xFFFFEFFFFFC2FULL * 2 * (
m + 1) >= 0xFFFFFFFFFFFFFULL * 2 *
m);
320 VERIFY_CHECK(0xFFFFFFFFFFFFFULL * 2 * (
m + 1) >= 0xFFFFFFFFFFFFFULL * 2 *
m);
321 VERIFY_CHECK(0x0FFFFFFFFFFFFULL * 2 * (
m + 1) >= 0x0FFFFFFFFFFFFULL * 2 *
m);
325 r->
n[0] = 0xFFFFEFFFFFC2FULL * 2 * (
m + 1) - a->
n[0];
326 r->
n[1] = 0xFFFFFFFFFFFFFULL * 2 * (
m + 1) - a->
n[1];
327 r->
n[2] = 0xFFFFFFFFFFFFFULL * 2 * (
m + 1) - a->
n[2];
328 r->
n[3] = 0xFFFFFFFFFFFFFULL * 2 * (
m + 1) - a->
n[3];
329 r->
n[4] = 0x0FFFFFFFFFFFFULL * 2 * (
m + 1) - a->
n[4];
361 uint64_t mask0, mask1;
362 volatile int vflag = flag;
364 mask0 = vflag + ~((uint64_t)0);
366 r->
n[0] = (r->
n[0] & mask0) | (a->
n[0] & mask1);
367 r->
n[1] = (r->
n[1] & mask0) | (a->
n[1] & mask1);
368 r->
n[2] = (r->
n[2] & mask0) | (a->
n[2] & mask1);
369 r->
n[3] = (r->
n[3] & mask0) | (a->
n[3] & mask1);
370 r->
n[4] = (r->
n[4] & mask0) | (a->
n[4] & mask1);
374 uint64_t t0 = r->
n[0], t1 = r->
n[1], t2 = r->
n[2], t3 = r->
n[3], t4 = r->
n[4];
375 uint64_t one = (uint64_t)1;
376 uint64_t mask = -(t0 & one) >> 12;
388 t0 += 0xFFFFEFFFFFC2FULL & mask;
403 r->
n[0] = (t0 >> 1) + ((t1 & one) << 51);
404 r->
n[1] = (t1 >> 1) + ((t2 & one) << 51);
405 r->
n[2] = (t2 >> 1) + ((t3 & one) << 51);
406 r->
n[3] = (t3 >> 1) + ((t4 & one) << 51);
428 uint64_t mask0, mask1;
429 volatile int vflag = flag;
431 mask0 = vflag + ~((uint64_t)0);
433 r->
n[0] = (r->
n[0] & mask0) | (a->
n[0] & mask1);
434 r->
n[1] = (r->
n[1] & mask0) | (a->
n[1] & mask1);
435 r->
n[2] = (r->
n[2] & mask0) | (a->
n[2] & mask1);
436 r->
n[3] = (r->
n[3] & mask0) | (a->
n[3] & mask1);
440 r->
n[0] = a->
n[0] | a->
n[1] << 52;
441 r->
n[1] = a->
n[1] >> 12 | a->
n[2] << 40;
442 r->
n[2] = a->
n[2] >> 24 | a->
n[3] << 28;
443 r->
n[3] = a->
n[3] >> 36 | a->
n[4] << 16;
447 r->
n[0] = a->
n[0] & 0xFFFFFFFFFFFFFULL;
448 r->
n[1] = a->
n[0] >> 52 | ((a->
n[1] << 12) & 0xFFFFFFFFFFFFFULL);
449 r->
n[2] = a->
n[1] >> 40 | ((a->
n[2] << 24) & 0xFFFFFFFFFFFFFULL);
450 r->
n[3] = a->
n[2] >> 28 | ((a->
n[3] << 36) & 0xFFFFFFFFFFFFFULL);
451 r->
n[4] = a->
n[3] >> 16;
455 const uint64_t M52 = UINT64_MAX >> 12;
456 const uint64_t a0 = a->
v[0], a1 = a->
v[1], a2 = a->
v[2], a3 = a->
v[3], a4 = a->
v[4];
468 r->
n[1] = (a0 >> 52 | a1 << 10) & M52;
469 r->
n[2] = (a1 >> 42 | a2 << 20) & M52;
470 r->
n[3] = (a2 >> 32 | a3 << 30) & M52;
471 r->
n[4] = (a3 >> 22 | a4 << 40);
475 const uint64_t M62 = UINT64_MAX >> 2;
476 const uint64_t a0 = a->
n[0], a1 = a->
n[1], a2 = a->
n[2], a3 = a->
n[3], a4 = a->
n[4];
478 r->
v[0] = (a0 | a1 << 52) & M62;
479 r->
v[1] = (a1 >> 10 | a2 << 42) & M62;
480 r->
v[2] = (a2 >> 20 | a3 << 32) & M62;
481 r->
v[3] = (a3 >> 30 | a4 << 22) & M62;
486 {{-0x1000003D1LL, 0, 0, 0, 256}},
#define SECP256K1_CHECKMEM_CHECK_VERIFY(p, len)
static int secp256k1_fe_sqrt(secp256k1_fe *SECP256K1_RESTRICT r, const secp256k1_fe *SECP256K1_RESTRICT a)
Compute a square root of a field element.
#define secp256k1_fe_normalize_var
#define secp256k1_fe_is_zero
#define secp256k1_fe_normalize
static SECP256K1_INLINE void secp256k1_fe_sqr_inner(uint32_t *r, const uint32_t *a)
static SECP256K1_INLINE void secp256k1_fe_mul_inner(uint32_t *r, const uint32_t *a, const uint32_t *SECP256K1_RESTRICT b)
static SECP256K1_INLINE void secp256k1_fe_impl_half(secp256k1_fe *r)
static void secp256k1_fe_impl_set_b32_mod(secp256k1_fe *r, const unsigned char *a)
static void secp256k1_fe_impl_normalize_weak(secp256k1_fe *r)
static int secp256k1_fe_impl_is_square_var(const secp256k1_fe *x)
static void secp256k1_fe_impl_get_b32(unsigned char *r, const secp256k1_fe *a)
Convert a field element to a 32-byte big endian value.
static SECP256K1_INLINE void secp256k1_fe_impl_add(secp256k1_fe *r, const secp256k1_fe *a)
static SECP256K1_INLINE void secp256k1_fe_impl_clear(secp256k1_fe *a)
static SECP256K1_INLINE void secp256k1_fe_impl_set_int(secp256k1_fe *r, int a)
static SECP256K1_INLINE int secp256k1_fe_impl_is_zero(const secp256k1_fe *a)
static void secp256k1_fe_impl_get_bounds(secp256k1_fe *r, int m)
static int secp256k1_fe_impl_set_b32_limit(secp256k1_fe *r, const unsigned char *a)
static SECP256K1_INLINE void secp256k1_fe_storage_cmov(secp256k1_fe_storage *r, const secp256k1_fe_storage *a, int flag)
static SECP256K1_INLINE void secp256k1_fe_impl_negate_unchecked(secp256k1_fe *r, const secp256k1_fe *a, int m)
static SECP256K1_INLINE void secp256k1_fe_impl_mul_int_unchecked(secp256k1_fe *r, int a)
static int secp256k1_fe_impl_cmp_var(const secp256k1_fe *a, const secp256k1_fe *b)
static void secp256k1_fe_to_signed62(secp256k1_modinv64_signed62 *r, const secp256k1_fe *a)
static int secp256k1_fe_impl_normalizes_to_zero(const secp256k1_fe *r)
static void secp256k1_fe_impl_inv_var(secp256k1_fe *r, const secp256k1_fe *x)
static SECP256K1_INLINE void secp256k1_fe_impl_sqr(secp256k1_fe *r, const secp256k1_fe *a)
static SECP256K1_INLINE void secp256k1_fe_impl_from_storage(secp256k1_fe *r, const secp256k1_fe_storage *a)
static void secp256k1_fe_impl_to_storage(secp256k1_fe_storage *r, const secp256k1_fe *a)
static SECP256K1_INLINE void secp256k1_fe_impl_add_int(secp256k1_fe *r, int a)
static int secp256k1_fe_impl_normalizes_to_zero_var(const secp256k1_fe *r)
static void secp256k1_fe_impl_normalize(secp256k1_fe *r)
static SECP256K1_INLINE void secp256k1_fe_impl_cmov(secp256k1_fe *r, const secp256k1_fe *a, int flag)
static void secp256k1_fe_impl_inv(secp256k1_fe *r, const secp256k1_fe *x)
static void secp256k1_fe_impl_normalize_var(secp256k1_fe *r)
static void secp256k1_fe_from_signed62(secp256k1_fe *r, const secp256k1_modinv64_signed62 *a)
static const secp256k1_modinv64_modinfo secp256k1_const_modinfo_fe
static SECP256K1_INLINE int secp256k1_fe_impl_is_odd(const secp256k1_fe *a)
static SECP256K1_INLINE void secp256k1_fe_impl_mul(secp256k1_fe *r, const secp256k1_fe *a, const secp256k1_fe *SECP256K1_RESTRICT b)
static void secp256k1_modinv64(secp256k1_modinv64_signed62 *x, const secp256k1_modinv64_modinfo *modinfo)
static void secp256k1_modinv64_var(secp256k1_modinv64_signed62 *x, const secp256k1_modinv64_modinfo *modinfo)
static int secp256k1_jacobi64_maybe_var(const secp256k1_modinv64_signed62 *x, const secp256k1_modinv64_modinfo *modinfo)
#define VERIFY_CHECK(cond)
#define SECP256K1_RESTRICT
This field implementation represents the value as 10 uint32_t limbs in base 2^26.