7 #ifndef SECP256K1_FIELD_REPR_IMPL_H 8 #define SECP256K1_FIELD_REPR_IMPL_H 16 static void secp256k1_fe_impl_verify(
const secp256k1_fe *a) {
17 const uint32_t *d = a->
n;
18 int m = a->normalized ? 1 : 2 * a->magnitude;
30 if (d[9] == 0x03FFFFFUL) {
31 uint32_t mid = d[8] & d[7] & d[6] & d[5] & d[4] & d[3] & d[2];
32 if (mid == 0x3FFFFFFUL) {
33 VERIFY_CHECK((d[1] + 0x40UL + ((d[0] + 0x3D1UL) >> 26)) <= 0x3FFFFFFUL);
41 r->
n[0] = 0x3FFFFFFUL * 2 *
m;
42 r->
n[1] = 0x3FFFFFFUL * 2 *
m;
43 r->
n[2] = 0x3FFFFFFUL * 2 *
m;
44 r->
n[3] = 0x3FFFFFFUL * 2 *
m;
45 r->
n[4] = 0x3FFFFFFUL * 2 *
m;
46 r->
n[5] = 0x3FFFFFFUL * 2 *
m;
47 r->
n[6] = 0x3FFFFFFUL * 2 *
m;
48 r->
n[7] = 0x3FFFFFFUL * 2 *
m;
49 r->
n[8] = 0x3FFFFFFUL * 2 *
m;
50 r->
n[9] = 0x03FFFFFUL * 2 *
m;
54 uint32_t t0 = r->
n[0], t1 = r->
n[1], t2 = r->
n[2], t3 = r->
n[3], t4 = r->
n[4],
55 t5 = r->
n[5], t6 = r->
n[6], t7 = r->
n[7], t8 = r->
n[8], t9 = r->
n[9];
59 uint32_t x = t9 >> 22; t9 &= 0x03FFFFFUL;
62 t0 += x * 0x3D1UL; t1 += (x << 6);
63 t1 += (t0 >> 26); t0 &= 0x3FFFFFFUL;
64 t2 += (t1 >> 26); t1 &= 0x3FFFFFFUL;
65 t3 += (t2 >> 26); t2 &= 0x3FFFFFFUL;
m = t2;
66 t4 += (t3 >> 26); t3 &= 0x3FFFFFFUL;
m &= t3;
67 t5 += (t4 >> 26); t4 &= 0x3FFFFFFUL;
m &= t4;
68 t6 += (t5 >> 26); t5 &= 0x3FFFFFFUL;
m &= t5;
69 t7 += (t6 >> 26); t6 &= 0x3FFFFFFUL;
m &= t6;
70 t8 += (t7 >> 26); t7 &= 0x3FFFFFFUL;
m &= t7;
71 t9 += (t8 >> 26); t8 &= 0x3FFFFFFUL;
m &= t8;
77 x = (t9 >> 22) | ((t9 == 0x03FFFFFUL) & (
m == 0x3FFFFFFUL)
78 & ((t1 + 0x40UL + ((t0 + 0x3D1UL) >> 26)) > 0x3FFFFFFUL));
81 t0 += x * 0x3D1UL; t1 += (x << 6);
82 t1 += (t0 >> 26); t0 &= 0x3FFFFFFUL;
83 t2 += (t1 >> 26); t1 &= 0x3FFFFFFUL;
84 t3 += (t2 >> 26); t2 &= 0x3FFFFFFUL;
85 t4 += (t3 >> 26); t3 &= 0x3FFFFFFUL;
86 t5 += (t4 >> 26); t4 &= 0x3FFFFFFUL;
87 t6 += (t5 >> 26); t5 &= 0x3FFFFFFUL;
88 t7 += (t6 >> 26); t6 &= 0x3FFFFFFUL;
89 t8 += (t7 >> 26); t7 &= 0x3FFFFFFUL;
90 t9 += (t8 >> 26); t8 &= 0x3FFFFFFUL;
98 r->
n[0] = t0; r->
n[1] = t1; r->
n[2] = t2; r->
n[3] = t3; r->
n[4] = t4;
99 r->
n[5] = t5; r->
n[6] = t6; r->
n[7] = t7; r->
n[8] = t8; r->
n[9] = t9;
103 uint32_t t0 = r->
n[0], t1 = r->
n[1], t2 = r->
n[2], t3 = r->
n[3], t4 = r->
n[4],
104 t5 = r->
n[5], t6 = r->
n[6], t7 = r->
n[7], t8 = r->
n[8], t9 = r->
n[9];
107 uint32_t x = t9 >> 22; t9 &= 0x03FFFFFUL;
110 t0 += x * 0x3D1UL; t1 += (x << 6);
111 t1 += (t0 >> 26); t0 &= 0x3FFFFFFUL;
112 t2 += (t1 >> 26); t1 &= 0x3FFFFFFUL;
113 t3 += (t2 >> 26); t2 &= 0x3FFFFFFUL;
114 t4 += (t3 >> 26); t3 &= 0x3FFFFFFUL;
115 t5 += (t4 >> 26); t4 &= 0x3FFFFFFUL;
116 t6 += (t5 >> 26); t5 &= 0x3FFFFFFUL;
117 t7 += (t6 >> 26); t6 &= 0x3FFFFFFUL;
118 t8 += (t7 >> 26); t7 &= 0x3FFFFFFUL;
119 t9 += (t8 >> 26); t8 &= 0x3FFFFFFUL;
124 r->
n[0] = t0; r->
n[1] = t1; r->
n[2] = t2; r->
n[3] = t3; r->
n[4] = t4;
125 r->
n[5] = t5; r->
n[6] = t6; r->
n[7] = t7; r->
n[8] = t8; r->
n[9] = t9;
129 uint32_t t0 = r->
n[0], t1 = r->
n[1], t2 = r->
n[2], t3 = r->
n[3], t4 = r->
n[4],
130 t5 = r->
n[5], t6 = r->
n[6], t7 = r->
n[7], t8 = r->
n[8], t9 = r->
n[9];
134 uint32_t x = t9 >> 22; t9 &= 0x03FFFFFUL;
137 t0 += x * 0x3D1UL; t1 += (x << 6);
138 t1 += (t0 >> 26); t0 &= 0x3FFFFFFUL;
139 t2 += (t1 >> 26); t1 &= 0x3FFFFFFUL;
140 t3 += (t2 >> 26); t2 &= 0x3FFFFFFUL;
m = t2;
141 t4 += (t3 >> 26); t3 &= 0x3FFFFFFUL;
m &= t3;
142 t5 += (t4 >> 26); t4 &= 0x3FFFFFFUL;
m &= t4;
143 t6 += (t5 >> 26); t5 &= 0x3FFFFFFUL;
m &= t5;
144 t7 += (t6 >> 26); t6 &= 0x3FFFFFFUL;
m &= t6;
145 t8 += (t7 >> 26); t7 &= 0x3FFFFFFUL;
m &= t7;
146 t9 += (t8 >> 26); t8 &= 0x3FFFFFFUL;
m &= t8;
152 x = (t9 >> 22) | ((t9 == 0x03FFFFFUL) & (
m == 0x3FFFFFFUL)
153 & ((t1 + 0x40UL + ((t0 + 0x3D1UL) >> 26)) > 0x3FFFFFFUL));
156 t0 += 0x3D1UL; t1 += (x << 6);
157 t1 += (t0 >> 26); t0 &= 0x3FFFFFFUL;
158 t2 += (t1 >> 26); t1 &= 0x3FFFFFFUL;
159 t3 += (t2 >> 26); t2 &= 0x3FFFFFFUL;
160 t4 += (t3 >> 26); t3 &= 0x3FFFFFFUL;
161 t5 += (t4 >> 26); t4 &= 0x3FFFFFFUL;
162 t6 += (t5 >> 26); t5 &= 0x3FFFFFFUL;
163 t7 += (t6 >> 26); t6 &= 0x3FFFFFFUL;
164 t8 += (t7 >> 26); t7 &= 0x3FFFFFFUL;
165 t9 += (t8 >> 26); t8 &= 0x3FFFFFFUL;
174 r->
n[0] = t0; r->
n[1] = t1; r->
n[2] = t2; r->
n[3] = t3; r->
n[4] = t4;
175 r->
n[5] = t5; r->
n[6] = t6; r->
n[7] = t7; r->
n[8] = t8; r->
n[9] = t9;
179 uint32_t t0 = r->
n[0], t1 = r->
n[1], t2 = r->
n[2], t3 = r->
n[3], t4 = r->
n[4],
180 t5 = r->
n[5], t6 = r->
n[6], t7 = r->
n[7], t8 = r->
n[8], t9 = r->
n[9];
186 uint32_t x = t9 >> 22; t9 &= 0x03FFFFFUL;
189 t0 += x * 0x3D1UL; t1 += (x << 6);
190 t1 += (t0 >> 26); t0 &= 0x3FFFFFFUL; z0 = t0; z1 = t0 ^ 0x3D0UL;
191 t2 += (t1 >> 26); t1 &= 0x3FFFFFFUL; z0 |= t1; z1 &= t1 ^ 0x40UL;
192 t3 += (t2 >> 26); t2 &= 0x3FFFFFFUL; z0 |= t2; z1 &= t2;
193 t4 += (t3 >> 26); t3 &= 0x3FFFFFFUL; z0 |= t3; z1 &= t3;
194 t5 += (t4 >> 26); t4 &= 0x3FFFFFFUL; z0 |= t4; z1 &= t4;
195 t6 += (t5 >> 26); t5 &= 0x3FFFFFFUL; z0 |= t5; z1 &= t5;
196 t7 += (t6 >> 26); t6 &= 0x3FFFFFFUL; z0 |= t6; z1 &= t6;
197 t8 += (t7 >> 26); t7 &= 0x3FFFFFFUL; z0 |= t7; z1 &= t7;
198 t9 += (t8 >> 26); t8 &= 0x3FFFFFFUL; z0 |= t8; z1 &= t8;
199 z0 |= t9; z1 &= t9 ^ 0x3C00000UL;
204 return (z0 == 0) | (z1 == 0x3FFFFFFUL);
208 uint32_t t0, t1, t2, t3, t4, t5, t6, t7, t8, t9;
222 z0 = t0 & 0x3FFFFFFUL;
226 if ((z0 != 0UL) & (z1 != 0x3FFFFFFUL)) {
243 t2 += (t1 >> 26); t1 &= 0x3FFFFFFUL; z0 |= t1; z1 &= t1 ^ 0x40UL;
244 t3 += (t2 >> 26); t2 &= 0x3FFFFFFUL; z0 |= t2; z1 &= t2;
245 t4 += (t3 >> 26); t3 &= 0x3FFFFFFUL; z0 |= t3; z1 &= t3;
246 t5 += (t4 >> 26); t4 &= 0x3FFFFFFUL; z0 |= t4; z1 &= t4;
247 t6 += (t5 >> 26); t5 &= 0x3FFFFFFUL; z0 |= t5; z1 &= t5;
248 t7 += (t6 >> 26); t6 &= 0x3FFFFFFUL; z0 |= t6; z1 &= t6;
249 t8 += (t7 >> 26); t7 &= 0x3FFFFFFUL; z0 |= t7; z1 &= t7;
250 t9 += (t8 >> 26); t8 &= 0x3FFFFFFUL; z0 |= t8; z1 &= t8;
251 z0 |= t9; z1 &= t9 ^ 0x3C00000UL;
256 return (z0 == 0) | (z1 == 0x3FFFFFFUL);
261 r->
n[1] = r->
n[2] = r->
n[3] = r->
n[4] = r->
n[5] = r->
n[6] = r->
n[7] = r->
n[8] = r->
n[9] = 0;
265 const uint32_t *
t = a->
n;
266 return (
t[0] |
t[1] |
t[2] |
t[3] |
t[4] |
t[5] |
t[6] |
t[7] |
t[8] |
t[9]) == 0;
275 for (i=0; i<10; i++) {
282 for (i = 9; i >= 0; i--) {
283 if (a->
n[i] > b->
n[i]) {
286 if (a->
n[i] < b->
n[i]) {
294 r->
n[0] = (uint32_t)a[31] | ((uint32_t)a[30] << 8) | ((uint32_t)a[29] << 16) | ((uint32_t)(a[28] & 0x3) << 24);
295 r->
n[1] = (uint32_t)((a[28] >> 2) & 0x3f) | ((uint32_t)a[27] << 6) | ((uint32_t)a[26] << 14) | ((uint32_t)(a[25] & 0xf) << 22);
296 r->
n[2] = (uint32_t)((a[25] >> 4) & 0xf) | ((uint32_t)a[24] << 4) | ((uint32_t)a[23] << 12) | ((uint32_t)(a[22] & 0x3f) << 20);
297 r->
n[3] = (uint32_t)((a[22] >> 6) & 0x3) | ((uint32_t)a[21] << 2) | ((uint32_t)a[20] << 10) | ((uint32_t)a[19] << 18);
298 r->
n[4] = (uint32_t)a[18] | ((uint32_t)a[17] << 8) | ((uint32_t)a[16] << 16) | ((uint32_t)(a[15] & 0x3) << 24);
299 r->
n[5] = (uint32_t)((a[15] >> 2) & 0x3f) | ((uint32_t)a[14] << 6) | ((uint32_t)a[13] << 14) | ((uint32_t)(a[12] & 0xf) << 22);
300 r->
n[6] = (uint32_t)((a[12] >> 4) & 0xf) | ((uint32_t)a[11] << 4) | ((uint32_t)a[10] << 12) | ((uint32_t)(a[9] & 0x3f) << 20);
301 r->
n[7] = (uint32_t)((a[9] >> 6) & 0x3) | ((uint32_t)a[8] << 2) | ((uint32_t)a[7] << 10) | ((uint32_t)a[6] << 18);
302 r->
n[8] = (uint32_t)a[5] | ((uint32_t)a[4] << 8) | ((uint32_t)a[3] << 16) | ((uint32_t)(a[2] & 0x3) << 24);
303 r->
n[9] = (uint32_t)((a[2] >> 2) & 0x3f) | ((uint32_t)a[1] << 6) | ((uint32_t)a[0] << 14);
308 return !((r->
n[9] == 0x3FFFFFUL) & ((r->
n[8] & r->
n[7] & r->
n[6] & r->
n[5] & r->
n[4] & r->
n[3] & r->
n[2]) == 0x3FFFFFFUL) & ((r->
n[1] + 0x40UL + ((r->
n[0] + 0x3D1UL) >> 26)) > 0x3FFFFFFUL));
313 r[0] = (a->
n[9] >> 14) & 0xff;
314 r[1] = (a->
n[9] >> 6) & 0xff;
315 r[2] = ((a->
n[9] & 0x3F) << 2) | ((a->
n[8] >> 24) & 0x3);
316 r[3] = (a->
n[8] >> 16) & 0xff;
317 r[4] = (a->
n[8] >> 8) & 0xff;
318 r[5] = a->
n[8] & 0xff;
319 r[6] = (a->
n[7] >> 18) & 0xff;
320 r[7] = (a->
n[7] >> 10) & 0xff;
321 r[8] = (a->
n[7] >> 2) & 0xff;
322 r[9] = ((a->
n[7] & 0x3) << 6) | ((a->
n[6] >> 20) & 0x3f);
323 r[10] = (a->
n[6] >> 12) & 0xff;
324 r[11] = (a->
n[6] >> 4) & 0xff;
325 r[12] = ((a->
n[6] & 0xf) << 4) | ((a->
n[5] >> 22) & 0xf);
326 r[13] = (a->
n[5] >> 14) & 0xff;
327 r[14] = (a->
n[5] >> 6) & 0xff;
328 r[15] = ((a->
n[5] & 0x3f) << 2) | ((a->
n[4] >> 24) & 0x3);
329 r[16] = (a->
n[4] >> 16) & 0xff;
330 r[17] = (a->
n[4] >> 8) & 0xff;
331 r[18] = a->
n[4] & 0xff;
332 r[19] = (a->
n[3] >> 18) & 0xff;
333 r[20] = (a->
n[3] >> 10) & 0xff;
334 r[21] = (a->
n[3] >> 2) & 0xff;
335 r[22] = ((a->
n[3] & 0x3) << 6) | ((a->
n[2] >> 20) & 0x3f);
336 r[23] = (a->
n[2] >> 12) & 0xff;
337 r[24] = (a->
n[2] >> 4) & 0xff;
338 r[25] = ((a->
n[2] & 0xf) << 4) | ((a->
n[1] >> 22) & 0xf);
339 r[26] = (a->
n[1] >> 14) & 0xff;
340 r[27] = (a->
n[1] >> 6) & 0xff;
341 r[28] = ((a->
n[1] & 0x3f) << 2) | ((a->
n[0] >> 24) & 0x3);
342 r[29] = (a->
n[0] >> 16) & 0xff;
343 r[30] = (a->
n[0] >> 8) & 0xff;
344 r[31] = a->
n[0] & 0xff;
356 r->
n[0] = 0x3FFFC2FUL * 2 * (
m + 1) - a->
n[0];
357 r->
n[1] = 0x3FFFFBFUL * 2 * (
m + 1) - a->
n[1];
358 r->
n[2] = 0x3FFFFFFUL * 2 * (
m + 1) - a->
n[2];
359 r->
n[3] = 0x3FFFFFFUL * 2 * (
m + 1) - a->
n[3];
360 r->
n[4] = 0x3FFFFFFUL * 2 * (
m + 1) - a->
n[4];
361 r->
n[5] = 0x3FFFFFFUL * 2 * (
m + 1) - a->
n[5];
362 r->
n[6] = 0x3FFFFFFUL * 2 * (
m + 1) - a->
n[6];
363 r->
n[7] = 0x3FFFFFFUL * 2 * (
m + 1) - a->
n[7];
364 r->
n[8] = 0x3FFFFFFUL * 2 * (
m + 1) - a->
n[8];
365 r->
n[9] = 0x03FFFFFUL * 2 * (
m + 1) - a->
n[9];
398 #if defined(USE_EXTERNAL_ASM) 407 #define VERIFY_BITS(x, n) VERIFY_CHECK(((x) >> (n)) == 0) 409 #define VERIFY_BITS(x, n) do { } while(0) 414 uint64_t u0, u1, u2, u3, u4, u5, u6, u7,
u8;
415 uint32_t t9, t1, t0, t2, t3, t4, t5, t6, t7;
416 const uint32_t
M = 0x3FFFFFFUL, R0 = 0x3D10UL, R1 = 0x400UL;
445 d = (uint64_t)a[0] * b[9]
446 + (uint64_t)a[1] * b[8]
447 + (uint64_t)a[2] * b[7]
448 + (uint64_t)a[3] * b[6]
449 + (uint64_t)a[4] * b[5]
450 + (uint64_t)a[5] * b[4]
451 + (uint64_t)a[6] * b[3]
452 + (uint64_t)a[7] * b[2]
453 + (uint64_t)a[8] * b[1]
454 + (uint64_t)a[9] * b[0];
457 t9 = d &
M; d >>= 26;
462 c = (uint64_t)a[0] * b[0];
465 d += (uint64_t)a[1] * b[9]
466 + (uint64_t)a[2] * b[8]
467 + (uint64_t)a[3] * b[7]
468 + (uint64_t)a[4] * b[6]
469 + (uint64_t)a[5] * b[5]
470 + (uint64_t)a[6] * b[4]
471 + (uint64_t)a[7] * b[3]
472 + (uint64_t)a[8] * b[2]
473 + (uint64_t)a[9] * b[1];
476 u0 = d &
M; d >>= 26; c += u0 * R0;
481 t0 = c &
M; c >>= 26; c += u0 * R1;
487 c += (uint64_t)a[0] * b[1]
488 + (uint64_t)a[1] * b[0];
491 d += (uint64_t)a[2] * b[9]
492 + (uint64_t)a[3] * b[8]
493 + (uint64_t)a[4] * b[7]
494 + (uint64_t)a[5] * b[6]
495 + (uint64_t)a[6] * b[5]
496 + (uint64_t)a[7] * b[4]
497 + (uint64_t)a[8] * b[3]
498 + (uint64_t)a[9] * b[2];
501 u1 = d &
M; d >>= 26; c += u1 * R0;
506 t1 = c &
M; c >>= 26; c += u1 * R1;
512 c += (uint64_t)a[0] * b[2]
513 + (uint64_t)a[1] * b[1]
514 + (uint64_t)a[2] * b[0];
517 d += (uint64_t)a[3] * b[9]
518 + (uint64_t)a[4] * b[8]
519 + (uint64_t)a[5] * b[7]
520 + (uint64_t)a[6] * b[6]
521 + (uint64_t)a[7] * b[5]
522 + (uint64_t)a[8] * b[4]
523 + (uint64_t)a[9] * b[3];
526 u2 = d &
M; d >>= 26; c += u2 * R0;
531 t2 = c &
M; c >>= 26; c += u2 * R1;
537 c += (uint64_t)a[0] * b[3]
538 + (uint64_t)a[1] * b[2]
539 + (uint64_t)a[2] * b[1]
540 + (uint64_t)a[3] * b[0];
543 d += (uint64_t)a[4] * b[9]
544 + (uint64_t)a[5] * b[8]
545 + (uint64_t)a[6] * b[7]
546 + (uint64_t)a[7] * b[6]
547 + (uint64_t)a[8] * b[5]
548 + (uint64_t)a[9] * b[4];
551 u3 = d &
M; d >>= 26; c += u3 * R0;
556 t3 = c &
M; c >>= 26; c += u3 * R1;
562 c += (uint64_t)a[0] * b[4]
563 + (uint64_t)a[1] * b[3]
564 + (uint64_t)a[2] * b[2]
565 + (uint64_t)a[3] * b[1]
566 + (uint64_t)a[4] * b[0];
569 d += (uint64_t)a[5] * b[9]
570 + (uint64_t)a[6] * b[8]
571 + (uint64_t)a[7] * b[7]
572 + (uint64_t)a[8] * b[6]
573 + (uint64_t)a[9] * b[5];
576 u4 = d &
M; d >>= 26; c += u4 * R0;
581 t4 = c &
M; c >>= 26; c += u4 * R1;
587 c += (uint64_t)a[0] * b[5]
588 + (uint64_t)a[1] * b[4]
589 + (uint64_t)a[2] * b[3]
590 + (uint64_t)a[3] * b[2]
591 + (uint64_t)a[4] * b[1]
592 + (uint64_t)a[5] * b[0];
595 d += (uint64_t)a[6] * b[9]
596 + (uint64_t)a[7] * b[8]
597 + (uint64_t)a[8] * b[7]
598 + (uint64_t)a[9] * b[6];
601 u5 = d &
M; d >>= 26; c += u5 * R0;
606 t5 = c &
M; c >>= 26; c += u5 * R1;
612 c += (uint64_t)a[0] * b[6]
613 + (uint64_t)a[1] * b[5]
614 + (uint64_t)a[2] * b[4]
615 + (uint64_t)a[3] * b[3]
616 + (uint64_t)a[4] * b[2]
617 + (uint64_t)a[5] * b[1]
618 + (uint64_t)a[6] * b[0];
621 d += (uint64_t)a[7] * b[9]
622 + (uint64_t)a[8] * b[8]
623 + (uint64_t)a[9] * b[7];
626 u6 = d &
M; d >>= 26; c += u6 * R0;
631 t6 = c &
M; c >>= 26; c += u6 * R1;
637 c += (uint64_t)a[0] * b[7]
638 + (uint64_t)a[1] * b[6]
639 + (uint64_t)a[2] * b[5]
640 + (uint64_t)a[3] * b[4]
641 + (uint64_t)a[4] * b[3]
642 + (uint64_t)a[5] * b[2]
643 + (uint64_t)a[6] * b[1]
644 + (uint64_t)a[7] * b[0];
648 d += (uint64_t)a[8] * b[9]
649 + (uint64_t)a[9] * b[8];
652 u7 = d &
M; d >>= 26; c += u7 * R0;
658 t7 = c &
M; c >>= 26; c += u7 * R1;
664 c += (uint64_t)a[0] * b[8]
665 + (uint64_t)a[1] * b[7]
666 + (uint64_t)a[2] * b[6]
667 + (uint64_t)a[3] * b[5]
668 + (uint64_t)a[4] * b[4]
669 + (uint64_t)a[5] * b[3]
670 + (uint64_t)a[6] * b[2]
671 + (uint64_t)a[7] * b[1]
672 + (uint64_t)a[8] * b[0];
676 d += (uint64_t)a[9] * b[9];
679 u8 = d &
M; d >>= 26; c +=
u8 * R0;
702 r[8] = c &
M; c >>= 26; c +=
u8 * R1;
710 r[9] = c & (
M >> 4); c >>= 22; c += d * (R1 << 4);
717 d = c * (R0 >> 4) + t0;
720 r[0] = d &
M; d >>= 26;
724 d += c * (R1 >> 4) + t1;
729 r[1] = d &
M; d >>= 26;
744 uint64_t u0, u1, u2, u3, u4, u5, u6, u7,
u8;
745 uint32_t t9, t0, t1, t2, t3, t4, t5, t6, t7;
746 const uint32_t
M = 0x3FFFFFFUL, R0 = 0x3D10UL, R1 = 0x400UL;
764 d = (uint64_t)(a[0]*2) * a[9]
765 + (uint64_t)(a[1]*2) * a[8]
766 + (uint64_t)(a[2]*2) * a[7]
767 + (uint64_t)(a[3]*2) * a[6]
768 + (uint64_t)(a[4]*2) * a[5];
771 t9 = d &
M; d >>= 26;
776 c = (uint64_t)a[0] * a[0];
779 d += (uint64_t)(a[1]*2) * a[9]
780 + (uint64_t)(a[2]*2) * a[8]
781 + (uint64_t)(a[3]*2) * a[7]
782 + (uint64_t)(a[4]*2) * a[6]
783 + (uint64_t)a[5] * a[5];
786 u0 = d &
M; d >>= 26; c += u0 * R0;
791 t0 = c &
M; c >>= 26; c += u0 * R1;
797 c += (uint64_t)(a[0]*2) * a[1];
800 d += (uint64_t)(a[2]*2) * a[9]
801 + (uint64_t)(a[3]*2) * a[8]
802 + (uint64_t)(a[4]*2) * a[7]
803 + (uint64_t)(a[5]*2) * a[6];
806 u1 = d &
M; d >>= 26; c += u1 * R0;
811 t1 = c &
M; c >>= 26; c += u1 * R1;
817 c += (uint64_t)(a[0]*2) * a[2]
818 + (uint64_t)a[1] * a[1];
821 d += (uint64_t)(a[3]*2) * a[9]
822 + (uint64_t)(a[4]*2) * a[8]
823 + (uint64_t)(a[5]*2) * a[7]
824 + (uint64_t)a[6] * a[6];
827 u2 = d &
M; d >>= 26; c += u2 * R0;
832 t2 = c &
M; c >>= 26; c += u2 * R1;
838 c += (uint64_t)(a[0]*2) * a[3]
839 + (uint64_t)(a[1]*2) * a[2];
842 d += (uint64_t)(a[4]*2) * a[9]
843 + (uint64_t)(a[5]*2) * a[8]
844 + (uint64_t)(a[6]*2) * a[7];
847 u3 = d &
M; d >>= 26; c += u3 * R0;
852 t3 = c &
M; c >>= 26; c += u3 * R1;
858 c += (uint64_t)(a[0]*2) * a[4]
859 + (uint64_t)(a[1]*2) * a[3]
860 + (uint64_t)a[2] * a[2];
863 d += (uint64_t)(a[5]*2) * a[9]
864 + (uint64_t)(a[6]*2) * a[8]
865 + (uint64_t)a[7] * a[7];
868 u4 = d &
M; d >>= 26; c += u4 * R0;
873 t4 = c &
M; c >>= 26; c += u4 * R1;
879 c += (uint64_t)(a[0]*2) * a[5]
880 + (uint64_t)(a[1]*2) * a[4]
881 + (uint64_t)(a[2]*2) * a[3];
884 d += (uint64_t)(a[6]*2) * a[9]
885 + (uint64_t)(a[7]*2) * a[8];
888 u5 = d &
M; d >>= 26; c += u5 * R0;
893 t5 = c &
M; c >>= 26; c += u5 * R1;
899 c += (uint64_t)(a[0]*2) * a[6]
900 + (uint64_t)(a[1]*2) * a[5]
901 + (uint64_t)(a[2]*2) * a[4]
902 + (uint64_t)a[3] * a[3];
905 d += (uint64_t)(a[7]*2) * a[9]
906 + (uint64_t)a[8] * a[8];
909 u6 = d &
M; d >>= 26; c += u6 * R0;
914 t6 = c &
M; c >>= 26; c += u6 * R1;
920 c += (uint64_t)(a[0]*2) * a[7]
921 + (uint64_t)(a[1]*2) * a[6]
922 + (uint64_t)(a[2]*2) * a[5]
923 + (uint64_t)(a[3]*2) * a[4];
927 d += (uint64_t)(a[8]*2) * a[9];
930 u7 = d &
M; d >>= 26; c += u7 * R0;
936 t7 = c &
M; c >>= 26; c += u7 * R1;
942 c += (uint64_t)(a[0]*2) * a[8]
943 + (uint64_t)(a[1]*2) * a[7]
944 + (uint64_t)(a[2]*2) * a[6]
945 + (uint64_t)(a[3]*2) * a[5]
946 + (uint64_t)a[4] * a[4];
950 d += (uint64_t)a[9] * a[9];
953 u8 = d &
M; d >>= 26; c +=
u8 * R0;
976 r[8] = c &
M; c >>= 26; c +=
u8 * R1;
984 r[9] = c & (
M >> 4); c >>= 22; c += d * (R1 << 4);
991 d = c * (R0 >> 4) + t0;
994 r[0] = d &
M; d >>= 26;
998 d += c * (R1 >> 4) + t1;
1003 r[1] = d &
M; d >>= 26;
1026 uint32_t mask0, mask1;
1027 volatile int vflag = flag;
1029 mask0 = vflag + ~((uint32_t)0);
1031 r->
n[0] = (r->
n[0] & mask0) | (a->
n[0] & mask1);
1032 r->
n[1] = (r->
n[1] & mask0) | (a->
n[1] & mask1);
1033 r->
n[2] = (r->
n[2] & mask0) | (a->
n[2] & mask1);
1034 r->
n[3] = (r->
n[3] & mask0) | (a->
n[3] & mask1);
1035 r->
n[4] = (r->
n[4] & mask0) | (a->
n[4] & mask1);
1036 r->
n[5] = (r->
n[5] & mask0) | (a->
n[5] & mask1);
1037 r->
n[6] = (r->
n[6] & mask0) | (a->
n[6] & mask1);
1038 r->
n[7] = (r->
n[7] & mask0) | (a->
n[7] & mask1);
1039 r->
n[8] = (r->
n[8] & mask0) | (a->
n[8] & mask1);
1040 r->
n[9] = (r->
n[9] & mask0) | (a->
n[9] & mask1);
1044 uint32_t t0 = r->
n[0], t1 = r->
n[1], t2 = r->
n[2], t3 = r->
n[3], t4 = r->
n[4],
1045 t5 = r->
n[5], t6 = r->
n[6], t7 = r->
n[7], t8 = r->
n[8], t9 = r->
n[9];
1046 uint32_t one = (uint32_t)1;
1047 uint32_t mask = -(t0 & one) >> 6;
1059 t0 += 0x3FFFC2FUL & mask;
1060 t1 += 0x3FFFFBFUL & mask;
1079 r->
n[0] = (t0 >> 1) + ((t1 & one) << 25);
1080 r->
n[1] = (t1 >> 1) + ((t2 & one) << 25);
1081 r->
n[2] = (t2 >> 1) + ((t3 & one) << 25);
1082 r->
n[3] = (t3 >> 1) + ((t4 & one) << 25);
1083 r->
n[4] = (t4 >> 1) + ((t5 & one) << 25);
1084 r->
n[5] = (t5 >> 1) + ((t6 & one) << 25);
1085 r->
n[6] = (t6 >> 1) + ((t7 & one) << 25);
1086 r->
n[7] = (t7 >> 1) + ((t8 & one) << 25);
1087 r->
n[8] = (t8 >> 1) + ((t9 & one) << 25);
1088 r->
n[9] = (t9 >> 1);
1109 uint32_t mask0, mask1;
1110 volatile int vflag = flag;
1112 mask0 = vflag + ~((uint32_t)0);
1114 r->
n[0] = (r->
n[0] & mask0) | (a->
n[0] & mask1);
1115 r->
n[1] = (r->
n[1] & mask0) | (a->
n[1] & mask1);
1116 r->
n[2] = (r->
n[2] & mask0) | (a->
n[2] & mask1);
1117 r->
n[3] = (r->
n[3] & mask0) | (a->
n[3] & mask1);
1118 r->
n[4] = (r->
n[4] & mask0) | (a->
n[4] & mask1);
1119 r->
n[5] = (r->
n[5] & mask0) | (a->
n[5] & mask1);
1120 r->
n[6] = (r->
n[6] & mask0) | (a->
n[6] & mask1);
1121 r->
n[7] = (r->
n[7] & mask0) | (a->
n[7] & mask1);
1125 r->
n[0] = a->
n[0] | a->
n[1] << 26;
1126 r->
n[1] = a->
n[1] >> 6 | a->
n[2] << 20;
1127 r->
n[2] = a->
n[2] >> 12 | a->
n[3] << 14;
1128 r->
n[3] = a->
n[3] >> 18 | a->
n[4] << 8;
1129 r->
n[4] = a->
n[4] >> 24 | a->
n[5] << 2 | a->
n[6] << 28;
1130 r->
n[5] = a->
n[6] >> 4 | a->
n[7] << 22;
1131 r->
n[6] = a->
n[7] >> 10 | a->
n[8] << 16;
1132 r->
n[7] = a->
n[8] >> 16 | a->
n[9] << 10;
1136 r->
n[0] = a->
n[0] & 0x3FFFFFFUL;
1137 r->
n[1] = a->
n[0] >> 26 | ((a->
n[1] << 6) & 0x3FFFFFFUL);
1138 r->
n[2] = a->
n[1] >> 20 | ((a->
n[2] << 12) & 0x3FFFFFFUL);
1139 r->
n[3] = a->
n[2] >> 14 | ((a->
n[3] << 18) & 0x3FFFFFFUL);
1140 r->
n[4] = a->
n[3] >> 8 | ((a->
n[4] << 24) & 0x3FFFFFFUL);
1141 r->
n[5] = (a->
n[4] >> 2) & 0x3FFFFFFUL;
1142 r->
n[6] = a->
n[4] >> 28 | ((a->
n[5] << 4) & 0x3FFFFFFUL);
1143 r->
n[7] = a->
n[5] >> 22 | ((a->
n[6] << 10) & 0x3FFFFFFUL);
1144 r->
n[8] = a->
n[6] >> 16 | ((a->
n[7] << 16) & 0x3FFFFFFUL);
1145 r->
n[9] = a->
n[7] >> 10;
1149 const uint32_t M26 = UINT32_MAX >> 6;
1150 const uint32_t a0 = a->
v[0], a1 = a->
v[1], a2 = a->
v[2], a3 = a->
v[3], a4 = a->
v[4],
1151 a5 = a->
v[5], a6 = a->
v[6], a7 = a->
v[7], a8 = a->
v[8];
1167 r->
n[1] = (a0 >> 26 | a1 << 4) & M26;
1168 r->
n[2] = (a1 >> 22 | a2 << 8) & M26;
1169 r->
n[3] = (a2 >> 18 | a3 << 12) & M26;
1170 r->
n[4] = (a3 >> 14 | a4 << 16) & M26;
1171 r->
n[5] = (a4 >> 10 | a5 << 20) & M26;
1172 r->
n[6] = (a5 >> 6 | a6 << 24) & M26;
1173 r->
n[7] = (a6 >> 2 ) & M26;
1174 r->
n[8] = (a6 >> 28 | a7 << 2) & M26;
1175 r->
n[9] = (a7 >> 24 | a8 << 6);
1179 const uint32_t M30 = UINT32_MAX >> 2;
1180 const uint64_t a0 = a->
n[0], a1 = a->
n[1], a2 = a->
n[2], a3 = a->
n[3], a4 = a->
n[4],
1181 a5 = a->
n[5], a6 = a->
n[6], a7 = a->
n[7], a8 = a->
n[8], a9 = a->
n[9];
1183 r->
v[0] = (a0 | a1 << 26) & M30;
1184 r->
v[1] = (a1 >> 4 | a2 << 22) & M30;
1185 r->
v[2] = (a2 >> 8 | a3 << 18) & M30;
1186 r->
v[3] = (a3 >> 12 | a4 << 14) & M30;
1187 r->
v[4] = (a4 >> 16 | a5 << 10) & M30;
1188 r->
v[5] = (a5 >> 20 | a6 << 6) & M30;
1189 r->
v[6] = (a6 >> 24 | a7 << 2
1191 r->
v[7] = (a8 >> 2 | a9 << 24) & M30;
1196 {{-0x3D1, -4, 0, 0, 0, 0, 0, 0, 65536}},
#define VERIFY_CHECK(cond)
static SECP256K1_INLINE void secp256k1_fe_impl_half(secp256k1_fe *r)
static SECP256K1_INLINE void secp256k1_fe_storage_cmov(secp256k1_fe_storage *r, const secp256k1_fe_storage *a, int flag)
This field implementation represents the value as 10 uint32_t limbs in base 2^26. ...
static SECP256K1_INLINE void secp256k1_fe_impl_cmov(secp256k1_fe *r, const secp256k1_fe *a, int flag)
static int secp256k1_fe_sqrt(secp256k1_fe *SECP256K1_RESTRICT r, const secp256k1_fe *SECP256K1_RESTRICT a)
Compute a square root of a field element.
static SECP256K1_INLINE void secp256k1_fe_impl_sqr(secp256k1_fe *r, const secp256k1_fe *a)
static int secp256k1_fe_impl_normalizes_to_zero_var(const secp256k1_fe *r)
#define secp256k1_fe_is_zero
static SECP256K1_INLINE void secp256k1_fe_sqr_inner(uint32_t *r, const uint32_t *a)
static int secp256k1_jacobi32_maybe_var(const secp256k1_modinv32_signed30 *x, const secp256k1_modinv32_modinfo *modinfo)
static int secp256k1_fe_impl_normalizes_to_zero(const secp256k1_fe *r)
static SECP256K1_INLINE void secp256k1_fe_impl_add_int(secp256k1_fe *r, int a)
static SECP256K1_INLINE void secp256k1_fe_impl_set_int(secp256k1_fe *r, int a)
static void secp256k1_fe_impl_set_b32_mod(secp256k1_fe *r, const unsigned char *a)
static void secp256k1_fe_impl_get_bounds(secp256k1_fe *r, int m)
static void secp256k1_fe_impl_normalize(secp256k1_fe *r)
static SECP256K1_INLINE int secp256k1_fe_impl_is_zero(const secp256k1_fe *a)
static void secp256k1_fe_impl_normalize_weak(secp256k1_fe *r)
static void secp256k1_fe_impl_inv_var(secp256k1_fe *r, const secp256k1_fe *x)
static SECP256K1_INLINE void secp256k1_fe_impl_from_storage(secp256k1_fe *r, const secp256k1_fe_storage *a)
static int secp256k1_fe_impl_is_square_var(const secp256k1_fe *x)
static SECP256K1_INLINE void secp256k1_fe_mul_inner(uint32_t *r, const uint32_t *a, const uint32_t *SECP256K1_RESTRICT b)
static int secp256k1_fe_impl_cmp_var(const secp256k1_fe *a, const secp256k1_fe *b)
static SECP256K1_INLINE void secp256k1_fe_impl_mul(secp256k1_fe *r, const secp256k1_fe *a, const secp256k1_fe *SECP256K1_RESTRICT b)
static const secp256k1_modinv32_modinfo secp256k1_const_modinfo_fe
static void secp256k1_fe_impl_normalize_var(secp256k1_fe *r)
static void secp256k1_modinv32_var(secp256k1_modinv32_signed30 *x, const secp256k1_modinv32_modinfo *modinfo)
static SECP256K1_INLINE int secp256k1_fe_impl_is_odd(const secp256k1_fe *a)
static void secp256k1_fe_from_signed30(secp256k1_fe *r, const secp256k1_modinv32_signed30 *a)
#define SECP256K1_RESTRICT
static SECP256K1_INLINE void secp256k1_fe_impl_negate_unchecked(secp256k1_fe *r, const secp256k1_fe *a, int m)
static void secp256k1_fe_impl_to_storage(secp256k1_fe_storage *r, const secp256k1_fe *a)
static void secp256k1_fe_impl_get_b32(unsigned char *r, const secp256k1_fe *a)
Convert a field element to a 32-byte big endian value.
static void secp256k1_fe_impl_inv(secp256k1_fe *r, const secp256k1_fe *x)
static void secp256k1_modinv32(secp256k1_modinv32_signed30 *x, const secp256k1_modinv32_modinfo *modinfo)
#define VERIFY_BITS(x, n)
#define secp256k1_fe_normalize
#define secp256k1_fe_normalize_var
static SECP256K1_INLINE void secp256k1_fe_impl_add(secp256k1_fe *r, const secp256k1_fe *a)
static void secp256k1_fe_to_signed30(secp256k1_modinv32_signed30 *r, const secp256k1_fe *a)
#define SECP256K1_CHECKMEM_CHECK_VERIFY(p, len)
static SECP256K1_INLINE void secp256k1_fe_impl_clear(secp256k1_fe *a)
static SECP256K1_INLINE void secp256k1_fe_impl_mul_int_unchecked(secp256k1_fe *r, int a)
static int secp256k1_fe_impl_set_b32_limit(secp256k1_fe *r, const unsigned char *a)