tls: P256: sp_256_sub_8_p256_mod always subtracts in-place, use that
i386:
function old new delta
sp_256_mont_reduce_8 245 243 -2
sp_256_mont_dbl_8 26 24 -2
sp_256_ecc_mulmod_8 1161 1157 -4
sp_256_proj_point_dbl_8 359 353 -6
sp_256_sub_8_p256_mod 71 32 -39
------------------------------------------------------------------------------
(add/remove: 0/0 grow/shrink: 0/5 up/down: 0/-53) Total: -53 bytes
non-asm code:
function old new delta
sp_256_sub_8_p256_mod - 12 +12
sp_256_mont_reduce_8 250 243 -7
sp_256_mont_dbl_8 31 24 -7
sp_256_ecc_mulmod_8 1171 1157 -14
sp_256_proj_point_dbl_8 374 353 -21
------------------------------------------------------------------------------
(add/remove: 1/0 grow/shrink: 0/4 up/down: 12/-49) Total: -37 bytes
Signed-off-by: Denys Vlasenko <vda.linux@googlemail.com>
diff --git a/networking/tls_sp_c32.c b/networking/tls_sp_c32.c
index 1391cb4..b3828d8 100644
--- a/networking/tls_sp_c32.c
+++ b/networking/tls_sp_c32.c
@@ -346,82 +346,55 @@
#endif
}
-/* Sub p256_mod from a into r. (r = a - p256_mod). */
+/* Sub p256_mod from r. (r = r - p256_mod). */
#if ALLOW_ASM && defined(__GNUC__) && defined(__i386__)
-static void sp_256_sub_8_p256_mod(sp_digit* r, const sp_digit* a)
+static void sp_256_sub_8_p256_mod(sp_digit* r)
{
- sp_digit reg;
//p256_mod[7..0] = ffffffff 00000001 00000000 00000000 00000000 ffffffff ffffffff ffffffff
asm volatile (
-"\n movl (%0), %2"
-"\n subl $0xffffffff, %2"
-"\n movl %2, (%1)"
+"\n subl $0xffffffff, (%0)"
+"\n sbbl $0xffffffff, 1*4(%0)"
+"\n sbbl $0xffffffff, 2*4(%0)"
+"\n sbbl $0, 3*4(%0)"
+"\n sbbl $0, 4*4(%0)"
+"\n sbbl $0, 5*4(%0)"
+"\n sbbl $1, 6*4(%0)"
+"\n sbbl $0xffffffff, 7*4(%0)"
"\n"
-"\n movl 1*4(%0), %2"
-"\n sbbl $0xffffffff, %2"
-"\n movl %2, 1*4(%1)"
-"\n"
-"\n movl 2*4(%0), %2"
-"\n sbbl $0xffffffff, %2"
-"\n movl %2, 2*4(%1)"
-"\n"
-"\n movl 3*4(%0), %2"
-"\n sbbl $0, %2"
-"\n movl %2, 3*4(%1)"
-"\n"
-"\n movl 4*4(%0), %2"
-"\n sbbl $0, %2"
-"\n movl %2, 4*4(%1)"
-"\n"
-"\n movl 5*4(%0), %2"
-"\n sbbl $0, %2"
-"\n movl %2, 5*4(%1)"
-"\n"
-"\n movl 6*4(%0), %2"
-"\n sbbl $1, %2"
-"\n movl %2, 6*4(%1)"
-"\n"
-"\n movl 7*4(%0), %2"
-"\n sbbl $0xffffffff, %2"
-"\n movl %2, 7*4(%1)"
-"\n"
- : "=r" (a), "=r" (r), "=r" (reg)
- : "0" (a), "1" (r)
+ : "=r" (r)
+ : "0" (r)
: "memory"
);
}
#elif ALLOW_ASM && defined(__GNUC__) && defined(__x86_64__)
-static void sp_256_sub_8_p256_mod(sp_digit* r, const sp_digit* a)
+static void sp_256_sub_8_p256_mod(sp_digit* r)
{
uint64_t reg;
uint64_t ooff;
//p256_mod[3..0] = ffffffff00000001 0000000000000000 00000000ffffffff ffffffffffffffff
asm volatile (
-"\n movq (%0), %3"
-"\n addq $1, %3" // adding 1 is the same as subtracting ffffffffffffffff
-"\n movq %3, (%1)" //
+"\n addq $1, (%0)" // adding 1 is the same as subtracting ffffffffffffffff
"\n cmc" // only carry bit needs inverting
+
+"\n sbbq %1, 1*8(%0)" // %1 holds 00000000ffffffff
+
+"\n sbbq $0, 2*8(%0)"
"\n"
-"\n movq 1*8(%0), %3"
-"\n sbbq %2, %3" // %2 holds 00000000ffffffff
-"\n movq %3, 1*8(%1)"
+"\n movq 3*8(%0), %2"
+"\n sbbq $0, %2" // adding 00000000ffffffff (in %1)
+"\n addq %1, %2" // is the same as subtracting ffffffff00000001
+"\n movq %2, 3*8(%0)"
"\n"
-"\n movq 2*8(%0), %3"
-"\n sbbq $0, %3"
-"\n movq %3, 2*8(%1)"
-"\n"
-"\n movq 3*8(%0), %3"
-"\n sbbq $0, %3" // adding 00000000ffffffff (in %2)
-"\n addq %2, %3" // is the same as subtracting ffffffff00000001
-"\n movq %3, 3*8(%1)"
-"\n"
- : "=r" (a), "=r" (r), "=r" (ooff), "=r" (reg)
- : "0" (a), "1" (r), "2" (0x00000000ffffffff)
+ : "=r" (r), "=r" (ooff), "=r" (reg)
+ : "0" (r), "1" (0x00000000ffffffff)
: "memory"
);
}
#else
-# define sp_256_sub_8_p256_mod(r, a) sp_256_sub_8((r), (a), p256_mod)
+static void sp_256_sub_8_p256_mod(sp_digit* r)
+{
+ sp_256_sub_8(r, r, p256_mod);
+}
#endif
/* Multiply a and b into r. (r = a * b) */
@@ -609,7 +582,7 @@
int carry = sp_256_add_8(r, a, b);
sp_256_norm_8(r);
if (carry) {
- sp_256_sub_8_p256_mod(r, r /*, m*/);
+ sp_256_sub_8_p256_mod(r);
sp_256_norm_8(r);
}
}
@@ -637,7 +610,7 @@
int carry = sp_256_add_8(r, a, a);
sp_256_norm_8(r);
if (carry)
- sp_256_sub_8_p256_mod(r, r /*, m*/);
+ sp_256_sub_8_p256_mod(r);
sp_256_norm_8(r);
}
@@ -649,13 +622,13 @@
int carry = sp_256_add_8(r, a, a);
sp_256_norm_8(r);
if (carry) {
- sp_256_sub_8_p256_mod(r, r /*, m*/);
+ sp_256_sub_8_p256_mod(r);
sp_256_norm_8(r);
}
carry = sp_256_add_8(r, r, a);
sp_256_norm_8(r);
if (carry) {
- sp_256_sub_8_p256_mod(r, r /*, m*/);
+ sp_256_sub_8_p256_mod(r);
sp_256_norm_8(r);
}
}
@@ -829,7 +802,7 @@
}
sp_256_mont_shift_8(a, a);
if (word16th != 0)
- sp_256_sub_8_p256_mod(a, a /*, m*/);
+ sp_256_sub_8_p256_mod(a);
sp_256_norm_8(a);
}
else { /* Same code for explicit mp == 1 (which is always the case for P256) */
@@ -849,7 +822,7 @@
}
sp_256_mont_shift_8(a, a);
if (word16th != 0)
- sp_256_sub_8_p256_mod(a, a /*, m*/);
+ sp_256_sub_8_p256_mod(a);
sp_256_norm_8(a);
}
}
@@ -1104,7 +1077,7 @@
sp_256_mont_reduce_8(r->x /*, p256_mod, p256_mp_mod*/);
/* Reduce x to less than modulus */
if (sp_256_cmp_8(r->x, p256_mod) >= 0)
- sp_256_sub_8_p256_mod(r->x, r->x /*, p256_mod*/);
+ sp_256_sub_8_p256_mod(r->x);
sp_256_norm_8(r->x);
/* y /= z^3 */
@@ -1113,7 +1086,7 @@
sp_256_mont_reduce_8(r->y /*, p256_mod, p256_mp_mod*/);
/* Reduce y to less than modulus */
if (sp_256_cmp_8(r->y, p256_mod) >= 0)
- sp_256_sub_8_p256_mod(r->y, r->y /*, p256_mod*/);
+ sp_256_sub_8_p256_mod(r->y);
sp_256_norm_8(r->y);
memset(r->z, 0, sizeof(r->z));