diff options
author | heretic <heretic@yandex-team.ru> | 2022-02-10 16:45:46 +0300 |
---|---|---|
committer | Daniil Cherednik <dcherednik@yandex-team.ru> | 2022-02-10 16:45:46 +0300 |
commit | 81eddc8c0b55990194e112b02d127b87d54164a9 (patch) | |
tree | 9142afc54d335ea52910662635b898e79e192e49 /contrib/libs/openssl/asm/aarch64 | |
parent | 397cbe258b9e064f49c4ca575279f02f39fef76e (diff) | |
download | ydb-81eddc8c0b55990194e112b02d127b87d54164a9.tar.gz |
Restoring authorship annotation for <heretic@yandex-team.ru>. Commit 2 of 2.
Diffstat (limited to 'contrib/libs/openssl/asm/aarch64')
8 files changed, 79 insertions, 79 deletions
diff --git a/contrib/libs/openssl/asm/aarch64/crypto/aes/aesv8-armx.S b/contrib/libs/openssl/asm/aarch64/crypto/aes/aesv8-armx.S index 1006d71f8d..433b8020af 100644 --- a/contrib/libs/openssl/asm/aarch64/crypto/aes/aesv8-armx.S +++ b/contrib/libs/openssl/asm/aarch64/crypto/aes/aesv8-armx.S @@ -103,12 +103,12 @@ aes_v8_set_encrypt_key: .Loop192: tbl v6.16b,{v4.16b},v2.16b ext v5.16b,v0.16b,v3.16b,#12 -#ifdef __ARMEB__ - st1 {v4.4s},[x2],#16 - sub x2,x2,#8 -#else +#ifdef __ARMEB__ + st1 {v4.4s},[x2],#16 + sub x2,x2,#8 +#else st1 {v4.8b},[x2],#8 -#endif +#endif aese v6.16b,v0.16b subs w1,w1,#1 @@ -579,11 +579,11 @@ aes_v8_ctr32_encrypt_blocks: ldr w5,[x3,#240] ldr w8, [x4, #12] -#ifdef __ARMEB__ - ld1 {v0.16b},[x4] -#else +#ifdef __ARMEB__ + ld1 {v0.16b},[x4] +#else ld1 {v0.4s},[x4] -#endif +#endif ld1 {v16.4s,v17.4s},[x3] // load key schedule... sub w5,w5,#4 mov x12,#16 @@ -602,14 +602,14 @@ aes_v8_ctr32_encrypt_blocks: add w10, w8, #1 orr v6.16b,v0.16b,v0.16b rev w10, w10 - mov v6.s[3],w10 - add w8, w8, #2 - orr v1.16b,v6.16b,v6.16b + mov v6.s[3],w10 + add w8, w8, #2 + orr v1.16b,v6.16b,v6.16b b.ls .Lctr32_tail rev w12, w8 - mov v6.s[3],w12 + mov v6.s[3],w12 sub x2,x2,#3 // bias - orr v18.16b,v6.16b,v6.16b + orr v18.16b,v6.16b,v6.16b b .Loop3x_ctr32 .align 4 @@ -636,11 +636,11 @@ aes_v8_ctr32_encrypt_blocks: aese v1.16b,v16.16b aesmc v5.16b,v1.16b ld1 {v2.16b},[x0],#16 - add w9,w8,#1 + add w9,w8,#1 aese v18.16b,v16.16b aesmc v18.16b,v18.16b ld1 {v3.16b},[x0],#16 - rev w9,w9 + rev w9,w9 aese v4.16b,v17.16b aesmc v4.16b,v4.16b aese v5.16b,v17.16b @@ -664,22 +664,22 @@ aes_v8_ctr32_encrypt_blocks: aese v5.16b,v21.16b aesmc v5.16b,v5.16b eor v19.16b,v19.16b,v7.16b - mov v6.s[3], w9 + mov v6.s[3], w9 aese v17.16b,v21.16b aesmc v17.16b,v17.16b - orr v0.16b,v6.16b,v6.16b + orr v0.16b,v6.16b,v6.16b rev w10,w10 aese v4.16b,v22.16b aesmc v4.16b,v4.16b - mov v6.s[3], w10 - rev w12,w8 + mov v6.s[3], w10 + rev w12,w8 aese v5.16b,v22.16b aesmc v5.16b,v5.16b - orr v1.16b,v6.16b,v6.16b - mov v6.s[3], w12 + orr v1.16b,v6.16b,v6.16b + mov v6.s[3], w12 aese v17.16b,v22.16b aesmc v17.16b,v17.16b - orr v18.16b,v6.16b,v6.16b + orr v18.16b,v6.16b,v6.16b subs x2,x2,#3 aese v4.16b,v23.16b aese v5.16b,v23.16b diff --git a/contrib/libs/openssl/asm/aarch64/crypto/chacha/chacha-armv8.S b/contrib/libs/openssl/asm/aarch64/crypto/chacha/chacha-armv8.S index 0bccbaa665..f4676cbf68 100644 --- a/contrib/libs/openssl/asm/aarch64/crypto/chacha/chacha-armv8.S +++ b/contrib/libs/openssl/asm/aarch64/crypto/chacha/chacha-armv8.S @@ -3,7 +3,7 @@ .text -.hidden OPENSSL_armcap_P +.hidden OPENSSL_armcap_P .align 5 .Lsigma: diff --git a/contrib/libs/openssl/asm/aarch64/crypto/ec/ecp_nistz256-armv8.S b/contrib/libs/openssl/asm/aarch64/crypto/ec/ecp_nistz256-armv8.S index 28749d5bcf..7caa581608 100644 --- a/contrib/libs/openssl/asm/aarch64/crypto/ec/ecp_nistz256-armv8.S +++ b/contrib/libs/openssl/asm/aarch64/crypto/ec/ecp_nistz256-armv8.S @@ -3015,7 +3015,7 @@ __ecp_nistz256_div_by_2: .align 5 ecp_nistz256_point_double: .inst 0xd503233f // paciasp - stp x29,x30,[sp,#-96]! + stp x29,x30,[sp,#-96]! add x29,sp,#0 stp x19,x20,[sp,#16] stp x21,x22,[sp,#32] @@ -3148,7 +3148,7 @@ ecp_nistz256_point_double: add sp,x29,#0 // destroy frame ldp x19,x20,[x29,#16] ldp x21,x22,[x29,#32] - ldp x29,x30,[sp],#96 + ldp x29,x30,[sp],#96 .inst 0xd50323bf // autiasp ret .size ecp_nistz256_point_double,.-ecp_nistz256_point_double @@ -3157,13 +3157,13 @@ ecp_nistz256_point_double: .align 5 ecp_nistz256_point_add: .inst 0xd503233f // paciasp - stp x29,x30,[sp,#-96]! + stp x29,x30,[sp,#-96]! add x29,sp,#0 stp x19,x20,[sp,#16] stp x21,x22,[sp,#32] stp x23,x24,[sp,#48] stp x25,x26,[sp,#64] - stp x27,x28,[sp,#80] + stp x27,x28,[sp,#80] sub sp,sp,#32*12 ldp x4,x5,[x2,#64] // in2_z @@ -3177,7 +3177,7 @@ ecp_nistz256_point_add: orr x10,x6,x7 orr x25,x8,x10 cmp x25,#0 - csetm x25,ne // ~in2infty + csetm x25,ne // ~in2infty add x0,sp,#192 bl __ecp_nistz256_sqr_mont // p256_sqr_mont(Z2sqr, in2_z); @@ -3187,7 +3187,7 @@ ecp_nistz256_point_add: orr x10,x6,x7 orr x24,x8,x10 cmp x24,#0 - csetm x24,ne // ~in1infty + csetm x24,ne // ~in1infty add x0,sp,#128 bl __ecp_nistz256_sqr_mont // p256_sqr_mont(Z1sqr, in1_z); @@ -3228,7 +3228,7 @@ ecp_nistz256_point_add: orr x14,x14,x15 // see if result is zero orr x16,x16,x17 - orr x26,x14,x16 // ~is_equal(S1,S2) + orr x26,x14,x16 // ~is_equal(S1,S2) add x2,sp,#192 add x0,sp,#256 @@ -3249,21 +3249,21 @@ ecp_nistz256_point_add: orr x14,x14,x15 // see if result is zero orr x16,x16,x17 - orr x14,x14,x16 // ~is_equal(U1,U2) + orr x14,x14,x16 // ~is_equal(U1,U2) - mvn x27,x24 // -1/0 -> 0/-1 - mvn x28,x25 // -1/0 -> 0/-1 - orr x14,x14,x27 - orr x14,x14,x28 - orr x14,x14,x26 - cbnz x14,.Ladd_proceed // if(~is_equal(U1,U2) | in1infty | in2infty | ~is_equal(S1,S2)) + mvn x27,x24 // -1/0 -> 0/-1 + mvn x28,x25 // -1/0 -> 0/-1 + orr x14,x14,x27 + orr x14,x14,x28 + orr x14,x14,x26 + cbnz x14,.Ladd_proceed // if(~is_equal(U1,U2) | in1infty | in2infty | ~is_equal(S1,S2)) .Ladd_double: mov x1,x22 mov x0,x21 ldp x23,x24,[x29,#48] ldp x25,x26,[x29,#64] - ldp x27,x28,[x29,#80] + ldp x27,x28,[x29,#80] add sp,sp,#32*(12-4) // difference in stack frames b .Ldouble_shortcut @@ -3345,14 +3345,14 @@ ecp_nistz256_point_add: ldp x8,x9,[x23] // in2 ldp x10,x11,[x23,#16] ldp x14,x15,[x22,#0] // in1 - cmp x24,#0 // ~, remember? + cmp x24,#0 // ~, remember? ldp x16,x17,[x22,#0+16] csel x8,x4,x8,ne csel x9,x5,x9,ne ldp x4,x5,[sp,#0+0+32] // res csel x10,x6,x10,ne csel x11,x7,x11,ne - cmp x25,#0 // ~, remember? + cmp x25,#0 // ~, remember? ldp x6,x7,[sp,#0+0+48] csel x14,x8,x14,ne csel x15,x9,x15,ne @@ -3363,14 +3363,14 @@ ecp_nistz256_point_add: stp x14,x15,[x21,#0] stp x16,x17,[x21,#0+16] ldp x14,x15,[x22,#32] // in1 - cmp x24,#0 // ~, remember? + cmp x24,#0 // ~, remember? ldp x16,x17,[x22,#32+16] csel x8,x4,x8,ne csel x9,x5,x9,ne ldp x4,x5,[sp,#0+32+32] // res csel x10,x6,x10,ne csel x11,x7,x11,ne - cmp x25,#0 // ~, remember? + cmp x25,#0 // ~, remember? ldp x6,x7,[sp,#0+32+48] csel x14,x8,x14,ne csel x15,x9,x15,ne @@ -3381,13 +3381,13 @@ ecp_nistz256_point_add: stp x14,x15,[x21,#32] stp x16,x17,[x21,#32+16] ldp x14,x15,[x22,#64] // in1 - cmp x24,#0 // ~, remember? + cmp x24,#0 // ~, remember? ldp x16,x17,[x22,#64+16] csel x8,x4,x8,ne csel x9,x5,x9,ne csel x10,x6,x10,ne csel x11,x7,x11,ne - cmp x25,#0 // ~, remember? + cmp x25,#0 // ~, remember? csel x14,x8,x14,ne csel x15,x9,x15,ne csel x16,x10,x16,ne @@ -3401,8 +3401,8 @@ ecp_nistz256_point_add: ldp x21,x22,[x29,#32] ldp x23,x24,[x29,#48] ldp x25,x26,[x29,#64] - ldp x27,x28,[x29,#80] - ldp x29,x30,[sp],#96 + ldp x27,x28,[x29,#80] + ldp x29,x30,[sp],#96 .inst 0xd50323bf // autiasp ret .size ecp_nistz256_point_add,.-ecp_nistz256_point_add @@ -3431,7 +3431,7 @@ ecp_nistz256_point_add_affine: orr x10,x6,x7 orr x24,x8,x10 cmp x24,#0 - csetm x24,ne // ~in1infty + csetm x24,ne // ~in1infty ldp x14,x15,[x2] // in2_x ldp x16,x17,[x2,#16] @@ -3445,7 +3445,7 @@ ecp_nistz256_point_add_affine: orr x8,x8,x10 orr x25,x14,x8 cmp x25,#0 - csetm x25,ne // ~in2infty + csetm x25,ne // ~in2infty add x0,sp,#128 bl __ecp_nistz256_sqr_mont // p256_sqr_mont(Z1sqr, in1_z); @@ -3552,14 +3552,14 @@ ecp_nistz256_point_add_affine: ldp x8,x9,[x23] // in2 ldp x10,x11,[x23,#16] ldp x14,x15,[x22,#0] // in1 - cmp x24,#0 // ~, remember? + cmp x24,#0 // ~, remember? ldp x16,x17,[x22,#0+16] csel x8,x4,x8,ne csel x9,x5,x9,ne ldp x4,x5,[sp,#0+0+32] // res csel x10,x6,x10,ne csel x11,x7,x11,ne - cmp x25,#0 // ~, remember? + cmp x25,#0 // ~, remember? ldp x6,x7,[sp,#0+0+48] csel x14,x8,x14,ne csel x15,x9,x15,ne @@ -3571,14 +3571,14 @@ ecp_nistz256_point_add_affine: stp x16,x17,[x21,#0+16] adr x23,.Lone_mont-64 ldp x14,x15,[x22,#32] // in1 - cmp x24,#0 // ~, remember? + cmp x24,#0 // ~, remember? ldp x16,x17,[x22,#32+16] csel x8,x4,x8,ne csel x9,x5,x9,ne ldp x4,x5,[sp,#0+32+32] // res csel x10,x6,x10,ne csel x11,x7,x11,ne - cmp x25,#0 // ~, remember? + cmp x25,#0 // ~, remember? ldp x6,x7,[sp,#0+32+48] csel x14,x8,x14,ne csel x15,x9,x15,ne @@ -3589,13 +3589,13 @@ ecp_nistz256_point_add_affine: stp x14,x15,[x21,#32] stp x16,x17,[x21,#32+16] ldp x14,x15,[x22,#64] // in1 - cmp x24,#0 // ~, remember? + cmp x24,#0 // ~, remember? ldp x16,x17,[x22,#64+16] csel x8,x4,x8,ne csel x9,x5,x9,ne csel x10,x6,x10,ne csel x11,x7,x11,ne - cmp x25,#0 // ~, remember? + cmp x25,#0 // ~, remember? csel x14,x8,x14,ne csel x15,x9,x15,ne csel x16,x10,x16,ne diff --git a/contrib/libs/openssl/asm/aarch64/crypto/poly1305/poly1305-armv8.S b/contrib/libs/openssl/asm/aarch64/crypto/poly1305/poly1305-armv8.S index c482ec1b90..62271b488a 100644 --- a/contrib/libs/openssl/asm/aarch64/crypto/poly1305/poly1305-armv8.S +++ b/contrib/libs/openssl/asm/aarch64/crypto/poly1305/poly1305-armv8.S @@ -4,13 +4,13 @@ // forward "declarations" are required for Apple -.hidden OPENSSL_armcap_P -.globl poly1305_init -.hidden poly1305_init +.hidden OPENSSL_armcap_P +.globl poly1305_init +.hidden poly1305_init .globl poly1305_blocks -.hidden poly1305_blocks +.hidden poly1305_blocks .globl poly1305_emit -.hidden poly1305_emit +.hidden poly1305_emit .type poly1305_init,%function .align 5 @@ -799,7 +799,7 @@ poly1305_blocks_neon: st1 {v23.s}[0],[x0] .Lno_data_neon: - ldr x29,[sp],#80 + ldr x29,[sp],#80 .inst 0xd50323bf // autiasp ret .size poly1305_blocks_neon,.-poly1305_blocks_neon diff --git a/contrib/libs/openssl/asm/aarch64/crypto/sha/keccak1600-armv8.S b/contrib/libs/openssl/asm/aarch64/crypto/sha/keccak1600-armv8.S index f61b061455..d8bdc71aee 100644 --- a/contrib/libs/openssl/asm/aarch64/crypto/sha/keccak1600-armv8.S +++ b/contrib/libs/openssl/asm/aarch64/crypto/sha/keccak1600-armv8.S @@ -844,7 +844,7 @@ SHA3_absorb_cext: blo .Lprocess_block_ce ldr d31,[x1],#8 // *inp++ #ifdef __AARCH64EB__ - rev64 v31.16b,v31.16b + rev64 v31.16b,v31.16b #endif eor v1.16b,v1.16b,v31.16b beq .Lprocess_block_ce @@ -857,7 +857,7 @@ SHA3_absorb_cext: blo .Lprocess_block_ce ldr d31,[x1],#8 // *inp++ #ifdef __AARCH64EB__ - rev64 v31.16b,v31.16b + rev64 v31.16b,v31.16b #endif eor v3.16b,v3.16b,v31.16b beq .Lprocess_block_ce @@ -870,7 +870,7 @@ SHA3_absorb_cext: blo .Lprocess_block_ce ldr d31,[x1],#8 // *inp++ #ifdef __AARCH64EB__ - rev64 v31.16b,v31.16b + rev64 v31.16b,v31.16b #endif eor v5.16b,v5.16b,v31.16b beq .Lprocess_block_ce @@ -883,7 +883,7 @@ SHA3_absorb_cext: blo .Lprocess_block_ce ldr d31,[x1],#8 // *inp++ #ifdef __AARCH64EB__ - rev64 v31.16b,v31.16b + rev64 v31.16b,v31.16b #endif eor v7.16b,v7.16b,v31.16b beq .Lprocess_block_ce @@ -896,7 +896,7 @@ SHA3_absorb_cext: blo .Lprocess_block_ce ldr d31,[x1],#8 // *inp++ #ifdef __AARCH64EB__ - rev64 v31.16b,v31.16b + rev64 v31.16b,v31.16b #endif eor v9.16b,v9.16b,v31.16b beq .Lprocess_block_ce @@ -909,7 +909,7 @@ SHA3_absorb_cext: blo .Lprocess_block_ce ldr d31,[x1],#8 // *inp++ #ifdef __AARCH64EB__ - rev64 v31.16b,v31.16b + rev64 v31.16b,v31.16b #endif eor v11.16b,v11.16b,v31.16b beq .Lprocess_block_ce @@ -922,7 +922,7 @@ SHA3_absorb_cext: blo .Lprocess_block_ce ldr d31,[x1],#8 // *inp++ #ifdef __AARCH64EB__ - rev64 v31.16b,v31.16b + rev64 v31.16b,v31.16b #endif eor v13.16b,v13.16b,v31.16b beq .Lprocess_block_ce @@ -935,7 +935,7 @@ SHA3_absorb_cext: blo .Lprocess_block_ce ldr d31,[x1],#8 // *inp++ #ifdef __AARCH64EB__ - rev64 v31.16b,v31.16b + rev64 v31.16b,v31.16b #endif eor v15.16b,v15.16b,v31.16b beq .Lprocess_block_ce @@ -948,7 +948,7 @@ SHA3_absorb_cext: blo .Lprocess_block_ce ldr d31,[x1],#8 // *inp++ #ifdef __AARCH64EB__ - rev64 v31.16b,v31.16b + rev64 v31.16b,v31.16b #endif eor v17.16b,v17.16b,v31.16b beq .Lprocess_block_ce @@ -961,7 +961,7 @@ SHA3_absorb_cext: blo .Lprocess_block_ce ldr d31,[x1],#8 // *inp++ #ifdef __AARCH64EB__ - rev64 v31.16b,v31.16b + rev64 v31.16b,v31.16b #endif eor v19.16b,v19.16b,v31.16b beq .Lprocess_block_ce @@ -974,7 +974,7 @@ SHA3_absorb_cext: blo .Lprocess_block_ce ldr d31,[x1],#8 // *inp++ #ifdef __AARCH64EB__ - rev64 v31.16b,v31.16b + rev64 v31.16b,v31.16b #endif eor v21.16b,v21.16b,v31.16b beq .Lprocess_block_ce @@ -987,13 +987,13 @@ SHA3_absorb_cext: blo .Lprocess_block_ce ldr d31,[x1],#8 // *inp++ #ifdef __AARCH64EB__ - rev64 v31.16b,v31.16b + rev64 v31.16b,v31.16b #endif eor v23.16b,v23.16b,v31.16b beq .Lprocess_block_ce ldr d31,[x1],#8 // *inp++ #ifdef __AARCH64EB__ - rev64 v31.16b,v31.16b + rev64 v31.16b,v31.16b #endif eor v24.16b,v24.16b,v31.16b diff --git a/contrib/libs/openssl/asm/aarch64/crypto/sha/sha1-armv8.S b/contrib/libs/openssl/asm/aarch64/crypto/sha/sha1-armv8.S index ad299ffc56..d64f46a8f6 100644 --- a/contrib/libs/openssl/asm/aarch64/crypto/sha/sha1-armv8.S +++ b/contrib/libs/openssl/asm/aarch64/crypto/sha/sha1-armv8.S @@ -3,7 +3,7 @@ .text -.hidden OPENSSL_armcap_P +.hidden OPENSSL_armcap_P .globl sha1_block_data_order .type sha1_block_data_order,%function .align 6 diff --git a/contrib/libs/openssl/asm/aarch64/crypto/sha/sha256-armv8.S b/contrib/libs/openssl/asm/aarch64/crypto/sha/sha256-armv8.S index 0a33c04c29..52ca2ef1e0 100644 --- a/contrib/libs/openssl/asm/aarch64/crypto/sha/sha256-armv8.S +++ b/contrib/libs/openssl/asm/aarch64/crypto/sha/sha256-armv8.S @@ -1,4 +1,4 @@ -// Copyright 2014-2020 The OpenSSL Project Authors. All Rights Reserved. +// Copyright 2014-2020 The OpenSSL Project Authors. All Rights Reserved. // // Licensed under the OpenSSL license (the "License"). You may not use // this file except in compliance with the License. You can obtain a copy @@ -59,7 +59,7 @@ .text -.hidden OPENSSL_armcap_P +.hidden OPENSSL_armcap_P .globl sha256_block_data_order .type sha256_block_data_order,%function .align 6 diff --git a/contrib/libs/openssl/asm/aarch64/crypto/sha/sha512-armv8.S b/contrib/libs/openssl/asm/aarch64/crypto/sha/sha512-armv8.S index c93930773e..31e4550faf 100644 --- a/contrib/libs/openssl/asm/aarch64/crypto/sha/sha512-armv8.S +++ b/contrib/libs/openssl/asm/aarch64/crypto/sha/sha512-armv8.S @@ -1,4 +1,4 @@ -// Copyright 2014-2020 The OpenSSL Project Authors. All Rights Reserved. +// Copyright 2014-2020 The OpenSSL Project Authors. All Rights Reserved. // // Licensed under the OpenSSL license (the "License"). You may not use // this file except in compliance with the License. You can obtain a copy @@ -59,7 +59,7 @@ .text -.hidden OPENSSL_armcap_P +.hidden OPENSSL_armcap_P .globl sha512_block_data_order .type sha512_block_data_order,%function .align 6 |