aboutsummaryrefslogtreecommitdiffstats
path: root/contrib/libs/openssl/asm/aarch64
diff options
context:
space:
mode:
authorheretic <heretic@yandex-team.ru>2022-02-10 16:45:46 +0300
committerDaniil Cherednik <dcherednik@yandex-team.ru>2022-02-10 16:45:46 +0300
commit81eddc8c0b55990194e112b02d127b87d54164a9 (patch)
tree9142afc54d335ea52910662635b898e79e192e49 /contrib/libs/openssl/asm/aarch64
parent397cbe258b9e064f49c4ca575279f02f39fef76e (diff)
downloadydb-81eddc8c0b55990194e112b02d127b87d54164a9.tar.gz
Restoring authorship annotation for <heretic@yandex-team.ru>. Commit 2 of 2.
Diffstat (limited to 'contrib/libs/openssl/asm/aarch64')
-rw-r--r--contrib/libs/openssl/asm/aarch64/crypto/aes/aesv8-armx.S46
-rw-r--r--contrib/libs/openssl/asm/aarch64/crypto/chacha/chacha-armv8.S2
-rw-r--r--contrib/libs/openssl/asm/aarch64/crypto/ec/ecp_nistz256-armv8.S62
-rw-r--r--contrib/libs/openssl/asm/aarch64/crypto/poly1305/poly1305-armv8.S12
-rw-r--r--contrib/libs/openssl/asm/aarch64/crypto/sha/keccak1600-armv8.S26
-rw-r--r--contrib/libs/openssl/asm/aarch64/crypto/sha/sha1-armv8.S2
-rw-r--r--contrib/libs/openssl/asm/aarch64/crypto/sha/sha256-armv8.S4
-rw-r--r--contrib/libs/openssl/asm/aarch64/crypto/sha/sha512-armv8.S4
8 files changed, 79 insertions, 79 deletions
diff --git a/contrib/libs/openssl/asm/aarch64/crypto/aes/aesv8-armx.S b/contrib/libs/openssl/asm/aarch64/crypto/aes/aesv8-armx.S
index 1006d71f8d..433b8020af 100644
--- a/contrib/libs/openssl/asm/aarch64/crypto/aes/aesv8-armx.S
+++ b/contrib/libs/openssl/asm/aarch64/crypto/aes/aesv8-armx.S
@@ -103,12 +103,12 @@ aes_v8_set_encrypt_key:
.Loop192:
tbl v6.16b,{v4.16b},v2.16b
ext v5.16b,v0.16b,v3.16b,#12
-#ifdef __ARMEB__
- st1 {v4.4s},[x2],#16
- sub x2,x2,#8
-#else
+#ifdef __ARMEB__
+ st1 {v4.4s},[x2],#16
+ sub x2,x2,#8
+#else
st1 {v4.8b},[x2],#8
-#endif
+#endif
aese v6.16b,v0.16b
subs w1,w1,#1
@@ -579,11 +579,11 @@ aes_v8_ctr32_encrypt_blocks:
ldr w5,[x3,#240]
ldr w8, [x4, #12]
-#ifdef __ARMEB__
- ld1 {v0.16b},[x4]
-#else
+#ifdef __ARMEB__
+ ld1 {v0.16b},[x4]
+#else
ld1 {v0.4s},[x4]
-#endif
+#endif
ld1 {v16.4s,v17.4s},[x3] // load key schedule...
sub w5,w5,#4
mov x12,#16
@@ -602,14 +602,14 @@ aes_v8_ctr32_encrypt_blocks:
add w10, w8, #1
orr v6.16b,v0.16b,v0.16b
rev w10, w10
- mov v6.s[3],w10
- add w8, w8, #2
- orr v1.16b,v6.16b,v6.16b
+ mov v6.s[3],w10
+ add w8, w8, #2
+ orr v1.16b,v6.16b,v6.16b
b.ls .Lctr32_tail
rev w12, w8
- mov v6.s[3],w12
+ mov v6.s[3],w12
sub x2,x2,#3 // bias
- orr v18.16b,v6.16b,v6.16b
+ orr v18.16b,v6.16b,v6.16b
b .Loop3x_ctr32
.align 4
@@ -636,11 +636,11 @@ aes_v8_ctr32_encrypt_blocks:
aese v1.16b,v16.16b
aesmc v5.16b,v1.16b
ld1 {v2.16b},[x0],#16
- add w9,w8,#1
+ add w9,w8,#1
aese v18.16b,v16.16b
aesmc v18.16b,v18.16b
ld1 {v3.16b},[x0],#16
- rev w9,w9
+ rev w9,w9
aese v4.16b,v17.16b
aesmc v4.16b,v4.16b
aese v5.16b,v17.16b
@@ -664,22 +664,22 @@ aes_v8_ctr32_encrypt_blocks:
aese v5.16b,v21.16b
aesmc v5.16b,v5.16b
eor v19.16b,v19.16b,v7.16b
- mov v6.s[3], w9
+ mov v6.s[3], w9
aese v17.16b,v21.16b
aesmc v17.16b,v17.16b
- orr v0.16b,v6.16b,v6.16b
+ orr v0.16b,v6.16b,v6.16b
rev w10,w10
aese v4.16b,v22.16b
aesmc v4.16b,v4.16b
- mov v6.s[3], w10
- rev w12,w8
+ mov v6.s[3], w10
+ rev w12,w8
aese v5.16b,v22.16b
aesmc v5.16b,v5.16b
- orr v1.16b,v6.16b,v6.16b
- mov v6.s[3], w12
+ orr v1.16b,v6.16b,v6.16b
+ mov v6.s[3], w12
aese v17.16b,v22.16b
aesmc v17.16b,v17.16b
- orr v18.16b,v6.16b,v6.16b
+ orr v18.16b,v6.16b,v6.16b
subs x2,x2,#3
aese v4.16b,v23.16b
aese v5.16b,v23.16b
diff --git a/contrib/libs/openssl/asm/aarch64/crypto/chacha/chacha-armv8.S b/contrib/libs/openssl/asm/aarch64/crypto/chacha/chacha-armv8.S
index 0bccbaa665..f4676cbf68 100644
--- a/contrib/libs/openssl/asm/aarch64/crypto/chacha/chacha-armv8.S
+++ b/contrib/libs/openssl/asm/aarch64/crypto/chacha/chacha-armv8.S
@@ -3,7 +3,7 @@
.text
-.hidden OPENSSL_armcap_P
+.hidden OPENSSL_armcap_P
.align 5
.Lsigma:
diff --git a/contrib/libs/openssl/asm/aarch64/crypto/ec/ecp_nistz256-armv8.S b/contrib/libs/openssl/asm/aarch64/crypto/ec/ecp_nistz256-armv8.S
index 28749d5bcf..7caa581608 100644
--- a/contrib/libs/openssl/asm/aarch64/crypto/ec/ecp_nistz256-armv8.S
+++ b/contrib/libs/openssl/asm/aarch64/crypto/ec/ecp_nistz256-armv8.S
@@ -3015,7 +3015,7 @@ __ecp_nistz256_div_by_2:
.align 5
ecp_nistz256_point_double:
.inst 0xd503233f // paciasp
- stp x29,x30,[sp,#-96]!
+ stp x29,x30,[sp,#-96]!
add x29,sp,#0
stp x19,x20,[sp,#16]
stp x21,x22,[sp,#32]
@@ -3148,7 +3148,7 @@ ecp_nistz256_point_double:
add sp,x29,#0 // destroy frame
ldp x19,x20,[x29,#16]
ldp x21,x22,[x29,#32]
- ldp x29,x30,[sp],#96
+ ldp x29,x30,[sp],#96
.inst 0xd50323bf // autiasp
ret
.size ecp_nistz256_point_double,.-ecp_nistz256_point_double
@@ -3157,13 +3157,13 @@ ecp_nistz256_point_double:
.align 5
ecp_nistz256_point_add:
.inst 0xd503233f // paciasp
- stp x29,x30,[sp,#-96]!
+ stp x29,x30,[sp,#-96]!
add x29,sp,#0
stp x19,x20,[sp,#16]
stp x21,x22,[sp,#32]
stp x23,x24,[sp,#48]
stp x25,x26,[sp,#64]
- stp x27,x28,[sp,#80]
+ stp x27,x28,[sp,#80]
sub sp,sp,#32*12
ldp x4,x5,[x2,#64] // in2_z
@@ -3177,7 +3177,7 @@ ecp_nistz256_point_add:
orr x10,x6,x7
orr x25,x8,x10
cmp x25,#0
- csetm x25,ne // ~in2infty
+ csetm x25,ne // ~in2infty
add x0,sp,#192
bl __ecp_nistz256_sqr_mont // p256_sqr_mont(Z2sqr, in2_z);
@@ -3187,7 +3187,7 @@ ecp_nistz256_point_add:
orr x10,x6,x7
orr x24,x8,x10
cmp x24,#0
- csetm x24,ne // ~in1infty
+ csetm x24,ne // ~in1infty
add x0,sp,#128
bl __ecp_nistz256_sqr_mont // p256_sqr_mont(Z1sqr, in1_z);
@@ -3228,7 +3228,7 @@ ecp_nistz256_point_add:
orr x14,x14,x15 // see if result is zero
orr x16,x16,x17
- orr x26,x14,x16 // ~is_equal(S1,S2)
+ orr x26,x14,x16 // ~is_equal(S1,S2)
add x2,sp,#192
add x0,sp,#256
@@ -3249,21 +3249,21 @@ ecp_nistz256_point_add:
orr x14,x14,x15 // see if result is zero
orr x16,x16,x17
- orr x14,x14,x16 // ~is_equal(U1,U2)
+ orr x14,x14,x16 // ~is_equal(U1,U2)
- mvn x27,x24 // -1/0 -> 0/-1
- mvn x28,x25 // -1/0 -> 0/-1
- orr x14,x14,x27
- orr x14,x14,x28
- orr x14,x14,x26
- cbnz x14,.Ladd_proceed // if(~is_equal(U1,U2) | in1infty | in2infty | ~is_equal(S1,S2))
+ mvn x27,x24 // -1/0 -> 0/-1
+ mvn x28,x25 // -1/0 -> 0/-1
+ orr x14,x14,x27
+ orr x14,x14,x28
+ orr x14,x14,x26
+ cbnz x14,.Ladd_proceed // if(~is_equal(U1,U2) | in1infty | in2infty | ~is_equal(S1,S2))
.Ladd_double:
mov x1,x22
mov x0,x21
ldp x23,x24,[x29,#48]
ldp x25,x26,[x29,#64]
- ldp x27,x28,[x29,#80]
+ ldp x27,x28,[x29,#80]
add sp,sp,#32*(12-4) // difference in stack frames
b .Ldouble_shortcut
@@ -3345,14 +3345,14 @@ ecp_nistz256_point_add:
ldp x8,x9,[x23] // in2
ldp x10,x11,[x23,#16]
ldp x14,x15,[x22,#0] // in1
- cmp x24,#0 // ~, remember?
+ cmp x24,#0 // ~, remember?
ldp x16,x17,[x22,#0+16]
csel x8,x4,x8,ne
csel x9,x5,x9,ne
ldp x4,x5,[sp,#0+0+32] // res
csel x10,x6,x10,ne
csel x11,x7,x11,ne
- cmp x25,#0 // ~, remember?
+ cmp x25,#0 // ~, remember?
ldp x6,x7,[sp,#0+0+48]
csel x14,x8,x14,ne
csel x15,x9,x15,ne
@@ -3363,14 +3363,14 @@ ecp_nistz256_point_add:
stp x14,x15,[x21,#0]
stp x16,x17,[x21,#0+16]
ldp x14,x15,[x22,#32] // in1
- cmp x24,#0 // ~, remember?
+ cmp x24,#0 // ~, remember?
ldp x16,x17,[x22,#32+16]
csel x8,x4,x8,ne
csel x9,x5,x9,ne
ldp x4,x5,[sp,#0+32+32] // res
csel x10,x6,x10,ne
csel x11,x7,x11,ne
- cmp x25,#0 // ~, remember?
+ cmp x25,#0 // ~, remember?
ldp x6,x7,[sp,#0+32+48]
csel x14,x8,x14,ne
csel x15,x9,x15,ne
@@ -3381,13 +3381,13 @@ ecp_nistz256_point_add:
stp x14,x15,[x21,#32]
stp x16,x17,[x21,#32+16]
ldp x14,x15,[x22,#64] // in1
- cmp x24,#0 // ~, remember?
+ cmp x24,#0 // ~, remember?
ldp x16,x17,[x22,#64+16]
csel x8,x4,x8,ne
csel x9,x5,x9,ne
csel x10,x6,x10,ne
csel x11,x7,x11,ne
- cmp x25,#0 // ~, remember?
+ cmp x25,#0 // ~, remember?
csel x14,x8,x14,ne
csel x15,x9,x15,ne
csel x16,x10,x16,ne
@@ -3401,8 +3401,8 @@ ecp_nistz256_point_add:
ldp x21,x22,[x29,#32]
ldp x23,x24,[x29,#48]
ldp x25,x26,[x29,#64]
- ldp x27,x28,[x29,#80]
- ldp x29,x30,[sp],#96
+ ldp x27,x28,[x29,#80]
+ ldp x29,x30,[sp],#96
.inst 0xd50323bf // autiasp
ret
.size ecp_nistz256_point_add,.-ecp_nistz256_point_add
@@ -3431,7 +3431,7 @@ ecp_nistz256_point_add_affine:
orr x10,x6,x7
orr x24,x8,x10
cmp x24,#0
- csetm x24,ne // ~in1infty
+ csetm x24,ne // ~in1infty
ldp x14,x15,[x2] // in2_x
ldp x16,x17,[x2,#16]
@@ -3445,7 +3445,7 @@ ecp_nistz256_point_add_affine:
orr x8,x8,x10
orr x25,x14,x8
cmp x25,#0
- csetm x25,ne // ~in2infty
+ csetm x25,ne // ~in2infty
add x0,sp,#128
bl __ecp_nistz256_sqr_mont // p256_sqr_mont(Z1sqr, in1_z);
@@ -3552,14 +3552,14 @@ ecp_nistz256_point_add_affine:
ldp x8,x9,[x23] // in2
ldp x10,x11,[x23,#16]
ldp x14,x15,[x22,#0] // in1
- cmp x24,#0 // ~, remember?
+ cmp x24,#0 // ~, remember?
ldp x16,x17,[x22,#0+16]
csel x8,x4,x8,ne
csel x9,x5,x9,ne
ldp x4,x5,[sp,#0+0+32] // res
csel x10,x6,x10,ne
csel x11,x7,x11,ne
- cmp x25,#0 // ~, remember?
+ cmp x25,#0 // ~, remember?
ldp x6,x7,[sp,#0+0+48]
csel x14,x8,x14,ne
csel x15,x9,x15,ne
@@ -3571,14 +3571,14 @@ ecp_nistz256_point_add_affine:
stp x16,x17,[x21,#0+16]
adr x23,.Lone_mont-64
ldp x14,x15,[x22,#32] // in1
- cmp x24,#0 // ~, remember?
+ cmp x24,#0 // ~, remember?
ldp x16,x17,[x22,#32+16]
csel x8,x4,x8,ne
csel x9,x5,x9,ne
ldp x4,x5,[sp,#0+32+32] // res
csel x10,x6,x10,ne
csel x11,x7,x11,ne
- cmp x25,#0 // ~, remember?
+ cmp x25,#0 // ~, remember?
ldp x6,x7,[sp,#0+32+48]
csel x14,x8,x14,ne
csel x15,x9,x15,ne
@@ -3589,13 +3589,13 @@ ecp_nistz256_point_add_affine:
stp x14,x15,[x21,#32]
stp x16,x17,[x21,#32+16]
ldp x14,x15,[x22,#64] // in1
- cmp x24,#0 // ~, remember?
+ cmp x24,#0 // ~, remember?
ldp x16,x17,[x22,#64+16]
csel x8,x4,x8,ne
csel x9,x5,x9,ne
csel x10,x6,x10,ne
csel x11,x7,x11,ne
- cmp x25,#0 // ~, remember?
+ cmp x25,#0 // ~, remember?
csel x14,x8,x14,ne
csel x15,x9,x15,ne
csel x16,x10,x16,ne
diff --git a/contrib/libs/openssl/asm/aarch64/crypto/poly1305/poly1305-armv8.S b/contrib/libs/openssl/asm/aarch64/crypto/poly1305/poly1305-armv8.S
index c482ec1b90..62271b488a 100644
--- a/contrib/libs/openssl/asm/aarch64/crypto/poly1305/poly1305-armv8.S
+++ b/contrib/libs/openssl/asm/aarch64/crypto/poly1305/poly1305-armv8.S
@@ -4,13 +4,13 @@
// forward "declarations" are required for Apple
-.hidden OPENSSL_armcap_P
-.globl poly1305_init
-.hidden poly1305_init
+.hidden OPENSSL_armcap_P
+.globl poly1305_init
+.hidden poly1305_init
.globl poly1305_blocks
-.hidden poly1305_blocks
+.hidden poly1305_blocks
.globl poly1305_emit
-.hidden poly1305_emit
+.hidden poly1305_emit
.type poly1305_init,%function
.align 5
@@ -799,7 +799,7 @@ poly1305_blocks_neon:
st1 {v23.s}[0],[x0]
.Lno_data_neon:
- ldr x29,[sp],#80
+ ldr x29,[sp],#80
.inst 0xd50323bf // autiasp
ret
.size poly1305_blocks_neon,.-poly1305_blocks_neon
diff --git a/contrib/libs/openssl/asm/aarch64/crypto/sha/keccak1600-armv8.S b/contrib/libs/openssl/asm/aarch64/crypto/sha/keccak1600-armv8.S
index f61b061455..d8bdc71aee 100644
--- a/contrib/libs/openssl/asm/aarch64/crypto/sha/keccak1600-armv8.S
+++ b/contrib/libs/openssl/asm/aarch64/crypto/sha/keccak1600-armv8.S
@@ -844,7 +844,7 @@ SHA3_absorb_cext:
blo .Lprocess_block_ce
ldr d31,[x1],#8 // *inp++
#ifdef __AARCH64EB__
- rev64 v31.16b,v31.16b
+ rev64 v31.16b,v31.16b
#endif
eor v1.16b,v1.16b,v31.16b
beq .Lprocess_block_ce
@@ -857,7 +857,7 @@ SHA3_absorb_cext:
blo .Lprocess_block_ce
ldr d31,[x1],#8 // *inp++
#ifdef __AARCH64EB__
- rev64 v31.16b,v31.16b
+ rev64 v31.16b,v31.16b
#endif
eor v3.16b,v3.16b,v31.16b
beq .Lprocess_block_ce
@@ -870,7 +870,7 @@ SHA3_absorb_cext:
blo .Lprocess_block_ce
ldr d31,[x1],#8 // *inp++
#ifdef __AARCH64EB__
- rev64 v31.16b,v31.16b
+ rev64 v31.16b,v31.16b
#endif
eor v5.16b,v5.16b,v31.16b
beq .Lprocess_block_ce
@@ -883,7 +883,7 @@ SHA3_absorb_cext:
blo .Lprocess_block_ce
ldr d31,[x1],#8 // *inp++
#ifdef __AARCH64EB__
- rev64 v31.16b,v31.16b
+ rev64 v31.16b,v31.16b
#endif
eor v7.16b,v7.16b,v31.16b
beq .Lprocess_block_ce
@@ -896,7 +896,7 @@ SHA3_absorb_cext:
blo .Lprocess_block_ce
ldr d31,[x1],#8 // *inp++
#ifdef __AARCH64EB__
- rev64 v31.16b,v31.16b
+ rev64 v31.16b,v31.16b
#endif
eor v9.16b,v9.16b,v31.16b
beq .Lprocess_block_ce
@@ -909,7 +909,7 @@ SHA3_absorb_cext:
blo .Lprocess_block_ce
ldr d31,[x1],#8 // *inp++
#ifdef __AARCH64EB__
- rev64 v31.16b,v31.16b
+ rev64 v31.16b,v31.16b
#endif
eor v11.16b,v11.16b,v31.16b
beq .Lprocess_block_ce
@@ -922,7 +922,7 @@ SHA3_absorb_cext:
blo .Lprocess_block_ce
ldr d31,[x1],#8 // *inp++
#ifdef __AARCH64EB__
- rev64 v31.16b,v31.16b
+ rev64 v31.16b,v31.16b
#endif
eor v13.16b,v13.16b,v31.16b
beq .Lprocess_block_ce
@@ -935,7 +935,7 @@ SHA3_absorb_cext:
blo .Lprocess_block_ce
ldr d31,[x1],#8 // *inp++
#ifdef __AARCH64EB__
- rev64 v31.16b,v31.16b
+ rev64 v31.16b,v31.16b
#endif
eor v15.16b,v15.16b,v31.16b
beq .Lprocess_block_ce
@@ -948,7 +948,7 @@ SHA3_absorb_cext:
blo .Lprocess_block_ce
ldr d31,[x1],#8 // *inp++
#ifdef __AARCH64EB__
- rev64 v31.16b,v31.16b
+ rev64 v31.16b,v31.16b
#endif
eor v17.16b,v17.16b,v31.16b
beq .Lprocess_block_ce
@@ -961,7 +961,7 @@ SHA3_absorb_cext:
blo .Lprocess_block_ce
ldr d31,[x1],#8 // *inp++
#ifdef __AARCH64EB__
- rev64 v31.16b,v31.16b
+ rev64 v31.16b,v31.16b
#endif
eor v19.16b,v19.16b,v31.16b
beq .Lprocess_block_ce
@@ -974,7 +974,7 @@ SHA3_absorb_cext:
blo .Lprocess_block_ce
ldr d31,[x1],#8 // *inp++
#ifdef __AARCH64EB__
- rev64 v31.16b,v31.16b
+ rev64 v31.16b,v31.16b
#endif
eor v21.16b,v21.16b,v31.16b
beq .Lprocess_block_ce
@@ -987,13 +987,13 @@ SHA3_absorb_cext:
blo .Lprocess_block_ce
ldr d31,[x1],#8 // *inp++
#ifdef __AARCH64EB__
- rev64 v31.16b,v31.16b
+ rev64 v31.16b,v31.16b
#endif
eor v23.16b,v23.16b,v31.16b
beq .Lprocess_block_ce
ldr d31,[x1],#8 // *inp++
#ifdef __AARCH64EB__
- rev64 v31.16b,v31.16b
+ rev64 v31.16b,v31.16b
#endif
eor v24.16b,v24.16b,v31.16b
diff --git a/contrib/libs/openssl/asm/aarch64/crypto/sha/sha1-armv8.S b/contrib/libs/openssl/asm/aarch64/crypto/sha/sha1-armv8.S
index ad299ffc56..d64f46a8f6 100644
--- a/contrib/libs/openssl/asm/aarch64/crypto/sha/sha1-armv8.S
+++ b/contrib/libs/openssl/asm/aarch64/crypto/sha/sha1-armv8.S
@@ -3,7 +3,7 @@
.text
-.hidden OPENSSL_armcap_P
+.hidden OPENSSL_armcap_P
.globl sha1_block_data_order
.type sha1_block_data_order,%function
.align 6
diff --git a/contrib/libs/openssl/asm/aarch64/crypto/sha/sha256-armv8.S b/contrib/libs/openssl/asm/aarch64/crypto/sha/sha256-armv8.S
index 0a33c04c29..52ca2ef1e0 100644
--- a/contrib/libs/openssl/asm/aarch64/crypto/sha/sha256-armv8.S
+++ b/contrib/libs/openssl/asm/aarch64/crypto/sha/sha256-armv8.S
@@ -1,4 +1,4 @@
-// Copyright 2014-2020 The OpenSSL Project Authors. All Rights Reserved.
+// Copyright 2014-2020 The OpenSSL Project Authors. All Rights Reserved.
//
// Licensed under the OpenSSL license (the "License"). You may not use
// this file except in compliance with the License. You can obtain a copy
@@ -59,7 +59,7 @@
.text
-.hidden OPENSSL_armcap_P
+.hidden OPENSSL_armcap_P
.globl sha256_block_data_order
.type sha256_block_data_order,%function
.align 6
diff --git a/contrib/libs/openssl/asm/aarch64/crypto/sha/sha512-armv8.S b/contrib/libs/openssl/asm/aarch64/crypto/sha/sha512-armv8.S
index c93930773e..31e4550faf 100644
--- a/contrib/libs/openssl/asm/aarch64/crypto/sha/sha512-armv8.S
+++ b/contrib/libs/openssl/asm/aarch64/crypto/sha/sha512-armv8.S
@@ -1,4 +1,4 @@
-// Copyright 2014-2020 The OpenSSL Project Authors. All Rights Reserved.
+// Copyright 2014-2020 The OpenSSL Project Authors. All Rights Reserved.
//
// Licensed under the OpenSSL license (the "License"). You may not use
// this file except in compliance with the License. You can obtain a copy
@@ -59,7 +59,7 @@
.text
-.hidden OPENSSL_armcap_P
+.hidden OPENSSL_armcap_P
.globl sha512_block_data_order
.type sha512_block_data_order,%function
.align 6