#if defined(__arm__) && defined(__APCS_32__)
-#if defined(__ARM_ARCH_5E__) || defined(__ARM_ARCH_5TE__) || \
- defined(__ARM_ARCH_5TEJ__) || defined(__ARM_ARCH_6__) || \
- defined(__ARM_ARCH_6J__) || defined(__ARM_ARCH_6K__) || \
- defined(__ARM_ARCH_6M__) || defined(__ARM_ARCH_6T2__) || \
- defined(__ARM_ARCH_6Z__) || defined(__ARM_ARCH_6ZK__)
-#define __ARM_ARCH_5E_OR_6__
-#endif
-
-#if defined(__ARM_ARCH_5E_OR_6__) || defined(__ARM_ARCH_7__) || \
- defined(__ARM_ARCH_7A__) || defined(__ARM_ARCH_7R__) || \
- defined(__ARM_ARCH_7M__) || defined(__ARM_ARCH_7EM__)
-#define __ARM_ARCH_5E_OR_6_OR_7__
-#endif
-
-#ifdef __ARM_ARCH_5E_OR_6__
-
-.macro scrypt_shuffle
- add lr, r0, #9*4
- ldmia r0, {r2-r7}
- ldmia lr, {r2, r8-r12, lr}
- str r3, [r0, #5*4]
- str r5, [r0, #15*4]
- str r6, [r0, #12*4]
- str r7, [r0, #1*4]
- ldr r5, [r0, #7*4]
- str r2, [r0, #13*4]
- str r8, [r0, #2*4]
- strd r4, [r0, #10*4]
- str r9, [r0, #7*4]
- str r10, [r0, #4*4]
- str r11, [r0, #9*4]
- str lr, [r0, #3*4]
-
- add r2, r0, #64+0*4
- add lr, r0, #64+9*4
- ldmia r2, {r2-r7}
- ldmia lr, {r2, r8-r12, lr}
- str r3, [r0, #64+5*4]
- str r5, [r0, #64+15*4]
- str r6, [r0, #64+12*4]
- str r7, [r0, #64+1*4]
- ldr r5, [r0, #64+7*4]
- str r2, [r0, #64+13*4]
- str r8, [r0, #64+2*4]
- strd r4, [r0, #64+10*4]
- str r9, [r0, #64+7*4]
- str r10, [r0, #64+4*4]
- str r11, [r0, #64+9*4]
- str lr, [r0, #64+3*4]
-.endm
-
-.macro salsa8_core_doubleround_body
- add r6, r2, r6
- add r7, r3, r7
- eor r10, r10, r6, ror #25
- add r6, r0, r4
- eor r11, r11, r7, ror #25
- add r7, r1, r5
- strd r10, [sp, #14*4]
- eor r12, r12, r6, ror #25
- eor lr, lr, r7, ror #25
-
- ldrd r6, [sp, #10*4]
- add r2, r10, r2
- add r3, r11, r3
- eor r6, r6, r2, ror #23
- add r2, r12, r0
- eor r7, r7, r3, ror #23
- add r3, lr, r1
- strd r6, [sp, #10*4]
- eor r8, r8, r2, ror #23
- eor r9, r9, r3, ror #23
-
- ldrd r2, [sp, #6*4]
- add r10, r6, r10
- add r11, r7, r11
- eor r2, r2, r10, ror #19
- add r10, r8, r12
- eor r3, r3, r11, ror #19
- add r11, r9, lr
- eor r4, r4, r10, ror #19
- eor r5, r5, r11, ror #19
-
- ldrd r10, [sp, #2*4]
- add r6, r2, r6
- add r7, r3, r7
- eor r10, r10, r6, ror #14
- add r6, r4, r8
- eor r11, r11, r7, ror #14
- add r7, r5, r9
- eor r0, r0, r6, ror #14
- eor r1, r1, r7, ror #14
-
-
- ldrd r6, [sp, #14*4]
- strd r2, [sp, #6*4]
- strd r10, [sp, #2*4]
- add r6, r11, r6
- add r7, r0, r7
- eor r4, r4, r6, ror #25
- add r6, r1, r12
- eor r5, r5, r7, ror #25
- add r7, r10, lr
- eor r2, r2, r6, ror #25
- eor r3, r3, r7, ror #25
- strd r2, [sp, #6*4]
-
- add r10, r3, r10
- ldrd r6, [sp, #10*4]
- add r11, r4, r11
- eor r8, r8, r10, ror #23
- add r10, r5, r0
- eor r9, r9, r11, ror #23
- add r11, r2, r1
- eor r6, r6, r10, ror #23
- eor r7, r7, r11, ror #23
- strd r6, [sp, #10*4]
-
- add r2, r7, r2
- ldrd r10, [sp, #14*4]
- add r3, r8, r3
- eor r12, r12, r2, ror #19
- add r2, r9, r4
- eor lr, lr, r3, ror #19
- add r3, r6, r5
- eor r10, r10, r2, ror #19
- eor r11, r11, r3, ror #19
-
- ldrd r2, [sp, #2*4]
- add r6, r11, r6
- add r7, r12, r7
- eor r0, r0, r6, ror #14
- add r6, lr, r8
- eor r1, r1, r7, ror #14
- add r7, r10, r9
- eor r2, r2, r6, ror #14
- eor r3, r3, r7, ror #14
-.endm
-
-.macro salsa8_core
- ldmia sp, {r0-r12, lr}
-
- ldrd r10, [sp, #14*4]
- salsa8_core_doubleround_body
- ldrd r6, [sp, #6*4]
- strd r2, [sp, #2*4]
- strd r10, [sp, #14*4]
- salsa8_core_doubleround_body
- ldrd r6, [sp, #6*4]
- strd r2, [sp, #2*4]
- strd r10, [sp, #14*4]
- salsa8_core_doubleround_body
- ldrd r6, [sp, #6*4]
- strd r2, [sp, #2*4]
- strd r10, [sp, #14*4]
- salsa8_core_doubleround_body
-
- stmia sp, {r0-r5}
- strd r8, [sp, #8*4]
- str r12, [sp, #12*4]
- str lr, [sp, #13*4]
- strd r10, [sp, #14*4]
-.endm
-
-#else
-
-.macro scrypt_shuffle
-.endm
-
.macro salsa8_core_doubleround_body
ldr r8, [sp, #8*4]
add r11, r11, r10
ldr r12, [sp, #15*4]
add lr, r10, lr
eor r11, r11, r8, ror #14
- add r8, r3, r2
+ str r9, [sp, #9*4]
eor r12, r12, lr, ror #14
+ add r8, r3, r2
add lr, r4, r7
+ str r10, [sp, #14*4]
eor r0, r0, r8, ror #14
ldr r8, [sp, #11*4]
eor r5, r5, lr, ror #14
add r8, r5, r1
eor r4, r4, lr, ror #25
add lr, r11, r6
- str r9, [sp, #9*4]
eor r9, r9, r8, ror #25
- str r10, [sp, #14*4]
eor r10, r10, lr, ror #25
salsa8_core_doubleround_body
ldr r12, [sp, #15*4]
add lr, r10, lr
eor r11, r11, r8, ror #14
- add r8, r3, r2
+ str r9, [sp, #9*4]
eor r12, r12, lr, ror #14
+ add r8, r3, r2
add lr, r4, r7
+ str r10, [sp, #14*4]
eor r0, r0, r8, ror #14
ldr r8, [sp, #11*4]
eor r5, r5, lr, ror #14
add r8, r5, r1
eor r4, r4, lr, ror #25
add lr, r11, r6
- str r9, [sp, #9*4]
eor r9, r9, r8, ror #25
- str r10, [sp, #14*4]
eor r10, r10, lr, ror #25
salsa8_core_doubleround_body
ldr r12, [sp, #15*4]
add lr, r10, lr
eor r11, r11, r8, ror #14
- add r8, r3, r2
+ str r9, [sp, #9*4]
eor r12, r12, lr, ror #14
+ add r8, r3, r2
add lr, r4, r7
+ str r10, [sp, #14*4]
eor r0, r0, r8, ror #14
ldr r8, [sp, #11*4]
eor r5, r5, lr, ror #14
add r8, r5, r1
eor r4, r4, lr, ror #25
add lr, r11, r6
- str r9, [sp, #9*4]
eor r9, r9, r8, ror #25
- str r10, [sp, #14*4]
eor r10, r10, lr, ror #25
salsa8_core_doubleround_body
stmia sp, {r0-r7}
.endm
-#endif
-
.macro scrypt_core_macro1a_x4
ldmia r0, {r4-r7}
scrypt_core:
_scrypt_core:
stmfd sp!, {r4-r11, lr}
- mov r12, sp
- sub sp, sp, #21*4
- bic sp, sp, #63
- str r12, [sp, #20*4]
-
- scrypt_shuffle
+ sub sp, sp, #20*4
str r0, [sp, #16*4]
add r12, r1, #1024*32*4
cmp r1, r12
bne scrypt_core_loop1
- ldr r4, [r0, #16*4]
sub r1, r1, #1024*32*4
str r1, [sp, #17*4]
- mov r4, r4, lsl #32-10
mov r12, #1024
- add r1, r1, r4, lsr #32-10-7
scrypt_core_loop2:
+ str r12, [sp, #18*4]
+
+ ldr r4, [r0, #16*4]
+ mov r4, r4, lsl #32-10
+ add r1, r1, r4, lsr #32-10-7
+
add r2, r0, #16*4
add r3, r1, #16*4
- str r12, [sp, #18*4]
mov r12, sp
-#ifdef __ARM_ARCH_5E_OR_6_OR_7__
- pld [r1, #24*4]
- pld [r1, #8*4]
-#endif
scrypt_core_macro1b_x4
scrypt_core_macro1b_x4
scrypt_core_macro1b_x4
ldr r0, [sp, #16*4]
mov r1, sp
- ldr r3, [sp, #17*4]
add r0, r0, #16*4
- scrypt_core_macro3_x4
- mov r4, r4, lsl #32-10
- add r3, r3, r4, lsr #32-10-7
- str r3, [sp, #19*4]
-#ifdef __ARM_ARCH_5E_OR_6_OR_7__
- pld [r3, #16*4]
- pld [r3]
-#endif
scrypt_core_macro3_x6
scrypt_core_macro3_x6
+ scrypt_core_macro3_x4
ldr r12, [sp, #18*4]
sub r0, r0, #32*4
- ldr r1, [sp, #19*4]
+ ldr r1, [sp, #17*4]
subs r12, r12, #1
bne scrypt_core_loop2
- scrypt_shuffle
-
- ldr sp, [sp, #20*4]
+ add sp, sp, #20*4
#ifdef __thumb__
ldmfd sp!, {r4-r11, lr}
bx lr