2 * Copyright 2009 Colin Percival, 2011 ArtForz, 2011 pooler, 2013 Balthazar
5 * Redistribution and use in source and binary forms, with or without
6 * modification, are permitted provided that the following conditions
8 * 1. Redistributions of source code must retain the above copyright
9 * notice, this list of conditions and the following disclaimer.
10 * 2. Redistributions in binary form must reproduce the above copyright
11 * notice, this list of conditions and the following disclaimer in the
12 * documentation and/or other materials provided with the distribution.
14 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
15 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
16 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
17 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
18 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
19 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
20 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
21 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
22 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
23 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
26 * This file was originally written by Colin Percival as part of the Tarsnap
27 * online backup system.
39 #define SCRYPT_BUFFER_SIZE (131072 + 63)
41 #if defined (OPTIMIZED_SALSA) && ( defined (__x86_64__) || defined (__i386__) || defined(__arm__) )
42 extern "C" void scrypt_core(unsigned int *X, unsigned int *V);
44 // Generic scrypt_core implementation
46 static inline void xor_salsa8(unsigned int B[16], const unsigned int Bx[16])
48 unsigned int x00,x01,x02,x03,x04,x05,x06,x07,x08,x09,x10,x11,x12,x13,x14,x15;
51 x00 = (B[0] ^= Bx[0]);
52 x01 = (B[1] ^= Bx[1]);
53 x02 = (B[2] ^= Bx[2]);
54 x03 = (B[3] ^= Bx[3]);
55 x04 = (B[4] ^= Bx[4]);
56 x05 = (B[5] ^= Bx[5]);
57 x06 = (B[6] ^= Bx[6]);
58 x07 = (B[7] ^= Bx[7]);
59 x08 = (B[8] ^= Bx[8]);
60 x09 = (B[9] ^= Bx[9]);
61 x10 = (B[10] ^= Bx[10]);
62 x11 = (B[11] ^= Bx[11]);
63 x12 = (B[12] ^= Bx[12]);
64 x13 = (B[13] ^= Bx[13]);
65 x14 = (B[14] ^= Bx[14]);
66 x15 = (B[15] ^= Bx[15]);
67 for (i = 0; i < 8; i += 2) {
68 #define R(a, b) (((a) << (b)) | ((a) >> (32 - (b))))
69 /* Operate on columns. */
70 x04 ^= R(x00+x12, 7); x09 ^= R(x05+x01, 7);
71 x14 ^= R(x10+x06, 7); x03 ^= R(x15+x11, 7);
73 x08 ^= R(x04+x00, 9); x13 ^= R(x09+x05, 9);
74 x02 ^= R(x14+x10, 9); x07 ^= R(x03+x15, 9);
76 x12 ^= R(x08+x04,13); x01 ^= R(x13+x09,13);
77 x06 ^= R(x02+x14,13); x11 ^= R(x07+x03,13);
79 x00 ^= R(x12+x08,18); x05 ^= R(x01+x13,18);
80 x10 ^= R(x06+x02,18); x15 ^= R(x11+x07,18);
82 /* Operate on rows. */
83 x01 ^= R(x00+x03, 7); x06 ^= R(x05+x04, 7);
84 x11 ^= R(x10+x09, 7); x12 ^= R(x15+x14, 7);
86 x02 ^= R(x01+x00, 9); x07 ^= R(x06+x05, 9);
87 x08 ^= R(x11+x10, 9); x13 ^= R(x12+x15, 9);
89 x03 ^= R(x02+x01,13); x04 ^= R(x07+x06,13);
90 x09 ^= R(x08+x11,13); x14 ^= R(x13+x12,13);
92 x00 ^= R(x03+x02,18); x05 ^= R(x04+x07,18);
93 x10 ^= R(x09+x08,18); x15 ^= R(x14+x13,18);
114 static inline void scrypt_core(unsigned int *X, unsigned int *V)
116 unsigned int i, j, k;
118 for (i = 0; i < 1024; i++) {
119 memcpy(&V[i * 32], X, 128);
120 xor_salsa8(&X[0], &X[16]);
121 xor_salsa8(&X[16], &X[0]);
123 for (i = 0; i < 1024; i++) {
124 j = 32 * (X[16] & 1023);
125 for (k = 0; k < 32; k++)
127 xor_salsa8(&X[0], &X[16]);
128 xor_salsa8(&X[16], &X[0]);
134 /* cpu and memory intensive function to transform a 80 byte buffer into a 32 byte output
135 scratchpad size needs to be at least 63 + (128 * r * p) + (256 * r + 64) + (128 * r * N) bytes
136 r = 1, p = 1, N = 1024
139 uint256 scrypt_nosalt(const void* input, size_t inputlen, void *scratchpad)
144 V = (unsigned int *)(((uintptr_t)(scratchpad) + 63) & ~ (uintptr_t)(63));
146 PBKDF2_SHA256((const uint8_t*)input, inputlen, (const uint8_t*)input, inputlen, 1, (uint8_t *)X, 128);
148 PBKDF2_SHA256((const uint8_t*)input, inputlen, (uint8_t *)X, 128, 1, (uint8_t*)&result, 32);
153 uint256 scrypt(const void* data, size_t datalen, const void* salt, size_t saltlen, void *scratchpad)
158 V = (unsigned int *)(((uintptr_t)(scratchpad) + 63) & ~ (uintptr_t)(63));
160 PBKDF2_SHA256((const uint8_t*)data, datalen, (const uint8_t*)salt, saltlen, 1, (uint8_t *)X, 128);
162 PBKDF2_SHA256((const uint8_t*)data, datalen, (uint8_t *)X, 128, 1, (uint8_t*)&result, 32);
167 uint256 scrypt_hash(const void* input, size_t inputlen)
169 unsigned char scratchpad[SCRYPT_BUFFER_SIZE];
170 return scrypt_nosalt(input, inputlen, scratchpad);
173 uint256 scrypt_salted_hash(const void* input, size_t inputlen, const void* salt, size_t saltlen)
175 unsigned char scratchpad[SCRYPT_BUFFER_SIZE];
176 return scrypt(input, inputlen, salt, saltlen, scratchpad);
179 uint256 scrypt_salted_multiround_hash(const void* input, size_t inputlen, const void* salt, size_t saltlen, const unsigned int nRounds)
181 uint256 resultHash = scrypt_salted_hash(input, inputlen, salt, saltlen);
182 uint256 transitionalHash = resultHash;
184 for(unsigned int i = 1; i < nRounds; i++)
186 resultHash = scrypt_salted_hash(input, inputlen, (const void*)&transitionalHash, 32);
187 transitionalHash = resultHash;
193 uint256 scrypt_blockhash(const void* input)
195 unsigned char scratchpad[SCRYPT_BUFFER_SIZE];
196 return scrypt_nosalt(input, 80, scratchpad);