Searched refs:SHA256_BLOCK_SIZE (Results 1 – 8 of 8) sorted by relevance
157 if (md->curlen == 0 && inlen >= SHA256_BLOCK_SIZE) { in sha256_process()160 md->length += SHA256_BLOCK_SIZE * 8; in sha256_process()161 in += SHA256_BLOCK_SIZE; in sha256_process()162 inlen -= SHA256_BLOCK_SIZE; in sha256_process()164 n = MIN(inlen, (SHA256_BLOCK_SIZE - md->curlen)); in sha256_process()169 if (md->curlen == SHA256_BLOCK_SIZE) { in sha256_process()172 md->length += 8 * SHA256_BLOCK_SIZE; in sha256_process()206 while (md->curlen < SHA256_BLOCK_SIZE) { in sha256_done()
12 #define SHA256_BLOCK_SIZE 64 macro17 u8 buf[SHA256_BLOCK_SIZE];
48 #define SHA256_BLOCK_SIZE 64 macro187 unsigned char k_ipad[SHA256_BLOCK_SIZE]; in rpmb_hmac()188 unsigned char k_opad[SHA256_BLOCK_SIZE]; in rpmb_hmac()206 for ( ; i < SHA256_BLOCK_SIZE; i++) { in rpmb_hmac()210 sha256_update(&ctx, k_ipad, SHA256_BLOCK_SIZE); in rpmb_hmac()218 sha256_update(&ctx, k_opad, SHA256_BLOCK_SIZE); in rpmb_hmac()
213 SHA256_BLOCK_SIZE,216 SHA256_BLOCK_SIZE,219 SHA256_BLOCK_SIZE,222 SHA256_BLOCK_SIZE,
250 tmp_len = SHA256_BLOCK_SIZE - ctx->len; in SHA256_update()255 if (ctx->len + len < SHA256_BLOCK_SIZE) { in SHA256_update()261 block_nb = new_len / SHA256_BLOCK_SIZE; in SHA256_update()268 rem_len = new_len % SHA256_BLOCK_SIZE; in SHA256_update()285 block_nb = (1 + ((SHA256_BLOCK_SIZE - 9) in SHA256_final()286 < (ctx->len % SHA256_BLOCK_SIZE))); in SHA256_final()
78 #define SHA256_BLOCK_SIZE (64) macro89 uint8_t leftover[SHA256_BLOCK_SIZE];
21 #define SHA256_BLOCK_SIZE 64 macro43 uint8_t block[2 * SHA256_BLOCK_SIZE];
76 if (s->leftover_offset >= SHA256_BLOCK_SIZE) { in tc_sha256_update()79 s->bits_hashed += (SHA256_BLOCK_SIZE << 3); in tc_sha256_update()