/*
* Copyright (C) 2011-2016 Free Software Foundation, Inc.
* Copyright (C) 2016 Red Hat, Inc.
*
* Author: Nikos Mavrogiannopoulos
*
* This file is part of GnuTLS.
*
* The GnuTLS is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public License
* as published by the Free Software Foundation; either version 2.1 of
* the License, or (at your option) any later version.
*
* This library is distributed in the hope that it will be useful, but
* WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public License
* along with this program. If not, see
*
*/
#include "errors.h"
#include "gnutls_int.h"
#include
#include "errors.h"
#include
#include
#include
#include
#include
void sha1_block_data_order(void *c, const void *p, size_t len);
void sha256_block_data_order(void *c, const void *p, size_t len);
void sha512_block_data_order(void *c, const void *p, size_t len);
typedef void (*update_func) (void *, size_t, const uint8_t *);
typedef void (*digest_func) (void *, size_t, uint8_t *);
typedef void (*set_key_func) (void *, size_t, const uint8_t *);
typedef void (*init_func) (void *);
struct aarch64_hash_ctx {
union {
struct sha1_ctx sha1;
struct sha224_ctx sha224;
struct sha256_ctx sha256;
struct sha384_ctx sha384;
struct sha512_ctx sha512;
} ctx;
void *ctx_ptr;
gnutls_digest_algorithm_t algo;
size_t length;
update_func update;
digest_func digest;
init_func init;
};
static int
wrap_aarch64_hash_update(void *_ctx, const void *text, size_t textsize)
{
struct aarch64_hash_ctx *ctx = _ctx;
ctx->update(ctx->ctx_ptr, textsize, text);
return GNUTLS_E_SUCCESS;
}
static void wrap_aarch64_hash_deinit(void *hd)
{
gnutls_free(hd);
}
void aarch64_sha1_update(struct sha1_ctx *ctx, size_t length,
const uint8_t * data)
{
struct {
uint32_t h0, h1, h2, h3, h4;
uint32_t Nl, Nh;
uint32_t data[16];
unsigned int num;
} octx;
size_t res;
unsigned t2, i;
if ((res = ctx->index)) {
res = SHA1_DATA_SIZE - res;
if (length < res)
res = length;
sha1_update(ctx, res, data);
data += res;
length -= res;
}
octx.h0 = ctx->state[0];
octx.h1 = ctx->state[1];
octx.h2 = ctx->state[2];
octx.h3 = ctx->state[3];
octx.h4 = ctx->state[4];
memcpy(octx.data, ctx->block, SHA1_DATA_SIZE);
octx.num = ctx->index;
res = length % SHA1_DATA_SIZE;
length -= res;
if (length > 0) {
t2 = length / SHA1_DATA_SIZE;
sha1_block_data_order(&octx, data, t2);
for (i=0;icount++;
data += length;
}
ctx->state[0] = octx.h0;
ctx->state[1] = octx.h1;
ctx->state[2] = octx.h2;
ctx->state[3] = octx.h3;
ctx->state[4] = octx.h4;
memcpy(ctx->block, octx.data, octx.num);
ctx->index = octx.num;
if (res > 0) {
sha1_update(ctx, res, data);
}
}
void aarch64_sha256_update(struct sha256_ctx *ctx, size_t length,
const uint8_t * data)
{
struct {
uint32_t h[8];
uint32_t Nl, Nh;
uint32_t data[16];
unsigned int num;
unsigned md_len;
} octx;
size_t res;
unsigned t2, i;
if ((res = ctx->index)) {
res = SHA256_DATA_SIZE - res;
if (length < res)
res = length;
sha256_update(ctx, res, data);
data += res;
length -= res;
}
memcpy(octx.h, ctx->state, sizeof(octx.h));
memcpy(octx.data, ctx->block, SHA256_DATA_SIZE);
octx.num = ctx->index;
res = length % SHA256_DATA_SIZE;
length -= res;
if (length > 0) {
t2 = length / SHA1_DATA_SIZE;
sha256_block_data_order(&octx, data, t2);
for (i=0;icount++;
data += length;
}
memcpy(ctx->state, octx.h, sizeof(octx.h));
memcpy(ctx->block, octx.data, octx.num);
ctx->index = octx.num;
if (res > 0) {
sha256_update(ctx, res, data);
}
}
void aarch64_sha512_update(struct sha512_ctx *ctx, size_t length,
const uint8_t * data)
{
struct {
uint64_t h[8];
uint64_t Nl, Nh;
union {
uint64_t d[16];
uint8_t p[16*8];
} u;
unsigned int num;
unsigned md_len;
} octx;
size_t res;
unsigned t2, i;
if ((res = ctx->index)) {
res = SHA512_DATA_SIZE - res;
if (length < res)
res = length;
sha512_update(ctx, res, data);
data += res;
length -= res;
}
memcpy(octx.h, ctx->state, sizeof(octx.h));
memcpy(octx.u.p, ctx->block, SHA512_DATA_SIZE);
octx.num = ctx->index;
res = length % SHA512_DATA_SIZE;
length -= res;
if (length > 0) {
t2 = length / SHA512_DATA_SIZE;
sha512_block_data_order(&octx, data, t2);
for (i=0;istate, octx.h, sizeof(octx.h));
memcpy(ctx->block, octx.u.p, octx.num);
ctx->index = octx.num;
if (res > 0) {
sha512_update(ctx, res, data);
}
}
static int _ctx_init(gnutls_digest_algorithm_t algo,
struct aarch64_hash_ctx *ctx)
{
switch (algo) {
case GNUTLS_DIG_SHA1:
sha1_init(&ctx->ctx.sha1);
ctx->update = (update_func) aarch64_sha1_update;
ctx->digest = (digest_func) sha1_digest;
ctx->init = (init_func) sha1_init;
ctx->ctx_ptr = &ctx->ctx.sha1;
ctx->length = SHA1_DIGEST_SIZE;
break;
case GNUTLS_DIG_SHA224:
sha224_init(&ctx->ctx.sha224);
ctx->update = (update_func) aarch64_sha256_update;
ctx->digest = (digest_func) sha224_digest;
ctx->init = (init_func) sha224_init;
ctx->ctx_ptr = &ctx->ctx.sha224;
ctx->length = SHA224_DIGEST_SIZE;
break;
case GNUTLS_DIG_SHA256:
sha256_init(&ctx->ctx.sha256);
ctx->update = (update_func) aarch64_sha256_update;
ctx->digest = (digest_func) sha256_digest;
ctx->init = (init_func) sha256_init;
ctx->ctx_ptr = &ctx->ctx.sha256;
ctx->length = SHA256_DIGEST_SIZE;
break;
case GNUTLS_DIG_SHA384:
sha384_init(&ctx->ctx.sha384);
ctx->update = (update_func) aarch64_sha512_update;
ctx->digest = (digest_func) sha384_digest;
ctx->init = (init_func) sha384_init;
ctx->ctx_ptr = &ctx->ctx.sha384;
ctx->length = SHA384_DIGEST_SIZE;
break;
case GNUTLS_DIG_SHA512:
sha512_init(&ctx->ctx.sha512);
ctx->update = (update_func) aarch64_sha512_update;
ctx->digest = (digest_func) sha512_digest;
ctx->init = (init_func) sha512_init;
ctx->ctx_ptr = &ctx->ctx.sha512;
ctx->length = SHA512_DIGEST_SIZE;
break;
default:
gnutls_assert();
return GNUTLS_E_INVALID_REQUEST;
}
return 0;
}
static int wrap_aarch64_hash_init(gnutls_digest_algorithm_t algo, void **_ctx)
{
struct aarch64_hash_ctx *ctx;
int ret;
ctx = gnutls_malloc(sizeof(struct aarch64_hash_ctx));
if (ctx == NULL) {
gnutls_assert();
return GNUTLS_E_MEMORY_ERROR;
}
ctx->algo = algo;
if ((ret = _ctx_init(algo, ctx)) < 0) {
gnutls_assert();
return ret;
}
*_ctx = ctx;
return 0;
}
static void *
wrap_aarch64_hash_copy(const void *_ctx)
{
struct aarch64_hash_ctx *new_ctx;
const struct aarch64_hash_ctx *ctx=_ctx;
ptrdiff_t off = (uint8_t *)ctx->ctx_ptr - (uint8_t *)(&ctx->ctx);
new_ctx = gnutls_malloc(sizeof(struct aarch64_hash_ctx));
if (new_ctx == NULL) {
gnutls_assert();
return NULL;
}
memcpy(new_ctx, ctx, sizeof(*new_ctx));
new_ctx->ctx_ptr = (uint8_t *)&new_ctx->ctx + off;
return new_ctx;
}
static int
wrap_aarch64_hash_output(void *src_ctx, void *digest, size_t digestsize)
{
struct aarch64_hash_ctx *ctx;
ctx = src_ctx;
if (digestsize < ctx->length)
return gnutls_assert_val(GNUTLS_E_SHORT_MEMORY_BUFFER);
ctx->digest(ctx->ctx_ptr, digestsize, digest);
return 0;
}
static int wrap_aarch64_hash_fast(gnutls_digest_algorithm_t algo,
const void *text, size_t text_size,
void *digest)
{
struct aarch64_hash_ctx ctx;
int ret;
ret = _ctx_init(algo, &ctx);
if (ret < 0)
return gnutls_assert_val(ret);
ctx.update(&ctx, text_size, text);
ctx.digest(&ctx, ctx.length, digest);
return 0;
}
const struct nettle_hash aarch64_sha1 =
NN_HASH(sha1, aarch64_sha1_update, sha1_digest, SHA1);
const struct nettle_hash aarch64_sha224 =
NN_HASH(sha224, aarch64_sha256_update, sha224_digest, SHA224);
const struct nettle_hash aarch64_sha256 =
NN_HASH(sha256, aarch64_sha256_update, sha256_digest, SHA256);
const struct nettle_hash aarch64_sha384 =
NN_HASH(sha384, aarch64_sha512_update, sha384_digest, SHA384);
const struct nettle_hash aarch64_sha512 =
NN_HASH(sha512, aarch64_sha512_update, sha512_digest, SHA512);
const gnutls_crypto_digest_st _gnutls_sha_aarch64 = {
.init = wrap_aarch64_hash_init,
.hash = wrap_aarch64_hash_update,
.output = wrap_aarch64_hash_output,
.copy = wrap_aarch64_hash_copy,
.deinit = wrap_aarch64_hash_deinit,
.fast = wrap_aarch64_hash_fast,
};