mirror of
https://github.com/python/cpython.git
synced 2025-07-07 19:35:27 +00:00

Bumps the HACL* revision to include recent revisions that corrects issues building with legacy/cross-platform macOS SDKs. Signed-off-by: aeiouaeiouaeiouaeiouaeiouaeiou <aeioudev@outlook.com>
2522 lines
67 KiB
C
2522 lines
67 KiB
C
/* MIT License
|
|
*
|
|
* Copyright (c) 2016-2022 INRIA, CMU and Microsoft Corporation
|
|
* Copyright (c) 2022-2023 HACL* Contributors
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a copy
|
|
* of this software and associated documentation files (the "Software"), to deal
|
|
* in the Software without restriction, including without limitation the rights
|
|
* to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
|
* copies of the Software, and to permit persons to whom the Software is
|
|
* furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice shall be included in all
|
|
* copies or substantial portions of the Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
|
* AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
|
* OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
|
* SOFTWARE.
|
|
*/
|
|
|
|
|
|
#include "internal/Hacl_Streaming_HMAC.h"
|
|
|
|
#include "Hacl_Streaming_Types.h"
|
|
|
|
#include "Hacl_Hash_SHA3.h"
|
|
#include "Hacl_Hash_SHA2.h"
|
|
#include "Hacl_Hash_Blake2s_Simd128.h"
|
|
#include "Hacl_Hash_Blake2s.h"
|
|
#include "Hacl_Hash_Blake2b_Simd256.h"
|
|
#include "Hacl_Hash_Blake2b.h"
|
|
#include "internal/Hacl_Streaming_Types.h"
|
|
|
|
#include "internal/Hacl_Hash_SHA3.h"
|
|
#include "internal/Hacl_Hash_SHA2.h"
|
|
#include "internal/Hacl_Hash_SHA1.h"
|
|
#include "internal/Hacl_Hash_MD5.h"
|
|
#include "internal/Hacl_Hash_Blake2s_Simd128.h"
|
|
#include "internal/Hacl_Hash_Blake2s.h"
|
|
#include "internal/Hacl_Hash_Blake2b_Simd256.h"
|
|
#include "internal/Hacl_Hash_Blake2b.h"
|
|
|
|
static Spec_Hash_Definitions_hash_alg alg_of_impl(Hacl_Agile_Hash_impl i)
|
|
{
|
|
switch (i)
|
|
{
|
|
case Hacl_Agile_Hash_MD5:
|
|
{
|
|
return Spec_Hash_Definitions_MD5;
|
|
}
|
|
case Hacl_Agile_Hash_SHA1:
|
|
{
|
|
return Spec_Hash_Definitions_SHA1;
|
|
}
|
|
case Hacl_Agile_Hash_SHA2_224:
|
|
{
|
|
return Spec_Hash_Definitions_SHA2_224;
|
|
}
|
|
case Hacl_Agile_Hash_SHA2_256:
|
|
{
|
|
return Spec_Hash_Definitions_SHA2_256;
|
|
}
|
|
case Hacl_Agile_Hash_SHA2_384:
|
|
{
|
|
return Spec_Hash_Definitions_SHA2_384;
|
|
}
|
|
case Hacl_Agile_Hash_SHA2_512:
|
|
{
|
|
return Spec_Hash_Definitions_SHA2_512;
|
|
}
|
|
case Hacl_Agile_Hash_SHA3_224:
|
|
{
|
|
return Spec_Hash_Definitions_SHA3_224;
|
|
}
|
|
case Hacl_Agile_Hash_SHA3_256:
|
|
{
|
|
return Spec_Hash_Definitions_SHA3_256;
|
|
}
|
|
case Hacl_Agile_Hash_SHA3_384:
|
|
{
|
|
return Spec_Hash_Definitions_SHA3_384;
|
|
}
|
|
case Hacl_Agile_Hash_SHA3_512:
|
|
{
|
|
return Spec_Hash_Definitions_SHA3_512;
|
|
}
|
|
case Hacl_Agile_Hash_Blake2S_32:
|
|
{
|
|
return Spec_Hash_Definitions_Blake2S;
|
|
}
|
|
case Hacl_Agile_Hash_Blake2S_128:
|
|
{
|
|
return Spec_Hash_Definitions_Blake2S;
|
|
}
|
|
case Hacl_Agile_Hash_Blake2B_32:
|
|
{
|
|
return Spec_Hash_Definitions_Blake2B;
|
|
}
|
|
case Hacl_Agile_Hash_Blake2B_256:
|
|
{
|
|
return Spec_Hash_Definitions_Blake2B;
|
|
}
|
|
default:
|
|
{
|
|
KRML_HOST_EPRINTF("KaRaMeL incomplete match at %s:%d\n", __FILE__, __LINE__);
|
|
KRML_HOST_EXIT(253U);
|
|
}
|
|
}
|
|
}
|
|
|
|
static Hacl_Agile_Hash_impl impl_of_state_s(Hacl_Agile_Hash_state_s s)
|
|
{
|
|
if (s.tag == Hacl_Agile_Hash_MD5_a)
|
|
{
|
|
return Hacl_Agile_Hash_MD5;
|
|
}
|
|
if (s.tag == Hacl_Agile_Hash_SHA1_a)
|
|
{
|
|
return Hacl_Agile_Hash_SHA1;
|
|
}
|
|
if (s.tag == Hacl_Agile_Hash_SHA2_224_a)
|
|
{
|
|
return Hacl_Agile_Hash_SHA2_224;
|
|
}
|
|
if (s.tag == Hacl_Agile_Hash_SHA2_256_a)
|
|
{
|
|
return Hacl_Agile_Hash_SHA2_256;
|
|
}
|
|
if (s.tag == Hacl_Agile_Hash_SHA2_384_a)
|
|
{
|
|
return Hacl_Agile_Hash_SHA2_384;
|
|
}
|
|
if (s.tag == Hacl_Agile_Hash_SHA2_512_a)
|
|
{
|
|
return Hacl_Agile_Hash_SHA2_512;
|
|
}
|
|
if (s.tag == Hacl_Agile_Hash_SHA3_224_a)
|
|
{
|
|
return Hacl_Agile_Hash_SHA3_224;
|
|
}
|
|
if (s.tag == Hacl_Agile_Hash_SHA3_256_a)
|
|
{
|
|
return Hacl_Agile_Hash_SHA3_256;
|
|
}
|
|
if (s.tag == Hacl_Agile_Hash_SHA3_384_a)
|
|
{
|
|
return Hacl_Agile_Hash_SHA3_384;
|
|
}
|
|
if (s.tag == Hacl_Agile_Hash_SHA3_512_a)
|
|
{
|
|
return Hacl_Agile_Hash_SHA3_512;
|
|
}
|
|
if (s.tag == Hacl_Agile_Hash_Blake2S_a)
|
|
{
|
|
return Hacl_Agile_Hash_Blake2S_32;
|
|
}
|
|
if (s.tag == Hacl_Agile_Hash_Blake2S_128_a)
|
|
{
|
|
return Hacl_Agile_Hash_Blake2S_128;
|
|
}
|
|
if (s.tag == Hacl_Agile_Hash_Blake2B_a)
|
|
{
|
|
return Hacl_Agile_Hash_Blake2B_32;
|
|
}
|
|
if (s.tag == Hacl_Agile_Hash_Blake2B_256_a)
|
|
{
|
|
return Hacl_Agile_Hash_Blake2B_256;
|
|
}
|
|
KRML_HOST_EPRINTF("KaRaMeL abort at %s:%d\n%s\n",
|
|
__FILE__,
|
|
__LINE__,
|
|
"unreachable (pattern matches are exhaustive in F*)");
|
|
KRML_HOST_EXIT(255U);
|
|
}
|
|
|
|
static Hacl_Agile_Hash_impl impl_of_state(Hacl_Agile_Hash_state_s *s)
|
|
{
|
|
return impl_of_state_s(*s);
|
|
}
|
|
|
|
static Hacl_Agile_Hash_state_s *malloc_(Hacl_Agile_Hash_impl a)
|
|
{
|
|
switch (a)
|
|
{
|
|
case Hacl_Agile_Hash_MD5:
|
|
{
|
|
uint32_t *s = (uint32_t *)KRML_HOST_CALLOC(4U, sizeof (uint32_t));
|
|
if (s == NULL)
|
|
{
|
|
return NULL;
|
|
}
|
|
uint32_t *s1 = s;
|
|
Hacl_Agile_Hash_state_s
|
|
*st = (Hacl_Agile_Hash_state_s *)KRML_HOST_MALLOC(sizeof (Hacl_Agile_Hash_state_s));
|
|
if (st != NULL)
|
|
{
|
|
st[0U] = ((Hacl_Agile_Hash_state_s){ .tag = Hacl_Agile_Hash_MD5_a, { .case_MD5_a = s1 } });
|
|
}
|
|
if (st == NULL)
|
|
{
|
|
KRML_HOST_FREE(s1);
|
|
return NULL;
|
|
}
|
|
return st;
|
|
}
|
|
case Hacl_Agile_Hash_SHA1:
|
|
{
|
|
uint32_t *s = (uint32_t *)KRML_HOST_CALLOC(5U, sizeof (uint32_t));
|
|
if (s == NULL)
|
|
{
|
|
return NULL;
|
|
}
|
|
uint32_t *s1 = s;
|
|
Hacl_Agile_Hash_state_s
|
|
*st = (Hacl_Agile_Hash_state_s *)KRML_HOST_MALLOC(sizeof (Hacl_Agile_Hash_state_s));
|
|
if (st != NULL)
|
|
{
|
|
st[0U] =
|
|
((Hacl_Agile_Hash_state_s){ .tag = Hacl_Agile_Hash_SHA1_a, { .case_SHA1_a = s1 } });
|
|
}
|
|
if (st == NULL)
|
|
{
|
|
KRML_HOST_FREE(s1);
|
|
return NULL;
|
|
}
|
|
return st;
|
|
}
|
|
case Hacl_Agile_Hash_SHA2_224:
|
|
{
|
|
uint32_t *s = (uint32_t *)KRML_HOST_CALLOC(8U, sizeof (uint32_t));
|
|
if (s == NULL)
|
|
{
|
|
return NULL;
|
|
}
|
|
uint32_t *s1 = s;
|
|
Hacl_Agile_Hash_state_s
|
|
*st = (Hacl_Agile_Hash_state_s *)KRML_HOST_MALLOC(sizeof (Hacl_Agile_Hash_state_s));
|
|
if (st != NULL)
|
|
{
|
|
st[0U] =
|
|
(
|
|
(Hacl_Agile_Hash_state_s){
|
|
.tag = Hacl_Agile_Hash_SHA2_224_a,
|
|
{ .case_SHA2_224_a = s1 }
|
|
}
|
|
);
|
|
}
|
|
if (st == NULL)
|
|
{
|
|
KRML_HOST_FREE(s1);
|
|
return NULL;
|
|
}
|
|
return st;
|
|
}
|
|
case Hacl_Agile_Hash_SHA2_256:
|
|
{
|
|
uint32_t *s = (uint32_t *)KRML_HOST_CALLOC(8U, sizeof (uint32_t));
|
|
if (s == NULL)
|
|
{
|
|
return NULL;
|
|
}
|
|
uint32_t *s1 = s;
|
|
Hacl_Agile_Hash_state_s
|
|
*st = (Hacl_Agile_Hash_state_s *)KRML_HOST_MALLOC(sizeof (Hacl_Agile_Hash_state_s));
|
|
if (st != NULL)
|
|
{
|
|
st[0U] =
|
|
(
|
|
(Hacl_Agile_Hash_state_s){
|
|
.tag = Hacl_Agile_Hash_SHA2_256_a,
|
|
{ .case_SHA2_256_a = s1 }
|
|
}
|
|
);
|
|
}
|
|
if (st == NULL)
|
|
{
|
|
KRML_HOST_FREE(s1);
|
|
return NULL;
|
|
}
|
|
return st;
|
|
}
|
|
case Hacl_Agile_Hash_SHA2_384:
|
|
{
|
|
uint64_t *s = (uint64_t *)KRML_HOST_CALLOC(8U, sizeof (uint64_t));
|
|
if (s == NULL)
|
|
{
|
|
return NULL;
|
|
}
|
|
uint64_t *s1 = s;
|
|
Hacl_Agile_Hash_state_s
|
|
*st = (Hacl_Agile_Hash_state_s *)KRML_HOST_MALLOC(sizeof (Hacl_Agile_Hash_state_s));
|
|
if (st != NULL)
|
|
{
|
|
st[0U] =
|
|
(
|
|
(Hacl_Agile_Hash_state_s){
|
|
.tag = Hacl_Agile_Hash_SHA2_384_a,
|
|
{ .case_SHA2_384_a = s1 }
|
|
}
|
|
);
|
|
}
|
|
if (st == NULL)
|
|
{
|
|
KRML_HOST_FREE(s1);
|
|
return NULL;
|
|
}
|
|
return st;
|
|
}
|
|
case Hacl_Agile_Hash_SHA2_512:
|
|
{
|
|
uint64_t *s = (uint64_t *)KRML_HOST_CALLOC(8U, sizeof (uint64_t));
|
|
if (s == NULL)
|
|
{
|
|
return NULL;
|
|
}
|
|
uint64_t *s1 = s;
|
|
Hacl_Agile_Hash_state_s
|
|
*st = (Hacl_Agile_Hash_state_s *)KRML_HOST_MALLOC(sizeof (Hacl_Agile_Hash_state_s));
|
|
if (st != NULL)
|
|
{
|
|
st[0U] =
|
|
(
|
|
(Hacl_Agile_Hash_state_s){
|
|
.tag = Hacl_Agile_Hash_SHA2_512_a,
|
|
{ .case_SHA2_512_a = s1 }
|
|
}
|
|
);
|
|
}
|
|
if (st == NULL)
|
|
{
|
|
KRML_HOST_FREE(s1);
|
|
return NULL;
|
|
}
|
|
return st;
|
|
}
|
|
case Hacl_Agile_Hash_SHA3_224:
|
|
{
|
|
uint64_t *s = (uint64_t *)KRML_HOST_CALLOC(25U, sizeof (uint64_t));
|
|
if (s == NULL)
|
|
{
|
|
return NULL;
|
|
}
|
|
uint64_t *s1 = s;
|
|
Hacl_Agile_Hash_state_s
|
|
*st = (Hacl_Agile_Hash_state_s *)KRML_HOST_MALLOC(sizeof (Hacl_Agile_Hash_state_s));
|
|
if (st != NULL)
|
|
{
|
|
st[0U] =
|
|
(
|
|
(Hacl_Agile_Hash_state_s){
|
|
.tag = Hacl_Agile_Hash_SHA3_224_a,
|
|
{ .case_SHA3_224_a = s1 }
|
|
}
|
|
);
|
|
}
|
|
if (st == NULL)
|
|
{
|
|
KRML_HOST_FREE(s1);
|
|
return NULL;
|
|
}
|
|
return st;
|
|
}
|
|
case Hacl_Agile_Hash_SHA3_256:
|
|
{
|
|
uint64_t *s = (uint64_t *)KRML_HOST_CALLOC(25U, sizeof (uint64_t));
|
|
if (s == NULL)
|
|
{
|
|
return NULL;
|
|
}
|
|
uint64_t *s1 = s;
|
|
Hacl_Agile_Hash_state_s
|
|
*st = (Hacl_Agile_Hash_state_s *)KRML_HOST_MALLOC(sizeof (Hacl_Agile_Hash_state_s));
|
|
if (st != NULL)
|
|
{
|
|
st[0U] =
|
|
(
|
|
(Hacl_Agile_Hash_state_s){
|
|
.tag = Hacl_Agile_Hash_SHA3_256_a,
|
|
{ .case_SHA3_256_a = s1 }
|
|
}
|
|
);
|
|
}
|
|
if (st == NULL)
|
|
{
|
|
KRML_HOST_FREE(s1);
|
|
return NULL;
|
|
}
|
|
return st;
|
|
}
|
|
case Hacl_Agile_Hash_SHA3_384:
|
|
{
|
|
uint64_t *s = (uint64_t *)KRML_HOST_CALLOC(25U, sizeof (uint64_t));
|
|
if (s == NULL)
|
|
{
|
|
return NULL;
|
|
}
|
|
uint64_t *s1 = s;
|
|
Hacl_Agile_Hash_state_s
|
|
*st = (Hacl_Agile_Hash_state_s *)KRML_HOST_MALLOC(sizeof (Hacl_Agile_Hash_state_s));
|
|
if (st != NULL)
|
|
{
|
|
st[0U] =
|
|
(
|
|
(Hacl_Agile_Hash_state_s){
|
|
.tag = Hacl_Agile_Hash_SHA3_384_a,
|
|
{ .case_SHA3_384_a = s1 }
|
|
}
|
|
);
|
|
}
|
|
if (st == NULL)
|
|
{
|
|
KRML_HOST_FREE(s1);
|
|
return NULL;
|
|
}
|
|
return st;
|
|
}
|
|
case Hacl_Agile_Hash_SHA3_512:
|
|
{
|
|
uint64_t *s = (uint64_t *)KRML_HOST_CALLOC(25U, sizeof (uint64_t));
|
|
if (s == NULL)
|
|
{
|
|
return NULL;
|
|
}
|
|
uint64_t *s1 = s;
|
|
Hacl_Agile_Hash_state_s
|
|
*st = (Hacl_Agile_Hash_state_s *)KRML_HOST_MALLOC(sizeof (Hacl_Agile_Hash_state_s));
|
|
if (st != NULL)
|
|
{
|
|
st[0U] =
|
|
(
|
|
(Hacl_Agile_Hash_state_s){
|
|
.tag = Hacl_Agile_Hash_SHA3_512_a,
|
|
{ .case_SHA3_512_a = s1 }
|
|
}
|
|
);
|
|
}
|
|
if (st == NULL)
|
|
{
|
|
KRML_HOST_FREE(s1);
|
|
return NULL;
|
|
}
|
|
return st;
|
|
}
|
|
case Hacl_Agile_Hash_Blake2S_32:
|
|
{
|
|
uint32_t *s = (uint32_t *)KRML_HOST_CALLOC(16U, sizeof (uint32_t));
|
|
if (s == NULL)
|
|
{
|
|
return NULL;
|
|
}
|
|
uint32_t *s1 = s;
|
|
Hacl_Agile_Hash_state_s
|
|
*st = (Hacl_Agile_Hash_state_s *)KRML_HOST_MALLOC(sizeof (Hacl_Agile_Hash_state_s));
|
|
if (st != NULL)
|
|
{
|
|
st[0U] =
|
|
(
|
|
(Hacl_Agile_Hash_state_s){
|
|
.tag = Hacl_Agile_Hash_Blake2S_a,
|
|
{ .case_Blake2S_a = s1 }
|
|
}
|
|
);
|
|
}
|
|
if (st == NULL)
|
|
{
|
|
KRML_HOST_FREE(s1);
|
|
return NULL;
|
|
}
|
|
return st;
|
|
}
|
|
case Hacl_Agile_Hash_Blake2S_128:
|
|
{
|
|
#if HACL_CAN_COMPILE_VEC128
|
|
Lib_IntVector_Intrinsics_vec128
|
|
*s = Hacl_Hash_Blake2s_Simd128_malloc_internal_state_with_key();
|
|
if (s == NULL)
|
|
{
|
|
return NULL;
|
|
}
|
|
Hacl_Agile_Hash_state_s
|
|
*st = (Hacl_Agile_Hash_state_s *)KRML_HOST_MALLOC(sizeof (Hacl_Agile_Hash_state_s));
|
|
if (st != NULL)
|
|
{
|
|
st[0U] =
|
|
(
|
|
(Hacl_Agile_Hash_state_s){
|
|
.tag = Hacl_Agile_Hash_Blake2S_128_a,
|
|
{ .case_Blake2S_128_a = s }
|
|
}
|
|
);
|
|
}
|
|
if (st == NULL)
|
|
{
|
|
KRML_ALIGNED_FREE(s);
|
|
return NULL;
|
|
}
|
|
return st;
|
|
#else
|
|
KRML_HOST_EPRINTF("KaRaMeL abort at %s:%d\n%s\n",
|
|
__FILE__,
|
|
__LINE__,
|
|
"provably unreachable code: did an unverified caller violate a precondition\?");
|
|
KRML_HOST_EXIT(255U);
|
|
#endif
|
|
break;
|
|
}
|
|
case Hacl_Agile_Hash_Blake2B_32:
|
|
{
|
|
uint64_t *s = (uint64_t *)KRML_HOST_CALLOC(16U, sizeof (uint64_t));
|
|
if (s == NULL)
|
|
{
|
|
return NULL;
|
|
}
|
|
uint64_t *s1 = s;
|
|
Hacl_Agile_Hash_state_s
|
|
*st = (Hacl_Agile_Hash_state_s *)KRML_HOST_MALLOC(sizeof (Hacl_Agile_Hash_state_s));
|
|
if (st != NULL)
|
|
{
|
|
st[0U] =
|
|
(
|
|
(Hacl_Agile_Hash_state_s){
|
|
.tag = Hacl_Agile_Hash_Blake2B_a,
|
|
{ .case_Blake2B_a = s1 }
|
|
}
|
|
);
|
|
}
|
|
if (st == NULL)
|
|
{
|
|
KRML_HOST_FREE(s1);
|
|
return NULL;
|
|
}
|
|
return st;
|
|
}
|
|
case Hacl_Agile_Hash_Blake2B_256:
|
|
{
|
|
#if HACL_CAN_COMPILE_VEC256
|
|
Lib_IntVector_Intrinsics_vec256
|
|
*s = Hacl_Hash_Blake2b_Simd256_malloc_internal_state_with_key();
|
|
if (s == NULL)
|
|
{
|
|
return NULL;
|
|
}
|
|
Hacl_Agile_Hash_state_s
|
|
*st = (Hacl_Agile_Hash_state_s *)KRML_HOST_MALLOC(sizeof (Hacl_Agile_Hash_state_s));
|
|
if (st != NULL)
|
|
{
|
|
st[0U] =
|
|
(
|
|
(Hacl_Agile_Hash_state_s){
|
|
.tag = Hacl_Agile_Hash_Blake2B_256_a,
|
|
{ .case_Blake2B_256_a = s }
|
|
}
|
|
);
|
|
}
|
|
if (st == NULL)
|
|
{
|
|
KRML_ALIGNED_FREE(s);
|
|
return NULL;
|
|
}
|
|
return st;
|
|
#else
|
|
KRML_HOST_EPRINTF("KaRaMeL abort at %s:%d\n%s\n",
|
|
__FILE__,
|
|
__LINE__,
|
|
"provably unreachable code: did an unverified caller violate a precondition\?");
|
|
KRML_HOST_EXIT(255U);
|
|
#endif
|
|
break;
|
|
}
|
|
default:
|
|
{
|
|
KRML_HOST_EPRINTF("KaRaMeL incomplete match at %s:%d\n", __FILE__, __LINE__);
|
|
KRML_HOST_EXIT(253U);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void init(Hacl_Agile_Hash_state_s *s)
|
|
{
|
|
Hacl_Agile_Hash_state_s scrut = *s;
|
|
if (scrut.tag == Hacl_Agile_Hash_MD5_a)
|
|
{
|
|
uint32_t *p1 = scrut.case_MD5_a;
|
|
Hacl_Hash_MD5_init(p1);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA1_a)
|
|
{
|
|
uint32_t *p1 = scrut.case_SHA1_a;
|
|
Hacl_Hash_SHA1_init(p1);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA2_224_a)
|
|
{
|
|
uint32_t *p1 = scrut.case_SHA2_224_a;
|
|
Hacl_Hash_SHA2_sha224_init(p1);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA2_256_a)
|
|
{
|
|
uint32_t *p1 = scrut.case_SHA2_256_a;
|
|
Hacl_Hash_SHA2_sha256_init(p1);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA2_384_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA2_384_a;
|
|
Hacl_Hash_SHA2_sha384_init(p1);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA2_512_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA2_512_a;
|
|
Hacl_Hash_SHA2_sha512_init(p1);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA3_224_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA3_224_a;
|
|
Hacl_Hash_SHA3_init_(Spec_Hash_Definitions_SHA3_224, p1);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA3_256_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA3_256_a;
|
|
Hacl_Hash_SHA3_init_(Spec_Hash_Definitions_SHA3_256, p1);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA3_384_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA3_384_a;
|
|
Hacl_Hash_SHA3_init_(Spec_Hash_Definitions_SHA3_384, p1);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA3_512_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA3_512_a;
|
|
Hacl_Hash_SHA3_init_(Spec_Hash_Definitions_SHA3_512, p1);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_Blake2S_a)
|
|
{
|
|
uint32_t *p1 = scrut.case_Blake2S_a;
|
|
Hacl_Hash_Blake2s_init(p1, 0U, 32U);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_Blake2S_128_a)
|
|
{
|
|
Lib_IntVector_Intrinsics_vec128 *p1 = scrut.case_Blake2S_128_a;
|
|
#if HACL_CAN_COMPILE_VEC128
|
|
Hacl_Hash_Blake2s_Simd128_init(p1, 0U, 32U);
|
|
return;
|
|
#else
|
|
KRML_MAYBE_UNUSED_VAR(p1);
|
|
return;
|
|
#endif
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_Blake2B_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_Blake2B_a;
|
|
Hacl_Hash_Blake2b_init(p1, 0U, 64U);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_Blake2B_256_a)
|
|
{
|
|
Lib_IntVector_Intrinsics_vec256 *p1 = scrut.case_Blake2B_256_a;
|
|
#if HACL_CAN_COMPILE_VEC256
|
|
Hacl_Hash_Blake2b_Simd256_init(p1, 0U, 64U);
|
|
return;
|
|
#else
|
|
KRML_MAYBE_UNUSED_VAR(p1);
|
|
return;
|
|
#endif
|
|
}
|
|
KRML_HOST_EPRINTF("KaRaMeL abort at %s:%d\n%s\n",
|
|
__FILE__,
|
|
__LINE__,
|
|
"unreachable (pattern matches are exhaustive in F*)");
|
|
KRML_HOST_EXIT(255U);
|
|
}
|
|
|
|
static void
|
|
update_multi(Hacl_Agile_Hash_state_s *s, uint64_t prevlen, uint8_t *blocks, uint32_t len)
|
|
{
|
|
Hacl_Agile_Hash_state_s scrut = *s;
|
|
if (scrut.tag == Hacl_Agile_Hash_MD5_a)
|
|
{
|
|
uint32_t *p1 = scrut.case_MD5_a;
|
|
uint32_t n = len / 64U;
|
|
Hacl_Hash_MD5_update_multi(p1, blocks, n);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA1_a)
|
|
{
|
|
uint32_t *p1 = scrut.case_SHA1_a;
|
|
uint32_t n = len / 64U;
|
|
Hacl_Hash_SHA1_update_multi(p1, blocks, n);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA2_224_a)
|
|
{
|
|
uint32_t *p1 = scrut.case_SHA2_224_a;
|
|
uint32_t n = len / 64U;
|
|
Hacl_Hash_SHA2_sha224_update_nblocks(n * 64U, blocks, p1);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA2_256_a)
|
|
{
|
|
uint32_t *p1 = scrut.case_SHA2_256_a;
|
|
uint32_t n = len / 64U;
|
|
Hacl_Hash_SHA2_sha256_update_nblocks(n * 64U, blocks, p1);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA2_384_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA2_384_a;
|
|
uint32_t n = len / 128U;
|
|
Hacl_Hash_SHA2_sha384_update_nblocks(n * 128U, blocks, p1);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA2_512_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA2_512_a;
|
|
uint32_t n = len / 128U;
|
|
Hacl_Hash_SHA2_sha512_update_nblocks(n * 128U, blocks, p1);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA3_224_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA3_224_a;
|
|
uint32_t n = len / 144U;
|
|
Hacl_Hash_SHA3_update_multi_sha3(Spec_Hash_Definitions_SHA3_224, p1, blocks, n);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA3_256_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA3_256_a;
|
|
uint32_t n = len / 136U;
|
|
Hacl_Hash_SHA3_update_multi_sha3(Spec_Hash_Definitions_SHA3_256, p1, blocks, n);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA3_384_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA3_384_a;
|
|
uint32_t n = len / 104U;
|
|
Hacl_Hash_SHA3_update_multi_sha3(Spec_Hash_Definitions_SHA3_384, p1, blocks, n);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA3_512_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA3_512_a;
|
|
uint32_t n = len / 72U;
|
|
Hacl_Hash_SHA3_update_multi_sha3(Spec_Hash_Definitions_SHA3_512, p1, blocks, n);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_Blake2S_a)
|
|
{
|
|
uint32_t *p1 = scrut.case_Blake2S_a;
|
|
uint32_t n = len / 64U;
|
|
uint32_t wv[16U] = { 0U };
|
|
Hacl_Hash_Blake2s_update_multi(n * 64U, wv, p1, prevlen, blocks, n);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_Blake2S_128_a)
|
|
{
|
|
Lib_IntVector_Intrinsics_vec128 *p1 = scrut.case_Blake2S_128_a;
|
|
#if HACL_CAN_COMPILE_VEC128
|
|
uint32_t n = len / 64U;
|
|
Hacl_Hash_Blake2s_Simd128_update_multi_no_inline(p1, prevlen, blocks, n);
|
|
return;
|
|
#else
|
|
KRML_MAYBE_UNUSED_VAR(p1);
|
|
return;
|
|
#endif
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_Blake2B_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_Blake2B_a;
|
|
uint32_t n = len / 128U;
|
|
uint64_t wv[16U] = { 0U };
|
|
Hacl_Hash_Blake2b_update_multi(n * 128U,
|
|
wv,
|
|
p1,
|
|
FStar_UInt128_uint64_to_uint128(prevlen),
|
|
blocks,
|
|
n);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_Blake2B_256_a)
|
|
{
|
|
Lib_IntVector_Intrinsics_vec256 *p1 = scrut.case_Blake2B_256_a;
|
|
#if HACL_CAN_COMPILE_VEC256
|
|
uint32_t n = len / 128U;
|
|
Hacl_Hash_Blake2b_Simd256_update_multi_no_inline(p1,
|
|
FStar_UInt128_uint64_to_uint128(prevlen),
|
|
blocks,
|
|
n);
|
|
return;
|
|
#else
|
|
KRML_MAYBE_UNUSED_VAR(p1);
|
|
return;
|
|
#endif
|
|
}
|
|
KRML_HOST_EPRINTF("KaRaMeL abort at %s:%d\n%s\n",
|
|
__FILE__,
|
|
__LINE__,
|
|
"unreachable (pattern matches are exhaustive in F*)");
|
|
KRML_HOST_EXIT(255U);
|
|
}
|
|
|
|
static void
|
|
update_last(Hacl_Agile_Hash_state_s *s, uint64_t prev_len, uint8_t *last, uint32_t last_len)
|
|
{
|
|
Hacl_Agile_Hash_state_s scrut = *s;
|
|
if (scrut.tag == Hacl_Agile_Hash_MD5_a)
|
|
{
|
|
uint32_t *p1 = scrut.case_MD5_a;
|
|
Hacl_Hash_MD5_update_last(p1, prev_len, last, last_len);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA1_a)
|
|
{
|
|
uint32_t *p1 = scrut.case_SHA1_a;
|
|
Hacl_Hash_SHA1_update_last(p1, prev_len, last, last_len);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA2_224_a)
|
|
{
|
|
uint32_t *p1 = scrut.case_SHA2_224_a;
|
|
Hacl_Hash_SHA2_sha224_update_last(prev_len + (uint64_t)last_len, last_len, last, p1);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA2_256_a)
|
|
{
|
|
uint32_t *p1 = scrut.case_SHA2_256_a;
|
|
Hacl_Hash_SHA2_sha256_update_last(prev_len + (uint64_t)last_len, last_len, last, p1);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA2_384_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA2_384_a;
|
|
Hacl_Hash_SHA2_sha384_update_last(FStar_UInt128_add(FStar_UInt128_uint64_to_uint128(prev_len),
|
|
FStar_UInt128_uint64_to_uint128((uint64_t)last_len)),
|
|
last_len,
|
|
last,
|
|
p1);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA2_512_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA2_512_a;
|
|
Hacl_Hash_SHA2_sha512_update_last(FStar_UInt128_add(FStar_UInt128_uint64_to_uint128(prev_len),
|
|
FStar_UInt128_uint64_to_uint128((uint64_t)last_len)),
|
|
last_len,
|
|
last,
|
|
p1);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA3_224_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA3_224_a;
|
|
Hacl_Hash_SHA3_update_last_sha3(Spec_Hash_Definitions_SHA3_224, p1, last, last_len);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA3_256_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA3_256_a;
|
|
Hacl_Hash_SHA3_update_last_sha3(Spec_Hash_Definitions_SHA3_256, p1, last, last_len);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA3_384_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA3_384_a;
|
|
Hacl_Hash_SHA3_update_last_sha3(Spec_Hash_Definitions_SHA3_384, p1, last, last_len);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA3_512_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA3_512_a;
|
|
Hacl_Hash_SHA3_update_last_sha3(Spec_Hash_Definitions_SHA3_512, p1, last, last_len);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_Blake2S_a)
|
|
{
|
|
uint32_t *p1 = scrut.case_Blake2S_a;
|
|
uint32_t wv[16U] = { 0U };
|
|
Hacl_Hash_Blake2s_update_last(last_len, wv, p1, false, prev_len, last_len, last);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_Blake2S_128_a)
|
|
{
|
|
Lib_IntVector_Intrinsics_vec128 *p1 = scrut.case_Blake2S_128_a;
|
|
#if HACL_CAN_COMPILE_VEC128
|
|
Hacl_Hash_Blake2s_Simd128_update_last_no_inline(p1, prev_len, last, last_len);
|
|
return;
|
|
#else
|
|
KRML_MAYBE_UNUSED_VAR(p1);
|
|
return;
|
|
#endif
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_Blake2B_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_Blake2B_a;
|
|
uint64_t wv[16U] = { 0U };
|
|
Hacl_Hash_Blake2b_update_last(last_len,
|
|
wv,
|
|
p1,
|
|
false,
|
|
FStar_UInt128_uint64_to_uint128(prev_len),
|
|
last_len,
|
|
last);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_Blake2B_256_a)
|
|
{
|
|
Lib_IntVector_Intrinsics_vec256 *p1 = scrut.case_Blake2B_256_a;
|
|
#if HACL_CAN_COMPILE_VEC256
|
|
Hacl_Hash_Blake2b_Simd256_update_last_no_inline(p1,
|
|
FStar_UInt128_uint64_to_uint128(prev_len),
|
|
last,
|
|
last_len);
|
|
return;
|
|
#else
|
|
KRML_MAYBE_UNUSED_VAR(p1);
|
|
return;
|
|
#endif
|
|
}
|
|
KRML_HOST_EPRINTF("KaRaMeL abort at %s:%d\n%s\n",
|
|
__FILE__,
|
|
__LINE__,
|
|
"unreachable (pattern matches are exhaustive in F*)");
|
|
KRML_HOST_EXIT(255U);
|
|
}
|
|
|
|
static void finish(Hacl_Agile_Hash_state_s *s, uint8_t *dst)
|
|
{
|
|
Hacl_Agile_Hash_state_s scrut = *s;
|
|
if (scrut.tag == Hacl_Agile_Hash_MD5_a)
|
|
{
|
|
uint32_t *p1 = scrut.case_MD5_a;
|
|
Hacl_Hash_MD5_finish(p1, dst);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA1_a)
|
|
{
|
|
uint32_t *p1 = scrut.case_SHA1_a;
|
|
Hacl_Hash_SHA1_finish(p1, dst);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA2_224_a)
|
|
{
|
|
uint32_t *p1 = scrut.case_SHA2_224_a;
|
|
Hacl_Hash_SHA2_sha224_finish(p1, dst);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA2_256_a)
|
|
{
|
|
uint32_t *p1 = scrut.case_SHA2_256_a;
|
|
Hacl_Hash_SHA2_sha256_finish(p1, dst);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA2_384_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA2_384_a;
|
|
Hacl_Hash_SHA2_sha384_finish(p1, dst);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA2_512_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA2_512_a;
|
|
Hacl_Hash_SHA2_sha512_finish(p1, dst);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA3_224_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA3_224_a;
|
|
uint32_t remOut = 28U;
|
|
uint8_t hbuf[256U] = { 0U };
|
|
uint64_t ws[32U] = { 0U };
|
|
memcpy(ws, p1, 25U * sizeof (uint64_t));
|
|
for (uint32_t i = 0U; i < 32U; i++)
|
|
{
|
|
store64_le(hbuf + i * 8U, ws[i]);
|
|
}
|
|
memcpy(dst + 28U - remOut, hbuf, remOut * sizeof (uint8_t));
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA3_256_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA3_256_a;
|
|
uint32_t remOut = 32U;
|
|
uint8_t hbuf[256U] = { 0U };
|
|
uint64_t ws[32U] = { 0U };
|
|
memcpy(ws, p1, 25U * sizeof (uint64_t));
|
|
for (uint32_t i = 0U; i < 32U; i++)
|
|
{
|
|
store64_le(hbuf + i * 8U, ws[i]);
|
|
}
|
|
memcpy(dst + 32U - remOut, hbuf, remOut * sizeof (uint8_t));
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA3_384_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA3_384_a;
|
|
uint32_t remOut = 48U;
|
|
uint8_t hbuf[256U] = { 0U };
|
|
uint64_t ws[32U] = { 0U };
|
|
memcpy(ws, p1, 25U * sizeof (uint64_t));
|
|
for (uint32_t i = 0U; i < 32U; i++)
|
|
{
|
|
store64_le(hbuf + i * 8U, ws[i]);
|
|
}
|
|
memcpy(dst + 48U - remOut, hbuf, remOut * sizeof (uint8_t));
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA3_512_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA3_512_a;
|
|
uint32_t remOut = 64U;
|
|
uint8_t hbuf[256U] = { 0U };
|
|
uint64_t ws[32U] = { 0U };
|
|
memcpy(ws, p1, 25U * sizeof (uint64_t));
|
|
for (uint32_t i = 0U; i < 32U; i++)
|
|
{
|
|
store64_le(hbuf + i * 8U, ws[i]);
|
|
}
|
|
memcpy(dst + 64U - remOut, hbuf, remOut * sizeof (uint8_t));
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_Blake2S_a)
|
|
{
|
|
uint32_t *p1 = scrut.case_Blake2S_a;
|
|
Hacl_Hash_Blake2s_finish(32U, dst, p1);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_Blake2S_128_a)
|
|
{
|
|
Lib_IntVector_Intrinsics_vec128 *p1 = scrut.case_Blake2S_128_a;
|
|
#if HACL_CAN_COMPILE_VEC128
|
|
Hacl_Hash_Blake2s_Simd128_finish(32U, dst, p1);
|
|
return;
|
|
#else
|
|
KRML_MAYBE_UNUSED_VAR(p1);
|
|
return;
|
|
#endif
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_Blake2B_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_Blake2B_a;
|
|
Hacl_Hash_Blake2b_finish(64U, dst, p1);
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_Blake2B_256_a)
|
|
{
|
|
Lib_IntVector_Intrinsics_vec256 *p1 = scrut.case_Blake2B_256_a;
|
|
#if HACL_CAN_COMPILE_VEC256
|
|
Hacl_Hash_Blake2b_Simd256_finish(64U, dst, p1);
|
|
return;
|
|
#else
|
|
KRML_MAYBE_UNUSED_VAR(p1);
|
|
return;
|
|
#endif
|
|
}
|
|
KRML_HOST_EPRINTF("KaRaMeL abort at %s:%d\n%s\n",
|
|
__FILE__,
|
|
__LINE__,
|
|
"unreachable (pattern matches are exhaustive in F*)");
|
|
KRML_HOST_EXIT(255U);
|
|
}
|
|
|
|
static void free_(Hacl_Agile_Hash_state_s *s)
|
|
{
|
|
Hacl_Agile_Hash_state_s scrut = *s;
|
|
if (scrut.tag == Hacl_Agile_Hash_MD5_a)
|
|
{
|
|
uint32_t *p1 = scrut.case_MD5_a;
|
|
KRML_HOST_FREE(p1);
|
|
}
|
|
else if (scrut.tag == Hacl_Agile_Hash_SHA1_a)
|
|
{
|
|
uint32_t *p1 = scrut.case_SHA1_a;
|
|
KRML_HOST_FREE(p1);
|
|
}
|
|
else if (scrut.tag == Hacl_Agile_Hash_SHA2_224_a)
|
|
{
|
|
uint32_t *p1 = scrut.case_SHA2_224_a;
|
|
KRML_HOST_FREE(p1);
|
|
}
|
|
else if (scrut.tag == Hacl_Agile_Hash_SHA2_256_a)
|
|
{
|
|
uint32_t *p1 = scrut.case_SHA2_256_a;
|
|
KRML_HOST_FREE(p1);
|
|
}
|
|
else if (scrut.tag == Hacl_Agile_Hash_SHA2_384_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA2_384_a;
|
|
KRML_HOST_FREE(p1);
|
|
}
|
|
else if (scrut.tag == Hacl_Agile_Hash_SHA2_512_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA2_512_a;
|
|
KRML_HOST_FREE(p1);
|
|
}
|
|
else if (scrut.tag == Hacl_Agile_Hash_SHA3_224_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA3_224_a;
|
|
KRML_HOST_FREE(p1);
|
|
}
|
|
else if (scrut.tag == Hacl_Agile_Hash_SHA3_256_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA3_256_a;
|
|
KRML_HOST_FREE(p1);
|
|
}
|
|
else if (scrut.tag == Hacl_Agile_Hash_SHA3_384_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA3_384_a;
|
|
KRML_HOST_FREE(p1);
|
|
}
|
|
else if (scrut.tag == Hacl_Agile_Hash_SHA3_512_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_SHA3_512_a;
|
|
KRML_HOST_FREE(p1);
|
|
}
|
|
else if (scrut.tag == Hacl_Agile_Hash_Blake2S_a)
|
|
{
|
|
uint32_t *p1 = scrut.case_Blake2S_a;
|
|
KRML_HOST_FREE(p1);
|
|
}
|
|
else if (scrut.tag == Hacl_Agile_Hash_Blake2S_128_a)
|
|
{
|
|
Lib_IntVector_Intrinsics_vec128 *p1 = scrut.case_Blake2S_128_a;
|
|
KRML_ALIGNED_FREE(p1);
|
|
}
|
|
else if (scrut.tag == Hacl_Agile_Hash_Blake2B_a)
|
|
{
|
|
uint64_t *p1 = scrut.case_Blake2B_a;
|
|
KRML_HOST_FREE(p1);
|
|
}
|
|
else if (scrut.tag == Hacl_Agile_Hash_Blake2B_256_a)
|
|
{
|
|
Lib_IntVector_Intrinsics_vec256 *p1 = scrut.case_Blake2B_256_a;
|
|
KRML_ALIGNED_FREE(p1);
|
|
}
|
|
else
|
|
{
|
|
KRML_HOST_EPRINTF("KaRaMeL abort at %s:%d\n%s\n",
|
|
__FILE__,
|
|
__LINE__,
|
|
"unreachable (pattern matches are exhaustive in F*)");
|
|
KRML_HOST_EXIT(255U);
|
|
}
|
|
KRML_HOST_FREE(s);
|
|
}
|
|
|
|
static void copy(Hacl_Agile_Hash_state_s *s_src, Hacl_Agile_Hash_state_s *s_dst)
|
|
{
|
|
Hacl_Agile_Hash_state_s scrut = *s_src;
|
|
if (scrut.tag == Hacl_Agile_Hash_MD5_a)
|
|
{
|
|
uint32_t *p_src = scrut.case_MD5_a;
|
|
Hacl_Agile_Hash_state_s x1 = *s_dst;
|
|
uint32_t *p_dst;
|
|
if (x1.tag == Hacl_Agile_Hash_MD5_a)
|
|
{
|
|
p_dst = x1.case_MD5_a;
|
|
}
|
|
else
|
|
{
|
|
p_dst = KRML_EABORT(uint32_t *, "unreachable (pattern matches are exhaustive in F*)");
|
|
}
|
|
memcpy(p_dst, p_src, 4U * sizeof (uint32_t));
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA1_a)
|
|
{
|
|
uint32_t *p_src = scrut.case_SHA1_a;
|
|
Hacl_Agile_Hash_state_s x1 = *s_dst;
|
|
uint32_t *p_dst;
|
|
if (x1.tag == Hacl_Agile_Hash_SHA1_a)
|
|
{
|
|
p_dst = x1.case_SHA1_a;
|
|
}
|
|
else
|
|
{
|
|
p_dst = KRML_EABORT(uint32_t *, "unreachable (pattern matches are exhaustive in F*)");
|
|
}
|
|
memcpy(p_dst, p_src, 5U * sizeof (uint32_t));
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA2_224_a)
|
|
{
|
|
uint32_t *p_src = scrut.case_SHA2_224_a;
|
|
Hacl_Agile_Hash_state_s x1 = *s_dst;
|
|
uint32_t *p_dst;
|
|
if (x1.tag == Hacl_Agile_Hash_SHA2_224_a)
|
|
{
|
|
p_dst = x1.case_SHA2_224_a;
|
|
}
|
|
else
|
|
{
|
|
p_dst = KRML_EABORT(uint32_t *, "unreachable (pattern matches are exhaustive in F*)");
|
|
}
|
|
memcpy(p_dst, p_src, 8U * sizeof (uint32_t));
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA2_256_a)
|
|
{
|
|
uint32_t *p_src = scrut.case_SHA2_256_a;
|
|
Hacl_Agile_Hash_state_s x1 = *s_dst;
|
|
uint32_t *p_dst;
|
|
if (x1.tag == Hacl_Agile_Hash_SHA2_256_a)
|
|
{
|
|
p_dst = x1.case_SHA2_256_a;
|
|
}
|
|
else
|
|
{
|
|
p_dst = KRML_EABORT(uint32_t *, "unreachable (pattern matches are exhaustive in F*)");
|
|
}
|
|
memcpy(p_dst, p_src, 8U * sizeof (uint32_t));
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA2_384_a)
|
|
{
|
|
uint64_t *p_src = scrut.case_SHA2_384_a;
|
|
Hacl_Agile_Hash_state_s x1 = *s_dst;
|
|
uint64_t *p_dst;
|
|
if (x1.tag == Hacl_Agile_Hash_SHA2_384_a)
|
|
{
|
|
p_dst = x1.case_SHA2_384_a;
|
|
}
|
|
else
|
|
{
|
|
p_dst = KRML_EABORT(uint64_t *, "unreachable (pattern matches are exhaustive in F*)");
|
|
}
|
|
memcpy(p_dst, p_src, 8U * sizeof (uint64_t));
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA2_512_a)
|
|
{
|
|
uint64_t *p_src = scrut.case_SHA2_512_a;
|
|
Hacl_Agile_Hash_state_s x1 = *s_dst;
|
|
uint64_t *p_dst;
|
|
if (x1.tag == Hacl_Agile_Hash_SHA2_512_a)
|
|
{
|
|
p_dst = x1.case_SHA2_512_a;
|
|
}
|
|
else
|
|
{
|
|
p_dst = KRML_EABORT(uint64_t *, "unreachable (pattern matches are exhaustive in F*)");
|
|
}
|
|
memcpy(p_dst, p_src, 8U * sizeof (uint64_t));
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA3_224_a)
|
|
{
|
|
uint64_t *p_src = scrut.case_SHA3_224_a;
|
|
Hacl_Agile_Hash_state_s x1 = *s_dst;
|
|
uint64_t *p_dst;
|
|
if (x1.tag == Hacl_Agile_Hash_SHA3_224_a)
|
|
{
|
|
p_dst = x1.case_SHA3_224_a;
|
|
}
|
|
else
|
|
{
|
|
p_dst = KRML_EABORT(uint64_t *, "unreachable (pattern matches are exhaustive in F*)");
|
|
}
|
|
memcpy(p_dst, p_src, 25U * sizeof (uint64_t));
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA3_256_a)
|
|
{
|
|
uint64_t *p_src = scrut.case_SHA3_256_a;
|
|
Hacl_Agile_Hash_state_s x1 = *s_dst;
|
|
uint64_t *p_dst;
|
|
if (x1.tag == Hacl_Agile_Hash_SHA3_256_a)
|
|
{
|
|
p_dst = x1.case_SHA3_256_a;
|
|
}
|
|
else
|
|
{
|
|
p_dst = KRML_EABORT(uint64_t *, "unreachable (pattern matches are exhaustive in F*)");
|
|
}
|
|
memcpy(p_dst, p_src, 25U * sizeof (uint64_t));
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA3_384_a)
|
|
{
|
|
uint64_t *p_src = scrut.case_SHA3_384_a;
|
|
Hacl_Agile_Hash_state_s x1 = *s_dst;
|
|
uint64_t *p_dst;
|
|
if (x1.tag == Hacl_Agile_Hash_SHA3_384_a)
|
|
{
|
|
p_dst = x1.case_SHA3_384_a;
|
|
}
|
|
else
|
|
{
|
|
p_dst = KRML_EABORT(uint64_t *, "unreachable (pattern matches are exhaustive in F*)");
|
|
}
|
|
memcpy(p_dst, p_src, 25U * sizeof (uint64_t));
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_SHA3_512_a)
|
|
{
|
|
uint64_t *p_src = scrut.case_SHA3_512_a;
|
|
Hacl_Agile_Hash_state_s x1 = *s_dst;
|
|
uint64_t *p_dst;
|
|
if (x1.tag == Hacl_Agile_Hash_SHA3_512_a)
|
|
{
|
|
p_dst = x1.case_SHA3_512_a;
|
|
}
|
|
else
|
|
{
|
|
p_dst = KRML_EABORT(uint64_t *, "unreachable (pattern matches are exhaustive in F*)");
|
|
}
|
|
memcpy(p_dst, p_src, 25U * sizeof (uint64_t));
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_Blake2S_a)
|
|
{
|
|
uint32_t *p_src = scrut.case_Blake2S_a;
|
|
Hacl_Agile_Hash_state_s x1 = *s_dst;
|
|
uint32_t *p_dst;
|
|
if (x1.tag == Hacl_Agile_Hash_Blake2S_a)
|
|
{
|
|
p_dst = x1.case_Blake2S_a;
|
|
}
|
|
else
|
|
{
|
|
p_dst = KRML_EABORT(uint32_t *, "unreachable (pattern matches are exhaustive in F*)");
|
|
}
|
|
memcpy(p_dst, p_src, 16U * sizeof (uint32_t));
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_Blake2S_128_a)
|
|
{
|
|
Lib_IntVector_Intrinsics_vec128 *p_src = scrut.case_Blake2S_128_a;
|
|
KRML_MAYBE_UNUSED_VAR(p_src);
|
|
#if HACL_CAN_COMPILE_VEC128
|
|
Hacl_Agile_Hash_state_s x1 = *s_dst;
|
|
Lib_IntVector_Intrinsics_vec128 *p_dst;
|
|
if (x1.tag == Hacl_Agile_Hash_Blake2S_128_a)
|
|
{
|
|
p_dst = x1.case_Blake2S_128_a;
|
|
}
|
|
else
|
|
{
|
|
p_dst =
|
|
KRML_EABORT(Lib_IntVector_Intrinsics_vec128 *,
|
|
"unreachable (pattern matches are exhaustive in F*)");
|
|
}
|
|
Hacl_Hash_Blake2s_Simd128_copy_internal_state(p_src, p_dst);
|
|
return;
|
|
#else
|
|
return;
|
|
#endif
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_Blake2B_a)
|
|
{
|
|
uint64_t *p_src = scrut.case_Blake2B_a;
|
|
Hacl_Agile_Hash_state_s x1 = *s_dst;
|
|
uint64_t *p_dst;
|
|
if (x1.tag == Hacl_Agile_Hash_Blake2B_a)
|
|
{
|
|
p_dst = x1.case_Blake2B_a;
|
|
}
|
|
else
|
|
{
|
|
p_dst = KRML_EABORT(uint64_t *, "unreachable (pattern matches are exhaustive in F*)");
|
|
}
|
|
memcpy(p_dst, p_src, 16U * sizeof (uint64_t));
|
|
return;
|
|
}
|
|
if (scrut.tag == Hacl_Agile_Hash_Blake2B_256_a)
|
|
{
|
|
Lib_IntVector_Intrinsics_vec256 *p_src = scrut.case_Blake2B_256_a;
|
|
KRML_MAYBE_UNUSED_VAR(p_src);
|
|
#if HACL_CAN_COMPILE_VEC256
|
|
Hacl_Agile_Hash_state_s x1 = *s_dst;
|
|
Lib_IntVector_Intrinsics_vec256 *p_dst;
|
|
if (x1.tag == Hacl_Agile_Hash_Blake2B_256_a)
|
|
{
|
|
p_dst = x1.case_Blake2B_256_a;
|
|
}
|
|
else
|
|
{
|
|
p_dst =
|
|
KRML_EABORT(Lib_IntVector_Intrinsics_vec256 *,
|
|
"unreachable (pattern matches are exhaustive in F*)");
|
|
}
|
|
Hacl_Hash_Blake2b_Simd256_copy_internal_state(p_src, p_dst);
|
|
return;
|
|
#else
|
|
return;
|
|
#endif
|
|
}
|
|
KRML_HOST_EPRINTF("KaRaMeL abort at %s:%d\n%s\n",
|
|
__FILE__,
|
|
__LINE__,
|
|
"unreachable (pattern matches are exhaustive in F*)");
|
|
KRML_HOST_EXIT(255U);
|
|
}
|
|
|
|
static void hash(Hacl_Agile_Hash_impl i, uint8_t *dst, uint8_t *input, uint32_t input_len)
|
|
{
|
|
switch (i)
|
|
{
|
|
case Hacl_Agile_Hash_MD5:
|
|
{
|
|
Hacl_Hash_MD5_hash_oneshot(dst, input, input_len);
|
|
break;
|
|
}
|
|
case Hacl_Agile_Hash_SHA1:
|
|
{
|
|
Hacl_Hash_SHA1_hash_oneshot(dst, input, input_len);
|
|
break;
|
|
}
|
|
case Hacl_Agile_Hash_SHA2_224:
|
|
{
|
|
Hacl_Hash_SHA2_hash_224(dst, input, input_len);
|
|
break;
|
|
}
|
|
case Hacl_Agile_Hash_SHA2_256:
|
|
{
|
|
Hacl_Hash_SHA2_hash_256(dst, input, input_len);
|
|
break;
|
|
}
|
|
case Hacl_Agile_Hash_SHA2_384:
|
|
{
|
|
Hacl_Hash_SHA2_hash_384(dst, input, input_len);
|
|
break;
|
|
}
|
|
case Hacl_Agile_Hash_SHA2_512:
|
|
{
|
|
Hacl_Hash_SHA2_hash_512(dst, input, input_len);
|
|
break;
|
|
}
|
|
case Hacl_Agile_Hash_SHA3_224:
|
|
{
|
|
Hacl_Hash_SHA3_sha3_224(dst, input, input_len);
|
|
break;
|
|
}
|
|
case Hacl_Agile_Hash_SHA3_256:
|
|
{
|
|
Hacl_Hash_SHA3_sha3_256(dst, input, input_len);
|
|
break;
|
|
}
|
|
case Hacl_Agile_Hash_SHA3_384:
|
|
{
|
|
Hacl_Hash_SHA3_sha3_384(dst, input, input_len);
|
|
break;
|
|
}
|
|
case Hacl_Agile_Hash_SHA3_512:
|
|
{
|
|
Hacl_Hash_SHA3_sha3_512(dst, input, input_len);
|
|
break;
|
|
}
|
|
case Hacl_Agile_Hash_Blake2S_32:
|
|
{
|
|
Hacl_Hash_Blake2s_hash_with_key(dst, 32U, input, input_len, NULL, 0U);
|
|
break;
|
|
}
|
|
case Hacl_Agile_Hash_Blake2S_128:
|
|
{
|
|
#if HACL_CAN_COMPILE_VEC128
|
|
Hacl_Hash_Blake2s_Simd128_hash_with_key(dst, 32U, input, input_len, NULL, 0U);
|
|
#endif
|
|
break;
|
|
}
|
|
case Hacl_Agile_Hash_Blake2B_32:
|
|
{
|
|
Hacl_Hash_Blake2b_hash_with_key(dst, 64U, input, input_len, NULL, 0U);
|
|
break;
|
|
}
|
|
case Hacl_Agile_Hash_Blake2B_256:
|
|
{
|
|
#if HACL_CAN_COMPILE_VEC256
|
|
Hacl_Hash_Blake2b_Simd256_hash_with_key(dst, 64U, input, input_len, NULL, 0U);
|
|
#endif
|
|
break;
|
|
}
|
|
default:
|
|
{
|
|
KRML_HOST_EPRINTF("KaRaMeL incomplete match at %s:%d\n", __FILE__, __LINE__);
|
|
KRML_HOST_EXIT(253U);
|
|
}
|
|
}
|
|
}
|
|
|
|
static uint32_t hash_len(Spec_Hash_Definitions_hash_alg a)
|
|
{
|
|
switch (a)
|
|
{
|
|
case Spec_Hash_Definitions_MD5:
|
|
{
|
|
return 16U;
|
|
}
|
|
case Spec_Hash_Definitions_SHA1:
|
|
{
|
|
return 20U;
|
|
}
|
|
case Spec_Hash_Definitions_SHA2_224:
|
|
{
|
|
return 28U;
|
|
}
|
|
case Spec_Hash_Definitions_SHA2_256:
|
|
{
|
|
return 32U;
|
|
}
|
|
case Spec_Hash_Definitions_SHA2_384:
|
|
{
|
|
return 48U;
|
|
}
|
|
case Spec_Hash_Definitions_SHA2_512:
|
|
{
|
|
return 64U;
|
|
}
|
|
case Spec_Hash_Definitions_Blake2S:
|
|
{
|
|
return 32U;
|
|
}
|
|
case Spec_Hash_Definitions_Blake2B:
|
|
{
|
|
return 64U;
|
|
}
|
|
case Spec_Hash_Definitions_SHA3_224:
|
|
{
|
|
return 28U;
|
|
}
|
|
case Spec_Hash_Definitions_SHA3_256:
|
|
{
|
|
return 32U;
|
|
}
|
|
case Spec_Hash_Definitions_SHA3_384:
|
|
{
|
|
return 48U;
|
|
}
|
|
case Spec_Hash_Definitions_SHA3_512:
|
|
{
|
|
return 64U;
|
|
}
|
|
default:
|
|
{
|
|
KRML_HOST_EPRINTF("KaRaMeL incomplete match at %s:%d\n", __FILE__, __LINE__);
|
|
KRML_HOST_EXIT(253U);
|
|
}
|
|
}
|
|
}
|
|
|
|
static uint32_t block_len(Spec_Hash_Definitions_hash_alg a)
|
|
{
|
|
switch (a)
|
|
{
|
|
case Spec_Hash_Definitions_MD5:
|
|
{
|
|
return 64U;
|
|
}
|
|
case Spec_Hash_Definitions_SHA1:
|
|
{
|
|
return 64U;
|
|
}
|
|
case Spec_Hash_Definitions_SHA2_224:
|
|
{
|
|
return 64U;
|
|
}
|
|
case Spec_Hash_Definitions_SHA2_256:
|
|
{
|
|
return 64U;
|
|
}
|
|
case Spec_Hash_Definitions_SHA2_384:
|
|
{
|
|
return 128U;
|
|
}
|
|
case Spec_Hash_Definitions_SHA2_512:
|
|
{
|
|
return 128U;
|
|
}
|
|
case Spec_Hash_Definitions_SHA3_224:
|
|
{
|
|
return 144U;
|
|
}
|
|
case Spec_Hash_Definitions_SHA3_256:
|
|
{
|
|
return 136U;
|
|
}
|
|
case Spec_Hash_Definitions_SHA3_384:
|
|
{
|
|
return 104U;
|
|
}
|
|
case Spec_Hash_Definitions_SHA3_512:
|
|
{
|
|
return 72U;
|
|
}
|
|
case Spec_Hash_Definitions_Shake128:
|
|
{
|
|
return 168U;
|
|
}
|
|
case Spec_Hash_Definitions_Shake256:
|
|
{
|
|
return 136U;
|
|
}
|
|
case Spec_Hash_Definitions_Blake2S:
|
|
{
|
|
return 64U;
|
|
}
|
|
case Spec_Hash_Definitions_Blake2B:
|
|
{
|
|
return 128U;
|
|
}
|
|
default:
|
|
{
|
|
KRML_HOST_EPRINTF("KaRaMeL incomplete match at %s:%d\n", __FILE__, __LINE__);
|
|
KRML_HOST_EXIT(253U);
|
|
}
|
|
}
|
|
}
|
|
|
|
static uint64_t max_input_len64(Spec_Hash_Definitions_hash_alg a)
|
|
{
|
|
switch (a)
|
|
{
|
|
case Spec_Hash_Definitions_MD5:
|
|
{
|
|
return 2305843009213693951ULL;
|
|
}
|
|
case Spec_Hash_Definitions_SHA1:
|
|
{
|
|
return 2305843009213693951ULL;
|
|
}
|
|
case Spec_Hash_Definitions_SHA2_224:
|
|
{
|
|
return 2305843009213693951ULL;
|
|
}
|
|
case Spec_Hash_Definitions_SHA2_256:
|
|
{
|
|
return 2305843009213693951ULL;
|
|
}
|
|
case Spec_Hash_Definitions_SHA2_384:
|
|
{
|
|
return 18446744073709551615ULL;
|
|
}
|
|
case Spec_Hash_Definitions_SHA2_512:
|
|
{
|
|
return 18446744073709551615ULL;
|
|
}
|
|
case Spec_Hash_Definitions_Blake2S:
|
|
{
|
|
return 18446744073709551615ULL;
|
|
}
|
|
case Spec_Hash_Definitions_Blake2B:
|
|
{
|
|
return 18446744073709551615ULL;
|
|
}
|
|
case Spec_Hash_Definitions_SHA3_224:
|
|
{
|
|
return 18446744073709551615ULL;
|
|
}
|
|
case Spec_Hash_Definitions_SHA3_256:
|
|
{
|
|
return 18446744073709551615ULL;
|
|
}
|
|
case Spec_Hash_Definitions_SHA3_384:
|
|
{
|
|
return 18446744073709551615ULL;
|
|
}
|
|
case Spec_Hash_Definitions_SHA3_512:
|
|
{
|
|
return 18446744073709551615ULL;
|
|
}
|
|
case Spec_Hash_Definitions_Shake128:
|
|
{
|
|
return 18446744073709551615ULL;
|
|
}
|
|
case Spec_Hash_Definitions_Shake256:
|
|
{
|
|
return 18446744073709551615ULL;
|
|
}
|
|
default:
|
|
{
|
|
KRML_HOST_EPRINTF("KaRaMeL incomplete match at %s:%d\n", __FILE__, __LINE__);
|
|
KRML_HOST_EXIT(253U);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void wrap_key(Hacl_Agile_Hash_impl impl, uint8_t *output, uint8_t *key, uint32_t len)
|
|
{
|
|
uint8_t *nkey = output;
|
|
uint32_t ite;
|
|
if (len <= block_len(alg_of_impl(impl)))
|
|
{
|
|
ite = len;
|
|
}
|
|
else
|
|
{
|
|
ite = hash_len(alg_of_impl(impl));
|
|
}
|
|
uint8_t *zeroes = output + ite;
|
|
KRML_MAYBE_UNUSED_VAR(zeroes);
|
|
if (len <= block_len(alg_of_impl(impl)))
|
|
{
|
|
if (len > 0U)
|
|
{
|
|
memcpy(nkey, key, len * sizeof (uint8_t));
|
|
return;
|
|
}
|
|
return;
|
|
}
|
|
hash(impl, nkey, key, len);
|
|
}
|
|
|
|
static void init0(uint8_t *k, uint8_t *buf, Hacl_Streaming_HMAC_Definitions_two_state s)
|
|
{
|
|
uint32_t k_len = s.fst;
|
|
Hacl_Agile_Hash_state_s *s1 = s.snd;
|
|
Hacl_Agile_Hash_state_s *s2 = s.thd;
|
|
init(s1);
|
|
init(s2);
|
|
Hacl_Agile_Hash_impl i1 = impl_of_state(s1);
|
|
Spec_Hash_Definitions_hash_alg a = alg_of_impl(i1);
|
|
uint8_t b0[168U] = { 0U };
|
|
uint8_t *block = b0;
|
|
wrap_key(i1, block, k, k_len);
|
|
uint8_t b1[168U];
|
|
memset(b1, 0x36U, 168U * sizeof (uint8_t));
|
|
uint8_t *ipad = b1;
|
|
uint8_t b[168U];
|
|
memset(b, 0x5cU, 168U * sizeof (uint8_t));
|
|
uint8_t *opad = b;
|
|
for (uint32_t i = 0U; i < block_len(a); i++)
|
|
{
|
|
uint8_t xi = ipad[i];
|
|
uint8_t yi = block[i];
|
|
buf[i] = (uint32_t)xi ^ (uint32_t)yi;
|
|
}
|
|
for (uint32_t i = 0U; i < block_len(a); i++)
|
|
{
|
|
uint8_t xi = opad[i];
|
|
uint8_t yi = block[i];
|
|
opad[i] = (uint32_t)xi ^ (uint32_t)yi;
|
|
}
|
|
update_multi(s2, 0ULL, opad, block_len(a));
|
|
}
|
|
|
|
static void finish0(Hacl_Streaming_HMAC_Definitions_two_state s, uint8_t *dst)
|
|
{
|
|
Hacl_Agile_Hash_state_s *s2 = s.thd;
|
|
Hacl_Agile_Hash_state_s *s1 = s.snd;
|
|
Hacl_Agile_Hash_impl i1 = impl_of_state(s1);
|
|
Spec_Hash_Definitions_hash_alg a = alg_of_impl(i1);
|
|
finish(s1, dst);
|
|
update_last(s2, (uint64_t)block_len(a), dst, hash_len(a));
|
|
finish(s2, dst);
|
|
}
|
|
|
|
Hacl_Agile_Hash_state_s
|
|
*Hacl_Streaming_HMAC_s1(
|
|
Hacl_Streaming_HMAC_Definitions_index i,
|
|
Hacl_Streaming_HMAC_Definitions_two_state s
|
|
)
|
|
{
|
|
KRML_MAYBE_UNUSED_VAR(i);
|
|
return s.snd;
|
|
}
|
|
|
|
Hacl_Agile_Hash_state_s
|
|
*Hacl_Streaming_HMAC_s2(
|
|
Hacl_Streaming_HMAC_Definitions_index i,
|
|
Hacl_Streaming_HMAC_Definitions_two_state s
|
|
)
|
|
{
|
|
KRML_MAYBE_UNUSED_VAR(i);
|
|
return s.thd;
|
|
}
|
|
|
|
Hacl_Streaming_HMAC_Definitions_index
|
|
Hacl_Streaming_HMAC_index_of_state(Hacl_Streaming_HMAC_Definitions_two_state s)
|
|
{
|
|
Hacl_Agile_Hash_state_s *s11 = s.snd;
|
|
uint32_t kl = s.fst;
|
|
Hacl_Agile_Hash_impl i1 = impl_of_state(s11);
|
|
return ((Hacl_Streaming_HMAC_Definitions_index){ .fst = i1, .snd = kl });
|
|
}
|
|
|
|
static Hacl_Agile_Hash_impl
|
|
__proj__Mkdtuple2__item___1__Hacl_Agile_Hash_impl_uint32_t(
|
|
Hacl_Streaming_HMAC_Definitions_index pair
|
|
)
|
|
{
|
|
return pair.fst;
|
|
}
|
|
|
|
static Hacl_Agile_Hash_impl
|
|
dfst__Hacl_Agile_Hash_impl_uint32_t(Hacl_Streaming_HMAC_Definitions_index t)
|
|
{
|
|
return __proj__Mkdtuple2__item___1__Hacl_Agile_Hash_impl_uint32_t(t);
|
|
}
|
|
|
|
static uint32_t
|
|
__proj__Mkdtuple2__item___2__Hacl_Agile_Hash_impl_uint32_t(
|
|
Hacl_Streaming_HMAC_Definitions_index pair
|
|
)
|
|
{
|
|
return pair.snd;
|
|
}
|
|
|
|
static uint32_t dsnd__Hacl_Agile_Hash_impl_uint32_t(Hacl_Streaming_HMAC_Definitions_index t)
|
|
{
|
|
return __proj__Mkdtuple2__item___2__Hacl_Agile_Hash_impl_uint32_t(t);
|
|
}
|
|
|
|
typedef struct option___uint32_t____Hacl_Agile_Hash_state_s_____Hacl_Agile_Hash_state_s___s
|
|
{
|
|
Hacl_Streaming_Types_optional tag;
|
|
Hacl_Streaming_HMAC_Definitions_two_state v;
|
|
}
|
|
option___uint32_t____Hacl_Agile_Hash_state_s_____Hacl_Agile_Hash_state_s__;
|
|
|
|
KRML_MAYBE_UNUSED static Hacl_Streaming_HMAC_agile_state
|
|
*malloc_internal(Hacl_Streaming_HMAC_Definitions_index i, uint8_t *key)
|
|
{
|
|
KRML_CHECK_SIZE(sizeof (uint8_t),
|
|
block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i))));
|
|
uint8_t
|
|
*buf =
|
|
(uint8_t *)KRML_HOST_CALLOC(block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i))),
|
|
sizeof (uint8_t));
|
|
if (buf == NULL)
|
|
{
|
|
return NULL;
|
|
}
|
|
uint8_t *buf1 = buf;
|
|
Hacl_Agile_Hash_state_s *s110 = malloc_(dfst__Hacl_Agile_Hash_impl_uint32_t(i));
|
|
option___uint32_t____Hacl_Agile_Hash_state_s_____Hacl_Agile_Hash_state_s__ block_state;
|
|
if (s110 == NULL)
|
|
{
|
|
block_state =
|
|
(
|
|
(option___uint32_t____Hacl_Agile_Hash_state_s_____Hacl_Agile_Hash_state_s__){
|
|
.tag = Hacl_Streaming_Types_None
|
|
}
|
|
);
|
|
}
|
|
else
|
|
{
|
|
Hacl_Agile_Hash_state_s *s21 = malloc_(dfst__Hacl_Agile_Hash_impl_uint32_t(i));
|
|
if (s21 == NULL)
|
|
{
|
|
KRML_HOST_FREE(s110);
|
|
block_state =
|
|
(
|
|
(option___uint32_t____Hacl_Agile_Hash_state_s_____Hacl_Agile_Hash_state_s__){
|
|
.tag = Hacl_Streaming_Types_None
|
|
}
|
|
);
|
|
}
|
|
else
|
|
{
|
|
block_state =
|
|
(
|
|
(option___uint32_t____Hacl_Agile_Hash_state_s_____Hacl_Agile_Hash_state_s__){
|
|
.tag = Hacl_Streaming_Types_Some,
|
|
.v = { .fst = dsnd__Hacl_Agile_Hash_impl_uint32_t(i), .snd = s110, .thd = s21 }
|
|
}
|
|
);
|
|
}
|
|
}
|
|
if (block_state.tag == Hacl_Streaming_Types_None)
|
|
{
|
|
KRML_HOST_FREE(buf1);
|
|
return NULL;
|
|
}
|
|
if (block_state.tag == Hacl_Streaming_Types_Some)
|
|
{
|
|
Hacl_Streaming_HMAC_Definitions_two_state block_state1 = block_state.v;
|
|
Hacl_Streaming_Types_optional k_ = Hacl_Streaming_Types_Some;
|
|
switch (k_)
|
|
{
|
|
case Hacl_Streaming_Types_None:
|
|
{
|
|
return NULL;
|
|
}
|
|
case Hacl_Streaming_Types_Some:
|
|
{
|
|
Hacl_Streaming_HMAC_agile_state
|
|
s =
|
|
{
|
|
.block_state = block_state1,
|
|
.buf = buf1,
|
|
.total_len = (uint64_t)block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i)))
|
|
};
|
|
Hacl_Streaming_HMAC_agile_state
|
|
*p =
|
|
(Hacl_Streaming_HMAC_agile_state *)KRML_HOST_MALLOC(sizeof (
|
|
Hacl_Streaming_HMAC_agile_state
|
|
));
|
|
if (p != NULL)
|
|
{
|
|
p[0U] = s;
|
|
}
|
|
if (p == NULL)
|
|
{
|
|
Hacl_Agile_Hash_state_s *s21 = block_state1.thd;
|
|
Hacl_Agile_Hash_state_s *s11 = block_state1.snd;
|
|
free_(s11);
|
|
free_(s21);
|
|
KRML_HOST_FREE(buf1);
|
|
return NULL;
|
|
}
|
|
init0(key, buf1, block_state1);
|
|
return p;
|
|
}
|
|
default:
|
|
{
|
|
KRML_HOST_EPRINTF("KaRaMeL incomplete match at %s:%d\n", __FILE__, __LINE__);
|
|
KRML_HOST_EXIT(253U);
|
|
}
|
|
}
|
|
}
|
|
KRML_HOST_EPRINTF("KaRaMeL abort at %s:%d\n%s\n",
|
|
__FILE__,
|
|
__LINE__,
|
|
"unreachable (pattern matches are exhaustive in F*)");
|
|
KRML_HOST_EXIT(255U);
|
|
}
|
|
|
|
KRML_MAYBE_UNUSED static bool is_blake2b_256(Hacl_Agile_Hash_impl uu___)
|
|
{
|
|
switch (uu___)
|
|
{
|
|
case Hacl_Agile_Hash_Blake2B_256:
|
|
{
|
|
return true;
|
|
}
|
|
default:
|
|
{
|
|
return false;
|
|
}
|
|
}
|
|
}
|
|
|
|
KRML_MAYBE_UNUSED static bool is_blake2s_128(Hacl_Agile_Hash_impl uu___)
|
|
{
|
|
switch (uu___)
|
|
{
|
|
case Hacl_Agile_Hash_Blake2S_128:
|
|
{
|
|
return true;
|
|
}
|
|
default:
|
|
{
|
|
return false;
|
|
}
|
|
}
|
|
}
|
|
|
|
Hacl_Streaming_Types_error_code
|
|
Hacl_Streaming_HMAC_malloc_(
|
|
Hacl_Agile_Hash_impl impl,
|
|
uint8_t *key,
|
|
uint32_t key_length,
|
|
Hacl_Streaming_HMAC_agile_state **dst
|
|
)
|
|
{
|
|
KRML_MAYBE_UNUSED_VAR(key);
|
|
KRML_MAYBE_UNUSED_VAR(key_length);
|
|
KRML_MAYBE_UNUSED_VAR(dst);
|
|
#if !HACL_CAN_COMPILE_VEC256
|
|
if (is_blake2b_256(impl))
|
|
{
|
|
return Hacl_Streaming_Types_InvalidAlgorithm;
|
|
}
|
|
#endif
|
|
#if !HACL_CAN_COMPILE_VEC128
|
|
if (is_blake2s_128(impl))
|
|
{
|
|
return Hacl_Streaming_Types_InvalidAlgorithm;
|
|
}
|
|
#endif
|
|
Hacl_Streaming_HMAC_agile_state
|
|
*st =
|
|
malloc_internal(((Hacl_Streaming_HMAC_Definitions_index){ .fst = impl, .snd = key_length }),
|
|
key);
|
|
if (st == NULL)
|
|
{
|
|
return Hacl_Streaming_Types_OutOfMemory;
|
|
}
|
|
*dst = st;
|
|
return Hacl_Streaming_Types_Success;
|
|
}
|
|
|
|
Hacl_Streaming_HMAC_Definitions_index
|
|
Hacl_Streaming_HMAC_get_impl(Hacl_Streaming_HMAC_agile_state *s)
|
|
{
|
|
Hacl_Streaming_HMAC_Definitions_two_state block_state = (*s).block_state;
|
|
return Hacl_Streaming_HMAC_index_of_state(block_state);
|
|
}
|
|
|
|
static void reset_internal(Hacl_Streaming_HMAC_agile_state *state, uint8_t *key)
|
|
{
|
|
Hacl_Streaming_HMAC_agile_state scrut = *state;
|
|
uint8_t *buf = scrut.buf;
|
|
Hacl_Streaming_HMAC_Definitions_two_state block_state = scrut.block_state;
|
|
Hacl_Streaming_HMAC_Definitions_index i1 = Hacl_Streaming_HMAC_index_of_state(block_state);
|
|
KRML_MAYBE_UNUSED_VAR(i1);
|
|
init0(key, buf, block_state);
|
|
Hacl_Streaming_HMAC_agile_state
|
|
tmp =
|
|
{
|
|
.block_state = block_state,
|
|
.buf = buf,
|
|
.total_len = (uint64_t)block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1)))
|
|
};
|
|
state[0U] = tmp;
|
|
}
|
|
|
|
Hacl_Streaming_Types_error_code
|
|
Hacl_Streaming_HMAC_reset(
|
|
Hacl_Streaming_HMAC_agile_state *state,
|
|
uint8_t *key,
|
|
uint32_t key_length
|
|
)
|
|
{
|
|
uint32_t k_len = Hacl_Streaming_HMAC_get_impl(state).snd;
|
|
if (key_length != k_len)
|
|
{
|
|
return Hacl_Streaming_Types_InvalidLength;
|
|
}
|
|
reset_internal(state, key);
|
|
return Hacl_Streaming_Types_Success;
|
|
}
|
|
|
|
Hacl_Streaming_Types_error_code
|
|
Hacl_Streaming_HMAC_update(
|
|
Hacl_Streaming_HMAC_agile_state *state,
|
|
uint8_t *chunk,
|
|
uint32_t chunk_len
|
|
)
|
|
{
|
|
Hacl_Streaming_HMAC_agile_state s = *state;
|
|
Hacl_Streaming_HMAC_Definitions_two_state block_state = s.block_state;
|
|
uint64_t total_len = s.total_len;
|
|
Hacl_Streaming_HMAC_Definitions_index i1 = Hacl_Streaming_HMAC_index_of_state(block_state);
|
|
if
|
|
(
|
|
(uint64_t)chunk_len >
|
|
max_input_len64(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1))) - total_len
|
|
)
|
|
{
|
|
return Hacl_Streaming_Types_MaximumLengthExceeded;
|
|
}
|
|
uint32_t sz;
|
|
if
|
|
(
|
|
total_len % (uint64_t)block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1))) == 0ULL
|
|
&& total_len > 0ULL
|
|
)
|
|
{
|
|
sz = block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1)));
|
|
}
|
|
else
|
|
{
|
|
sz =
|
|
(uint32_t)(total_len %
|
|
(uint64_t)block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1))));
|
|
}
|
|
if (chunk_len <= block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1))) - sz)
|
|
{
|
|
Hacl_Streaming_HMAC_agile_state s3 = *state;
|
|
Hacl_Streaming_HMAC_Definitions_two_state block_state1 = s3.block_state;
|
|
uint8_t *buf = s3.buf;
|
|
uint64_t total_len1 = s3.total_len;
|
|
uint32_t sz1;
|
|
if
|
|
(
|
|
total_len1 % (uint64_t)block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1))) == 0ULL
|
|
&& total_len1 > 0ULL
|
|
)
|
|
{
|
|
sz1 = block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1)));
|
|
}
|
|
else
|
|
{
|
|
sz1 =
|
|
(uint32_t)(total_len1 %
|
|
(uint64_t)block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1))));
|
|
}
|
|
uint8_t *buf2 = buf + sz1;
|
|
memcpy(buf2, chunk, chunk_len * sizeof (uint8_t));
|
|
uint64_t total_len2 = total_len1 + (uint64_t)chunk_len;
|
|
*state =
|
|
(
|
|
(Hacl_Streaming_HMAC_agile_state){
|
|
.block_state = block_state1,
|
|
.buf = buf,
|
|
.total_len = total_len2
|
|
}
|
|
);
|
|
}
|
|
else if (sz == 0U)
|
|
{
|
|
Hacl_Streaming_HMAC_agile_state s3 = *state;
|
|
Hacl_Streaming_HMAC_Definitions_two_state block_state1 = s3.block_state;
|
|
uint8_t *buf = s3.buf;
|
|
uint64_t total_len1 = s3.total_len;
|
|
uint32_t sz1;
|
|
if
|
|
(
|
|
total_len1 % (uint64_t)block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1))) == 0ULL
|
|
&& total_len1 > 0ULL
|
|
)
|
|
{
|
|
sz1 = block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1)));
|
|
}
|
|
else
|
|
{
|
|
sz1 =
|
|
(uint32_t)(total_len1 %
|
|
(uint64_t)block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1))));
|
|
}
|
|
if (!(sz1 == 0U))
|
|
{
|
|
uint64_t prevlen = total_len1 - (uint64_t)sz1;
|
|
Hacl_Agile_Hash_state_s *s11 = block_state1.snd;
|
|
update_multi(s11,
|
|
prevlen,
|
|
buf,
|
|
block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1))));
|
|
}
|
|
uint32_t ite;
|
|
if
|
|
(
|
|
(uint64_t)chunk_len %
|
|
(uint64_t)block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1)))
|
|
== 0ULL
|
|
&& (uint64_t)chunk_len > 0ULL
|
|
)
|
|
{
|
|
ite = block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1)));
|
|
}
|
|
else
|
|
{
|
|
ite =
|
|
(uint32_t)((uint64_t)chunk_len %
|
|
(uint64_t)block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1))));
|
|
}
|
|
uint32_t
|
|
n_blocks = (chunk_len - ite) / block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1)));
|
|
uint32_t
|
|
data1_len = n_blocks * block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1)));
|
|
uint32_t data2_len = chunk_len - data1_len;
|
|
uint8_t *data1 = chunk;
|
|
uint8_t *data2 = chunk + data1_len;
|
|
Hacl_Agile_Hash_state_s *s11 = block_state1.snd;
|
|
update_multi(s11, total_len1, data1, data1_len);
|
|
uint8_t *dst = buf;
|
|
memcpy(dst, data2, data2_len * sizeof (uint8_t));
|
|
*state =
|
|
(
|
|
(Hacl_Streaming_HMAC_agile_state){
|
|
.block_state = block_state1,
|
|
.buf = buf,
|
|
.total_len = total_len1 + (uint64_t)chunk_len
|
|
}
|
|
);
|
|
}
|
|
else
|
|
{
|
|
uint32_t diff = block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1))) - sz;
|
|
uint8_t *chunk1 = chunk;
|
|
uint8_t *chunk2 = chunk + diff;
|
|
Hacl_Streaming_HMAC_agile_state s3 = *state;
|
|
Hacl_Streaming_HMAC_Definitions_two_state block_state10 = s3.block_state;
|
|
uint8_t *buf0 = s3.buf;
|
|
uint64_t total_len10 = s3.total_len;
|
|
uint32_t sz10;
|
|
if
|
|
(
|
|
total_len10 % (uint64_t)block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1))) ==
|
|
0ULL
|
|
&& total_len10 > 0ULL
|
|
)
|
|
{
|
|
sz10 = block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1)));
|
|
}
|
|
else
|
|
{
|
|
sz10 =
|
|
(uint32_t)(total_len10 %
|
|
(uint64_t)block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1))));
|
|
}
|
|
uint8_t *buf2 = buf0 + sz10;
|
|
memcpy(buf2, chunk1, diff * sizeof (uint8_t));
|
|
uint64_t total_len2 = total_len10 + (uint64_t)diff;
|
|
*state =
|
|
(
|
|
(Hacl_Streaming_HMAC_agile_state){
|
|
.block_state = block_state10,
|
|
.buf = buf0,
|
|
.total_len = total_len2
|
|
}
|
|
);
|
|
Hacl_Streaming_HMAC_agile_state s30 = *state;
|
|
Hacl_Streaming_HMAC_Definitions_two_state block_state1 = s30.block_state;
|
|
uint8_t *buf = s30.buf;
|
|
uint64_t total_len1 = s30.total_len;
|
|
uint32_t sz1;
|
|
if
|
|
(
|
|
total_len1 % (uint64_t)block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1))) == 0ULL
|
|
&& total_len1 > 0ULL
|
|
)
|
|
{
|
|
sz1 = block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1)));
|
|
}
|
|
else
|
|
{
|
|
sz1 =
|
|
(uint32_t)(total_len1 %
|
|
(uint64_t)block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1))));
|
|
}
|
|
if (!(sz1 == 0U))
|
|
{
|
|
uint64_t prevlen = total_len1 - (uint64_t)sz1;
|
|
Hacl_Agile_Hash_state_s *s11 = block_state1.snd;
|
|
update_multi(s11,
|
|
prevlen,
|
|
buf,
|
|
block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1))));
|
|
}
|
|
uint32_t ite;
|
|
if
|
|
(
|
|
(uint64_t)(chunk_len - diff) %
|
|
(uint64_t)block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1)))
|
|
== 0ULL
|
|
&& (uint64_t)(chunk_len - diff) > 0ULL
|
|
)
|
|
{
|
|
ite = block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1)));
|
|
}
|
|
else
|
|
{
|
|
ite =
|
|
(uint32_t)((uint64_t)(chunk_len - diff) %
|
|
(uint64_t)block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1))));
|
|
}
|
|
uint32_t
|
|
n_blocks =
|
|
(chunk_len - diff - ite) / block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1)));
|
|
uint32_t
|
|
data1_len = n_blocks * block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1)));
|
|
uint32_t data2_len = chunk_len - diff - data1_len;
|
|
uint8_t *data1 = chunk2;
|
|
uint8_t *data2 = chunk2 + data1_len;
|
|
Hacl_Agile_Hash_state_s *s11 = block_state1.snd;
|
|
update_multi(s11, total_len1, data1, data1_len);
|
|
uint8_t *dst = buf;
|
|
memcpy(dst, data2, data2_len * sizeof (uint8_t));
|
|
*state =
|
|
(
|
|
(Hacl_Streaming_HMAC_agile_state){
|
|
.block_state = block_state1,
|
|
.buf = buf,
|
|
.total_len = total_len1 + (uint64_t)(chunk_len - diff)
|
|
}
|
|
);
|
|
}
|
|
return Hacl_Streaming_Types_Success;
|
|
}
|
|
|
|
typedef struct
|
|
___uint32_t____Hacl_Agile_Hash_state_s_____Hacl_Agile_Hash_state_s____uint32_t____Hacl_Agile_Hash_state_s_____Hacl_Agile_Hash_state_s___s
|
|
{
|
|
Hacl_Streaming_HMAC_Definitions_two_state fst;
|
|
Hacl_Streaming_HMAC_Definitions_two_state snd;
|
|
}
|
|
___uint32_t____Hacl_Agile_Hash_state_s_____Hacl_Agile_Hash_state_s____uint32_t____Hacl_Agile_Hash_state_s_____Hacl_Agile_Hash_state_s__;
|
|
|
|
Hacl_Streaming_Types_error_code
|
|
Hacl_Streaming_HMAC_digest(
|
|
Hacl_Streaming_HMAC_agile_state *state,
|
|
uint8_t *output,
|
|
uint32_t digest_length
|
|
)
|
|
{
|
|
KRML_MAYBE_UNUSED_VAR(digest_length);
|
|
Hacl_Streaming_HMAC_Definitions_two_state block_state0 = (*state).block_state;
|
|
Hacl_Streaming_HMAC_Definitions_index i1 = Hacl_Streaming_HMAC_index_of_state(block_state0);
|
|
Hacl_Streaming_HMAC_agile_state scrut0 = *state;
|
|
Hacl_Streaming_HMAC_Definitions_two_state block_state = scrut0.block_state;
|
|
uint8_t *buf_ = scrut0.buf;
|
|
uint64_t total_len = scrut0.total_len;
|
|
uint32_t r;
|
|
if
|
|
(
|
|
total_len % (uint64_t)block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1))) == 0ULL
|
|
&& total_len > 0ULL
|
|
)
|
|
{
|
|
r = block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1)));
|
|
}
|
|
else
|
|
{
|
|
r =
|
|
(uint32_t)(total_len %
|
|
(uint64_t)block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1))));
|
|
}
|
|
uint8_t *buf_1 = buf_;
|
|
Hacl_Agile_Hash_state_s *s110 = malloc_(dfst__Hacl_Agile_Hash_impl_uint32_t(i1));
|
|
option___uint32_t____Hacl_Agile_Hash_state_s_____Hacl_Agile_Hash_state_s__ tmp_block_state;
|
|
if (s110 == NULL)
|
|
{
|
|
tmp_block_state =
|
|
(
|
|
(option___uint32_t____Hacl_Agile_Hash_state_s_____Hacl_Agile_Hash_state_s__){
|
|
.tag = Hacl_Streaming_Types_None
|
|
}
|
|
);
|
|
}
|
|
else
|
|
{
|
|
Hacl_Agile_Hash_state_s *s21 = malloc_(dfst__Hacl_Agile_Hash_impl_uint32_t(i1));
|
|
if (s21 == NULL)
|
|
{
|
|
KRML_HOST_FREE(s110);
|
|
tmp_block_state =
|
|
(
|
|
(option___uint32_t____Hacl_Agile_Hash_state_s_____Hacl_Agile_Hash_state_s__){
|
|
.tag = Hacl_Streaming_Types_None
|
|
}
|
|
);
|
|
}
|
|
else
|
|
{
|
|
tmp_block_state =
|
|
(
|
|
(option___uint32_t____Hacl_Agile_Hash_state_s_____Hacl_Agile_Hash_state_s__){
|
|
.tag = Hacl_Streaming_Types_Some,
|
|
.v = { .fst = dsnd__Hacl_Agile_Hash_impl_uint32_t(i1), .snd = s110, .thd = s21 }
|
|
}
|
|
);
|
|
}
|
|
}
|
|
if (tmp_block_state.tag == Hacl_Streaming_Types_None)
|
|
{
|
|
return Hacl_Streaming_Types_OutOfMemory;
|
|
}
|
|
if (tmp_block_state.tag == Hacl_Streaming_Types_Some)
|
|
{
|
|
Hacl_Streaming_HMAC_Definitions_two_state tmp_block_state1 = tmp_block_state.v;
|
|
___uint32_t____Hacl_Agile_Hash_state_s_____Hacl_Agile_Hash_state_s____uint32_t____Hacl_Agile_Hash_state_s_____Hacl_Agile_Hash_state_s__
|
|
scrut = { .fst = block_state, .snd = tmp_block_state1 };
|
|
Hacl_Agile_Hash_state_s *s2_ = scrut.snd.thd;
|
|
Hacl_Agile_Hash_state_s *s1_ = scrut.snd.snd;
|
|
Hacl_Agile_Hash_state_s *s21 = scrut.fst.thd;
|
|
Hacl_Agile_Hash_state_s *s111 = scrut.fst.snd;
|
|
copy(s111, s1_);
|
|
copy(s21, s2_);
|
|
uint64_t prev_len = total_len - (uint64_t)r;
|
|
uint32_t ite;
|
|
if (r % block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1))) == 0U && r > 0U)
|
|
{
|
|
ite = block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1)));
|
|
}
|
|
else
|
|
{
|
|
ite = r % block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1)));
|
|
}
|
|
uint8_t *buf_last = buf_1 + r - ite;
|
|
uint8_t *buf_multi = buf_1;
|
|
Hacl_Agile_Hash_state_s *s112 = tmp_block_state1.snd;
|
|
update_multi(s112, prev_len, buf_multi, 0U);
|
|
uint64_t prev_len_last = total_len - (uint64_t)r;
|
|
Hacl_Agile_Hash_state_s *s11 = tmp_block_state1.snd;
|
|
update_last(s11, prev_len_last, buf_last, r);
|
|
finish0(tmp_block_state1, output);
|
|
return Hacl_Streaming_Types_Success;
|
|
}
|
|
KRML_HOST_EPRINTF("KaRaMeL abort at %s:%d\n%s\n",
|
|
__FILE__,
|
|
__LINE__,
|
|
"unreachable (pattern matches are exhaustive in F*)");
|
|
KRML_HOST_EXIT(255U);
|
|
}
|
|
|
|
void Hacl_Streaming_HMAC_free(Hacl_Streaming_HMAC_agile_state *state)
|
|
{
|
|
Hacl_Streaming_HMAC_Definitions_two_state block_state0 = (*state).block_state;
|
|
Hacl_Streaming_HMAC_Definitions_index i1 = Hacl_Streaming_HMAC_index_of_state(block_state0);
|
|
KRML_MAYBE_UNUSED_VAR(i1);
|
|
Hacl_Streaming_HMAC_agile_state scrut = *state;
|
|
uint8_t *buf = scrut.buf;
|
|
Hacl_Streaming_HMAC_Definitions_two_state block_state = scrut.block_state;
|
|
Hacl_Agile_Hash_state_s *s21 = block_state.thd;
|
|
Hacl_Agile_Hash_state_s *s11 = block_state.snd;
|
|
free_(s11);
|
|
free_(s21);
|
|
KRML_HOST_FREE(buf);
|
|
KRML_HOST_FREE(state);
|
|
}
|
|
|
|
Hacl_Streaming_HMAC_agile_state
|
|
*Hacl_Streaming_HMAC_copy(Hacl_Streaming_HMAC_agile_state *state)
|
|
{
|
|
Hacl_Streaming_HMAC_agile_state scrut0 = *state;
|
|
Hacl_Streaming_HMAC_Definitions_two_state block_state0 = scrut0.block_state;
|
|
uint8_t *buf0 = scrut0.buf;
|
|
uint64_t total_len0 = scrut0.total_len;
|
|
Hacl_Streaming_HMAC_Definitions_index i1 = Hacl_Streaming_HMAC_index_of_state(block_state0);
|
|
KRML_CHECK_SIZE(sizeof (uint8_t),
|
|
block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1))));
|
|
uint8_t
|
|
*buf =
|
|
(uint8_t *)KRML_HOST_CALLOC(block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1))),
|
|
sizeof (uint8_t));
|
|
if (buf == NULL)
|
|
{
|
|
return NULL;
|
|
}
|
|
memcpy(buf,
|
|
buf0,
|
|
block_len(alg_of_impl(dfst__Hacl_Agile_Hash_impl_uint32_t(i1))) * sizeof (uint8_t));
|
|
Hacl_Agile_Hash_state_s *s110 = malloc_(dfst__Hacl_Agile_Hash_impl_uint32_t(i1));
|
|
option___uint32_t____Hacl_Agile_Hash_state_s_____Hacl_Agile_Hash_state_s__ block_state;
|
|
if (s110 == NULL)
|
|
{
|
|
block_state =
|
|
(
|
|
(option___uint32_t____Hacl_Agile_Hash_state_s_____Hacl_Agile_Hash_state_s__){
|
|
.tag = Hacl_Streaming_Types_None
|
|
}
|
|
);
|
|
}
|
|
else
|
|
{
|
|
Hacl_Agile_Hash_state_s *s21 = malloc_(dfst__Hacl_Agile_Hash_impl_uint32_t(i1));
|
|
if (s21 == NULL)
|
|
{
|
|
KRML_HOST_FREE(s110);
|
|
block_state =
|
|
(
|
|
(option___uint32_t____Hacl_Agile_Hash_state_s_____Hacl_Agile_Hash_state_s__){
|
|
.tag = Hacl_Streaming_Types_None
|
|
}
|
|
);
|
|
}
|
|
else
|
|
{
|
|
block_state =
|
|
(
|
|
(option___uint32_t____Hacl_Agile_Hash_state_s_____Hacl_Agile_Hash_state_s__){
|
|
.tag = Hacl_Streaming_Types_Some,
|
|
.v = { .fst = dsnd__Hacl_Agile_Hash_impl_uint32_t(i1), .snd = s110, .thd = s21 }
|
|
}
|
|
);
|
|
}
|
|
}
|
|
if (block_state.tag == Hacl_Streaming_Types_None)
|
|
{
|
|
KRML_HOST_FREE(buf);
|
|
return NULL;
|
|
}
|
|
if (block_state.tag == Hacl_Streaming_Types_Some)
|
|
{
|
|
Hacl_Streaming_HMAC_Definitions_two_state block_state1 = block_state.v;
|
|
___uint32_t____Hacl_Agile_Hash_state_s_____Hacl_Agile_Hash_state_s____uint32_t____Hacl_Agile_Hash_state_s_____Hacl_Agile_Hash_state_s__
|
|
scrut = { .fst = block_state0, .snd = block_state1 };
|
|
Hacl_Agile_Hash_state_s *s2_ = scrut.snd.thd;
|
|
Hacl_Agile_Hash_state_s *s1_ = scrut.snd.snd;
|
|
Hacl_Agile_Hash_state_s *s21 = scrut.fst.thd;
|
|
Hacl_Agile_Hash_state_s *s111 = scrut.fst.snd;
|
|
copy(s111, s1_);
|
|
copy(s21, s2_);
|
|
Hacl_Streaming_Types_optional k_ = Hacl_Streaming_Types_Some;
|
|
switch (k_)
|
|
{
|
|
case Hacl_Streaming_Types_None:
|
|
{
|
|
return NULL;
|
|
}
|
|
case Hacl_Streaming_Types_Some:
|
|
{
|
|
Hacl_Streaming_HMAC_agile_state
|
|
s = { .block_state = block_state1, .buf = buf, .total_len = total_len0 };
|
|
Hacl_Streaming_HMAC_agile_state
|
|
*p =
|
|
(Hacl_Streaming_HMAC_agile_state *)KRML_HOST_MALLOC(sizeof (
|
|
Hacl_Streaming_HMAC_agile_state
|
|
));
|
|
if (p != NULL)
|
|
{
|
|
p[0U] = s;
|
|
}
|
|
if (p == NULL)
|
|
{
|
|
Hacl_Agile_Hash_state_s *s210 = block_state1.thd;
|
|
Hacl_Agile_Hash_state_s *s11 = block_state1.snd;
|
|
free_(s11);
|
|
free_(s210);
|
|
KRML_HOST_FREE(buf);
|
|
return NULL;
|
|
}
|
|
return p;
|
|
}
|
|
default:
|
|
{
|
|
KRML_HOST_EPRINTF("KaRaMeL incomplete match at %s:%d\n", __FILE__, __LINE__);
|
|
KRML_HOST_EXIT(253U);
|
|
}
|
|
}
|
|
}
|
|
KRML_HOST_EPRINTF("KaRaMeL abort at %s:%d\n%s\n",
|
|
__FILE__,
|
|
__LINE__,
|
|
"unreachable (pattern matches are exhaustive in F*)");
|
|
KRML_HOST_EXIT(255U);
|
|
}
|
|
|