/* MIT License * * Copyright (c) 2016-2020 INRIA, CMU and Microsoft Corporation * * Permission is hereby granted, free of charge, to any person obtaining a copy * of this software and associated documentation files (the "Software"), to deal * in the Software without restriction, including without limitation the rights * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell * copies of the Software, and to permit persons to whom the Software is * furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice shall be included in all * copies or substantial portions of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE * SOFTWARE. */ #include "Hacl_Streaming_Poly1305_256.h" Hacl_Streaming_Poly1305_256_poly1305_256_state *Hacl_Streaming_Poly1305_256_create_in(uint8_t *k) { uint8_t *buf = (uint8_t *)KRML_HOST_CALLOC((uint32_t)64U, sizeof (uint8_t)); Lib_IntVector_Intrinsics_vec256 *r1 = (Lib_IntVector_Intrinsics_vec256 *)KRML_HOST_MALLOC(sizeof (Lib_IntVector_Intrinsics_vec256) * (uint32_t)25U); { uint32_t _i; for (_i = 0U; _i < (uint32_t)25U; ++_i) r1[_i] = Lib_IntVector_Intrinsics_vec256_zero; } { Lib_IntVector_Intrinsics_vec256 *block_state = r1; uint8_t *k_ = (uint8_t *)KRML_HOST_CALLOC((uint32_t)32U, sizeof (uint8_t)); uint8_t *k_0; memcpy(k_, k, (uint32_t)32U * sizeof (uint8_t)); k_0 = k_; { Hacl_Streaming_Poly1305_256_poly1305_256_state s; s.block_state = block_state; s.buf = buf; s.total_len = (uint64_t)0U; s.p_key = k_0; KRML_CHECK_SIZE(sizeof (Hacl_Streaming_Poly1305_256_poly1305_256_state), (uint32_t)1U); { Hacl_Streaming_Poly1305_256_poly1305_256_state *p = (Hacl_Streaming_Poly1305_256_poly1305_256_state *)KRML_HOST_MALLOC(sizeof ( Hacl_Streaming_Poly1305_256_poly1305_256_state )); p[0U] = s; Hacl_Poly1305_256_poly1305_init(block_state, k); return p; } } } } void Hacl_Streaming_Poly1305_256_init(uint8_t *k, Hacl_Streaming_Poly1305_256_poly1305_256_state *s) { Hacl_Streaming_Poly1305_256_poly1305_256_state scrut = *s; uint8_t *k_ = scrut.p_key; uint8_t *buf = scrut.buf; Lib_IntVector_Intrinsics_vec256 *block_state = scrut.block_state; uint8_t *k_1; Hacl_Poly1305_256_poly1305_init(block_state, k); memcpy(k_, k, (uint32_t)32U * sizeof (uint8_t)); k_1 = k_; { Hacl_Streaming_Poly1305_256_poly1305_256_state lit; lit.block_state = block_state; lit.buf = buf; lit.total_len = (uint64_t)0U; lit.p_key = k_1; s[0U] = lit; } } void Hacl_Streaming_Poly1305_256_update( Hacl_Streaming_Poly1305_256_poly1305_256_state *p, uint8_t *data, uint32_t len ) { Hacl_Streaming_Poly1305_256_poly1305_256_state s = *p; uint64_t total_len = s.total_len; uint32_t sz; if (total_len % (uint64_t)(uint32_t)64U == (uint64_t)0U && total_len > (uint64_t)0U) { sz = (uint32_t)64U; } else { sz = (uint32_t)(total_len % (uint64_t)(uint32_t)64U); } if (len <= (uint32_t)64U - sz) { Hacl_Streaming_Poly1305_256_poly1305_256_state s1 = *p; Lib_IntVector_Intrinsics_vec256 *block_state1 = s1.block_state; uint8_t *buf = s1.buf; uint64_t total_len1 = s1.total_len; uint8_t *k_1 = s1.p_key; uint32_t sz1; if (total_len1 % (uint64_t)(uint32_t)64U == (uint64_t)0U && total_len1 > (uint64_t)0U) { sz1 = (uint32_t)64U; } else { sz1 = (uint32_t)(total_len1 % (uint64_t)(uint32_t)64U); } { uint8_t *buf2 = buf + sz1; uint64_t total_len2; memcpy(buf2, data, len * sizeof (uint8_t)); total_len2 = total_len1 + (uint64_t)len; { Hacl_Streaming_Poly1305_256_poly1305_256_state lit; lit.block_state = block_state1; lit.buf = buf; lit.total_len = total_len2; lit.p_key = k_1; *p = lit; return; } } } if (sz == (uint32_t)0U) { Hacl_Streaming_Poly1305_256_poly1305_256_state s1 = *p; Lib_IntVector_Intrinsics_vec256 *block_state1 = s1.block_state; uint8_t *buf = s1.buf; uint64_t total_len1 = s1.total_len; uint8_t *k_1 = s1.p_key; uint32_t sz1; if (total_len1 % (uint64_t)(uint32_t)64U == (uint64_t)0U && total_len1 > (uint64_t)0U) { sz1 = (uint32_t)64U; } else { sz1 = (uint32_t)(total_len1 % (uint64_t)(uint32_t)64U); } { uint32_t ite; uint32_t n_blocks; uint32_t data1_len; uint32_t data2_len; uint8_t *data1; uint8_t *data2; uint8_t *dst; if (!(sz1 == (uint32_t)0U)) { Hacl_Poly1305_256_poly1305_update(block_state1, (uint32_t)64U, buf); } if ((uint64_t)len % (uint64_t)(uint32_t)64U == (uint64_t)0U && (uint64_t)len > (uint64_t)0U) { ite = (uint32_t)64U; } else { ite = (uint32_t)((uint64_t)len % (uint64_t)(uint32_t)64U); } n_blocks = (len - ite) / (uint32_t)64U; data1_len = n_blocks * (uint32_t)64U; data2_len = len - data1_len; data1 = data; data2 = data + data1_len; Hacl_Poly1305_256_poly1305_update(block_state1, data1_len, data1); dst = buf; memcpy(dst, data2, data2_len * sizeof (uint8_t)); { Hacl_Streaming_Poly1305_256_poly1305_256_state lit; lit.block_state = block_state1; lit.buf = buf; lit.total_len = total_len1 + (uint64_t)len; lit.p_key = k_1; *p = lit; return; } } } { uint32_t diff = (uint32_t)64U - sz; uint8_t *data1 = data; uint8_t *data2 = data + diff; Hacl_Streaming_Poly1305_256_poly1305_256_state s10 = *p; Lib_IntVector_Intrinsics_vec256 *block_state10 = s10.block_state; uint8_t *buf0 = s10.buf; uint64_t total_len10 = s10.total_len; uint8_t *k_1 = s10.p_key; uint32_t sz10; if (total_len10 % (uint64_t)(uint32_t)64U == (uint64_t)0U && total_len10 > (uint64_t)0U) { sz10 = (uint32_t)64U; } else { sz10 = (uint32_t)(total_len10 % (uint64_t)(uint32_t)64U); } { uint8_t *buf2 = buf0 + sz10; uint64_t total_len2; memcpy(buf2, data1, diff * sizeof (uint8_t)); total_len2 = total_len10 + (uint64_t)diff; { Hacl_Streaming_Poly1305_256_poly1305_256_state lit; Hacl_Streaming_Poly1305_256_poly1305_256_state s1; Lib_IntVector_Intrinsics_vec256 *block_state1; uint8_t *buf; uint64_t total_len1; uint8_t *k_10; uint32_t sz1; uint32_t ite; uint32_t n_blocks; uint32_t data1_len; uint32_t data2_len; uint8_t *data11; uint8_t *data21; uint8_t *dst; lit.block_state = block_state10; lit.buf = buf0; lit.total_len = total_len2; lit.p_key = k_1; *p = lit; s1 = *p; block_state1 = s1.block_state; buf = s1.buf; total_len1 = s1.total_len; k_10 = s1.p_key; if (total_len1 % (uint64_t)(uint32_t)64U == (uint64_t)0U && total_len1 > (uint64_t)0U) { sz1 = (uint32_t)64U; } else { sz1 = (uint32_t)(total_len1 % (uint64_t)(uint32_t)64U); } if (!(sz1 == (uint32_t)0U)) { Hacl_Poly1305_256_poly1305_update(block_state1, (uint32_t)64U, buf); } if ( (uint64_t)(len - diff) % (uint64_t)(uint32_t)64U == (uint64_t)0U && (uint64_t)(len - diff) > (uint64_t)0U ) { ite = (uint32_t)64U; } else { ite = (uint32_t)((uint64_t)(len - diff) % (uint64_t)(uint32_t)64U); } n_blocks = (len - diff - ite) / (uint32_t)64U; data1_len = n_blocks * (uint32_t)64U; data2_len = len - diff - data1_len; data11 = data2; data21 = data2 + data1_len; Hacl_Poly1305_256_poly1305_update(block_state1, data1_len, data11); dst = buf; memcpy(dst, data21, data2_len * sizeof (uint8_t)); { Hacl_Streaming_Poly1305_256_poly1305_256_state lit0; lit0.block_state = block_state1; lit0.buf = buf; lit0.total_len = total_len1 + (uint64_t)(len - diff); lit0.p_key = k_10; *p = lit0; } } } } } void Hacl_Streaming_Poly1305_256_finish( Hacl_Streaming_Poly1305_256_poly1305_256_state *p, uint8_t *dst ) { Hacl_Streaming_Poly1305_256_poly1305_256_state scrut = *p; Lib_IntVector_Intrinsics_vec256 *block_state = scrut.block_state; uint8_t *buf_ = scrut.buf; uint64_t total_len = scrut.total_len; uint8_t *k_ = scrut.p_key; uint32_t r; if (total_len % (uint64_t)(uint32_t)64U == (uint64_t)0U && total_len > (uint64_t)0U) { r = (uint32_t)64U; } else { r = (uint32_t)(total_len % (uint64_t)(uint32_t)64U); } { uint8_t *buf_1 = buf_; Lib_IntVector_Intrinsics_vec256 r1[25U]; { uint32_t _i; for (_i = 0U; _i < (uint32_t)25U; ++_i) r1[_i] = Lib_IntVector_Intrinsics_vec256_zero; } { Lib_IntVector_Intrinsics_vec256 *tmp_block_state = r1; uint32_t ite0; uint8_t *buf_last; uint8_t *buf_multi; uint32_t ite1; uint32_t ite2; uint64_t prev_len_last; uint32_t ite; memcpy(tmp_block_state, block_state, (uint32_t)25U * sizeof (Lib_IntVector_Intrinsics_vec256)); if (r % (uint32_t)16U == (uint32_t)0U && r > (uint32_t)0U) { ite0 = (uint32_t)16U; } else { ite0 = r % (uint32_t)16U; } buf_last = buf_1 + r - ite0; buf_multi = buf_1; if (r % (uint32_t)16U == (uint32_t)0U && r > (uint32_t)0U) { ite1 = (uint32_t)16U; } else { ite1 = r % (uint32_t)16U; } Hacl_Poly1305_256_poly1305_update(tmp_block_state, r - ite1, buf_multi); if (r % (uint32_t)16U == (uint32_t)0U && r > (uint32_t)0U) { ite2 = (uint32_t)16U; } else { ite2 = r % (uint32_t)16U; } prev_len_last = total_len - (uint64_t)ite2; if (r % (uint32_t)16U == (uint32_t)0U && r > (uint32_t)0U) { ite = (uint32_t)16U; } else { ite = r % (uint32_t)16U; } Hacl_Poly1305_256_poly1305_update(tmp_block_state, ite, buf_last); { Lib_IntVector_Intrinsics_vec256 tmp[25U]; { uint32_t _i; for (_i = 0U; _i < (uint32_t)25U; ++_i) tmp[_i] = Lib_IntVector_Intrinsics_vec256_zero; } memcpy(tmp, tmp_block_state, (uint32_t)25U * sizeof (Lib_IntVector_Intrinsics_vec256)); Hacl_Poly1305_256_poly1305_finish(dst, k_, tmp); } } } } void Hacl_Streaming_Poly1305_256_free(Hacl_Streaming_Poly1305_256_poly1305_256_state *s) { Hacl_Streaming_Poly1305_256_poly1305_256_state scrut = *s; uint8_t *k_ = scrut.p_key; uint8_t *buf = scrut.buf; Lib_IntVector_Intrinsics_vec256 *block_state = scrut.block_state; KRML_HOST_FREE(k_); KRML_HOST_FREE(block_state); KRML_HOST_FREE(buf); KRML_HOST_FREE(s); }