
Type: improvement Change-Id: I707399b8ba617a659476bfd7d793f04a1283e694 Signed-off-by: Dave Wallace <dwallacelf@gmail.com>
562 lines
16 KiB
C
562 lines
16 KiB
C
/*
|
|
* Copyright (c) 2021 Cisco and/or its affiliates.
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at:
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
#include <quic/quic.h>
|
|
#include <quic/quic_crypto.h>
|
|
#include <vnet/session/session.h>
|
|
|
|
#include <quicly.h>
|
|
#include <picotls/openssl.h>
|
|
#include <pthread.h>
|
|
|
|
#define QUICLY_EPOCH_1RTT 3
|
|
|
|
extern quic_main_t quic_main;
|
|
extern quic_ctx_t *quic_get_conn_ctx (quicly_conn_t *conn);
|
|
vnet_crypto_main_t *cm = &crypto_main;
|
|
|
|
typedef struct crypto_key_
|
|
{
|
|
vnet_crypto_alg_t algo;
|
|
u8 key[32];
|
|
u16 key_len;
|
|
} crypto_key_t;
|
|
|
|
struct cipher_context_t
|
|
{
|
|
ptls_cipher_context_t super;
|
|
vnet_crypto_op_t op;
|
|
vnet_crypto_op_id_t id;
|
|
crypto_key_t key;
|
|
};
|
|
|
|
struct aead_crypto_context_t
|
|
{
|
|
ptls_aead_context_t super;
|
|
EVP_CIPHER_CTX *evp_ctx;
|
|
uint8_t static_iv[PTLS_MAX_IV_SIZE];
|
|
vnet_crypto_op_t op;
|
|
crypto_key_t key;
|
|
|
|
vnet_crypto_op_id_t id;
|
|
uint8_t iv[PTLS_MAX_IV_SIZE];
|
|
};
|
|
|
|
static int
|
|
quic_crypto_setup_cipher (quicly_crypto_engine_t *engine, quicly_conn_t *conn,
|
|
size_t epoch, int is_enc,
|
|
ptls_cipher_context_t **header_protect_ctx,
|
|
ptls_aead_context_t **packet_protect_ctx,
|
|
ptls_aead_algorithm_t *aead,
|
|
ptls_hash_algorithm_t *hash, const void *secret)
|
|
{
|
|
uint8_t hpkey[PTLS_MAX_SECRET_SIZE];
|
|
int ret;
|
|
|
|
*packet_protect_ctx = NULL;
|
|
/* generate new header protection key */
|
|
if (header_protect_ctx != NULL)
|
|
{
|
|
*header_protect_ctx = NULL;
|
|
ret =
|
|
ptls_hkdf_expand_label (hash, hpkey, aead->ctr_cipher->key_size,
|
|
ptls_iovec_init (secret, hash->digest_size),
|
|
"quic hp", ptls_iovec_init (NULL, 0), NULL);
|
|
if (ret)
|
|
goto Exit;
|
|
*header_protect_ctx = ptls_cipher_new (aead->ctr_cipher, is_enc, hpkey);
|
|
if (NULL == *header_protect_ctx)
|
|
{
|
|
ret = PTLS_ERROR_NO_MEMORY;
|
|
goto Exit;
|
|
}
|
|
}
|
|
|
|
/* generate new AEAD context */
|
|
*packet_protect_ctx =
|
|
ptls_aead_new (aead, hash, is_enc, secret, QUICLY_AEAD_BASE_LABEL);
|
|
if (NULL == *packet_protect_ctx)
|
|
{
|
|
ret = PTLS_ERROR_NO_MEMORY;
|
|
goto Exit;
|
|
}
|
|
|
|
if (epoch == QUICLY_EPOCH_1RTT && !is_enc)
|
|
{
|
|
quic_ctx_t *qctx = quic_get_conn_ctx (conn);
|
|
if (qctx->ingress_keys.aead_ctx != NULL)
|
|
qctx->key_phase_ingress++;
|
|
|
|
qctx->ingress_keys.aead_ctx = *packet_protect_ctx;
|
|
if (header_protect_ctx != NULL)
|
|
qctx->ingress_keys.hp_ctx = *header_protect_ctx;
|
|
}
|
|
|
|
ret = 0;
|
|
|
|
Exit:
|
|
if (ret)
|
|
{
|
|
if (*packet_protect_ctx != NULL)
|
|
{
|
|
ptls_aead_free (*packet_protect_ctx);
|
|
*packet_protect_ctx = NULL;
|
|
}
|
|
if (header_protect_ctx && *header_protect_ctx != NULL)
|
|
{
|
|
ptls_cipher_free (*header_protect_ctx);
|
|
*header_protect_ctx = NULL;
|
|
}
|
|
}
|
|
ptls_clear_memory (hpkey, sizeof (hpkey));
|
|
return ret;
|
|
}
|
|
|
|
static u32
|
|
quic_crypto_set_key (crypto_key_t *key)
|
|
{
|
|
u8 thread_index = vlib_get_thread_index ();
|
|
u32 key_id = quic_main.per_thread_crypto_key_indices[thread_index];
|
|
vnet_crypto_key_t *vnet_key = vnet_crypto_get_key (key_id);
|
|
vlib_main_t *vm = vlib_get_main ();
|
|
vnet_crypto_engine_t *engine;
|
|
|
|
vec_foreach (engine, cm->engines)
|
|
if (engine->key_op_handler)
|
|
engine->key_op_handler (vm, VNET_CRYPTO_KEY_OP_DEL, key_id);
|
|
|
|
vnet_key->alg = key->algo;
|
|
clib_memcpy (vnet_key->data, key->key, key->key_len);
|
|
|
|
vec_foreach (engine, cm->engines)
|
|
if (engine->key_op_handler)
|
|
engine->key_op_handler (vm, VNET_CRYPTO_KEY_OP_ADD, key_id);
|
|
|
|
return key_id;
|
|
}
|
|
|
|
static size_t
|
|
quic_crypto_aead_decrypt (quic_ctx_t *qctx, ptls_aead_context_t *_ctx,
|
|
void *_output, const void *input, size_t inlen,
|
|
uint64_t decrypted_pn, const void *aad,
|
|
size_t aadlen)
|
|
{
|
|
vlib_main_t *vm = vlib_get_main ();
|
|
|
|
struct aead_crypto_context_t *ctx = (struct aead_crypto_context_t *) _ctx;
|
|
|
|
vnet_crypto_op_init (&ctx->op, ctx->id);
|
|
ctx->op.aad = (u8 *) aad;
|
|
ctx->op.aad_len = aadlen;
|
|
ctx->op.iv = ctx->iv;
|
|
ptls_aead__build_iv (ctx->super.algo, ctx->op.iv, ctx->static_iv,
|
|
decrypted_pn);
|
|
ctx->op.src = (u8 *) input;
|
|
ctx->op.dst = _output;
|
|
ctx->op.key_index = quic_crypto_set_key (&ctx->key);
|
|
ctx->op.len = inlen - ctx->super.algo->tag_size;
|
|
ctx->op.tag_len = ctx->super.algo->tag_size;
|
|
ctx->op.tag = ctx->op.src + ctx->op.len;
|
|
|
|
vnet_crypto_process_ops (vm, &(ctx->op), 1);
|
|
|
|
return ctx->op.len;
|
|
}
|
|
|
|
void
|
|
quic_crypto_decrypt_packet (quic_ctx_t *qctx, quic_rx_packet_ctx_t *pctx)
|
|
{
|
|
ptls_cipher_context_t *header_protection = NULL;
|
|
ptls_aead_context_t *aead = NULL;
|
|
int pn;
|
|
|
|
/* Long Header packets are not decrypted by vpp */
|
|
if (QUICLY_PACKET_IS_LONG_HEADER (pctx->packet.octets.base[0]))
|
|
return;
|
|
|
|
uint64_t next_expected_packet_number =
|
|
quicly_get_next_expected_packet_number (qctx->conn);
|
|
if (next_expected_packet_number == UINT64_MAX)
|
|
return;
|
|
|
|
aead = qctx->ingress_keys.aead_ctx;
|
|
header_protection = qctx->ingress_keys.hp_ctx;
|
|
|
|
if (!aead || !header_protection)
|
|
return;
|
|
|
|
size_t encrypted_len = pctx->packet.octets.len - pctx->packet.encrypted_off;
|
|
uint8_t hpmask[5] = { 0 };
|
|
uint32_t pnbits = 0;
|
|
size_t pnlen, ptlen, i;
|
|
|
|
/* decipher the header protection, as well as obtaining pnbits, pnlen */
|
|
if (encrypted_len < header_protection->algo->iv_size + QUICLY_MAX_PN_SIZE)
|
|
return;
|
|
ptls_cipher_init (header_protection, pctx->packet.octets.base +
|
|
pctx->packet.encrypted_off +
|
|
QUICLY_MAX_PN_SIZE);
|
|
ptls_cipher_encrypt (header_protection, hpmask, hpmask, sizeof (hpmask));
|
|
pctx->packet.octets.base[0] ^=
|
|
hpmask[0] &
|
|
(QUICLY_PACKET_IS_LONG_HEADER (pctx->packet.octets.base[0]) ? 0xf : 0x1f);
|
|
pnlen = (pctx->packet.octets.base[0] & 0x3) + 1;
|
|
for (i = 0; i != pnlen; ++i)
|
|
{
|
|
pctx->packet.octets.base[pctx->packet.encrypted_off + i] ^=
|
|
hpmask[i + 1];
|
|
pnbits = (pnbits << 8) |
|
|
pctx->packet.octets.base[pctx->packet.encrypted_off + i];
|
|
}
|
|
|
|
size_t aead_off = pctx->packet.encrypted_off + pnlen;
|
|
|
|
pn = quicly_determine_packet_number (pnbits, pnlen * 8,
|
|
next_expected_packet_number);
|
|
|
|
int key_phase_bit =
|
|
(pctx->packet.octets.base[0] & QUICLY_KEY_PHASE_BIT) != 0;
|
|
|
|
if (key_phase_bit != (qctx->key_phase_ingress & 1))
|
|
{
|
|
pctx->packet.octets.base[0] ^=
|
|
hpmask[0] &
|
|
(QUICLY_PACKET_IS_LONG_HEADER (pctx->packet.octets.base[0]) ? 0xf :
|
|
0x1f);
|
|
for (i = 0; i != pnlen; ++i)
|
|
{
|
|
pctx->packet.octets.base[pctx->packet.encrypted_off + i] ^=
|
|
hpmask[i + 1];
|
|
}
|
|
return;
|
|
}
|
|
|
|
if ((ptlen = quic_crypto_aead_decrypt (
|
|
qctx, aead, pctx->packet.octets.base + aead_off,
|
|
pctx->packet.octets.base + aead_off,
|
|
pctx->packet.octets.len - aead_off, pn, pctx->packet.octets.base,
|
|
aead_off)) == SIZE_MAX)
|
|
{
|
|
fprintf (stderr, "%s: aead decryption failure (pn: %d)\n", __FUNCTION__,
|
|
pn);
|
|
return;
|
|
}
|
|
|
|
pctx->packet.encrypted_off = aead_off;
|
|
pctx->packet.octets.len = ptlen + aead_off;
|
|
|
|
pctx->packet.decrypted.pn = pn;
|
|
pctx->packet.decrypted.key_phase = qctx->key_phase_ingress;
|
|
}
|
|
|
|
void
|
|
quic_crypto_encrypt_packet (struct st_quicly_crypto_engine_t *engine,
|
|
quicly_conn_t *conn,
|
|
ptls_cipher_context_t *header_protect_ctx,
|
|
ptls_aead_context_t *packet_protect_ctx,
|
|
ptls_iovec_t datagram, size_t first_byte_at,
|
|
size_t payload_from, uint64_t packet_number,
|
|
int coalesced)
|
|
{
|
|
vlib_main_t *vm = vlib_get_main ();
|
|
|
|
struct cipher_context_t *hp_ctx =
|
|
(struct cipher_context_t *) header_protect_ctx;
|
|
struct aead_crypto_context_t *aead_ctx =
|
|
(struct aead_crypto_context_t *) packet_protect_ctx;
|
|
|
|
void *input = datagram.base + payload_from;
|
|
void *output = input;
|
|
size_t inlen =
|
|
datagram.len - payload_from - packet_protect_ctx->algo->tag_size;
|
|
const void *aad = datagram.base + first_byte_at;
|
|
size_t aadlen = payload_from - first_byte_at;
|
|
|
|
/* Build AEAD encrypt crypto operation */
|
|
vnet_crypto_op_init (&aead_ctx->op, aead_ctx->id);
|
|
aead_ctx->op.aad = (u8 *) aad;
|
|
aead_ctx->op.aad_len = aadlen;
|
|
aead_ctx->op.iv = aead_ctx->iv;
|
|
ptls_aead__build_iv (aead_ctx->super.algo, aead_ctx->op.iv,
|
|
aead_ctx->static_iv, packet_number);
|
|
aead_ctx->op.key_index = quic_crypto_set_key (&aead_ctx->key);
|
|
aead_ctx->op.src = (u8 *) input;
|
|
aead_ctx->op.dst = output;
|
|
aead_ctx->op.len = inlen;
|
|
aead_ctx->op.tag_len = aead_ctx->super.algo->tag_size;
|
|
aead_ctx->op.tag = aead_ctx->op.src + inlen;
|
|
vnet_crypto_process_ops (vm, &(aead_ctx->op), 1);
|
|
assert (aead_ctx->op.status == VNET_CRYPTO_OP_STATUS_COMPLETED);
|
|
|
|
/* Build Header protection crypto operation */
|
|
ptls_aead_supplementary_encryption_t supp = {
|
|
.ctx = header_protect_ctx,
|
|
.input =
|
|
datagram.base + payload_from - QUICLY_SEND_PN_SIZE + QUICLY_MAX_PN_SIZE
|
|
};
|
|
|
|
/* Build Header protection crypto operation */
|
|
vnet_crypto_op_init (&hp_ctx->op, hp_ctx->id);
|
|
memset (supp.output, 0, sizeof (supp.output));
|
|
hp_ctx->op.iv = (u8 *) supp.input;
|
|
hp_ctx->op.key_index = quic_crypto_set_key (&hp_ctx->key);
|
|
;
|
|
hp_ctx->op.src = (u8 *) supp.output;
|
|
hp_ctx->op.dst = (u8 *) supp.output;
|
|
hp_ctx->op.len = sizeof (supp.output);
|
|
vnet_crypto_process_ops (vm, &(hp_ctx->op), 1);
|
|
assert (hp_ctx->op.status == VNET_CRYPTO_OP_STATUS_COMPLETED);
|
|
|
|
datagram.base[first_byte_at] ^=
|
|
supp.output[0] &
|
|
(QUICLY_PACKET_IS_LONG_HEADER (datagram.base[first_byte_at]) ? 0xf : 0x1f);
|
|
for (size_t i = 0; i != QUICLY_SEND_PN_SIZE; ++i)
|
|
datagram.base[payload_from + i - QUICLY_SEND_PN_SIZE] ^=
|
|
supp.output[i + 1];
|
|
}
|
|
|
|
static int
|
|
quic_crypto_cipher_setup_crypto (ptls_cipher_context_t *_ctx, int is_enc,
|
|
const void *key, const EVP_CIPHER *cipher)
|
|
{
|
|
struct cipher_context_t *ctx = (struct cipher_context_t *) _ctx;
|
|
|
|
vnet_crypto_alg_t algo;
|
|
if (!strcmp (ctx->super.algo->name, "AES128-CTR"))
|
|
{
|
|
algo = VNET_CRYPTO_ALG_AES_128_CTR;
|
|
ctx->id = is_enc ? VNET_CRYPTO_OP_AES_128_CTR_ENC :
|
|
VNET_CRYPTO_OP_AES_128_CTR_DEC;
|
|
ptls_openssl_aes128ctr.setup_crypto (_ctx, is_enc, key);
|
|
}
|
|
else if (!strcmp (ctx->super.algo->name, "AES256-CTR"))
|
|
{
|
|
algo = VNET_CRYPTO_ALG_AES_256_CTR;
|
|
ctx->id = is_enc ? VNET_CRYPTO_OP_AES_256_CTR_ENC :
|
|
VNET_CRYPTO_OP_AES_256_CTR_DEC;
|
|
ptls_openssl_aes256ctr.setup_crypto (_ctx, is_enc, key);
|
|
}
|
|
else
|
|
{
|
|
QUIC_DBG (1, "%s, Invalid crypto cipher : ", __FUNCTION__,
|
|
_ctx->algo->name);
|
|
assert (0);
|
|
}
|
|
|
|
if (quic_main.vnet_crypto_enabled)
|
|
{
|
|
// ctx->key_index =
|
|
// quic_crypto_go_setup_key (algo, key, _ctx->algo->key_size);
|
|
ctx->key.algo = algo;
|
|
ctx->key.key_len = _ctx->algo->key_size;
|
|
assert (ctx->key.key_len <= 32);
|
|
clib_memcpy (&ctx->key.key, key, ctx->key.key_len);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
quic_crypto_aes128ctr_setup_crypto (ptls_cipher_context_t *ctx, int is_enc,
|
|
const void *key)
|
|
{
|
|
return quic_crypto_cipher_setup_crypto (ctx, 1, key, EVP_aes_128_ctr ());
|
|
}
|
|
|
|
static int
|
|
quic_crypto_aes256ctr_setup_crypto (ptls_cipher_context_t *ctx, int is_enc,
|
|
const void *key)
|
|
{
|
|
return quic_crypto_cipher_setup_crypto (ctx, 1, key, EVP_aes_256_ctr ());
|
|
}
|
|
|
|
static int
|
|
quic_crypto_aead_setup_crypto (ptls_aead_context_t *_ctx, int is_enc,
|
|
const void *key, const void *iv,
|
|
const EVP_CIPHER *cipher)
|
|
{
|
|
struct aead_crypto_context_t *ctx = (struct aead_crypto_context_t *) _ctx;
|
|
|
|
vnet_crypto_alg_t algo;
|
|
if (!strcmp (ctx->super.algo->name, "AES128-GCM"))
|
|
{
|
|
algo = VNET_CRYPTO_ALG_AES_128_GCM;
|
|
ctx->id = is_enc ? VNET_CRYPTO_OP_AES_128_GCM_ENC :
|
|
VNET_CRYPTO_OP_AES_128_GCM_DEC;
|
|
ptls_openssl_aes128gcm.setup_crypto (_ctx, is_enc, key, iv);
|
|
}
|
|
else if (!strcmp (ctx->super.algo->name, "AES256-GCM"))
|
|
{
|
|
algo = VNET_CRYPTO_ALG_AES_256_GCM;
|
|
ctx->id = is_enc ? VNET_CRYPTO_OP_AES_256_GCM_ENC :
|
|
VNET_CRYPTO_OP_AES_256_GCM_DEC;
|
|
ptls_openssl_aes256gcm.setup_crypto (_ctx, is_enc, key, iv);
|
|
}
|
|
else
|
|
{
|
|
QUIC_DBG (1, "%s, invalied aead cipher %s", __FUNCTION__,
|
|
_ctx->algo->name);
|
|
assert (0);
|
|
}
|
|
|
|
if (quic_main.vnet_crypto_enabled)
|
|
{
|
|
clib_memcpy (ctx->static_iv, iv, ctx->super.algo->iv_size);
|
|
// ctx->key_index =
|
|
// quic_crypto_go_setup_key (algo, key, _ctx->algo->key_size);
|
|
ctx->key.algo = algo;
|
|
ctx->key.key_len = _ctx->algo->key_size;
|
|
assert (ctx->key.key_len <= 32);
|
|
clib_memcpy (&ctx->key.key, key, ctx->key.key_len);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
quic_crypto_aead_aes128gcm_setup_crypto (ptls_aead_context_t *ctx, int is_enc,
|
|
const void *key, const void *iv)
|
|
{
|
|
return quic_crypto_aead_setup_crypto (ctx, is_enc, key, iv,
|
|
EVP_aes_128_gcm ());
|
|
}
|
|
|
|
static int
|
|
quic_crypto_aead_aes256gcm_setup_crypto (ptls_aead_context_t *ctx, int is_enc,
|
|
const void *key, const void *iv)
|
|
{
|
|
return quic_crypto_aead_setup_crypto (ctx, is_enc, key, iv,
|
|
EVP_aes_256_gcm ());
|
|
}
|
|
|
|
int
|
|
quic_encrypt_ticket_cb (ptls_encrypt_ticket_t *_self, ptls_t *tls,
|
|
int is_encrypt, ptls_buffer_t *dst, ptls_iovec_t src)
|
|
{
|
|
quic_session_cache_t *self = (void *) _self;
|
|
int ret;
|
|
|
|
if (is_encrypt)
|
|
{
|
|
|
|
/* replace the cached entry along with a newly generated session id */
|
|
clib_mem_free (self->data.base);
|
|
if ((self->data.base = clib_mem_alloc (src.len)) == NULL)
|
|
return PTLS_ERROR_NO_MEMORY;
|
|
|
|
ptls_get_context (tls)->random_bytes (self->id, sizeof (self->id));
|
|
clib_memcpy (self->data.base, src.base, src.len);
|
|
self->data.len = src.len;
|
|
|
|
/* store the session id in buffer */
|
|
if ((ret = ptls_buffer_reserve (dst, sizeof (self->id))) != 0)
|
|
return ret;
|
|
clib_memcpy (dst->base + dst->off, self->id, sizeof (self->id));
|
|
dst->off += sizeof (self->id);
|
|
}
|
|
else
|
|
{
|
|
/* check if session id is the one stored in cache */
|
|
if (src.len != sizeof (self->id))
|
|
return PTLS_ERROR_SESSION_NOT_FOUND;
|
|
if (clib_memcmp (self->id, src.base, sizeof (self->id)) != 0)
|
|
return PTLS_ERROR_SESSION_NOT_FOUND;
|
|
|
|
/* return the cached value */
|
|
if ((ret = ptls_buffer_reserve (dst, self->data.len)) != 0)
|
|
return ret;
|
|
clib_memcpy (dst->base + dst->off, self->data.base, self->data.len);
|
|
dst->off += self->data.len;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
ptls_cipher_algorithm_t quic_crypto_aes128ctr = {
|
|
"AES128-CTR",
|
|
PTLS_AES128_KEY_SIZE,
|
|
1,
|
|
PTLS_AES_IV_SIZE,
|
|
sizeof (struct cipher_context_t),
|
|
quic_crypto_aes128ctr_setup_crypto
|
|
};
|
|
|
|
ptls_cipher_algorithm_t quic_crypto_aes256ctr = {
|
|
"AES256-CTR",
|
|
PTLS_AES256_KEY_SIZE,
|
|
1 /* block size */,
|
|
PTLS_AES_IV_SIZE,
|
|
sizeof (struct cipher_context_t),
|
|
quic_crypto_aes256ctr_setup_crypto
|
|
};
|
|
|
|
#define PTLS_X86_CACHE_LINE_ALIGN_BITS 6
|
|
ptls_aead_algorithm_t quic_crypto_aes128gcm = {
|
|
"AES128-GCM",
|
|
PTLS_AESGCM_CONFIDENTIALITY_LIMIT,
|
|
PTLS_AESGCM_INTEGRITY_LIMIT,
|
|
&quic_crypto_aes128ctr,
|
|
&ptls_openssl_aes128ecb,
|
|
PTLS_AES128_KEY_SIZE,
|
|
PTLS_AESGCM_IV_SIZE,
|
|
PTLS_AESGCM_TAG_SIZE,
|
|
{ PTLS_TLS12_AESGCM_FIXED_IV_SIZE, PTLS_TLS12_AESGCM_RECORD_IV_SIZE },
|
|
1,
|
|
PTLS_X86_CACHE_LINE_ALIGN_BITS,
|
|
sizeof (struct aead_crypto_context_t),
|
|
quic_crypto_aead_aes128gcm_setup_crypto
|
|
};
|
|
|
|
ptls_aead_algorithm_t quic_crypto_aes256gcm = {
|
|
"AES256-GCM",
|
|
PTLS_AESGCM_CONFIDENTIALITY_LIMIT,
|
|
PTLS_AESGCM_INTEGRITY_LIMIT,
|
|
&quic_crypto_aes256ctr,
|
|
&ptls_openssl_aes256ecb,
|
|
PTLS_AES256_KEY_SIZE,
|
|
PTLS_AESGCM_IV_SIZE,
|
|
PTLS_AESGCM_TAG_SIZE,
|
|
{ PTLS_TLS12_AESGCM_FIXED_IV_SIZE, PTLS_TLS12_AESGCM_RECORD_IV_SIZE },
|
|
1,
|
|
PTLS_X86_CACHE_LINE_ALIGN_BITS,
|
|
sizeof (struct aead_crypto_context_t),
|
|
quic_crypto_aead_aes256gcm_setup_crypto
|
|
};
|
|
|
|
ptls_cipher_suite_t quic_crypto_aes128gcmsha256 = {
|
|
PTLS_CIPHER_SUITE_AES_128_GCM_SHA256, &quic_crypto_aes128gcm,
|
|
&ptls_openssl_sha256
|
|
};
|
|
|
|
ptls_cipher_suite_t quic_crypto_aes256gcmsha384 = {
|
|
PTLS_CIPHER_SUITE_AES_256_GCM_SHA384, &quic_crypto_aes256gcm,
|
|
&ptls_openssl_sha384
|
|
};
|
|
|
|
ptls_cipher_suite_t *quic_crypto_cipher_suites[] = {
|
|
&quic_crypto_aes256gcmsha384, &quic_crypto_aes128gcmsha256, NULL
|
|
};
|
|
|
|
quicly_crypto_engine_t quic_crypto_engine = { quic_crypto_setup_cipher,
|
|
quic_crypto_encrypt_packet };
|
|
|
|
/*
|
|
* fd.io coding-style-patch-verification: ON
|
|
*
|
|
* Local Variables:
|
|
* eval: (c-set-style "gnu")
|
|
* End:
|
|
*/
|