mirror of
https://github.com/zama-ai/tfhe-rs.git
synced 2026-01-08 22:28:01 -05:00
chore(gpu): no crash with aes benches if oom error
This commit is contained in:
committed by
Agnès Leroy
parent
164fc26025
commit
54c8c5e020
@@ -3,7 +3,7 @@ pub mod cuda {
|
|||||||
use benchmark::params_aliases::BENCH_PARAM_GPU_MULTI_BIT_GROUP_4_MESSAGE_2_CARRY_2_KS_PBS_TUNIFORM_2M128;
|
use benchmark::params_aliases::BENCH_PARAM_GPU_MULTI_BIT_GROUP_4_MESSAGE_2_CARRY_2_KS_PBS_TUNIFORM_2M128;
|
||||||
use benchmark::utilities::{write_to_json, OperatorType};
|
use benchmark::utilities::{write_to_json, OperatorType};
|
||||||
use criterion::{black_box, Criterion};
|
use criterion::{black_box, Criterion};
|
||||||
use tfhe::core_crypto::gpu::CudaStreams;
|
use tfhe::core_crypto::gpu::{check_valid_cuda_malloc, CudaStreams};
|
||||||
use tfhe::integer::gpu::ciphertext::CudaUnsignedRadixCiphertext;
|
use tfhe::integer::gpu::ciphertext::CudaUnsignedRadixCiphertext;
|
||||||
use tfhe::integer::gpu::CudaServerKey;
|
use tfhe::integer::gpu::CudaServerKey;
|
||||||
use tfhe::integer::keycache::KEY_CACHE;
|
use tfhe::integer::keycache::KEY_CACHE;
|
||||||
@@ -102,37 +102,52 @@ pub mod cuda {
|
|||||||
let sks = CudaServerKey::new(&cpu_cks, &streams);
|
let sks = CudaServerKey::new(&cpu_cks, &streams);
|
||||||
let cks = RadixClientKey::from((cpu_cks, 1));
|
let cks = RadixClientKey::from((cpu_cks, 1));
|
||||||
|
|
||||||
let ct_key = cks.encrypt_u128_for_aes_ctr(key);
|
//
|
||||||
let ct_iv = cks.encrypt_u128_for_aes_ctr(iv);
|
// Memory checks
|
||||||
|
//
|
||||||
|
let gpu_index = streams.gpu_indexes[0];
|
||||||
|
|
||||||
let d_key = CudaUnsignedRadixCiphertext::from_radix_ciphertext(&ct_key, &streams);
|
let key_expansion_size = sks.get_key_expansion_size_on_gpu(&streams);
|
||||||
let d_iv = CudaUnsignedRadixCiphertext::from_radix_ciphertext(&ct_iv, &streams);
|
let aes_encrypt_size =
|
||||||
|
sks.get_aes_encrypt_size_on_gpu(NUM_AES_INPUTS, SBOX_PARALLELISM, &streams);
|
||||||
|
|
||||||
let round_keys = sks.key_expansion(&d_key, &streams);
|
if check_valid_cuda_malloc(key_expansion_size, gpu_index)
|
||||||
|
&& check_valid_cuda_malloc(aes_encrypt_size, gpu_index)
|
||||||
|
{
|
||||||
|
let ct_key = cks.encrypt_u128_for_aes_ctr(key);
|
||||||
|
let ct_iv = cks.encrypt_u128_for_aes_ctr(iv);
|
||||||
|
|
||||||
println!("{bench_id}");
|
let d_key = CudaUnsignedRadixCiphertext::from_radix_ciphertext(&ct_key, &streams);
|
||||||
bench_group.bench_function(&bench_id, |b| {
|
let d_iv = CudaUnsignedRadixCiphertext::from_radix_ciphertext(&ct_iv, &streams);
|
||||||
b.iter(|| {
|
|
||||||
black_box(sks.aes_encrypt(
|
|
||||||
&d_iv,
|
|
||||||
&round_keys,
|
|
||||||
0,
|
|
||||||
NUM_AES_INPUTS,
|
|
||||||
SBOX_PARALLELISM,
|
|
||||||
&streams,
|
|
||||||
));
|
|
||||||
})
|
|
||||||
});
|
|
||||||
|
|
||||||
write_to_json::<u64, _>(
|
let round_keys = sks.key_expansion(&d_key, &streams);
|
||||||
&bench_id,
|
|
||||||
atomic_param,
|
println!("{bench_id}");
|
||||||
param.name(),
|
bench_group.bench_function(&bench_id, |b| {
|
||||||
"aes_encryption",
|
b.iter(|| {
|
||||||
&OperatorType::Atomic,
|
black_box(sks.aes_encrypt(
|
||||||
aes_op_bit_size,
|
&d_iv,
|
||||||
vec![atomic_param.message_modulus().0.ilog2(); aes_op_bit_size as usize],
|
&round_keys,
|
||||||
);
|
0,
|
||||||
|
NUM_AES_INPUTS,
|
||||||
|
SBOX_PARALLELISM,
|
||||||
|
&streams,
|
||||||
|
));
|
||||||
|
})
|
||||||
|
});
|
||||||
|
|
||||||
|
write_to_json::<u64, _>(
|
||||||
|
&bench_id,
|
||||||
|
atomic_param,
|
||||||
|
param.name(),
|
||||||
|
"aes_encryption",
|
||||||
|
&OperatorType::Atomic,
|
||||||
|
aes_op_bit_size,
|
||||||
|
vec![atomic_param.message_modulus().0.ilog2(); aes_op_bit_size as usize],
|
||||||
|
);
|
||||||
|
} else {
|
||||||
|
println!("{} skipped: Not enough memory in GPU", bench_id);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
bench_group.finish();
|
bench_group.finish();
|
||||||
|
|||||||
@@ -3,7 +3,7 @@ pub mod cuda {
|
|||||||
use benchmark::params_aliases::BENCH_PARAM_GPU_MULTI_BIT_GROUP_4_MESSAGE_2_CARRY_2_KS_PBS_TUNIFORM_2M128;
|
use benchmark::params_aliases::BENCH_PARAM_GPU_MULTI_BIT_GROUP_4_MESSAGE_2_CARRY_2_KS_PBS_TUNIFORM_2M128;
|
||||||
use benchmark::utilities::{write_to_json, OperatorType};
|
use benchmark::utilities::{write_to_json, OperatorType};
|
||||||
use criterion::{black_box, Criterion};
|
use criterion::{black_box, Criterion};
|
||||||
use tfhe::core_crypto::gpu::CudaStreams;
|
use tfhe::core_crypto::gpu::{check_valid_cuda_malloc, CudaStreams};
|
||||||
use tfhe::integer::gpu::ciphertext::CudaUnsignedRadixCiphertext;
|
use tfhe::integer::gpu::ciphertext::CudaUnsignedRadixCiphertext;
|
||||||
use tfhe::integer::gpu::CudaServerKey;
|
use tfhe::integer::gpu::CudaServerKey;
|
||||||
use tfhe::integer::keycache::KEY_CACHE;
|
use tfhe::integer::keycache::KEY_CACHE;
|
||||||
@@ -106,38 +106,52 @@ pub mod cuda {
|
|||||||
let sks = CudaServerKey::new(&cpu_cks, &streams);
|
let sks = CudaServerKey::new(&cpu_cks, &streams);
|
||||||
let cks = RadixClientKey::from((cpu_cks, 1));
|
let cks = RadixClientKey::from((cpu_cks, 1));
|
||||||
|
|
||||||
let ct_key = cks.encrypt_2u128_for_aes_ctr_256(key_hi, key_lo);
|
//
|
||||||
|
// Memory checks
|
||||||
|
//
|
||||||
|
let gpu_index = streams.gpu_indexes[0];
|
||||||
|
|
||||||
let ct_iv = cks.encrypt_u128_for_aes_ctr(iv);
|
let key_expansion_size = sks.get_key_expansion_256_size_on_gpu(&streams);
|
||||||
|
let aes_encrypt_size =
|
||||||
|
sks.get_aes_encrypt_size_on_gpu(NUM_AES_INPUTS, SBOX_PARALLELISM, &streams);
|
||||||
|
|
||||||
let d_key = CudaUnsignedRadixCiphertext::from_radix_ciphertext(&ct_key, &streams);
|
if check_valid_cuda_malloc(key_expansion_size, gpu_index)
|
||||||
let d_iv = CudaUnsignedRadixCiphertext::from_radix_ciphertext(&ct_iv, &streams);
|
&& check_valid_cuda_malloc(aes_encrypt_size, gpu_index)
|
||||||
|
{
|
||||||
|
let ct_key = cks.encrypt_2u128_for_aes_ctr_256(key_hi, key_lo);
|
||||||
|
let ct_iv = cks.encrypt_u128_for_aes_ctr(iv);
|
||||||
|
|
||||||
let round_keys = sks.key_expansion_256(&d_key, &streams);
|
let d_key = CudaUnsignedRadixCiphertext::from_radix_ciphertext(&ct_key, &streams);
|
||||||
|
let d_iv = CudaUnsignedRadixCiphertext::from_radix_ciphertext(&ct_iv, &streams);
|
||||||
|
|
||||||
println!("{bench_id}");
|
let round_keys = sks.key_expansion_256(&d_key, &streams);
|
||||||
bench_group.bench_function(&bench_id, |b| {
|
|
||||||
b.iter(|| {
|
|
||||||
black_box(sks.aes_256_encrypt(
|
|
||||||
&d_iv,
|
|
||||||
&round_keys,
|
|
||||||
0,
|
|
||||||
NUM_AES_INPUTS,
|
|
||||||
SBOX_PARALLELISM,
|
|
||||||
&streams,
|
|
||||||
));
|
|
||||||
})
|
|
||||||
});
|
|
||||||
|
|
||||||
write_to_json::<u64, _>(
|
println!("{bench_id}");
|
||||||
&bench_id,
|
bench_group.bench_function(&bench_id, |b| {
|
||||||
atomic_param,
|
b.iter(|| {
|
||||||
param.name(),
|
black_box(sks.aes_256_encrypt(
|
||||||
"aes_256_encryption",
|
&d_iv,
|
||||||
&OperatorType::Atomic,
|
&round_keys,
|
||||||
aes_block_op_bit_size,
|
0,
|
||||||
vec![atomic_param.message_modulus().0.ilog2(); aes_block_op_bit_size as usize],
|
NUM_AES_INPUTS,
|
||||||
);
|
SBOX_PARALLELISM,
|
||||||
|
&streams,
|
||||||
|
));
|
||||||
|
})
|
||||||
|
});
|
||||||
|
|
||||||
|
write_to_json::<u64, _>(
|
||||||
|
&bench_id,
|
||||||
|
atomic_param,
|
||||||
|
param.name(),
|
||||||
|
"aes_256_encryption",
|
||||||
|
&OperatorType::Atomic,
|
||||||
|
aes_block_op_bit_size,
|
||||||
|
vec![atomic_param.message_modulus().0.ilog2(); aes_block_op_bit_size as usize],
|
||||||
|
);
|
||||||
|
} else {
|
||||||
|
println!("{} skipped: Not enough memory in GPU", bench_id);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
bench_group.finish();
|
bench_group.finish();
|
||||||
|
|||||||
Reference in New Issue
Block a user