use crate::cpu_spqlios::{FFT64Spqlios, ffi::vec_znx}; use poulpy_hal::{ api::{TakeSlice, VecZnxBigNormalizeTmpBytes}, layouts::{ Backend, Module, Scratch, VecZnx, VecZnxBig, VecZnxBigOwned, VecZnxBigToMut, VecZnxBigToRef, VecZnxToMut, VecZnxToRef, ZnxInfos, ZnxSliceSize, ZnxView, ZnxViewMut, }, oep::{ TakeSliceImpl, VecZnxBigAddImpl, VecZnxBigAddInplaceImpl, VecZnxBigAddNormalImpl, VecZnxBigAddSmallImpl, VecZnxBigAddSmallInplaceImpl, VecZnxBigAllocBytesImpl, VecZnxBigAllocImpl, VecZnxBigAutomorphismImpl, VecZnxBigAutomorphismInplaceImpl, VecZnxBigAutomorphismInplaceTmpBytesImpl, VecZnxBigFromBytesImpl, VecZnxBigFromSmallImpl, VecZnxBigNegateImpl, VecZnxBigNegateInplaceImpl, VecZnxBigNormalizeImpl, VecZnxBigNormalizeTmpBytesImpl, VecZnxBigSubImpl, VecZnxBigSubInplaceImpl, VecZnxBigSubNegateInplaceImpl, VecZnxBigSubSmallAImpl, VecZnxBigSubSmallBImpl, VecZnxBigSubSmallInplaceImpl, VecZnxBigSubSmallNegateInplaceImpl, }, reference::{ fft64::vec_znx_big::vec_znx_big_normalize, vec_znx::{vec_znx_add_normal_ref, vec_znx_normalize_tmp_bytes}, znx::{znx_copy_ref, znx_zero_ref}, }, source::Source, }; unsafe impl VecZnxBigAllocBytesImpl for FFT64Spqlios { fn vec_znx_big_alloc_bytes_impl(n: usize, cols: usize, size: usize) -> usize { Self::layout_big_word_count() * n * cols * size * size_of::() } } unsafe impl VecZnxBigAllocImpl for FFT64Spqlios { fn vec_znx_big_alloc_impl(n: usize, cols: usize, size: usize) -> VecZnxBigOwned { VecZnxBig::alloc(n, cols, size) } } unsafe impl VecZnxBigFromBytesImpl for FFT64Spqlios { fn vec_znx_big_from_bytes_impl(n: usize, cols: usize, size: usize, bytes: Vec) -> VecZnxBigOwned { VecZnxBig::from_bytes(n, cols, size, bytes) } } unsafe impl VecZnxBigFromSmallImpl for FFT64Spqlios { fn vec_znx_big_from_small_impl(res: &mut R, res_col: usize, a: &A, a_col: usize) where R: VecZnxBigToMut, A: VecZnxToRef, { let mut res: VecZnxBig<&mut [u8], Self> = res.to_mut(); let a: VecZnx<&[u8]> = a.to_ref(); #[cfg(debug_assertions)] { assert_eq!(res.n(), a.n()); } let res_size: usize = res.size(); let a_size: usize = a.size(); let min_size: usize = res_size.min(a_size); for j in 0..min_size { znx_copy_ref(res.at_mut(res_col, j), a.at(a_col, j)); } for j in min_size..res_size { znx_zero_ref(res.at_mut(res_col, j)); } } } unsafe impl VecZnxBigAddNormalImpl for FFT64Spqlios { fn add_normal_impl>( _module: &Module, base2k: usize, res: &mut R, res_col: usize, k: usize, source: &mut Source, sigma: f64, bound: f64, ) { let res: VecZnxBig<&mut [u8], FFT64Spqlios> = res.to_mut(); let mut res_znx: VecZnx<&mut [u8]> = VecZnx { data: res.data, n: res.n, cols: res.cols, size: res.size, max_size: res.max_size, }; vec_znx_add_normal_ref(base2k, &mut res_znx, res_col, k, sigma, bound, source); } } unsafe impl VecZnxBigAddImpl for FFT64Spqlios { /// Adds `a` to `b` and stores the result on `c`. fn vec_znx_big_add_impl(module: &Module, res: &mut R, res_col: usize, a: &A, a_col: usize, b: &B, b_col: usize) where R: VecZnxBigToMut, A: VecZnxBigToRef, B: VecZnxBigToRef, { let a: VecZnxBig<&[u8], Self> = a.to_ref(); let b: VecZnxBig<&[u8], Self> = b.to_ref(); let mut res: VecZnxBig<&mut [u8], Self> = res.to_mut(); #[cfg(debug_assertions)] { assert_eq!(a.n(), res.n()); assert_eq!(b.n(), res.n()); assert_ne!(a.as_ptr(), b.as_ptr()); } unsafe { vec_znx::vec_znx_add( module.ptr(), res.at_mut_ptr(res_col, 0), res.size() as u64, res.sl() as u64, a.at_ptr(a_col, 0), a.size() as u64, a.sl() as u64, b.at_ptr(b_col, 0), b.size() as u64, b.sl() as u64, ) } } } unsafe impl VecZnxBigAddInplaceImpl for FFT64Spqlios { /// Adds `a` to `b` and stores the result on `b`. fn vec_znx_big_add_inplace_impl(module: &Module, res: &mut R, res_col: usize, a: &A, a_col: usize) where R: VecZnxBigToMut, A: VecZnxBigToRef, { let a: VecZnxBig<&[u8], Self> = a.to_ref(); let mut res: VecZnxBig<&mut [u8], Self> = res.to_mut(); #[cfg(debug_assertions)] { assert_eq!(a.n(), res.n()); } unsafe { vec_znx::vec_znx_add( module.ptr(), res.at_mut_ptr(res_col, 0), res.size() as u64, res.sl() as u64, a.at_ptr(a_col, 0), a.size() as u64, a.sl() as u64, res.at_ptr(res_col, 0), res.size() as u64, res.sl() as u64, ) } } } unsafe impl VecZnxBigAddSmallImpl for FFT64Spqlios { /// Adds `a` to `b` and stores the result on `c`. fn vec_znx_big_add_small_impl( module: &Module, res: &mut R, res_col: usize, a: &A, a_col: usize, b: &B, b_col: usize, ) where R: VecZnxBigToMut, A: VecZnxBigToRef, B: VecZnxToRef, { let a: VecZnxBig<&[u8], Self> = a.to_ref(); let b: VecZnx<&[u8]> = b.to_ref(); let mut res: VecZnxBig<&mut [u8], Self> = res.to_mut(); #[cfg(debug_assertions)] { assert_eq!(a.n(), res.n()); assert_eq!(b.n(), res.n()); assert_ne!(a.as_ptr(), b.as_ptr()); } unsafe { vec_znx::vec_znx_add( module.ptr(), res.at_mut_ptr(res_col, 0), res.size() as u64, res.sl() as u64, a.at_ptr(a_col, 0), a.size() as u64, a.sl() as u64, b.at_ptr(b_col, 0), b.size() as u64, b.sl() as u64, ) } } } unsafe impl VecZnxBigAddSmallInplaceImpl for FFT64Spqlios { /// Adds `a` to `b` and stores the result on `b`. fn vec_znx_big_add_small_inplace_impl(module: &Module, res: &mut R, res_col: usize, a: &A, a_col: usize) where R: VecZnxBigToMut, A: VecZnxToRef, { let a: VecZnx<&[u8]> = a.to_ref(); let mut res: VecZnxBig<&mut [u8], Self> = res.to_mut(); #[cfg(debug_assertions)] { assert_eq!(a.n(), res.n()); } unsafe { vec_znx::vec_znx_add( module.ptr(), res.at_mut_ptr(res_col, 0), res.size() as u64, res.sl() as u64, res.at_ptr(res_col, 0), res.size() as u64, res.sl() as u64, a.at_ptr(a_col, 0), a.size() as u64, a.sl() as u64, ) } } } unsafe impl VecZnxBigSubImpl for FFT64Spqlios { /// Subtracts `a` to `b` and stores the result on `c`. fn vec_znx_big_sub_impl(module: &Module, res: &mut R, res_col: usize, a: &A, a_col: usize, b: &B, b_col: usize) where R: VecZnxBigToMut, A: VecZnxBigToRef, B: VecZnxBigToRef, { let a: VecZnxBig<&[u8], Self> = a.to_ref(); let b: VecZnxBig<&[u8], Self> = b.to_ref(); let mut res: VecZnxBig<&mut [u8], Self> = res.to_mut(); #[cfg(debug_assertions)] { assert_eq!(a.n(), res.n()); assert_eq!(b.n(), res.n()); assert_ne!(a.as_ptr(), b.as_ptr()); } unsafe { vec_znx::vec_znx_sub( module.ptr(), res.at_mut_ptr(res_col, 0), res.size() as u64, res.sl() as u64, a.at_ptr(a_col, 0), a.size() as u64, a.sl() as u64, b.at_ptr(b_col, 0), b.size() as u64, b.sl() as u64, ) } } } unsafe impl VecZnxBigSubInplaceImpl for FFT64Spqlios { /// Subtracts `a` from `b` and stores the result on `b`. fn vec_znx_big_sub_inplace_impl(module: &Module, res: &mut R, res_col: usize, a: &A, a_col: usize) where R: VecZnxBigToMut, A: VecZnxBigToRef, { let a: VecZnxBig<&[u8], Self> = a.to_ref(); let mut res: VecZnxBig<&mut [u8], Self> = res.to_mut(); #[cfg(debug_assertions)] { assert_eq!(a.n(), res.n()); } unsafe { vec_znx::vec_znx_sub( module.ptr(), res.at_mut_ptr(res_col, 0), res.size() as u64, res.sl() as u64, res.at_ptr(res_col, 0), res.size() as u64, res.sl() as u64, a.at_ptr(a_col, 0), a.size() as u64, a.sl() as u64, ) } } } unsafe impl VecZnxBigSubNegateInplaceImpl for FFT64Spqlios { /// Subtracts `b` from `a` and stores the result on `b`. fn vec_znx_big_sub_negate_inplace_impl(module: &Module, res: &mut R, res_col: usize, a: &A, a_col: usize) where R: VecZnxBigToMut, A: VecZnxBigToRef, { let a: VecZnxBig<&[u8], Self> = a.to_ref(); let mut res: VecZnxBig<&mut [u8], Self> = res.to_mut(); #[cfg(debug_assertions)] { assert_eq!(a.n(), res.n()); } unsafe { vec_znx::vec_znx_sub( module.ptr(), res.at_mut_ptr(res_col, 0), res.size() as u64, res.sl() as u64, a.at_ptr(a_col, 0), a.size() as u64, a.sl() as u64, res.at_ptr(res_col, 0), res.size() as u64, res.sl() as u64, ) } } } unsafe impl VecZnxBigSubSmallAImpl for FFT64Spqlios { /// Subtracts `b` from `a` and stores the result on `c`. fn vec_znx_big_sub_small_a_impl( module: &Module, res: &mut R, res_col: usize, a: &A, a_col: usize, b: &B, b_col: usize, ) where R: VecZnxBigToMut, A: VecZnxToRef, B: VecZnxBigToRef, { let a: VecZnx<&[u8]> = a.to_ref(); let b: VecZnxBig<&[u8], Self> = b.to_ref(); let mut res: VecZnxBig<&mut [u8], Self> = res.to_mut(); #[cfg(debug_assertions)] { assert_eq!(a.n(), res.n()); assert_eq!(b.n(), res.n()); assert_ne!(a.as_ptr(), b.as_ptr()); } unsafe { vec_znx::vec_znx_sub( module.ptr(), res.at_mut_ptr(res_col, 0), res.size() as u64, res.sl() as u64, a.at_ptr(a_col, 0), a.size() as u64, a.sl() as u64, b.at_ptr(b_col, 0), b.size() as u64, b.sl() as u64, ) } } } unsafe impl VecZnxBigSubSmallInplaceImpl for FFT64Spqlios { /// Subtracts `a` from `res` and stores the result on `res`. fn vec_znx_big_sub_small_inplace_impl(module: &Module, res: &mut R, res_col: usize, a: &A, a_col: usize) where R: VecZnxBigToMut, A: VecZnxToRef, { let a: VecZnx<&[u8]> = a.to_ref(); let mut res: VecZnxBig<&mut [u8], Self> = res.to_mut(); #[cfg(debug_assertions)] { assert_eq!(a.n(), res.n()); } unsafe { vec_znx::vec_znx_sub( module.ptr(), res.at_mut_ptr(res_col, 0), res.size() as u64, res.sl() as u64, res.at_ptr(res_col, 0), res.size() as u64, res.sl() as u64, a.at_ptr(a_col, 0), a.size() as u64, a.sl() as u64, ) } } } unsafe impl VecZnxBigSubSmallBImpl for FFT64Spqlios { /// Subtracts `b` from `a` and stores the result on `c`. fn vec_znx_big_sub_small_b_impl( module: &Module, res: &mut R, res_col: usize, a: &A, a_col: usize, b: &B, b_col: usize, ) where R: VecZnxBigToMut, A: VecZnxBigToRef, B: VecZnxToRef, { let a: VecZnxBig<&[u8], Self> = a.to_ref(); let b: VecZnx<&[u8]> = b.to_ref(); let mut res: VecZnxBig<&mut [u8], Self> = res.to_mut(); #[cfg(debug_assertions)] { assert_eq!(a.n(), res.n()); assert_eq!(b.n(), res.n()); assert_ne!(a.as_ptr(), b.as_ptr()); } unsafe { vec_znx::vec_znx_sub( module.ptr(), res.at_mut_ptr(res_col, 0), res.size() as u64, res.sl() as u64, a.at_ptr(a_col, 0), a.size() as u64, a.sl() as u64, b.at_ptr(b_col, 0), b.size() as u64, b.sl() as u64, ) } } } unsafe impl VecZnxBigSubSmallNegateInplaceImpl for FFT64Spqlios { /// Subtracts `res` from `a` and stores the result on `res`. fn vec_znx_big_sub_small_negate_inplace_impl(module: &Module, res: &mut R, res_col: usize, a: &A, a_col: usize) where R: VecZnxBigToMut, A: VecZnxToRef, { let a: VecZnx<&[u8]> = a.to_ref(); let mut res: VecZnxBig<&mut [u8], Self> = res.to_mut(); #[cfg(debug_assertions)] { assert_eq!(a.n(), res.n()); } unsafe { vec_znx::vec_znx_sub( module.ptr(), res.at_mut_ptr(res_col, 0), res.size() as u64, res.sl() as u64, a.at_ptr(a_col, 0), a.size() as u64, a.sl() as u64, res.at_ptr(res_col, 0), res.size() as u64, res.sl() as u64, ) } } } unsafe impl VecZnxBigNegateImpl for FFT64Spqlios { fn vec_znx_big_negate_impl(module: &Module, res: &mut R, res_col: usize, a: &A, a_col: usize) where R: VecZnxBigToMut, A: VecZnxBigToRef, { let mut res: VecZnxBig<&mut [u8], Self> = res.to_mut(); let a: VecZnxBig<&[u8], Self> = a.to_ref(); unsafe { vec_znx::vec_znx_negate( module.ptr(), res.at_mut_ptr(res_col, 0), res.size() as u64, res.sl() as u64, a.at_ptr(a_col, 0), a.size() as u64, a.sl() as u64, ) } } } unsafe impl VecZnxBigNegateInplaceImpl for FFT64Spqlios { fn vec_znx_big_negate_inplace_impl(module: &Module, a: &mut A, a_col: usize) where A: VecZnxBigToMut, { let mut a: VecZnxBig<&mut [u8], Self> = a.to_mut(); unsafe { vec_znx::vec_znx_negate( module.ptr(), a.at_mut_ptr(a_col, 0), a.size() as u64, a.sl() as u64, a.at_ptr(a_col, 0), a.size() as u64, a.sl() as u64, ) } } } unsafe impl VecZnxBigNormalizeTmpBytesImpl for FFT64Spqlios { fn vec_znx_big_normalize_tmp_bytes_impl(module: &Module) -> usize { vec_znx_normalize_tmp_bytes(module.n()) } } unsafe impl VecZnxBigNormalizeImpl for FFT64Spqlios where Self: TakeSliceImpl, { fn vec_znx_big_normalize_impl( module: &Module, res_basek: usize, res: &mut R, res_col: usize, a_basek: usize, a: &A, a_col: usize, scratch: &mut Scratch, ) where R: VecZnxToMut, A: VecZnxBigToRef, { let (carry, _) = scratch.take_slice(module.vec_znx_big_normalize_tmp_bytes() / size_of::()); // unsafe { // vec_znx::vec_znx_normalize_base2k( // module.ptr(), // base2k as u64, // res.at_mut_ptr(res_col, 0), // res.size() as u64, // res.sl() as u64, // a.at_ptr(a_col, 0), // a.size() as u64, // a.sl() as u64, // tmp_bytes.as_mut_ptr(), // ); // } vec_znx_big_normalize(res_basek, res, res_col, a_basek, a, a_col, carry); } } unsafe impl VecZnxBigAutomorphismImpl for FFT64Spqlios { /// Applies the automorphism X^i -> X^ik on `a` and stores the result on `b`. fn vec_znx_big_automorphism_impl(module: &Module, k: i64, res: &mut R, res_col: usize, a: &A, a_col: usize) where R: VecZnxBigToMut, A: VecZnxBigToRef, { let a: VecZnxBig<&[u8], Self> = a.to_ref(); let mut res: VecZnxBig<&mut [u8], Self> = res.to_mut(); #[cfg(debug_assertions)] { assert_eq!(a.n(), res.n()); } unsafe { vec_znx::vec_znx_automorphism( module.ptr(), k, res.at_mut_ptr(res_col, 0), res.size() as u64, res.sl() as u64, a.at_ptr(a_col, 0), a.size() as u64, a.sl() as u64, ) } } } unsafe impl VecZnxBigAutomorphismInplaceTmpBytesImpl for FFT64Spqlios { fn vec_znx_big_automorphism_inplace_tmp_bytes_impl(_module: &Module) -> usize { 0 } } unsafe impl VecZnxBigAutomorphismInplaceImpl for FFT64Spqlios { /// Applies the automorphism X^i -> X^ik on `a` and stores the result on `a`. fn vec_znx_big_automorphism_inplace_impl( module: &Module, k: i64, a: &mut A, a_col: usize, _scratch: &mut Scratch, ) where A: VecZnxBigToMut, { let mut a: VecZnxBig<&mut [u8], Self> = a.to_mut(); unsafe { vec_znx::vec_znx_automorphism( module.ptr(), k, a.at_mut_ptr(a_col, 0), a.size() as u64, a.sl() as u64, a.at_ptr(a_col, 0), a.size() as u64, a.sl() as u64, ) } } }