chore: cleanup

This commit is contained in:
Daniel Tehrani
2023-07-29 17:38:27 -07:00
parent a58a290b4d
commit a465129225
13 changed files with 44 additions and 59 deletions

View File

@@ -3,3 +3,4 @@ members = [
"tensor_pcs",
"shockwave_plus"
]
resolver = "2"

1
rust-toolchain Normal file
View File

@@ -0,0 +1 @@
1.71.0

View File

@@ -2,7 +2,6 @@
mod polynomial;
mod r1cs;
mod sumcheck;
mod utils;
use ark_std::{end_timer, start_timer};
use serde::{Deserialize, Serialize};

View File

@@ -9,12 +9,14 @@ pub struct MlPoly<F> {
}
impl<F: FieldExt> MlPoly<F> {
#[allow(dead_code)]
pub fn new(evals: Vec<F>) -> Self {
assert!(evals.len().is_power_of_two());
let num_vars = (evals.len() as f64).log2() as usize;
Self { evals, num_vars }
}
#[allow(dead_code)]
fn dot_prod(x: &[F], y: &[F]) -> F {
assert_eq!(x.len(), y.len());
let mut result = F::ZERO;
@@ -27,6 +29,7 @@ impl<F: FieldExt> MlPoly<F> {
// Evaluate the multilinear extension of the polynomial `a`, at point `t`.
// `a` is in evaluation form.
// `t` should be in big-endian form.
#[allow(dead_code)]
pub fn eval(&self, t: &[F]) -> F {
let n = self.evals.len();
debug_assert_eq!((n as f64).log2() as usize, t.len());

View File

@@ -256,12 +256,29 @@ mod tests {
use halo2curves::ff::Field;
use crate::utils::boolean_hypercube;
use super::*;
type F = halo2curves::secp256k1::Fp;
use crate::polynomial::ml_poly::MlPoly;
// Returns a vector of vectors of length m, where each vector is a boolean vector (big endian)
fn boolean_hypercube<F: FieldExt>(m: usize) -> Vec<Vec<F>> {
let n = 2usize.pow(m as u32);
let mut boolean_hypercube = Vec::<Vec<F>>::with_capacity(n);
for i in 0..n {
let mut tmp = Vec::with_capacity(m);
for j in 0..m {
let i_b = F::from((i >> j & 1) as u64);
tmp.push(i_b);
}
tmp.reverse();
boolean_hypercube.push(tmp);
}
boolean_hypercube
}
#[test]
fn test_r1cs() {
let num_cons = 10;

View File

@@ -1,20 +0,0 @@
use crate::FieldExt;
// Returns a vector of vectors of length m, where each vector is a boolean vector (big endian)
pub fn boolean_hypercube<F: FieldExt>(m: usize) -> Vec<Vec<F>> {
let n = 2usize.pow(m as u32);
let mut boolean_hypercube = Vec::<Vec<F>>::with_capacity(n);
for i in 0..n {
let mut tmp = Vec::with_capacity(m);
for j in 0..m {
let i_b = F::from((i >> j & 1) as u64);
tmp.push(i_b);
}
tmp.reverse();
boolean_hypercube.push(tmp);
}
boolean_hypercube
}

View File

@@ -1,6 +1,4 @@
use crate::FieldExt;
use halo2curves::ff::Field;
use std::vec;
pub fn fft<F>(coeffs: &[F], domain: &[F]) -> Vec<F>
where
@@ -11,7 +9,6 @@ where
return coeffs.to_vec();
}
// TODO: Just borrow the values
// Split into evens and odds
let L = coeffs
.iter()
@@ -46,7 +43,14 @@ where
return evals_L;
}
pub fn ifft<F: FieldExt + Field>(domain: &[F], evals: &[F]) -> Vec<F> {
#[cfg(test)]
mod tests {
use halo2curves::ff::Field;
use halo2curves::ff::PrimeField;
use halo2curves::pasta::Fp;
// Test the fft function by running the inverse fft
fn ifft<F: FieldExt + Field>(domain: &[F], evals: &[F]) -> Vec<F> {
let mut coeffs = vec![];
let len_mod_inv = F::from(domain.len() as u64).invert().unwrap();
let vals = fft(&evals, &domain);
@@ -57,12 +61,7 @@ pub fn ifft<F: FieldExt + Field>(domain: &[F], evals: &[F]) -> Vec<F> {
}
coeffs
}
#[cfg(test)]
mod tests {
use halo2curves::ff::PrimeField;
use halo2curves::pasta::Fp;
}
use super::*;
#[test]

View File

@@ -1,3 +1,4 @@
#![allow(non_snake_case)]
mod fft;
mod polynomial;
pub mod rs_config;

View File

@@ -32,7 +32,7 @@ pub struct CommittedTensorCode<F: FieldExt> {
}
impl<F: FieldExt> CommittedTensorCode<F> {
pub fn query_column(&self, column: usize, num_cols: usize) -> Vec<F> {
pub fn query_column(&self, column: usize) -> Vec<F> {
let num_rows = self.tensor_codeword.0.len();
let leaves =

View File

@@ -297,14 +297,12 @@ impl<F: FieldExt> TensorMultilinearPCS<F> {
}
fn test_phase(&self, indices: &[usize], u_hat_comm: &CommittedTensorCode<F>) -> Vec<Vec<F>> {
let num_cols = self.config.num_cols() * 2;
// Query the columns of u_hat
let num_indices = self.config.l;
let u_hat_openings = indices
.iter()
.map(|index| u_hat_comm.query_column(*index, num_cols))
.map(|index| u_hat_comm.query_column(*index))
.collect::<Vec<Vec<F>>>();
debug_assert_eq!(u_hat_openings.len(), num_indices);

View File

@@ -1,7 +1,6 @@
use crate::FieldExt;
use halo2curves::ff::PrimeField;
use merlin::Transcript as MerlinTranscript;
use std::{io::Repeat, marker::PhantomData, panic::UnwindSafe};
use std::marker::PhantomData;
#[derive(Clone)]
pub struct Transcript<F: FieldExt> {

View File

@@ -1,7 +1,3 @@
use core::num;
use std::marker::PhantomData;
use super::utils::hash_two;
use crate::{utils::hash_all, FieldExt};
use serde::{Deserialize, Serialize};

View File

@@ -23,15 +23,6 @@ pub fn dot_prod<F: FieldExt>(x: &[F], y: &[F]) -> F {
result
}
pub fn hash_two(values: &[[u8; 32]; 2]) -> [u8; 32] {
let mut hasher = Keccak::v256();
hasher.update(&values[0]);
hasher.update(&values[1]);
let mut hash = [0u8; 32];
hasher.finalize(&mut hash);
hash
}
pub fn hash_all(values: &[[u8; 32]]) -> [u8; 32] {
let mut hasher = Keccak::v256();
for value in values {