mirror of
https://github.com/vacp2p/zerokit.git
synced 2026-01-07 21:04:02 -05:00
Compare commits
63 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
2071346174 | ||
|
|
c0769395bd | ||
|
|
2fc079d633 | ||
|
|
0ebeea50fd | ||
|
|
c890bc83ad | ||
|
|
77a8d28965 | ||
|
|
5c73af1130 | ||
|
|
c74ab11c82 | ||
|
|
a52cf84f46 | ||
|
|
3160d9504d | ||
|
|
0b30ba112f | ||
|
|
a2f9aaeeee | ||
|
|
a198960cf3 | ||
|
|
7f6f66bb13 | ||
|
|
a4bb3feb50 | ||
|
|
2386e8732f | ||
|
|
44c6cf3cdd | ||
|
|
eb8eedfdb4 | ||
|
|
57b694db5d | ||
|
|
0b00c639a0 | ||
|
|
7c801a804e | ||
|
|
9da80dd807 | ||
|
|
bcbd6a97af | ||
|
|
6965cf2852 | ||
|
|
578e0507b3 | ||
|
|
bf1e184da9 | ||
|
|
4473688efa | ||
|
|
c80569d518 | ||
|
|
fd99b6af74 | ||
|
|
65f53e3da3 | ||
|
|
042f8a9739 | ||
|
|
baf474e747 | ||
|
|
dc0b31752c | ||
|
|
36013bf4ba | ||
|
|
211b2d4830 | ||
|
|
5f4bcb74ce | ||
|
|
de5fd36add | ||
|
|
19c0f551c8 | ||
|
|
4133f1f8c3 | ||
|
|
149096f7a6 | ||
|
|
7023e85fce | ||
|
|
a4cafa6adc | ||
|
|
4077357e3f | ||
|
|
84d9799d09 | ||
|
|
c576af8e62 | ||
|
|
81470b9678 | ||
|
|
9d4198c205 | ||
|
|
c60e0c33fc | ||
|
|
ba467d370c | ||
|
|
ffd5851d7d | ||
|
|
759d312680 | ||
|
|
fb0ffd74a3 | ||
|
|
9d8372be39 | ||
|
|
de9c0d5072 | ||
|
|
5c60ec7cce | ||
|
|
8793965650 | ||
|
|
1930ca1610 | ||
|
|
4b4169d7a7 | ||
|
|
8a3e33be41 | ||
|
|
7bb2444ba4 | ||
|
|
00f8d039a8 | ||
|
|
e39f156fff | ||
|
|
8b04930583 |
5
.github/labels.yml
vendored
5
.github/labels.yml
vendored
@@ -90,11 +90,6 @@
|
||||
description: go-waku-productionization track (Waku Product)
|
||||
color: 9DEA79
|
||||
|
||||
# Tracks within zk-WASM project
|
||||
- name: track:kickoff
|
||||
description: Kickoff track (zk-WASM)
|
||||
color: 06B6C8
|
||||
|
||||
# Tracks within RAD project
|
||||
- name: track:waku-specs
|
||||
description: Waku specs track (RAD)
|
||||
|
||||
185
.github/workflows/ci.yml
vendored
185
.github/workflows/ci.yml
vendored
@@ -10,6 +10,7 @@ on:
|
||||
- "!rln/resources/**"
|
||||
- "!utils/src/**"
|
||||
pull_request:
|
||||
types: [opened, synchronize, reopened, ready_for_review]
|
||||
paths-ignore:
|
||||
- "**.md"
|
||||
- "!.github/workflows/*.yml"
|
||||
@@ -18,139 +19,179 @@ on:
|
||||
- "!rln/resources/**"
|
||||
- "!utils/src/**"
|
||||
|
||||
name: Tests
|
||||
name: CI
|
||||
|
||||
jobs:
|
||||
utils-test:
|
||||
# skip tests on draft PRs
|
||||
if: github.event_name == 'push' || (github.event_name == 'pull_request' && !github.event.pull_request.draft)
|
||||
strategy:
|
||||
matrix:
|
||||
platform: [ ubuntu-latest, macos-latest ]
|
||||
crate: [ utils ]
|
||||
platform: [ubuntu-latest, macos-latest]
|
||||
crate: [utils]
|
||||
runs-on: ${{ matrix.platform }}
|
||||
timeout-minutes: 60
|
||||
|
||||
name: test - ${{ matrix.crate }} - ${{ matrix.platform }}
|
||||
name: Test - ${{ matrix.crate }} - ${{ matrix.platform }}
|
||||
steps:
|
||||
- name: Checkout sources
|
||||
uses: actions/checkout@v3
|
||||
uses: actions/checkout@v4
|
||||
- name: Install stable toolchain
|
||||
uses: actions-rs/toolchain@v1
|
||||
with:
|
||||
profile: minimal
|
||||
toolchain: stable
|
||||
override: true
|
||||
uses: dtolnay/rust-toolchain@stable
|
||||
- uses: Swatinem/rust-cache@v2
|
||||
- name: Install dependencies
|
||||
run: make installdeps
|
||||
- name: cargo-make test
|
||||
- name: Test utils
|
||||
run: |
|
||||
cargo make test --release
|
||||
working-directory: ${{ matrix.crate }}
|
||||
|
||||
rln-test:
|
||||
# skip tests on draft PRs
|
||||
if: github.event_name == 'push' || (github.event_name == 'pull_request' && !github.event.pull_request.draft)
|
||||
strategy:
|
||||
matrix:
|
||||
platform: [ ubuntu-latest, macos-latest ]
|
||||
crate: [ rln ]
|
||||
feature: [ "default", "arkzkey", "stateless" ]
|
||||
platform: [ubuntu-latest, macos-latest]
|
||||
crate: [rln]
|
||||
feature: ["default", "stateless"]
|
||||
runs-on: ${{ matrix.platform }}
|
||||
timeout-minutes: 60
|
||||
|
||||
name: test - ${{ matrix.crate }} - ${{ matrix.platform }} - ${{ matrix.feature }}
|
||||
name: Test - ${{ matrix.crate }} - ${{ matrix.platform }} - ${{ matrix.feature }}
|
||||
steps:
|
||||
- name: Checkout sources
|
||||
uses: actions/checkout@v3
|
||||
uses: actions/checkout@v4
|
||||
- name: Install stable toolchain
|
||||
uses: actions-rs/toolchain@v1
|
||||
with:
|
||||
profile: minimal
|
||||
toolchain: stable
|
||||
override: true
|
||||
uses: dtolnay/rust-toolchain@stable
|
||||
- uses: Swatinem/rust-cache@v2
|
||||
- name: Install dependencies
|
||||
run: make installdeps
|
||||
- name: cargo-make test
|
||||
- name: Test rln
|
||||
run: |
|
||||
cargo make test_${{ matrix.feature }} --release
|
||||
if [ ${{ matrix.feature }} == default ]; then
|
||||
cargo make test --release
|
||||
else
|
||||
cargo make test_${{ matrix.feature }} --release
|
||||
fi
|
||||
working-directory: ${{ matrix.crate }}
|
||||
|
||||
rln-wasm:
|
||||
rln-wasm-test:
|
||||
# skip tests on draft PRs
|
||||
if: github.event_name == 'push' || (github.event_name == 'pull_request' && !github.event.pull_request.draft)
|
||||
strategy:
|
||||
matrix:
|
||||
platform: [ ubuntu-latest, macos-latest ]
|
||||
platform: [ubuntu-latest, macos-latest]
|
||||
crate: [rln-wasm]
|
||||
feature: ["default"]
|
||||
runs-on: ${{ matrix.platform }}
|
||||
timeout-minutes: 60
|
||||
|
||||
name: test - rln-wasm - ${{ matrix.platform }}
|
||||
name: Test - ${{ matrix.crate }} - ${{ matrix.platform }} - ${{ matrix.feature }}
|
||||
steps:
|
||||
- uses: actions/checkout@v3
|
||||
- uses: actions/checkout@v4
|
||||
- name: Install stable toolchain
|
||||
uses: actions-rs/toolchain@v1
|
||||
with:
|
||||
profile: minimal
|
||||
toolchain: stable
|
||||
override: true
|
||||
uses: dtolnay/rust-toolchain@stable
|
||||
- uses: Swatinem/rust-cache@v2
|
||||
- name: Install Dependencies
|
||||
- name: Install dependencies
|
||||
run: make installdeps
|
||||
- name: Install wasm-pack
|
||||
uses: jetli/wasm-pack-action@v0.3.0
|
||||
- run: cargo make build
|
||||
working-directory: rln-wasm
|
||||
- run: cargo make test --release
|
||||
working-directory: rln-wasm
|
||||
- name: Build rln-wasm
|
||||
run: cargo make build
|
||||
working-directory: ${{ matrix.crate }}
|
||||
- name: Test rln-wasm on node
|
||||
run: cargo make test --release
|
||||
working-directory: ${{ matrix.crate }}
|
||||
- name: Test rln-wasm on browser
|
||||
run: cargo make test_browser --release
|
||||
working-directory: ${{ matrix.crate }}
|
||||
|
||||
rln-wasm-parallel-test:
|
||||
# skip tests on draft PRs
|
||||
if: github.event_name == 'push' || (github.event_name == 'pull_request' && !github.event.pull_request.draft)
|
||||
strategy:
|
||||
matrix:
|
||||
platform: [ubuntu-latest, macos-latest]
|
||||
crate: [rln-wasm]
|
||||
feature: ["parallel"]
|
||||
runs-on: ${{ matrix.platform }}
|
||||
timeout-minutes: 60
|
||||
|
||||
name: Test - ${{ matrix.crate }} - ${{ matrix.platform }} - ${{ matrix.feature }}
|
||||
steps:
|
||||
- uses: actions/checkout@v4
|
||||
- name: Install nightly toolchain
|
||||
uses: dtolnay/rust-toolchain@nightly
|
||||
with:
|
||||
components: rust-src
|
||||
targets: wasm32-unknown-unknown
|
||||
- uses: Swatinem/rust-cache@v2
|
||||
- name: Install dependencies
|
||||
run: make installdeps
|
||||
- name: Build rln-wasm in parallel mode
|
||||
run: cargo make build_parallel
|
||||
working-directory: ${{ matrix.crate }}
|
||||
- name: Test rln-wasm in parallel mode on browser
|
||||
run: cargo make test_parallel --release
|
||||
working-directory: ${{ matrix.crate }}
|
||||
|
||||
lint:
|
||||
# run on both ready and draft PRs
|
||||
if: github.event_name == 'push' || (github.event_name == 'pull_request' && !github.event.pull_request.draft)
|
||||
strategy:
|
||||
matrix:
|
||||
# we run lint tests only on ubuntu
|
||||
platform: [ ubuntu-latest ]
|
||||
crate: [ rln, utils ]
|
||||
# run lint tests only on ubuntu
|
||||
platform: [ubuntu-latest]
|
||||
crate: [rln, rln-wasm, utils]
|
||||
runs-on: ${{ matrix.platform }}
|
||||
timeout-minutes: 60
|
||||
|
||||
name: lint - ${{ matrix.crate }} - ${{ matrix.platform }}
|
||||
name: Lint - ${{ matrix.crate }} - ${{ matrix.platform }}
|
||||
steps:
|
||||
- name: Checkout sources
|
||||
uses: actions/checkout@v3
|
||||
uses: actions/checkout@v4
|
||||
- name: Install stable toolchain
|
||||
uses: actions-rs/toolchain@v1
|
||||
uses: dtolnay/rust-toolchain@stable
|
||||
with:
|
||||
profile: minimal
|
||||
toolchain: stable
|
||||
override: true
|
||||
components: rustfmt, clippy
|
||||
- name: Install wasm32 target
|
||||
if: matrix.crate == 'rln-wasm'
|
||||
run: rustup target add wasm32-unknown-unknown
|
||||
- uses: Swatinem/rust-cache@v2
|
||||
- name: Install Dependencies
|
||||
- name: Install dependencies
|
||||
run: make installdeps
|
||||
- name: cargo fmt
|
||||
- name: Check formatting
|
||||
if: success() || failure()
|
||||
run: cargo fmt -- --check
|
||||
working-directory: ${{ matrix.crate }}
|
||||
- name: cargo clippy
|
||||
if: success() || failure()
|
||||
- name: Check clippy wasm target
|
||||
if: (success() || failure()) && (matrix.crate == 'rln-wasm')
|
||||
run: |
|
||||
cargo clippy --release -- -D warnings
|
||||
cargo clippy --target wasm32-unknown-unknown --tests --release -- -D warnings
|
||||
working-directory: ${{ matrix.crate }}
|
||||
- name: Check clippy default feature
|
||||
if: (success() || failure()) && (matrix.crate != 'rln-wasm')
|
||||
run: |
|
||||
cargo clippy --all-targets --tests --release -- -D warnings
|
||||
- name: Check clippy stateless feature
|
||||
if: (success() || failure()) && (matrix.crate == 'rln')
|
||||
run: |
|
||||
cargo clippy --all-targets --tests --release --features=stateless --no-default-features -- -D warnings
|
||||
working-directory: ${{ matrix.crate }}
|
||||
# We skip clippy on rln-wasm, since wasm target is managed by cargo make
|
||||
# Currently not treating warnings as error, too noisy
|
||||
# -- -D warnings
|
||||
|
||||
benchmark-utils:
|
||||
# run only in pull requests
|
||||
if: github.event_name == 'pull_request'
|
||||
# run only on ready PRs
|
||||
if: github.event_name == 'pull_request' && !github.event.pull_request.draft
|
||||
strategy:
|
||||
matrix:
|
||||
# we run benchmark tests only on ubuntu
|
||||
platform: [ ubuntu-latest ]
|
||||
crate: [ utils ]
|
||||
# run benchmark tests only on ubuntu
|
||||
platform: [ubuntu-latest]
|
||||
crate: [utils]
|
||||
runs-on: ${{ matrix.platform }}
|
||||
timeout-minutes: 60
|
||||
|
||||
name: benchmark - ${{ matrix.platform }} - ${{ matrix.crate }}
|
||||
name: Benchmark - ${{ matrix.crate }} - ${{ matrix.platform }}
|
||||
steps:
|
||||
- name: Checkout sources
|
||||
uses: actions/checkout@v3
|
||||
uses: actions/checkout@v4
|
||||
- uses: Swatinem/rust-cache@v2
|
||||
- uses: boa-dev/criterion-compare-action@v3
|
||||
with:
|
||||
@@ -158,24 +199,24 @@ jobs:
|
||||
cwd: ${{ matrix.crate }}
|
||||
|
||||
benchmark-rln:
|
||||
# run only in pull requests
|
||||
if: github.event_name == 'pull_request'
|
||||
# run only on ready PRs
|
||||
if: github.event_name == 'pull_request' && !github.event.pull_request.draft
|
||||
strategy:
|
||||
matrix:
|
||||
# we run benchmark tests only on ubuntu
|
||||
platform: [ ubuntu-latest ]
|
||||
crate: [ rln ]
|
||||
feature: [ "default", "arkzkey" ]
|
||||
# run benchmark tests only on ubuntu
|
||||
platform: [ubuntu-latest]
|
||||
crate: [rln]
|
||||
feature: ["default"]
|
||||
runs-on: ${{ matrix.platform }}
|
||||
timeout-minutes: 60
|
||||
|
||||
name: benchmark - ${{ matrix.platform }} - ${{ matrix.crate }} - ${{ matrix.feature }}
|
||||
name: Benchmark - ${{ matrix.crate }} - ${{ matrix.platform }} - ${{ matrix.feature }}
|
||||
steps:
|
||||
- name: Checkout sources
|
||||
uses: actions/checkout@v3
|
||||
uses: actions/checkout@v4
|
||||
- uses: Swatinem/rust-cache@v2
|
||||
- uses: boa-dev/criterion-compare-action@v3
|
||||
with:
|
||||
branchName: ${{ github.base_ref }}
|
||||
cwd: ${{ matrix.crate }}
|
||||
features: ${{ matrix.feature }}
|
||||
features: ${{ matrix.feature }}
|
||||
|
||||
153
.github/workflows/nightly-release.yml
vendored
153
.github/workflows/nightly-release.yml
vendored
@@ -6,43 +6,45 @@ on:
|
||||
|
||||
jobs:
|
||||
linux:
|
||||
strategy:
|
||||
matrix:
|
||||
feature: [ "default", "arkzkey", "stateless" ]
|
||||
target:
|
||||
- x86_64-unknown-linux-gnu
|
||||
- aarch64-unknown-linux-gnu
|
||||
- i686-unknown-linux-gnu
|
||||
include:
|
||||
- feature: stateless
|
||||
cargo_args: --exclude rln-cli
|
||||
name: Linux build
|
||||
runs-on: ubuntu-latest
|
||||
strategy:
|
||||
matrix:
|
||||
features:
|
||||
- ["stateless"]
|
||||
- ["stateless", "parallel"]
|
||||
- ["pmtree-ft"]
|
||||
- ["pmtree-ft", "parallel"]
|
||||
- ["fullmerkletree"]
|
||||
- ["fullmerkletree", "parallel"]
|
||||
- ["optimalmerkletree"]
|
||||
- ["optimalmerkletree", "parallel"]
|
||||
target: [x86_64-unknown-linux-gnu, aarch64-unknown-linux-gnu]
|
||||
env:
|
||||
FEATURES_CARGO: ${{ join(matrix.features, ',') }}
|
||||
FEATURES_TAG: ${{ join(matrix.features, '-') }}
|
||||
TARGET: ${{ matrix.target }}
|
||||
steps:
|
||||
- name: Checkout sources
|
||||
uses: actions/checkout@v3
|
||||
uses: actions/checkout@v4
|
||||
- name: Install stable toolchain
|
||||
uses: actions-rs/toolchain@v1
|
||||
uses: dtolnay/rust-toolchain@stable
|
||||
with:
|
||||
profile: minimal
|
||||
toolchain: stable
|
||||
override: true
|
||||
target: ${{ matrix.target }}
|
||||
target: ${{ env.TARGET }}
|
||||
- uses: Swatinem/rust-cache@v2
|
||||
- name: Install dependencies
|
||||
run: make installdeps
|
||||
- name: cross build
|
||||
- name: Cross build
|
||||
run: |
|
||||
cross build --release --target ${{ matrix.target }} --features ${{ matrix.feature }} --workspace --exclude rln-wasm ${{ matrix.cargo_args }}
|
||||
cross build --release --target $TARGET --no-default-features --features "$FEATURES_CARGO" --workspace
|
||||
mkdir release
|
||||
cp target/${{ matrix.target }}/release/librln* release/
|
||||
tar -czvf ${{ matrix.target }}-${{ matrix.feature }}-rln.tar.gz release/
|
||||
|
||||
cp target/$TARGET/release/librln* release/
|
||||
tar -czvf $TARGET-$FEATURES_TAG-rln.tar.gz release/
|
||||
- name: Upload archive artifact
|
||||
uses: actions/upload-artifact@v4
|
||||
with:
|
||||
name: ${{ matrix.target }}-${{ matrix.feature }}-archive
|
||||
path: ${{ matrix.target }}-${{ matrix.feature }}-rln.tar.gz
|
||||
name: ${{ env.TARGET }}-${{ env.FEATURES_TAG }}-archive
|
||||
path: ${{ env.TARGET }}-${{ env.FEATURES_TAG }}-rln.tar.gz
|
||||
retention-days: 2
|
||||
|
||||
macos:
|
||||
@@ -50,84 +52,107 @@ jobs:
|
||||
runs-on: macos-latest
|
||||
strategy:
|
||||
matrix:
|
||||
feature: [ "default", "arkzkey", "stateless" ]
|
||||
target:
|
||||
- x86_64-apple-darwin
|
||||
- aarch64-apple-darwin
|
||||
include:
|
||||
- feature: stateless
|
||||
cargo_args: --exclude rln-cli
|
||||
features:
|
||||
- ["stateless"]
|
||||
- ["stateless", "parallel"]
|
||||
- ["pmtree-ft"]
|
||||
- ["pmtree-ft", "parallel"]
|
||||
- ["fullmerkletree"]
|
||||
- ["fullmerkletree", "parallel"]
|
||||
- ["optimalmerkletree"]
|
||||
- ["optimalmerkletree", "parallel"]
|
||||
target: [x86_64-apple-darwin, aarch64-apple-darwin]
|
||||
env:
|
||||
FEATURES_CARGO: ${{ join(matrix.features, ',') }}
|
||||
FEATURES_TAG: ${{ join(matrix.features, '-') }}
|
||||
TARGET: ${{ matrix.target }}
|
||||
steps:
|
||||
- name: Checkout sources
|
||||
uses: actions/checkout@v3
|
||||
uses: actions/checkout@v4
|
||||
- name: Install stable toolchain
|
||||
uses: actions-rs/toolchain@v1
|
||||
uses: dtolnay/rust-toolchain@stable
|
||||
with:
|
||||
profile: minimal
|
||||
toolchain: stable
|
||||
override: true
|
||||
target: ${{ matrix.target }}
|
||||
target: ${{ env.TARGET }}
|
||||
- uses: Swatinem/rust-cache@v2
|
||||
- name: Install dependencies
|
||||
run: make installdeps
|
||||
- name: cross build
|
||||
- name: Cross build
|
||||
run: |
|
||||
cross build --release --target ${{ matrix.target }} --features ${{ matrix.feature }} --workspace --exclude rln-wasm ${{ matrix.cargo_args }}
|
||||
cross build --release --target $TARGET --no-default-features --features "$FEATURES_CARGO" --workspace
|
||||
mkdir release
|
||||
cp target/${{ matrix.target }}/release/librln* release/
|
||||
tar -czvf ${{ matrix.target }}-${{ matrix.feature }}-rln.tar.gz release/
|
||||
|
||||
cp target/$TARGET/release/librln* release/
|
||||
tar -czvf $TARGET-$FEATURES_TAG-rln.tar.gz release/
|
||||
- name: Upload archive artifact
|
||||
uses: actions/upload-artifact@v4
|
||||
with:
|
||||
name: ${{ matrix.target }}-${{ matrix.feature }}-archive
|
||||
path: ${{ matrix.target }}-${{ matrix.feature }}-rln.tar.gz
|
||||
name: ${{ env.TARGET }}-${{ env.FEATURES_TAG }}-archive
|
||||
path: ${{ env.TARGET }}-${{ env.FEATURES_TAG }}-rln.tar.gz
|
||||
retention-days: 2
|
||||
|
||||
browser-rln-wasm:
|
||||
name: Browser build (RLN WASM)
|
||||
rln-wasm:
|
||||
name: Build rln-wasm
|
||||
runs-on: ubuntu-latest
|
||||
strategy:
|
||||
matrix:
|
||||
feature:
|
||||
- "default"
|
||||
- "parallel"
|
||||
- "utils"
|
||||
steps:
|
||||
- name: Checkout sources
|
||||
uses: actions/checkout@v3
|
||||
uses: actions/checkout@v4
|
||||
- name: Install stable toolchain
|
||||
uses: actions-rs/toolchain@v1
|
||||
uses: dtolnay/rust-toolchain@stable
|
||||
with:
|
||||
profile: minimal
|
||||
toolchain: stable
|
||||
override: true
|
||||
targets: wasm32-unknown-unknown
|
||||
- name: Install nightly toolchain
|
||||
uses: dtolnay/rust-toolchain@nightly
|
||||
with:
|
||||
components: rust-src
|
||||
targets: wasm32-unknown-unknown
|
||||
- uses: Swatinem/rust-cache@v2
|
||||
- name: Install dependencies
|
||||
run: make installdeps
|
||||
- name: Install wasm-pack
|
||||
uses: jetli/wasm-pack-action@v0.3.0
|
||||
- name: cross make build
|
||||
- name: Build rln-wasm package
|
||||
run: |
|
||||
cross make build
|
||||
mkdir release
|
||||
cp pkg/** release/
|
||||
tar -czvf browser-rln-wasm.tar.gz release/
|
||||
working-directory: rln-wasm
|
||||
if [[ ${{ matrix.feature }} == "parallel" ]]; then
|
||||
env CARGO_TARGET_WASM32_UNKNOWN_UNKNOWN_RUSTFLAGS="-C target-feature=+atomics,+bulk-memory,+mutable-globals -C link-arg=--shared-memory -C link-arg=--max-memory=1073741824 -C link-arg=--import-memory -C link-arg=--export=__wasm_init_tls -C link-arg=--export=__tls_size -C link-arg=--export=__tls_align -C link-arg=--export=__tls_base" \
|
||||
rustup run nightly wasm-pack build --release --target web --scope waku \
|
||||
--features parallel -Z build-std=panic_abort,std
|
||||
sed -i.bak 's/rln-wasm/zerokit-rln-wasm-parallel/g' pkg/package.json && rm pkg/package.json.bak
|
||||
elif [[ ${{ matrix.feature }} == "utils" ]]; then
|
||||
wasm-pack build --release --target web --scope waku --no-default-features --features utils
|
||||
sed -i.bak 's/rln-wasm/zerokit-rln-wasm-utils/g' pkg/package.json && rm pkg/package.json.bak
|
||||
else
|
||||
wasm-pack build --release --target web --scope waku
|
||||
sed -i.bak 's/rln-wasm/zerokit-rln-wasm/g' pkg/package.json && rm pkg/package.json.bak
|
||||
fi
|
||||
|
||||
jq '. + {keywords: ["zerokit", "rln", "wasm"]}' pkg/package.json > pkg/package.json.tmp && \
|
||||
mv pkg/package.json.tmp pkg/package.json
|
||||
|
||||
mkdir release
|
||||
cp -r pkg/* release/
|
||||
tar -czvf rln-wasm-${{ matrix.feature }}.tar.gz release/
|
||||
working-directory: rln-wasm
|
||||
- name: Upload archive artifact
|
||||
uses: actions/upload-artifact@v4
|
||||
with:
|
||||
name: browser-rln-wasm-archive
|
||||
path: rln-wasm/browser-rln-wasm.tar.gz
|
||||
name: rln-wasm-${{ matrix.feature }}-archive
|
||||
path: rln-wasm/rln-wasm-${{ matrix.feature }}.tar.gz
|
||||
retention-days: 2
|
||||
|
||||
prepare-prerelease:
|
||||
name: Prepare pre-release
|
||||
needs: [ linux, macos, browser-rln-wasm ]
|
||||
needs: [linux, macos, rln-wasm]
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- name: Checkout code
|
||||
uses: actions/checkout@v3
|
||||
uses: actions/checkout@v4
|
||||
with:
|
||||
ref: master
|
||||
- name: Download artifacts
|
||||
uses: actions/download-artifact@v4
|
||||
|
||||
- name: Delete tag
|
||||
uses: dev-drprasad/delete-tag-and-release@v0.2.1
|
||||
with:
|
||||
@@ -135,7 +160,6 @@ jobs:
|
||||
tag_name: nightly
|
||||
env:
|
||||
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
||||
|
||||
- name: Create prerelease
|
||||
run: |
|
||||
start_tag=$(gh release list -L 2 --exclude-drafts | grep -v nightly | cut -d$'\t' -f3 | sed -n '1p')
|
||||
@@ -147,7 +171,6 @@ jobs:
|
||||
*-archive/*.tar.gz \
|
||||
env:
|
||||
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
||||
|
||||
- name: Delete artifacts
|
||||
uses: geekyeggo/delete-artifact@v5
|
||||
with:
|
||||
|
||||
2
.github/workflows/sync-labels.yml
vendored
2
.github/workflows/sync-labels.yml
vendored
@@ -9,7 +9,7 @@ jobs:
|
||||
build:
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- uses: actions/checkout@v3
|
||||
- uses: actions/checkout@v4
|
||||
- uses: micnncim/action-label-syncer@v1
|
||||
env:
|
||||
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
||||
|
||||
24
.gitignore
vendored
24
.gitignore
vendored
@@ -1,17 +1,29 @@
|
||||
# Common files to ignore in Rust projects
|
||||
.DS_Store
|
||||
.idea
|
||||
*.log
|
||||
tmp/
|
||||
rln/pmtree_db
|
||||
|
||||
# Generated by Cargo
|
||||
# will have compiled files and executables
|
||||
debug/
|
||||
target/
|
||||
wabt/
|
||||
# Generated by Cargo will have compiled files and executables
|
||||
/target
|
||||
|
||||
# Generated by Nix
|
||||
result
|
||||
|
||||
# These are backup files generated by rustfmt
|
||||
**/*.rs.bk
|
||||
|
||||
# MSVC Windows builds of rustc generate these, which store debugging information
|
||||
*.pdb
|
||||
|
||||
# FFI C examples
|
||||
rln/ffi_c_examples/main
|
||||
rln/ffi_c_examples/rln.h
|
||||
rln/ffi_c_examples/database
|
||||
|
||||
# FFI Nim examples
|
||||
rln/ffi_nim_examples/main
|
||||
rln/ffi_nim_examples/database
|
||||
|
||||
# Vscode
|
||||
.vscode
|
||||
|
||||
28
CHANGELOG.md
28
CHANGELOG.md
@@ -1,28 +0,0 @@
|
||||
## 2023-02-28 v0.2
|
||||
|
||||
This release contains:
|
||||
|
||||
- Improved code quality
|
||||
- Allows consumers of zerokit RLN to set leaves to the Merkle Tree from an arbitrary index. Useful for batching updates to the Merkle Tree.
|
||||
- Improved performance for proof generation and verification
|
||||
- rln_wasm which allows for the consumption of RLN through a WebAssembly interface
|
||||
- Refactored to generate Semaphore-compatible credentials
|
||||
- Dual License under Apache 2.0 and MIT
|
||||
- RLN compiles as a static library, which can be consumed through a C FFI
|
||||
|
||||
|
||||
## 2022-09-19 v0.1
|
||||
|
||||
Initial beta release.
|
||||
|
||||
This release contains:
|
||||
|
||||
- RLN Module with API to manage, compute and verify [RLN](https://rfc.vac.dev/spec/32/) zkSNARK proofs and RLN primitives.
|
||||
- This can be consumed either as a Rust API or as a C FFI. The latter means it can be easily consumed through other environments, such as [Go](https://github.com/status-im/go-zerokit-rln/blob/master/rln/librln.h) or [Nim](https://github.com/status-im/nwaku/blob/4745c7872c69b5fd5c6ddab36df9c5c3d55f57c3/waku/v2/protocol/waku_rln_relay/waku_rln_relay_types.nim).
|
||||
|
||||
It also contains the following examples and experiments:
|
||||
|
||||
- Basic [example wrapper](https://github.com/vacp2p/zerokit/tree/master/multiplier) around a simple Circom circuit to show Circom integration through ark-circom and FFI.
|
||||
- Experimental [Semaphore wrapper](https://github.com/vacp2p/zerokit/tree/master/semaphore).
|
||||
|
||||
Feedback welcome! You can either [open an issue](https://github.com/vacp2p/zerokit/issues) or come talk to us in our [Vac Discord](https://discord.gg/PQFdubGt6d) #zerokit channel.
|
||||
197
CONTRIBUTING.md
Normal file
197
CONTRIBUTING.md
Normal file
@@ -0,0 +1,197 @@
|
||||
# Contributing to Zerokit
|
||||
|
||||
Thank you for your interest in contributing to Zerokit!
|
||||
This guide will discuss how the Zerokit team handles [Commits](#commits),
|
||||
[Pull Requests](#pull-requests) and [Merging](#merging).
|
||||
|
||||
**Note:** We won't force external contributors to follow this verbatim.
|
||||
Following these guidelines definitely helps us in accepting your contributions.
|
||||
|
||||
## Getting Started
|
||||
|
||||
1. Fork the repository
|
||||
2. Create a feature branch: `git checkout -b fix/your-bug-fix` or `git checkout -b feat/your-feature-name`
|
||||
3. Make your changes following our guidelines
|
||||
4. Ensure relevant tests pass (see [testing guidelines](#building-and-testing))
|
||||
5. Commit your changes (signed commits are highly encouraged - see [commit guidelines](#commits))
|
||||
6. Push and create a Pull Request
|
||||
|
||||
## Development Setup
|
||||
|
||||
### Prerequisites
|
||||
|
||||
Install the required dependencies:
|
||||
|
||||
```bash
|
||||
make installdeps
|
||||
```
|
||||
|
||||
Or use Nix:
|
||||
|
||||
```bash
|
||||
nix develop
|
||||
```
|
||||
|
||||
### Building and Testing
|
||||
|
||||
```bash
|
||||
# Build all crates
|
||||
make build
|
||||
|
||||
# Run standard tests
|
||||
make test
|
||||
|
||||
# Module-specific testing
|
||||
cd rln && cargo make test_stateless # Test stateless features
|
||||
cd rln-wasm && cargo make test_browser # Test in browser headless mode
|
||||
cd rln-wasm && cargo make test_parallel # Test parallel features
|
||||
```
|
||||
|
||||
### Tools
|
||||
|
||||
We recommend using the [markdownlint extension](https://marketplace.visualstudio.com/items?itemName=DavidAnson.vscode-markdownlint)
|
||||
for VS Code to maintain consistent documentation formatting.
|
||||
|
||||
## Commits
|
||||
|
||||
We want to keep our commits small and focused.
|
||||
This allows for easily reviewing individual commits and/or
|
||||
splitting up pull requests when they grow too big.
|
||||
Additionally, this allows us to merge smaller changes quicker and release more often.
|
||||
|
||||
**All commits must be GPG signed.**
|
||||
This ensures the authenticity and integrity of contributions.
|
||||
|
||||
### Conventional Commits
|
||||
|
||||
When making the commit, write the commit message
|
||||
following the [Conventional Commits (v1.0.0)](https://www.conventionalcommits.org/en/v1.0.0/) specification.
|
||||
Following this convention allows us to provide an automated release process
|
||||
that also generates a detailed Changelog.
|
||||
|
||||
As described by the specification, our commit messages should be written as:
|
||||
|
||||
```markdown
|
||||
<type>[optional scope]: <description>
|
||||
|
||||
[optional body]
|
||||
|
||||
[optional footer(s)]
|
||||
```
|
||||
|
||||
Some examples of this pattern include:
|
||||
|
||||
```markdown
|
||||
feat(rln): add parallel witness calculation support
|
||||
```
|
||||
|
||||
```markdown
|
||||
fix(rln-wasm): resolve memory leak in browser threading
|
||||
```
|
||||
|
||||
```markdown
|
||||
docs: update RLN protocol flow documentation
|
||||
```
|
||||
|
||||
#### Scopes
|
||||
|
||||
Use scopes to improve the Changelog:
|
||||
|
||||
- `rln` - Core RLN implementation
|
||||
- `rln-cli` - Command-line interface
|
||||
- `rln-wasm` - WebAssembly bindings
|
||||
- `utils` - Cryptographic utilities (Merkle trees, Poseidon hash)
|
||||
- `ci` - Continuous integration
|
||||
|
||||
#### Breaking Changes
|
||||
|
||||
Mark breaking changes by adding `!` after the type:
|
||||
|
||||
```markdown
|
||||
feat(rln)!: change proof generation API
|
||||
```
|
||||
|
||||
## Pull Requests
|
||||
|
||||
Before creating a pull request, search for related issues.
|
||||
If none exist, create an issue describing the problem you're solving.
|
||||
|
||||
### CI Flow
|
||||
|
||||
Our continuous integration automatically runs when you create a Pull Request:
|
||||
|
||||
- **Build verification**: All crates compile successfully
|
||||
- **Test execution**: Comprehensive testing across all modules and feature combinations
|
||||
- **Code formatting**: `cargo fmt` compliance
|
||||
- **Linting**: `cargo clippy` checks
|
||||
- **Cross-platform builds**: Testing on multiple platforms
|
||||
|
||||
Ensure the following commands pass before submitting:
|
||||
|
||||
```bash
|
||||
# Format code
|
||||
cargo fmt --all
|
||||
|
||||
# Check for common mistakes
|
||||
cargo clippy --all-targets
|
||||
|
||||
# Run all tests
|
||||
make test
|
||||
```
|
||||
|
||||
### Adding Tests
|
||||
|
||||
Include tests for new functionality:
|
||||
|
||||
- **Unit tests** for specific functions
|
||||
- **Integration tests** for broader functionality
|
||||
- **WASM tests** for browser compatibility
|
||||
|
||||
### Typos and Small Changes
|
||||
|
||||
For minor fixes like typos, please report them as issues instead of opening PRs.
|
||||
This helps us manage resources effectively and ensures meaningful contributions.
|
||||
|
||||
## Merging
|
||||
|
||||
We use "squash merging" for all pull requests.
|
||||
This combines all commits into one commit, so keep pull requests small and focused.
|
||||
|
||||
### Requirements
|
||||
|
||||
- CI checks must pass
|
||||
- At least one maintainer review and approval
|
||||
- All review feedback addressed
|
||||
|
||||
### Squash Guidelines
|
||||
|
||||
When squashing, update the commit title to be a proper Conventional Commit and
|
||||
include any other relevant commits in the body:
|
||||
|
||||
```markdown
|
||||
feat(rln): implement parallel witness calculation (#123)
|
||||
|
||||
fix(tests): resolve memory leak in test suite
|
||||
chore(ci): update rust toolchain version
|
||||
```
|
||||
|
||||
## Roadmap Alignment
|
||||
|
||||
Please refer to our [project roadmap](https://roadmap.vac.dev/) for current development priorities.
|
||||
Consider how your changes align with these strategic goals when contributing.
|
||||
|
||||
## Getting Help
|
||||
|
||||
- **Issues**: Create a GitHub issue for bugs or feature requests
|
||||
- **Discussions**: Use GitHub Discussions for questions
|
||||
- **Documentation**: Check existing docs and unit tests for examples
|
||||
|
||||
## License
|
||||
|
||||
By contributing to Zerokit, you agree that your contributions will be licensed under both MIT and
|
||||
Apache 2.0 licenses, consistent with the project's dual licensing.
|
||||
|
||||
## Additional Resources
|
||||
|
||||
- [Conventional Commits Guide](https://www.conventionalcommits.org/en/v1.0.0/)
|
||||
- [Project GitHub Repository](https://github.com/vacp2p/zerokit)
|
||||
3076
Cargo.lock
generated
3076
Cargo.lock
generated
File diff suppressed because it is too large
Load Diff
12
Cargo.toml
12
Cargo.toml
@@ -1,14 +1,10 @@
|
||||
[workspace]
|
||||
members = ["rln", "rln-cli", "rln-wasm", "utils"]
|
||||
default-members = ["rln", "rln-cli", "utils"]
|
||||
members = ["rln", "utils"]
|
||||
exclude = ["rln-cli", "rln-wasm"]
|
||||
resolver = "2"
|
||||
|
||||
# Compilation profile for any non-workspace member.
|
||||
# Dependencies are optimized, even in a dev build. This improves dev performance
|
||||
# while having neglible impact on incremental build times.
|
||||
# Dependencies are optimized, even in a dev build.
|
||||
# This improves dev performance while having negligible impact on incremental build times.
|
||||
[profile.dev.package."*"]
|
||||
opt-level = 3
|
||||
|
||||
[profile.release.package."rln-wasm"]
|
||||
# Tell `rustc` to optimize for small code size.
|
||||
opt-level = "s"
|
||||
|
||||
32
Makefile
32
Makefile
@@ -1,6 +1,6 @@
|
||||
.PHONY: all installdeps build test clean
|
||||
.PHONY: all installdeps build test bench clean
|
||||
|
||||
all: .pre-build build
|
||||
all: installdeps build
|
||||
|
||||
.fetch-submodules:
|
||||
@git submodule update --init --recursive
|
||||
@@ -13,25 +13,27 @@ endif
|
||||
|
||||
installdeps: .pre-build
|
||||
ifeq ($(shell uname),Darwin)
|
||||
# commented due to https://github.com/orgs/Homebrew/discussions/4612
|
||||
# @brew update
|
||||
@brew install cmake ninja
|
||||
@brew install ninja
|
||||
else ifeq ($(shell uname),Linux)
|
||||
@sudo apt-get update
|
||||
@sudo apt-get install -y cmake ninja-build
|
||||
@if [ -f /etc/os-release ] && grep -q "ID=nixos" /etc/os-release; then \
|
||||
echo "Detected NixOS, skipping apt installation."; \
|
||||
else \
|
||||
sudo apt update; \
|
||||
sudo apt install -y cmake ninja-build; \
|
||||
fi
|
||||
endif
|
||||
@git -C "wabt" pull || git clone --recursive https://github.com/WebAssembly/wabt.git "wabt"
|
||||
@cd wabt && mkdir -p build && cd build && cmake .. -GNinja && ninja && sudo ninja install
|
||||
@which wasm-pack || cargo install wasm-pack
|
||||
# nvm already checks if it's installed, and no-ops if it is
|
||||
@curl -o- https://raw.githubusercontent.com/nvm-sh/nvm/v0.39.7/install.sh | bash
|
||||
@. ${HOME}/.nvm/nvm.sh && nvm install 18.20.2 && nvm use 18.20.2;
|
||||
@which wasm-pack > /dev/null && wasm-pack --version | grep -q "0.13.1" || cargo install wasm-pack --version=0.13.1
|
||||
@test -s "$$HOME/.nvm/nvm.sh" || curl -o- https://raw.githubusercontent.com/nvm-sh/nvm/v0.40.2/install.sh | bash
|
||||
@bash -c '. "$$HOME/.nvm/nvm.sh"; [ "$$(node -v 2>/dev/null)" = "v22.14.0" ] || nvm install 22.14.0; nvm use 22.14.0; nvm alias default 22.14.0'
|
||||
|
||||
build: .pre-build
|
||||
build: installdeps
|
||||
@cargo make build
|
||||
|
||||
test: .pre-build
|
||||
test: build
|
||||
@cargo make test
|
||||
|
||||
bench: build
|
||||
@cargo make bench
|
||||
|
||||
clean:
|
||||
@cargo clean
|
||||
|
||||
82
README.md
82
README.md
@@ -1,45 +1,87 @@
|
||||
# Zerokit
|
||||
|
||||
A set of Zero Knowledge modules, written in Rust and designed to be used in other system programming environments.
|
||||
[](https://crates.io/crates/rln)
|
||||
[](https://github.com/vacp2p/zerokit/actions)
|
||||
[](https://opensource.org/licenses/MIT)
|
||||
[](https://opensource.org/licenses/Apache-2.0)
|
||||
|
||||
## Initial scope
|
||||
A collection of Zero Knowledge modules written in Rust and designed to be used in other system programming environments.
|
||||
|
||||
Focus on RLN and being able to use [Circom](https://iden3.io/circom) based
|
||||
version through ark-circom, as opposed to the native one that currently exists
|
||||
in Rust.
|
||||
## Overview
|
||||
|
||||
## Acknowledgements
|
||||
Zerokit provides zero-knowledge cryptographic primitives with a focus on performance, security, and usability.
|
||||
The current focus is on Rate-Limiting Nullifier [RLN](https://github.com/Rate-Limiting-Nullifier) implementation.
|
||||
|
||||
- Uses [ark-circom](https://github.com/gakonst/ark-circom), Rust wrapper around Circom.
|
||||
Current implementation is based on the following
|
||||
[specification](https://rfc.vac.dev/vac/raw/rln-v2)
|
||||
and focused on RLNv2 which allows to set a rate limit for the number of messages that can be sent by a user.
|
||||
|
||||
- Inspired by Applied ZKP group work, e.g. [zk-kit](https://github.com/appliedzkp/zk-kit).
|
||||
## Features
|
||||
|
||||
- [RLN library](https://github.com/kilic/rln) written in Rust based on Bellman.
|
||||
- **RLN Implementation**: Efficient Rate-Limiting Nullifier using zkSNARK
|
||||
- **Circom Compatibility**: Uses Circom-based circuits for RLN
|
||||
- **Cross-Platform**: Support for multiple architectures with cross-compilation
|
||||
- **FFI-Friendly**: Easy to integrate with other languages
|
||||
- **WASM Support**: Can be compiled to WebAssembly for web applications
|
||||
|
||||
- [semaphore-rs](https://github.com/worldcoin/semaphore-rs) written in Rust based on ark-circom.
|
||||
## Architecture
|
||||
|
||||
## Users
|
||||
|
||||
Zerokit is used by -
|
||||
|
||||
- [nwaku](https://github.com/waku-org/nwaku)
|
||||
- [js-rln](https://github.com/waku-org/js-rln)
|
||||
Zerokit currently focuses on RLN (Rate-Limiting Nullifier) implementation using [Circom](https://iden3.io/circom)
|
||||
circuits through ark-circom, providing an alternative to existing native Rust implementations.
|
||||
|
||||
## Build and Test
|
||||
|
||||
To install missing dependencies, run the following commands from the root folder
|
||||
### Install Dependencies
|
||||
|
||||
```bash
|
||||
make installdeps
|
||||
```
|
||||
|
||||
To build and test all crates, run the following commands from the root folder
|
||||
#### Use Nix to install dependencies
|
||||
|
||||
```bash
|
||||
nix develop
|
||||
```
|
||||
|
||||
### Build and Test All Crates
|
||||
|
||||
```bash
|
||||
make build
|
||||
make test
|
||||
```
|
||||
|
||||
## Release assets
|
||||
## Release Assets
|
||||
|
||||
We use [`cross-rs`](https://github.com/cross-rs/cross) to cross-compile and generate release assets for rln.
|
||||
We use [`cross-rs`](https://github.com/cross-rs/cross) to cross-compile and generate release assets:
|
||||
|
||||
```bash
|
||||
# Example: Build for specific target
|
||||
cross build --target x86_64-unknown-linux-gnu --release -p rln
|
||||
```
|
||||
|
||||
## Used By
|
||||
|
||||
Zerokit powers zero-knowledge functionality in:
|
||||
|
||||
- [**nwaku**](https://github.com/waku-org/nwaku) - Nim implementation of the Waku v2 protocol
|
||||
- [**js-rln**](https://github.com/waku-org/js-rln) - JavaScript bindings for RLN
|
||||
|
||||
## Acknowledgements
|
||||
|
||||
- Inspired by [Applied ZKP](https://zkp.science/) group work, including [zk-kit](https://github.com/appliedzkp/zk-kit)
|
||||
- Uses [ark-circom](https://github.com/gakonst/ark-circom) for zkey and Groth16 proof generation
|
||||
- Witness calculation based on [circom-witnesscalc](https://github.com/iden3/circom-witnesscalc) by iden3.
|
||||
The execution graph file used by this code has been generated by means of the same iden3 software.
|
||||
|
||||
> [!IMPORTANT]
|
||||
> The circom-witnesscalc code fragments have been borrowed instead of depending on this crate,
|
||||
> because its types of input and output data were incompatible with the corresponding zerokit code fragments,
|
||||
> and circom-witnesscalc has some dependencies, which are redundant for our purpose.
|
||||
|
||||
## Documentation
|
||||
|
||||
For detailed documentation on each module:
|
||||
|
||||
```bash
|
||||
cargo doc --open
|
||||
```
|
||||
|
||||
48
flake.lock
generated
Normal file
48
flake.lock
generated
Normal file
@@ -0,0 +1,48 @@
|
||||
{
|
||||
"nodes": {
|
||||
"nixpkgs": {
|
||||
"locked": {
|
||||
"lastModified": 1757590060,
|
||||
"narHash": "sha256-EWwwdKLMZALkgHFyKW7rmyhxECO74+N+ZO5xTDnY/5c=",
|
||||
"owner": "NixOS",
|
||||
"repo": "nixpkgs",
|
||||
"rev": "0ef228213045d2cdb5a169a95d63ded38670b293",
|
||||
"type": "github"
|
||||
},
|
||||
"original": {
|
||||
"owner": "NixOS",
|
||||
"repo": "nixpkgs",
|
||||
"rev": "0ef228213045d2cdb5a169a95d63ded38670b293",
|
||||
"type": "github"
|
||||
}
|
||||
},
|
||||
"root": {
|
||||
"inputs": {
|
||||
"nixpkgs": "nixpkgs",
|
||||
"rust-overlay": "rust-overlay"
|
||||
}
|
||||
},
|
||||
"rust-overlay": {
|
||||
"inputs": {
|
||||
"nixpkgs": [
|
||||
"nixpkgs"
|
||||
]
|
||||
},
|
||||
"locked": {
|
||||
"lastModified": 1748399823,
|
||||
"narHash": "sha256-kahD8D5hOXOsGbNdoLLnqCL887cjHkx98Izc37nDjlA=",
|
||||
"owner": "oxalica",
|
||||
"repo": "rust-overlay",
|
||||
"rev": "d68a69dc71bc19beb3479800392112c2f6218159",
|
||||
"type": "github"
|
||||
},
|
||||
"original": {
|
||||
"owner": "oxalica",
|
||||
"repo": "rust-overlay",
|
||||
"type": "github"
|
||||
}
|
||||
}
|
||||
},
|
||||
"root": "root",
|
||||
"version": 7
|
||||
}
|
||||
83
flake.nix
Normal file
83
flake.nix
Normal file
@@ -0,0 +1,83 @@
|
||||
{
|
||||
description = "A flake for building zerokit";
|
||||
|
||||
nixConfig = {
|
||||
extra-substituters = [ "https://nix-cache.status.im/" ];
|
||||
extra-trusted-public-keys = [ "nix-cache.status.im-1:x/93lOfLU+duPplwMSBR+OlY4+mo+dCN7n0mr4oPwgY=" ];
|
||||
};
|
||||
|
||||
inputs = {
|
||||
# Version 24.11
|
||||
nixpkgs.url = "github:NixOS/nixpkgs?rev=0ef228213045d2cdb5a169a95d63ded38670b293";
|
||||
rust-overlay = {
|
||||
url = "github:oxalica/rust-overlay";
|
||||
inputs.nixpkgs.follows = "nixpkgs";
|
||||
};
|
||||
};
|
||||
|
||||
outputs = { self, nixpkgs, rust-overlay }:
|
||||
let
|
||||
stableSystems = [
|
||||
"x86_64-linux" "aarch64-linux"
|
||||
"x86_64-darwin" "aarch64-darwin"
|
||||
"x86_64-windows" "i686-linux"
|
||||
"i686-windows"
|
||||
];
|
||||
forAllSystems = nixpkgs.lib.genAttrs stableSystems;
|
||||
|
||||
pkgsFor = forAllSystems (
|
||||
system: import nixpkgs {
|
||||
inherit system;
|
||||
config = {
|
||||
android_sdk.accept_license = true;
|
||||
allowUnfree = true;
|
||||
};
|
||||
overlays = [
|
||||
(import rust-overlay)
|
||||
(f: p: { inherit rust-overlay; })
|
||||
];
|
||||
}
|
||||
);
|
||||
in rec
|
||||
{
|
||||
packages = forAllSystems (system: let
|
||||
pkgs = pkgsFor.${system};
|
||||
buildPackage = pkgs.callPackage ./nix/default.nix;
|
||||
buildRln = (buildPackage { src = self; project = "rln"; }).override;
|
||||
in rec {
|
||||
rln = buildRln { };
|
||||
|
||||
rln-linux-arm64 = buildRln {
|
||||
target-platform = "aarch64-multiplatform";
|
||||
rust-target = "aarch64-unknown-linux-gnu";
|
||||
};
|
||||
|
||||
rln-android-arm64 = buildRln {
|
||||
target-platform = "aarch64-android-prebuilt";
|
||||
rust-target = "aarch64-linux-android";
|
||||
};
|
||||
|
||||
rln-ios-arm64 = buildRln {
|
||||
target-platform = "aarch64-darwin";
|
||||
rust-target = "aarch64-apple-ios";
|
||||
};
|
||||
|
||||
# TODO: Remove legacy name for RLN android library
|
||||
zerokit-android-arm64 = rln-android-arm64;
|
||||
|
||||
default = rln;
|
||||
});
|
||||
|
||||
devShells = forAllSystems (system: let
|
||||
pkgs = pkgsFor.${system};
|
||||
in {
|
||||
default = pkgs.mkShell {
|
||||
buildInputs = with pkgs; [
|
||||
git cmake cargo-make rustup
|
||||
binaryen ninja gnuplot
|
||||
rust-bin.stable.latest.default
|
||||
];
|
||||
};
|
||||
});
|
||||
};
|
||||
}
|
||||
64
nix/default.nix
Normal file
64
nix/default.nix
Normal file
@@ -0,0 +1,64 @@
|
||||
{
|
||||
pkgs,
|
||||
rust-overlay,
|
||||
project,
|
||||
src ? ../.,
|
||||
release ? true,
|
||||
target-platform ? null,
|
||||
rust-target ? null,
|
||||
features ? null,
|
||||
}:
|
||||
|
||||
let
|
||||
# Use cross-compilation if target-platform is specified.
|
||||
targetPlatformPkgs = if target-platform != null
|
||||
then pkgs.pkgsCross.${target-platform}
|
||||
else pkgs;
|
||||
|
||||
rust-bin = rust-overlay.lib.mkRustBin { } targetPlatformPkgs.buildPackages;
|
||||
|
||||
# Use Rust and Cargo versions from rust-overlay.
|
||||
rustPlatform = targetPlatformPkgs.makeRustPlatform {
|
||||
cargo = rust-bin.stable.latest.minimal;
|
||||
rustc = rust-bin.stable.latest.minimal;
|
||||
};
|
||||
in rustPlatform.buildRustPackage {
|
||||
pname = "zerokit";
|
||||
version = if src ? rev then src.rev else "nightly";
|
||||
|
||||
# Improve caching of sources
|
||||
src = builtins.path { path = src; name = "zerokit"; };
|
||||
|
||||
cargoLock = {
|
||||
lockFile = src + "/Cargo.lock";
|
||||
allowBuiltinFetchGit = true;
|
||||
};
|
||||
|
||||
nativeBuildInputs = [ pkgs.rust-cbindgen ];
|
||||
|
||||
doCheck = false;
|
||||
|
||||
CARGO_HOME = "/tmp";
|
||||
|
||||
buildPhase = ''
|
||||
cargo build --lib \
|
||||
${if release then "--release" else ""} \
|
||||
${if rust-target != null then "--target=${rust-target}" else ""} \
|
||||
${if features != null then "--features=${features}" else ""} \
|
||||
--manifest-path ${project}/Cargo.toml
|
||||
'';
|
||||
|
||||
installPhase = ''
|
||||
set -eu
|
||||
mkdir -p $out/lib
|
||||
find target -type f -name 'librln.*' -not -path '*/deps/*' -exec cp -v '{}' "$out/lib/" \;
|
||||
mkdir -p $out/include
|
||||
cbindgen ${src}/rln -l c > "$out/include/rln.h"
|
||||
'';
|
||||
|
||||
|
||||
meta = with pkgs.lib; {
|
||||
description = "Zerokit";
|
||||
license = licenses.mit;
|
||||
};
|
||||
}
|
||||
20
rln-cli/.gitignore
vendored
Normal file
20
rln-cli/.gitignore
vendored
Normal file
@@ -0,0 +1,20 @@
|
||||
# Common files to ignore in Rust projects
|
||||
.DS_Store
|
||||
.idea
|
||||
*.log
|
||||
tmp/
|
||||
|
||||
# Generated by Cargo will have compiled files and executables
|
||||
/target
|
||||
|
||||
# Generated by rln-cli
|
||||
/database
|
||||
|
||||
# Generated by Nix
|
||||
result
|
||||
|
||||
# These are backup files generated by rustfmt
|
||||
**/*.rs.bk
|
||||
|
||||
# MSVC Windows builds of rustc generate these, which store debugging information
|
||||
*.pdb
|
||||
1647
rln-cli/Cargo.lock
generated
Normal file
1647
rln-cli/Cargo.lock
generated
Normal file
File diff suppressed because it is too large
Load Diff
@@ -1,13 +1,27 @@
|
||||
[package]
|
||||
name = "rln-cli"
|
||||
version = "0.3.0"
|
||||
version = "0.5.0"
|
||||
edition = "2021"
|
||||
|
||||
[[example]]
|
||||
name = "relay"
|
||||
path = "src/examples/relay.rs"
|
||||
|
||||
[[example]]
|
||||
name = "stateless"
|
||||
path = "src/examples/stateless.rs"
|
||||
required-features = ["stateless"]
|
||||
|
||||
[dependencies]
|
||||
rln = { path = "../rln", default-features = true, features = ["arkzkey"] }
|
||||
clap = { version = "4.2.7", features = ["cargo", "derive", "env"]}
|
||||
clap_derive = { version = "=4.2.0" }
|
||||
color-eyre = "=0.6.2"
|
||||
# serialization
|
||||
serde_json = "1.0.48"
|
||||
serde = { version = "1.0.130", features = ["derive"] }
|
||||
rln = { path = "../rln", version = "1.0.0", default-features = false }
|
||||
zerokit_utils = { path = "../utils", version = "1.0.0", default-features = false }
|
||||
clap = { version = "4.5.53", features = ["cargo", "derive", "env"] }
|
||||
serde_json = "1.0.145"
|
||||
serde = { version = "1.0.228", features = ["derive"] }
|
||||
|
||||
[features]
|
||||
default = ["rln/pmtree-ft", "rln/parallel"]
|
||||
stateless = ["rln/stateless", "rln/parallel"]
|
||||
|
||||
[package.metadata.docs.rs]
|
||||
all-features = true
|
||||
|
||||
9
rln-cli/Makefile.toml
Normal file
9
rln-cli/Makefile.toml
Normal file
@@ -0,0 +1,9 @@
|
||||
[tasks.build]
|
||||
command = "cargo"
|
||||
args = ["build"]
|
||||
|
||||
[tasks.test]
|
||||
disabled = true
|
||||
|
||||
[tasks.bench]
|
||||
disabled = true
|
||||
29
rln-cli/README.md
Normal file
29
rln-cli/README.md
Normal file
@@ -0,0 +1,29 @@
|
||||
# Zerokit RLN-CLI
|
||||
|
||||
The Zerokit RLN-CLI provides a command-line interface examples on how to use public API of the [Zerokit RLN Module](../rln/README.md).
|
||||
|
||||
## Relay Example
|
||||
|
||||
The following [Relay Example](src/examples/relay.rs) demonstrates how RLN enables spam prevention in anonymous environments for multple users.
|
||||
|
||||
You can run the example using the following command:
|
||||
|
||||
```bash
|
||||
cargo run --example relay
|
||||
```
|
||||
|
||||
You can also change **MESSAGE_LIMIT** and **TREE_DEPTH** in the [relay.rs](src/examples/relay.rs) file to see how the RLN instance behaves with different parameters.
|
||||
|
||||
The customize **TREE_DEPTH** constant differs from the default value of `20` should follow [Custom Circuit Compilation](../rln/README.md#advanced-custom-circuit-compilation) instructions.
|
||||
|
||||
## Stateless Example
|
||||
|
||||
The following [Stateless Example](src/examples/stateless.rs) demonstrates how RLN can be used for stateless features by creating the Merkle tree outside of RLN instance.
|
||||
|
||||
This example function similarly to the [Relay Example](#relay-example) but uses a stateless RLN and seperate Merkle tree.
|
||||
|
||||
You can run the example using the following command:
|
||||
|
||||
```bash
|
||||
cargo run --example stateless --no-default-features --features stateless
|
||||
```
|
||||
@@ -1,67 +0,0 @@
|
||||
use std::path::PathBuf;
|
||||
|
||||
use clap::Subcommand;
|
||||
|
||||
#[derive(Subcommand)]
|
||||
pub(crate) enum Commands {
|
||||
New {
|
||||
tree_height: usize,
|
||||
/// Sets a custom config file
|
||||
#[arg(short, long)]
|
||||
config: PathBuf,
|
||||
},
|
||||
NewWithParams {
|
||||
tree_height: usize,
|
||||
/// Sets a custom config file
|
||||
#[arg(short, long)]
|
||||
config: PathBuf,
|
||||
#[arg(short, long)]
|
||||
tree_config_input: PathBuf,
|
||||
},
|
||||
SetTree {
|
||||
tree_height: usize,
|
||||
},
|
||||
SetLeaf {
|
||||
index: usize,
|
||||
#[arg(short, long)]
|
||||
file: PathBuf,
|
||||
},
|
||||
SetMultipleLeaves {
|
||||
index: usize,
|
||||
#[arg(short, long)]
|
||||
file: PathBuf,
|
||||
},
|
||||
ResetMultipleLeaves {
|
||||
#[arg(short, long)]
|
||||
file: PathBuf,
|
||||
},
|
||||
SetNextLeaf {
|
||||
#[arg(short, long)]
|
||||
file: PathBuf,
|
||||
},
|
||||
DeleteLeaf {
|
||||
index: usize,
|
||||
},
|
||||
GetRoot,
|
||||
GetProof {
|
||||
index: usize,
|
||||
},
|
||||
Prove {
|
||||
#[arg(short, long)]
|
||||
input: PathBuf,
|
||||
},
|
||||
Verify {
|
||||
#[arg(short, long)]
|
||||
file: PathBuf,
|
||||
},
|
||||
GenerateProof {
|
||||
#[arg(short, long)]
|
||||
input: PathBuf,
|
||||
},
|
||||
VerifyWithRoots {
|
||||
#[arg(short, long)]
|
||||
input: PathBuf,
|
||||
#[arg(short, long)]
|
||||
roots: PathBuf,
|
||||
},
|
||||
}
|
||||
@@ -1,29 +0,0 @@
|
||||
use color_eyre::Result;
|
||||
use serde::{Deserialize, Serialize};
|
||||
use std::{fs::File, io::Read, path::PathBuf};
|
||||
|
||||
pub const RLN_STATE_PATH: &str = "RLN_STATE_PATH";
|
||||
|
||||
#[derive(Default, Serialize, Deserialize)]
|
||||
pub(crate) struct Config {
|
||||
pub inner: Option<InnerConfig>,
|
||||
}
|
||||
|
||||
#[derive(Default, Serialize, Deserialize)]
|
||||
pub(crate) struct InnerConfig {
|
||||
pub file: PathBuf,
|
||||
pub tree_height: usize,
|
||||
}
|
||||
|
||||
impl Config {
|
||||
pub(crate) fn load_config() -> Result<Config> {
|
||||
let path = PathBuf::from(std::env::var(RLN_STATE_PATH)?);
|
||||
|
||||
let mut file = File::open(path)?;
|
||||
|
||||
let mut contents = String::new();
|
||||
file.read_to_string(&mut contents)?;
|
||||
let state: Config = serde_json::from_str(&contents)?;
|
||||
Ok(state)
|
||||
}
|
||||
}
|
||||
302
rln-cli/src/examples/relay.rs
Normal file
302
rln-cli/src/examples/relay.rs
Normal file
@@ -0,0 +1,302 @@
|
||||
use std::{
|
||||
collections::HashMap,
|
||||
fs::File,
|
||||
io::{stdin, stdout, Read, Write},
|
||||
path::{Path, PathBuf},
|
||||
};
|
||||
|
||||
use clap::{Parser, Subcommand};
|
||||
use rln::prelude::{
|
||||
hash_to_field_le, keygen, poseidon_hash, recover_id_secret, Fr, IdSecret, PmtreeConfigBuilder,
|
||||
RLNProofValues, RLNWitnessInput, RLN,
|
||||
};
|
||||
use zerokit_utils::pm_tree::Mode;
|
||||
|
||||
const MESSAGE_LIMIT: u32 = 1;
|
||||
|
||||
const TREE_DEPTH: usize = 20;
|
||||
|
||||
type Result<T> = std::result::Result<T, Box<dyn std::error::Error>>;
|
||||
|
||||
#[derive(Parser)]
|
||||
#[command(author, version, about, long_about = None)]
|
||||
struct Cli {
|
||||
#[command(subcommand)]
|
||||
command: Commands,
|
||||
}
|
||||
|
||||
#[derive(Subcommand)]
|
||||
enum Commands {
|
||||
List,
|
||||
Register,
|
||||
Send {
|
||||
#[arg(short, long)]
|
||||
user_index: usize,
|
||||
#[arg(short, long)]
|
||||
message_id: u32,
|
||||
#[arg(short, long)]
|
||||
signal: String,
|
||||
},
|
||||
Clear,
|
||||
Exit,
|
||||
}
|
||||
|
||||
#[derive(Debug, Clone)]
|
||||
struct Identity {
|
||||
identity_secret: IdSecret,
|
||||
id_commitment: Fr,
|
||||
}
|
||||
|
||||
impl Identity {
|
||||
fn new() -> Self {
|
||||
let (identity_secret, id_commitment) = keygen().unwrap();
|
||||
Identity {
|
||||
identity_secret,
|
||||
id_commitment,
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
struct RLNSystem {
|
||||
rln: RLN,
|
||||
used_nullifiers: HashMap<Fr, RLNProofValues>,
|
||||
local_identities: HashMap<usize, Identity>,
|
||||
}
|
||||
|
||||
impl RLNSystem {
|
||||
fn new() -> Result<Self> {
|
||||
let mut resources: Vec<Vec<u8>> = Vec::new();
|
||||
let resources_path: PathBuf = format!("../rln/resources/tree_depth_{TREE_DEPTH}").into();
|
||||
let filenames = ["rln_final.arkzkey", "graph.bin"];
|
||||
for filename in filenames {
|
||||
let fullpath = resources_path.join(Path::new(filename));
|
||||
let mut file = File::open(&fullpath)?;
|
||||
let metadata = std::fs::metadata(&fullpath)?;
|
||||
let mut output_buffer = vec![0; metadata.len() as usize];
|
||||
file.read_exact(&mut output_buffer)?;
|
||||
resources.push(output_buffer);
|
||||
}
|
||||
let tree_config = PmtreeConfigBuilder::new()
|
||||
.path("./database")
|
||||
.temporary(false)
|
||||
.cache_capacity(1073741824)
|
||||
.flush_every_ms(500)
|
||||
.mode(Mode::HighThroughput)
|
||||
.use_compression(false)
|
||||
.build()?;
|
||||
let rln = RLN::new_with_params(
|
||||
TREE_DEPTH,
|
||||
resources[0].clone(),
|
||||
resources[1].clone(),
|
||||
tree_config,
|
||||
)?;
|
||||
println!("RLN instance initialized successfully");
|
||||
Ok(RLNSystem {
|
||||
rln,
|
||||
used_nullifiers: HashMap::new(),
|
||||
local_identities: HashMap::new(),
|
||||
})
|
||||
}
|
||||
|
||||
fn list_users(&self) {
|
||||
if self.local_identities.is_empty() {
|
||||
println!("No users registered yet.");
|
||||
return;
|
||||
}
|
||||
|
||||
println!("Registered users:");
|
||||
for (index, identity) in &self.local_identities {
|
||||
println!("User Index: {index}");
|
||||
println!("+ Identity secret: {}", *identity.identity_secret);
|
||||
println!("+ Identity commitment: {}", identity.id_commitment);
|
||||
println!();
|
||||
}
|
||||
}
|
||||
|
||||
fn register_user(&mut self) -> Result<usize> {
|
||||
let index = self.rln.leaves_set();
|
||||
let identity = Identity::new();
|
||||
|
||||
let rate_commitment =
|
||||
poseidon_hash(&[identity.id_commitment, Fr::from(MESSAGE_LIMIT)]).unwrap();
|
||||
match self.rln.set_next_leaf(rate_commitment) {
|
||||
Ok(_) => {
|
||||
println!("Registered User Index: {index}");
|
||||
println!("+ Identity secret: {}", *identity.identity_secret);
|
||||
println!("+ Identity commitment: {}", identity.id_commitment);
|
||||
self.local_identities.insert(index, identity);
|
||||
}
|
||||
Err(_) => {
|
||||
println!("Maximum user limit reached: 2^{TREE_DEPTH}");
|
||||
}
|
||||
};
|
||||
|
||||
Ok(index)
|
||||
}
|
||||
|
||||
fn generate_and_verify_proof(
|
||||
&mut self,
|
||||
user_index: usize,
|
||||
message_id: u32,
|
||||
signal: &str,
|
||||
external_nullifier: Fr,
|
||||
) -> Result<RLNProofValues> {
|
||||
let identity = match self.local_identities.get(&user_index) {
|
||||
Some(identity) => identity,
|
||||
None => return Err(format!("user index {user_index} not found").into()),
|
||||
};
|
||||
|
||||
let (path_elements, identity_path_index) = self.rln.get_merkle_proof(user_index)?;
|
||||
let x = hash_to_field_le(signal.as_bytes())?;
|
||||
|
||||
let witness = RLNWitnessInput::new(
|
||||
identity.identity_secret.clone(),
|
||||
Fr::from(MESSAGE_LIMIT),
|
||||
Fr::from(message_id),
|
||||
path_elements,
|
||||
identity_path_index,
|
||||
x,
|
||||
external_nullifier,
|
||||
)?;
|
||||
|
||||
let (proof, proof_values) = self.rln.generate_rln_proof(&witness)?;
|
||||
|
||||
println!("Proof generated successfully:");
|
||||
println!("+ User Index: {user_index}");
|
||||
println!("+ Message ID: {message_id}");
|
||||
println!("+ Signal: {signal}");
|
||||
|
||||
let verified = self.rln.verify_rln_proof(&proof, &proof_values, &x)?;
|
||||
if verified {
|
||||
println!("Proof verified successfully");
|
||||
}
|
||||
|
||||
Ok(proof_values)
|
||||
}
|
||||
|
||||
fn check_nullifier(&mut self, proof_values: RLNProofValues) -> Result<()> {
|
||||
if let Some(&previous_proof_values) = self.used_nullifiers.get(&proof_values.nullifier) {
|
||||
self.handle_duplicate_message_id(previous_proof_values, proof_values)?;
|
||||
return Ok(());
|
||||
}
|
||||
|
||||
self.used_nullifiers
|
||||
.insert(proof_values.nullifier, proof_values);
|
||||
println!("Message verified and accepted");
|
||||
Ok(())
|
||||
}
|
||||
|
||||
fn handle_duplicate_message_id(
|
||||
&mut self,
|
||||
previous_proof_values: RLNProofValues,
|
||||
current_proof_values: RLNProofValues,
|
||||
) -> Result<()> {
|
||||
if previous_proof_values.x == current_proof_values.x
|
||||
&& previous_proof_values.y == current_proof_values.y
|
||||
{
|
||||
return Err("this exact message and signal has already been sent".into());
|
||||
}
|
||||
|
||||
match recover_id_secret(&previous_proof_values, ¤t_proof_values) {
|
||||
Ok(leaked_identity_secret) => {
|
||||
if let Some((user_index, identity)) = self
|
||||
.local_identities
|
||||
.iter()
|
||||
.find(|(_, identity)| identity.identity_secret == leaked_identity_secret)
|
||||
.map(|(index, identity)| (*index, identity))
|
||||
{
|
||||
let real_identity_secret = identity.identity_secret.clone();
|
||||
if leaked_identity_secret != real_identity_secret {
|
||||
Err("Identity secret mismatch: leaked_identity_secret != real_identity_secret".into())
|
||||
} else {
|
||||
println!(
|
||||
"DUPLICATE message ID detected! Reveal identity secret: {}",
|
||||
*leaked_identity_secret
|
||||
);
|
||||
self.local_identities.remove(&user_index);
|
||||
self.rln.delete_leaf(user_index)?;
|
||||
println!("User index {user_index} has been SLASHED");
|
||||
Ok(())
|
||||
}
|
||||
} else {
|
||||
Err("user identity secret ******** not found".into())
|
||||
}
|
||||
}
|
||||
Err(err) => Err(format!("Failed to recover identity secret: {err}").into()),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
fn main() -> Result<()> {
|
||||
println!("Initializing RLN instance...");
|
||||
print!("\x1B[2J\x1B[1;1H");
|
||||
let mut rln_system = RLNSystem::new()?;
|
||||
let rln_epoch = hash_to_field_le(b"epoch")?;
|
||||
let rln_identifier = hash_to_field_le(b"rln-identifier")?;
|
||||
let external_nullifier = poseidon_hash(&[rln_epoch, rln_identifier]).unwrap();
|
||||
println!("RLN Relay Example:");
|
||||
println!("Message Limit: {MESSAGE_LIMIT}");
|
||||
println!("----------------------------------");
|
||||
println!();
|
||||
show_commands();
|
||||
loop {
|
||||
print!("\n> ");
|
||||
stdout().flush()?;
|
||||
let mut input = String::new();
|
||||
stdin().read_line(&mut input)?;
|
||||
let trimmed = input.trim();
|
||||
let args = std::iter::once("").chain(trimmed.split_whitespace());
|
||||
|
||||
match Cli::try_parse_from(args) {
|
||||
Ok(cli) => match cli.command {
|
||||
Commands::List => {
|
||||
rln_system.list_users();
|
||||
}
|
||||
Commands::Register => {
|
||||
rln_system.register_user()?;
|
||||
}
|
||||
Commands::Send {
|
||||
user_index,
|
||||
message_id,
|
||||
signal,
|
||||
} => {
|
||||
match rln_system.generate_and_verify_proof(
|
||||
user_index,
|
||||
message_id,
|
||||
&signal,
|
||||
external_nullifier,
|
||||
) {
|
||||
Ok(proof_values) => {
|
||||
if let Err(err) = rln_system.check_nullifier(proof_values) {
|
||||
println!("Check nullifier error: {err}");
|
||||
};
|
||||
}
|
||||
Err(err) => {
|
||||
println!("Proof generation error: {err}");
|
||||
}
|
||||
}
|
||||
}
|
||||
Commands::Clear => {
|
||||
print!("\x1B[2J\x1B[1;1H");
|
||||
show_commands();
|
||||
}
|
||||
Commands::Exit => {
|
||||
break;
|
||||
}
|
||||
},
|
||||
Err(err) => {
|
||||
eprintln!("Command error: {err}");
|
||||
}
|
||||
}
|
||||
}
|
||||
Ok(())
|
||||
}
|
||||
|
||||
fn show_commands() {
|
||||
println!("Available commands:");
|
||||
println!(" list - List registered users");
|
||||
println!(" register - Register a new user index");
|
||||
println!(" send -u <index> -m <message_id> -s <signal> - Send a message with proof");
|
||||
println!(" clear - Clear the screen");
|
||||
println!(" exit - Exit the program");
|
||||
}
|
||||
292
rln-cli/src/examples/stateless.rs
Normal file
292
rln-cli/src/examples/stateless.rs
Normal file
@@ -0,0 +1,292 @@
|
||||
#![cfg(feature = "stateless")]
|
||||
|
||||
use std::{
|
||||
collections::HashMap,
|
||||
io::{stdin, stdout, Write},
|
||||
};
|
||||
|
||||
use clap::{Parser, Subcommand};
|
||||
use rln::prelude::{
|
||||
hash_to_field_le, keygen, poseidon_hash, recover_id_secret, Fr, IdSecret, OptimalMerkleTree,
|
||||
PoseidonHash, RLNProofValues, RLNWitnessInput, ZerokitMerkleProof, ZerokitMerkleTree,
|
||||
DEFAULT_TREE_DEPTH, RLN,
|
||||
};
|
||||
|
||||
const MESSAGE_LIMIT: u32 = 1;
|
||||
|
||||
type Result<T> = std::result::Result<T, Box<dyn std::error::Error>>;
|
||||
type ConfigOf<T> = <T as ZerokitMerkleTree>::Config;
|
||||
|
||||
#[derive(Parser)]
|
||||
#[command(author, version, about, long_about = None)]
|
||||
struct Cli {
|
||||
#[command(subcommand)]
|
||||
command: Commands,
|
||||
}
|
||||
|
||||
#[derive(Subcommand)]
|
||||
enum Commands {
|
||||
List,
|
||||
Register,
|
||||
Send {
|
||||
#[arg(short, long)]
|
||||
user_index: usize,
|
||||
#[arg(short, long)]
|
||||
message_id: u32,
|
||||
#[arg(short, long)]
|
||||
signal: String,
|
||||
},
|
||||
Clear,
|
||||
Exit,
|
||||
}
|
||||
|
||||
#[derive(Debug, Clone)]
|
||||
struct Identity {
|
||||
identity_secret: IdSecret,
|
||||
id_commitment: Fr,
|
||||
}
|
||||
|
||||
impl Identity {
|
||||
fn new() -> Self {
|
||||
let (identity_secret, id_commitment) = keygen().unwrap();
|
||||
Identity {
|
||||
identity_secret,
|
||||
id_commitment,
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
struct RLNSystem {
|
||||
rln: RLN,
|
||||
tree: OptimalMerkleTree<PoseidonHash>,
|
||||
used_nullifiers: HashMap<Fr, RLNProofValues>,
|
||||
local_identities: HashMap<usize, Identity>,
|
||||
}
|
||||
|
||||
impl RLNSystem {
|
||||
fn new() -> Result<Self> {
|
||||
let rln = RLN::new()?;
|
||||
let default_leaf = Fr::from(0);
|
||||
let tree: OptimalMerkleTree<PoseidonHash> = OptimalMerkleTree::new(
|
||||
DEFAULT_TREE_DEPTH,
|
||||
default_leaf,
|
||||
ConfigOf::<OptimalMerkleTree<PoseidonHash>>::default(),
|
||||
)
|
||||
.unwrap();
|
||||
|
||||
Ok(RLNSystem {
|
||||
rln,
|
||||
tree,
|
||||
used_nullifiers: HashMap::new(),
|
||||
local_identities: HashMap::new(),
|
||||
})
|
||||
}
|
||||
|
||||
fn list_users(&self) {
|
||||
if self.local_identities.is_empty() {
|
||||
println!("No users registered yet.");
|
||||
return;
|
||||
}
|
||||
|
||||
println!("Registered users:");
|
||||
for (index, identity) in &self.local_identities {
|
||||
println!("User Index: {index}");
|
||||
println!("+ Identity secret: {}", *identity.identity_secret);
|
||||
println!("+ Identity commitment: {}", identity.id_commitment);
|
||||
println!();
|
||||
}
|
||||
}
|
||||
|
||||
fn register_user(&mut self) -> Result<usize> {
|
||||
let index = self.tree.leaves_set();
|
||||
let identity = Identity::new();
|
||||
|
||||
let rate_commitment =
|
||||
poseidon_hash(&[identity.id_commitment, Fr::from(MESSAGE_LIMIT)]).unwrap();
|
||||
self.tree.update_next(rate_commitment)?;
|
||||
|
||||
println!("Registered User Index: {index}");
|
||||
println!("+ Identity secret: {}", *identity.identity_secret);
|
||||
println!("+ Identity commitment: {}", identity.id_commitment);
|
||||
|
||||
self.local_identities.insert(index, identity);
|
||||
Ok(index)
|
||||
}
|
||||
|
||||
fn generate_and_verify_proof(
|
||||
&mut self,
|
||||
user_index: usize,
|
||||
message_id: u32,
|
||||
signal: &str,
|
||||
external_nullifier: Fr,
|
||||
) -> Result<RLNProofValues> {
|
||||
let identity = match self.local_identities.get(&user_index) {
|
||||
Some(identity) => identity,
|
||||
None => return Err(format!("user index {user_index} not found").into()),
|
||||
};
|
||||
|
||||
let merkle_proof = self.tree.proof(user_index)?;
|
||||
let x = hash_to_field_le(signal.as_bytes())?;
|
||||
|
||||
let witness = RLNWitnessInput::new(
|
||||
identity.identity_secret.clone(),
|
||||
Fr::from(MESSAGE_LIMIT),
|
||||
Fr::from(message_id),
|
||||
merkle_proof.get_path_elements(),
|
||||
merkle_proof.get_path_index(),
|
||||
x,
|
||||
external_nullifier,
|
||||
)?;
|
||||
|
||||
let (proof, proof_values) = self.rln.generate_rln_proof(&witness)?;
|
||||
|
||||
println!("Proof generated successfully:");
|
||||
println!("+ User Index: {user_index}");
|
||||
println!("+ Message ID: {message_id}");
|
||||
println!("+ Signal: {signal}");
|
||||
|
||||
let tree_root = self.tree.root();
|
||||
|
||||
let verified = self
|
||||
.rln
|
||||
.verify_with_roots(&proof, &proof_values, &x, &[tree_root])?;
|
||||
if verified {
|
||||
println!("Proof verified successfully");
|
||||
}
|
||||
|
||||
Ok(proof_values)
|
||||
}
|
||||
|
||||
fn check_nullifier(&mut self, proof_values: RLNProofValues) -> Result<()> {
|
||||
let tree_root = self.tree.root();
|
||||
|
||||
if proof_values.root != tree_root {
|
||||
println!("Check nullifier failed: invalid root");
|
||||
return Ok(());
|
||||
}
|
||||
|
||||
if let Some(&previous_proof_values) = self.used_nullifiers.get(&proof_values.nullifier) {
|
||||
self.handle_duplicate_message_id(previous_proof_values, proof_values)?;
|
||||
return Ok(());
|
||||
}
|
||||
|
||||
self.used_nullifiers
|
||||
.insert(proof_values.nullifier, proof_values);
|
||||
println!("Message verified and accepted");
|
||||
Ok(())
|
||||
}
|
||||
|
||||
fn handle_duplicate_message_id(
|
||||
&mut self,
|
||||
previous_proof_values: RLNProofValues,
|
||||
current_proof_values: RLNProofValues,
|
||||
) -> Result<()> {
|
||||
if previous_proof_values.x == current_proof_values.x
|
||||
&& previous_proof_values.y == current_proof_values.y
|
||||
{
|
||||
return Err("this exact message and signal has already been sent".into());
|
||||
}
|
||||
|
||||
match recover_id_secret(&previous_proof_values, ¤t_proof_values) {
|
||||
Ok(leaked_identity_secret) => {
|
||||
if let Some((user_index, identity)) = self
|
||||
.local_identities
|
||||
.iter()
|
||||
.find(|(_, identity)| identity.identity_secret == leaked_identity_secret)
|
||||
.map(|(index, identity)| (*index, identity))
|
||||
{
|
||||
let real_identity_secret = identity.identity_secret.clone();
|
||||
if leaked_identity_secret != real_identity_secret {
|
||||
Err("Identity secret mismatch: leaked_identity_secret != real_identity_secret".into())
|
||||
} else {
|
||||
println!(
|
||||
"DUPLICATE message ID detected! Reveal identity secret: {}",
|
||||
*leaked_identity_secret
|
||||
);
|
||||
self.local_identities.remove(&user_index);
|
||||
println!("User index {user_index} has been SLASHED");
|
||||
Ok(())
|
||||
}
|
||||
} else {
|
||||
Err("user identity secret ******** not found".into())
|
||||
}
|
||||
}
|
||||
Err(err) => Err(format!("Failed to recover identity secret: {err}").into()),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
fn main() -> Result<()> {
|
||||
println!("Initializing RLN instance...");
|
||||
print!("\x1B[2J\x1B[1;1H");
|
||||
let mut rln_system = RLNSystem::new()?;
|
||||
let rln_epoch = hash_to_field_le(b"epoch")?;
|
||||
let rln_identifier = hash_to_field_le(b"rln-identifier")?;
|
||||
let external_nullifier = poseidon_hash(&[rln_epoch, rln_identifier]).unwrap();
|
||||
println!("RLN Stateless Relay Example:");
|
||||
println!("Message Limit: {MESSAGE_LIMIT}");
|
||||
println!("----------------------------------");
|
||||
println!();
|
||||
show_commands();
|
||||
|
||||
loop {
|
||||
print!("\n> ");
|
||||
stdout().flush()?;
|
||||
let mut input = String::new();
|
||||
stdin().read_line(&mut input)?;
|
||||
let trimmed = input.trim();
|
||||
let args = std::iter::once("").chain(trimmed.split_whitespace());
|
||||
|
||||
match Cli::try_parse_from(args) {
|
||||
Ok(cli) => match cli.command {
|
||||
Commands::List => {
|
||||
rln_system.list_users();
|
||||
}
|
||||
Commands::Register => {
|
||||
rln_system.register_user()?;
|
||||
}
|
||||
Commands::Send {
|
||||
user_index,
|
||||
message_id,
|
||||
signal,
|
||||
} => {
|
||||
match rln_system.generate_and_verify_proof(
|
||||
user_index,
|
||||
message_id,
|
||||
&signal,
|
||||
external_nullifier,
|
||||
) {
|
||||
Ok(proof_values) => {
|
||||
if let Err(err) = rln_system.check_nullifier(proof_values) {
|
||||
println!("Check nullifier error: {err}");
|
||||
};
|
||||
}
|
||||
Err(err) => {
|
||||
println!("Proof generation error: {err}");
|
||||
}
|
||||
}
|
||||
}
|
||||
Commands::Clear => {
|
||||
print!("\x1B[2J\x1B[1;1H");
|
||||
show_commands();
|
||||
}
|
||||
Commands::Exit => {
|
||||
break;
|
||||
}
|
||||
},
|
||||
Err(err) => {
|
||||
eprintln!("Command error: {err}");
|
||||
}
|
||||
}
|
||||
}
|
||||
Ok(())
|
||||
}
|
||||
|
||||
fn show_commands() {
|
||||
println!("Available commands:");
|
||||
println!(" list - List registered users");
|
||||
println!(" register - Register a new user index");
|
||||
println!(" send -u <index> -m <message_id> -s <signal> - Send a message with proof");
|
||||
println!(" clear - Clear the screen");
|
||||
println!(" exit - Exit the program");
|
||||
}
|
||||
@@ -1,161 +0,0 @@
|
||||
use std::{fs::File, io::Read, path::Path};
|
||||
|
||||
use clap::Parser;
|
||||
use color_eyre::{Report, Result};
|
||||
use commands::Commands;
|
||||
use rln::public::RLN;
|
||||
use state::State;
|
||||
|
||||
mod commands;
|
||||
mod config;
|
||||
mod state;
|
||||
|
||||
#[derive(Parser)]
|
||||
#[command(author, version, about, long_about = None)]
|
||||
struct Cli {
|
||||
#[command(subcommand)]
|
||||
command: Option<Commands>,
|
||||
}
|
||||
|
||||
fn main() -> Result<()> {
|
||||
let cli = Cli::parse();
|
||||
|
||||
let mut state = State::load_state()?;
|
||||
|
||||
match &cli.command {
|
||||
Some(Commands::New {
|
||||
tree_height,
|
||||
config,
|
||||
}) => {
|
||||
let resources = File::open(&config)?;
|
||||
state.rln = Some(RLN::new(*tree_height, resources)?);
|
||||
Ok(())
|
||||
}
|
||||
Some(Commands::NewWithParams {
|
||||
tree_height,
|
||||
config,
|
||||
tree_config_input,
|
||||
}) => {
|
||||
let mut resources: Vec<Vec<u8>> = Vec::new();
|
||||
#[cfg(feature = "arkzkey")]
|
||||
let filenames = ["rln.wasm", "rln_final.arkzkey", "verification_key.arkvkey"];
|
||||
#[cfg(not(feature = "arkzkey"))]
|
||||
let filenames = ["rln.wasm", "rln_final.zkey", "verification_key.arkvkey"];
|
||||
for filename in filenames {
|
||||
let fullpath = config.join(Path::new(filename));
|
||||
let mut file = File::open(&fullpath)?;
|
||||
let metadata = std::fs::metadata(&fullpath)?;
|
||||
let mut buffer = vec![0; metadata.len() as usize];
|
||||
file.read_exact(&mut buffer)?;
|
||||
resources.push(buffer);
|
||||
}
|
||||
let tree_config_input_file = File::open(&tree_config_input)?;
|
||||
state.rln = Some(RLN::new_with_params(
|
||||
*tree_height,
|
||||
resources[0].clone(),
|
||||
resources[1].clone(),
|
||||
resources[2].clone(),
|
||||
tree_config_input_file,
|
||||
)?);
|
||||
Ok(())
|
||||
}
|
||||
Some(Commands::SetTree { tree_height }) => {
|
||||
state
|
||||
.rln
|
||||
.ok_or(Report::msg("no RLN instance initialized"))?
|
||||
.set_tree(*tree_height)?;
|
||||
Ok(())
|
||||
}
|
||||
Some(Commands::SetLeaf { index, file }) => {
|
||||
let input_data = File::open(&file)?;
|
||||
state
|
||||
.rln
|
||||
.ok_or(Report::msg("no RLN instance initialized"))?
|
||||
.set_leaf(*index, input_data)?;
|
||||
Ok(())
|
||||
}
|
||||
Some(Commands::SetMultipleLeaves { index, file }) => {
|
||||
let input_data = File::open(&file)?;
|
||||
state
|
||||
.rln
|
||||
.ok_or(Report::msg("no RLN instance initialized"))?
|
||||
.set_leaves_from(*index, input_data)?;
|
||||
Ok(())
|
||||
}
|
||||
Some(Commands::ResetMultipleLeaves { file }) => {
|
||||
let input_data = File::open(&file)?;
|
||||
state
|
||||
.rln
|
||||
.ok_or(Report::msg("no RLN instance initialized"))?
|
||||
.init_tree_with_leaves(input_data)?;
|
||||
Ok(())
|
||||
}
|
||||
Some(Commands::SetNextLeaf { file }) => {
|
||||
let input_data = File::open(&file)?;
|
||||
state
|
||||
.rln
|
||||
.ok_or(Report::msg("no RLN instance initialized"))?
|
||||
.set_next_leaf(input_data)?;
|
||||
Ok(())
|
||||
}
|
||||
Some(Commands::DeleteLeaf { index }) => {
|
||||
state
|
||||
.rln
|
||||
.ok_or(Report::msg("no RLN instance initialized"))?
|
||||
.delete_leaf(*index)?;
|
||||
Ok(())
|
||||
}
|
||||
Some(Commands::GetRoot) => {
|
||||
let writer = std::io::stdout();
|
||||
state
|
||||
.rln
|
||||
.ok_or(Report::msg("no RLN instance initialized"))?
|
||||
.get_root(writer)?;
|
||||
Ok(())
|
||||
}
|
||||
Some(Commands::GetProof { index }) => {
|
||||
let writer = std::io::stdout();
|
||||
state
|
||||
.rln
|
||||
.ok_or(Report::msg("no RLN instance initialized"))?
|
||||
.get_proof(*index, writer)?;
|
||||
Ok(())
|
||||
}
|
||||
Some(Commands::Prove { input }) => {
|
||||
let input_data = File::open(&input)?;
|
||||
let writer = std::io::stdout();
|
||||
state
|
||||
.rln
|
||||
.ok_or(Report::msg("no RLN instance initialized"))?
|
||||
.prove(input_data, writer)?;
|
||||
Ok(())
|
||||
}
|
||||
Some(Commands::Verify { file }) => {
|
||||
let input_data = File::open(&file)?;
|
||||
state
|
||||
.rln
|
||||
.ok_or(Report::msg("no RLN instance initialized"))?
|
||||
.verify(input_data)?;
|
||||
Ok(())
|
||||
}
|
||||
Some(Commands::GenerateProof { input }) => {
|
||||
let input_data = File::open(&input)?;
|
||||
let writer = std::io::stdout();
|
||||
state
|
||||
.rln
|
||||
.ok_or(Report::msg("no RLN instance initialized"))?
|
||||
.generate_rln_proof(input_data, writer)?;
|
||||
Ok(())
|
||||
}
|
||||
Some(Commands::VerifyWithRoots { input, roots }) => {
|
||||
let input_data = File::open(&input)?;
|
||||
let roots_data = File::open(&roots)?;
|
||||
state
|
||||
.rln
|
||||
.ok_or(Report::msg("no RLN instance initialized"))?
|
||||
.verify_with_roots(input_data, roots_data)?;
|
||||
Ok(())
|
||||
}
|
||||
None => Ok(()),
|
||||
}
|
||||
}
|
||||
@@ -1,23 +0,0 @@
|
||||
use color_eyre::Result;
|
||||
use rln::public::RLN;
|
||||
use std::fs::File;
|
||||
|
||||
use crate::config::{Config, InnerConfig};
|
||||
|
||||
#[derive(Default)]
|
||||
pub(crate) struct State {
|
||||
pub rln: Option<RLN>,
|
||||
}
|
||||
|
||||
impl State {
|
||||
pub(crate) fn load_state() -> Result<State> {
|
||||
let config = Config::load_config()?;
|
||||
let rln = if let Some(InnerConfig { file, tree_height }) = config.inner {
|
||||
let resources = File::open(&file)?;
|
||||
Some(RLN::new(tree_height, resources)?)
|
||||
} else {
|
||||
None
|
||||
};
|
||||
Ok(State { rln })
|
||||
}
|
||||
}
|
||||
24
rln-wasm/.gitignore
vendored
24
rln-wasm/.gitignore
vendored
@@ -1,6 +1,22 @@
|
||||
# Common files to ignore in Rust projects
|
||||
.DS_Store
|
||||
.idea
|
||||
*.log
|
||||
tmp/
|
||||
|
||||
# Generated by Cargo will have compiled files and executables
|
||||
/target
|
||||
|
||||
# Generated by rln-wasm
|
||||
/pkg
|
||||
/examples/node_modules
|
||||
/examples/package-lock.json
|
||||
|
||||
# Generated by Nix
|
||||
result
|
||||
|
||||
# These are backup files generated by rustfmt
|
||||
**/*.rs.bk
|
||||
Cargo.lock
|
||||
bin/
|
||||
pkg/
|
||||
wasm-pack.log
|
||||
|
||||
# MSVC Windows builds of rustc generate these, which store debugging information
|
||||
*.pdb
|
||||
|
||||
1792
rln-wasm/Cargo.lock
generated
Normal file
1792
rln-wasm/Cargo.lock
generated
Normal file
File diff suppressed because it is too large
Load Diff
@@ -1,39 +1,53 @@
|
||||
[package]
|
||||
name = "rln-wasm"
|
||||
version = "0.0.13"
|
||||
version = "1.0.0"
|
||||
edition = "2021"
|
||||
license = "MIT or Apache2"
|
||||
license = "MIT OR Apache-2.0"
|
||||
|
||||
[lib]
|
||||
crate-type = ["cdylib", "rlib"]
|
||||
|
||||
[features]
|
||||
default = ["console_error_panic_hook"]
|
||||
|
||||
[dependencies]
|
||||
rln = { path = "../rln", default-features = false, features = [
|
||||
"wasm",
|
||||
rln = { path = "../rln", version = "1.0.0", default-features = false, features = [
|
||||
"stateless",
|
||||
] }
|
||||
num-bigint = { version = "0.4", default-features = false, features = [
|
||||
"rand",
|
||||
"serde",
|
||||
] }
|
||||
wasmer = { version = "2.3", default-features = false, features = ["js", "std"] }
|
||||
web-sys = { version = "0.3", features = ["console"] }
|
||||
getrandom = { version = "0.2.7", default-features = false, features = ["js"] }
|
||||
wasm-bindgen = "0.2.63"
|
||||
serde-wasm-bindgen = "0.4"
|
||||
js-sys = "0.3.59"
|
||||
serde_json = "1.0.85"
|
||||
zerokit_utils = { path = "../utils", version = "1.0.0", default-features = false }
|
||||
num-bigint = { version = "0.4.6", default-features = false }
|
||||
js-sys = "0.3.83"
|
||||
wasm-bindgen = "0.2.106"
|
||||
serde-wasm-bindgen = "0.6.5"
|
||||
serde = "1.0.228"
|
||||
wasm-bindgen-rayon = { version = "1.3.0", features = [
|
||||
"no-bundler",
|
||||
], optional = true }
|
||||
ark-relations = { version = "0.5.1", features = ["std"] }
|
||||
ark-groth16 = { version = "0.5.0", default-features = false }
|
||||
rand = "0.8.5"
|
||||
|
||||
# The `console_error_panic_hook` crate provides better debugging of panics by
|
||||
# logging them with `console.error`. This is great for development, but requires
|
||||
# all the `std::fmt` and `std::panicking` infrastructure, so isn't great for
|
||||
# code size when deploying.
|
||||
console_error_panic_hook = { version = "0.1.7", optional = true }
|
||||
zerokit_utils = { version = "0.5.1", path = "../utils" }
|
||||
|
||||
[target.'cfg(target_arch = "wasm32")'.dependencies]
|
||||
getrandom = { version = "0.2.16", features = ["js"] }
|
||||
|
||||
[dev-dependencies]
|
||||
wasm-bindgen-test = "0.3.13"
|
||||
wasm-bindgen-futures = "0.4.33"
|
||||
serde_json = "1.0.145"
|
||||
wasm-bindgen-test = "0.3.56"
|
||||
wasm-bindgen-futures = "0.4.56"
|
||||
ark-std = { version = "0.5.0", default-features = false }
|
||||
|
||||
[dev-dependencies.web-sys]
|
||||
version = "0.3.83"
|
||||
features = ["Window", "Navigator"]
|
||||
|
||||
[features]
|
||||
default = []
|
||||
utils = []
|
||||
panic_hook = ["console_error_panic_hook"]
|
||||
parallel = ["rln/parallel", "wasm-bindgen-rayon", "ark-groth16/parallel"]
|
||||
|
||||
[package.metadata.docs.rs]
|
||||
all-features = true
|
||||
|
||||
@@ -1,31 +1,137 @@
|
||||
[tasks.pack-build]
|
||||
[tasks.build]
|
||||
clear = true
|
||||
dependencies = ["pack_build", "pack_rename", "pack_add_keywords"]
|
||||
|
||||
[tasks.build_parallel]
|
||||
clear = true
|
||||
dependencies = [
|
||||
"pack_build_parallel",
|
||||
"pack_rename_parallel",
|
||||
"pack_add_keywords",
|
||||
]
|
||||
|
||||
[tasks.build_utils]
|
||||
clear = true
|
||||
dependencies = ["pack_build_utils", "pack_rename_utils", "pack_add_keywords"]
|
||||
|
||||
[tasks.pack_build]
|
||||
command = "wasm-pack"
|
||||
args = ["build", "--release", "--target", "web", "--scope", "waku"]
|
||||
|
||||
[tasks.pack-rename]
|
||||
[tasks.pack_rename]
|
||||
script = "sed -i.bak 's/rln-wasm/zerokit-rln-wasm/g' pkg/package.json && rm pkg/package.json.bak"
|
||||
|
||||
[tasks.build]
|
||||
clear = true
|
||||
dependencies = ["pack-build", "pack-rename", "post-build"]
|
||||
[tasks.pack_build_parallel]
|
||||
command = "env"
|
||||
args = [
|
||||
"CARGO_TARGET_WASM32_UNKNOWN_UNKNOWN_RUSTFLAGS=-C target-feature=+atomics,+bulk-memory,+mutable-globals -C link-arg=--shared-memory -C link-arg=--max-memory=1073741824 -C link-arg=--import-memory -C link-arg=--export=__wasm_init_tls -C link-arg=--export=__tls_size -C link-arg=--export=__tls_align -C link-arg=--export=__tls_base",
|
||||
"rustup",
|
||||
"run",
|
||||
"nightly",
|
||||
"wasm-pack",
|
||||
"build",
|
||||
"--release",
|
||||
"--target",
|
||||
"web",
|
||||
"--scope",
|
||||
"waku",
|
||||
"--features",
|
||||
"parallel",
|
||||
"-Z",
|
||||
"build-std=panic_abort,std",
|
||||
]
|
||||
|
||||
[tasks.post-build]
|
||||
command = "wasm-strip"
|
||||
args = ["./pkg/rln_wasm_bg.wasm"]
|
||||
[tasks.pack_rename_parallel]
|
||||
script = "sed -i.bak 's/rln-wasm/zerokit-rln-wasm-parallel/g' pkg/package.json && rm pkg/package.json.bak"
|
||||
|
||||
[tasks.pack_build_utils]
|
||||
command = "wasm-pack"
|
||||
args = [
|
||||
"build",
|
||||
"--release",
|
||||
"--target",
|
||||
"web",
|
||||
"--scope",
|
||||
"waku",
|
||||
"--no-default-features",
|
||||
"--features",
|
||||
"utils",
|
||||
]
|
||||
|
||||
[tasks.pack_rename_utils]
|
||||
script = "sed -i.bak 's/rln-wasm/zerokit-rln-wasm-utils/g' pkg/package.json && rm pkg/package.json.bak"
|
||||
|
||||
[tasks.pack_add_keywords]
|
||||
script = """
|
||||
jq '. + {keywords: ["zerokit", "rln", "wasm"]}' pkg/package.json > pkg/package.json.tmp && \
|
||||
mv pkg/package.json.tmp pkg/package.json
|
||||
"""
|
||||
|
||||
[tasks.test]
|
||||
command = "wasm-pack"
|
||||
args = ["test", "--release", "--node"]
|
||||
args = [
|
||||
"test",
|
||||
"--release",
|
||||
"--node",
|
||||
"--target",
|
||||
"wasm32-unknown-unknown",
|
||||
"--",
|
||||
"--nocapture",
|
||||
]
|
||||
dependencies = ["build"]
|
||||
|
||||
[tasks.login]
|
||||
[tasks.test_browser]
|
||||
command = "wasm-pack"
|
||||
args = ["login"]
|
||||
args = [
|
||||
"test",
|
||||
"--release",
|
||||
"--chrome",
|
||||
"--headless",
|
||||
"--target",
|
||||
"wasm32-unknown-unknown",
|
||||
"--",
|
||||
"--nocapture",
|
||||
]
|
||||
dependencies = ["build"]
|
||||
|
||||
[tasks.publish]
|
||||
[tasks.test_parallel]
|
||||
command = "env"
|
||||
args = [
|
||||
"CARGO_TARGET_WASM32_UNKNOWN_UNKNOWN_RUSTFLAGS=-C target-feature=+atomics,+bulk-memory,+mutable-globals -C link-arg=--shared-memory -C link-arg=--max-memory=1073741824 -C link-arg=--import-memory -C link-arg=--export=__wasm_init_tls -C link-arg=--export=__tls_size -C link-arg=--export=__tls_align -C link-arg=--export=__tls_base",
|
||||
"rustup",
|
||||
"run",
|
||||
"nightly",
|
||||
"wasm-pack",
|
||||
"test",
|
||||
"--release",
|
||||
"--chrome",
|
||||
"--headless",
|
||||
"--target",
|
||||
"wasm32-unknown-unknown",
|
||||
"--features",
|
||||
"parallel",
|
||||
"-Z",
|
||||
"build-std=panic_abort,std",
|
||||
"--",
|
||||
"--nocapture",
|
||||
]
|
||||
dependencies = ["build_parallel"]
|
||||
|
||||
[tasks.test_utils]
|
||||
command = "wasm-pack"
|
||||
args = ["publish", "--access", "public", "--target", "web"]
|
||||
args = [
|
||||
"test",
|
||||
"--release",
|
||||
"--node",
|
||||
"--target",
|
||||
"wasm32-unknown-unknown",
|
||||
"--no-default-features",
|
||||
"--features",
|
||||
"utils",
|
||||
"--",
|
||||
"--nocapture",
|
||||
]
|
||||
dependencies = ["build_utils"]
|
||||
|
||||
[tasks.bench]
|
||||
command = "echo"
|
||||
args = ["'No benchmarks available for this project'"]
|
||||
disabled = true
|
||||
|
||||
@@ -1,41 +1,119 @@
|
||||
# RLN for WASM
|
||||
This library is used in [waku-org/js-rln](https://github.com/waku-org/js-rln/)
|
||||
|
||||
## Building the library
|
||||
1. Install `wasm-pack`
|
||||
```
|
||||
curl https://rustwasm.github.io/wasm-pack/installer/init.sh -sSf | sh
|
||||
```
|
||||
2. Install `cargo-make`
|
||||
```
|
||||
cargo install cargo-make
|
||||
```
|
||||
[](https://badge.fury.io/js/@waku%2Fzerokit-rln-wasm)
|
||||
[](https://opensource.org/licenses/MIT)
|
||||
[](https://opensource.org/licenses/Apache-2.0)
|
||||
|
||||
OR
|
||||
The Zerokit RLN WASM Module provides WebAssembly bindings for working with
|
||||
Rate-Limiting Nullifier [RLN](https://rfc.vac.dev/vac/raw/rln-v2) zkSNARK proofs and primitives.
|
||||
This module is used by [waku-org/js-rln](https://github.com/waku-org/js-rln/) to enable
|
||||
RLN functionality in JavaScript/TypeScript applications.
|
||||
|
||||
```
|
||||
## Install Dependencies
|
||||
|
||||
> [!NOTE]
|
||||
> This project requires the following tools:
|
||||
>
|
||||
> - `wasm-pack` (v0.13.1) - for compiling Rust to WebAssembly
|
||||
> - `cargo-make` - for running build commands
|
||||
> - `nvm` - to install and manage Node.js (v22.14.0+)
|
||||
|
||||
### Quick Install
|
||||
|
||||
```bash
|
||||
make installdeps
|
||||
```
|
||||
3. Compile zerokit for `wasm32-unknown-unknown`:
|
||||
```
|
||||
cd rln-wasm
|
||||
cargo make build
|
||||
```
|
||||
4. Compile a slimmer version of zerokit for `wasm32-unknown-unknown`:
|
||||
```
|
||||
cd rln-wasm
|
||||
cargo make post-build
|
||||
|
||||
### Manual Installation
|
||||
|
||||
```bash
|
||||
# Install wasm-pack
|
||||
cargo install wasm-pack --version=0.13.1
|
||||
|
||||
# Install cargo-make
|
||||
cargo install cargo-make
|
||||
|
||||
# Install Node.js via nvm
|
||||
nvm install 22.14.0
|
||||
nvm use 22.14.0
|
||||
nvm alias default 22.14.0
|
||||
```
|
||||
|
||||
## Running tests
|
||||
```
|
||||
## Building the Library
|
||||
|
||||
Navigate to the rln-wasm directory:
|
||||
|
||||
```bash
|
||||
cd rln-wasm
|
||||
cargo make test
|
||||
```
|
||||
|
||||
## Publishing a npm package
|
||||
Build commands:
|
||||
|
||||
```bash
|
||||
cargo make build # Default → @waku/zerokit-rln-wasm
|
||||
cargo make build_parallel # Parallel → @waku/zerokit-rln-wasm-parallel (requires nightly Rust)
|
||||
cargo make build_utils # Utils only → @waku/zerokit-rln-wasm-utils
|
||||
```
|
||||
|
||||
All packages output to `pkg/` directory.
|
||||
|
||||
## Running Tests and Benchmarks
|
||||
|
||||
```bash
|
||||
cargo make test # Standard tests
|
||||
cargo make test_browser # Browser headless mode
|
||||
cargo make test_utils # Utils-only tests
|
||||
cargo make test_parallel # Parallel tests
|
||||
```
|
||||
|
||||
## Examples
|
||||
|
||||
See [Node example](./examples/index.js) and [README](./examples/Readme.md) for proof generation, verification, and slashing.
|
||||
|
||||
## Parallel Computation
|
||||
|
||||
Enables multi-threaded browser execution using `wasm-bindgen-rayon`.
|
||||
|
||||
> [!NOTE]
|
||||
>
|
||||
> - Parallel support is not enabled by default due to WebAssembly and browser limitations.
|
||||
> - Requires `nightly` Rust: `rustup install nightly`
|
||||
> - Browser-only (not compatible with Node.js)
|
||||
> - Requires HTTP headers for `SharedArrayBuffer`:
|
||||
> - `Cross-Origin-Opener-Policy: same-origin`
|
||||
> - `Cross-Origin-Embedder-Policy: require-corp`
|
||||
|
||||
### Usage
|
||||
|
||||
Direct usage (modern browsers with WebAssembly threads support):
|
||||
|
||||
```js
|
||||
import * as wasmPkg from '@waku/zerokit-rln-wasm-parallel';
|
||||
|
||||
await wasmPkg.default();
|
||||
await wasmPkg.initThreadPool(navigator.hardwareConcurrency);
|
||||
wasmPkg.nowCallAnyExportedFuncs();
|
||||
```
|
||||
|
||||
### Feature Detection for Older Browsers
|
||||
|
||||
If you're targeting [older browser versions that didn't support WebAssembly threads yet](https://webassembly.org/roadmap/), you'll want to use both builds - the parallel version for modern browsers and the default version as a fallback. Use feature detection to choose the appropriate build on the JavaScript side.
|
||||
|
||||
You can use the [wasm-feature-detect](https://github.com/GoogleChromeLabs/wasm-feature-detect) library for this purpose:
|
||||
|
||||
```js
|
||||
import { threads } from 'wasm-feature-detect';
|
||||
|
||||
let wasmPkg;
|
||||
|
||||
if (await threads()) {
|
||||
wasmPkg = await import('@waku/zerokit-rln-wasm-parallel');
|
||||
await wasmPkg.default();
|
||||
await wasmPkg.initThreadPool(navigator.hardwareConcurrency);
|
||||
} else {
|
||||
wasmPkg = await import('@waku/zerokit-rln-wasm');
|
||||
await wasmPkg.default();
|
||||
}
|
||||
|
||||
wasmPkg.nowCallAnyExportedFuncs();
|
||||
```
|
||||
cd rln-wasm
|
||||
cargo make login
|
||||
cargo make publish
|
||||
```
|
||||
22
rln-wasm/examples/README.md
Normal file
22
rln-wasm/examples/README.md
Normal file
@@ -0,0 +1,22 @@
|
||||
# RLN WASM Node Examples
|
||||
|
||||
This example demonstrates how to use the RLN WASM package in a Node.js environment.
|
||||
|
||||
## Build the @waku/zerokit-rln-wasm package at the root of rln-wasm module
|
||||
|
||||
```bash
|
||||
cargo make build
|
||||
```
|
||||
|
||||
## Move into this directory and install dependencies
|
||||
|
||||
```bash
|
||||
cd examples
|
||||
npm install
|
||||
```
|
||||
|
||||
## Run
|
||||
|
||||
```bash
|
||||
npm start
|
||||
```
|
||||
484
rln-wasm/examples/index.js
Normal file
484
rln-wasm/examples/index.js
Normal file
@@ -0,0 +1,484 @@
|
||||
import { readFileSync } from "fs";
|
||||
import { fileURLToPath } from "url";
|
||||
import { dirname, join } from "path";
|
||||
|
||||
const __filename = fileURLToPath(import.meta.url);
|
||||
const __dirname = dirname(__filename);
|
||||
|
||||
function debugUint8Array(uint8Array) {
|
||||
return Array.from(uint8Array, (byte) =>
|
||||
byte.toString(16).padStart(2, "0")
|
||||
).join(", ");
|
||||
}
|
||||
|
||||
async function calculateWitness(circomPath, inputs, witnessCalculatorFile) {
|
||||
const wasmFile = readFileSync(circomPath);
|
||||
const wasmFileBuffer = wasmFile.buffer.slice(
|
||||
wasmFile.byteOffset,
|
||||
wasmFile.byteOffset + wasmFile.byteLength
|
||||
);
|
||||
const witnessCalculator = await witnessCalculatorFile(wasmFileBuffer);
|
||||
const calculatedWitness = await witnessCalculator.calculateWitness(
|
||||
inputs,
|
||||
false
|
||||
);
|
||||
return calculatedWitness;
|
||||
}
|
||||
|
||||
async function main() {
|
||||
const rlnWasm = await import("../pkg/rln_wasm.js");
|
||||
const wasmPath = join(__dirname, "../pkg/rln_wasm_bg.wasm");
|
||||
const wasmBytes = readFileSync(wasmPath);
|
||||
rlnWasm.initSync({ module: wasmBytes });
|
||||
|
||||
const zkeyPath = join(
|
||||
__dirname,
|
||||
"../../rln/resources/tree_depth_20/rln_final.arkzkey"
|
||||
);
|
||||
const circomPath = join(
|
||||
__dirname,
|
||||
"../../rln/resources/tree_depth_20/rln.wasm"
|
||||
);
|
||||
const witnessCalculatorPath = join(
|
||||
__dirname,
|
||||
"../resources/witness_calculator.js"
|
||||
);
|
||||
const { builder: witnessCalculatorFile } = await import(
|
||||
witnessCalculatorPath
|
||||
);
|
||||
|
||||
console.log("Creating RLN instance");
|
||||
const zkeyData = readFileSync(zkeyPath);
|
||||
let rlnInstance;
|
||||
try {
|
||||
rlnInstance = new rlnWasm.WasmRLN(new Uint8Array(zkeyData));
|
||||
} catch (error) {
|
||||
console.error("Initial RLN instance creation error:", error);
|
||||
return;
|
||||
}
|
||||
console.log("RLN instance created successfully");
|
||||
|
||||
console.log("\nGenerating identity keys");
|
||||
let identity;
|
||||
try {
|
||||
identity = rlnWasm.Identity.generate();
|
||||
} catch (error) {
|
||||
console.error("Key generation error:", error);
|
||||
return;
|
||||
}
|
||||
const identitySecret = identity.getSecretHash();
|
||||
const idCommitment = identity.getCommitment();
|
||||
console.log("Identity generated");
|
||||
console.log(" - identity_secret = " + identitySecret.debug());
|
||||
console.log(" - id_commitment = " + idCommitment.debug());
|
||||
|
||||
console.log("\nCreating message limit");
|
||||
const userMessageLimit = rlnWasm.WasmFr.fromUint(1);
|
||||
console.log(" - user_message_limit = " + userMessageLimit.debug());
|
||||
|
||||
console.log("\nComputing rate commitment");
|
||||
let rateCommitment;
|
||||
try {
|
||||
rateCommitment = rlnWasm.Hasher.poseidonHashPair(
|
||||
idCommitment,
|
||||
userMessageLimit
|
||||
);
|
||||
} catch (error) {
|
||||
console.error("Rate commitment hash error:", error);
|
||||
return;
|
||||
}
|
||||
console.log(" - rate_commitment = " + rateCommitment.debug());
|
||||
|
||||
console.log("\nWasmFr serialization: WasmFr <-> bytes");
|
||||
const serRateCommitment = rateCommitment.toBytesLE();
|
||||
console.log(
|
||||
" - serialized rate_commitment = [" +
|
||||
debugUint8Array(serRateCommitment) +
|
||||
"]"
|
||||
);
|
||||
|
||||
let deserRateCommitment;
|
||||
try {
|
||||
deserRateCommitment = rlnWasm.WasmFr.fromBytesLE(serRateCommitment);
|
||||
} catch (error) {
|
||||
console.error("Rate commitment deserialization error:", error);
|
||||
return;
|
||||
}
|
||||
console.log(
|
||||
" - deserialized rate_commitment = " + deserRateCommitment.debug()
|
||||
);
|
||||
|
||||
console.log("\nIdentity serialization: Identity <-> bytes");
|
||||
const serIdentity = identity.toBytesLE();
|
||||
console.log(
|
||||
" - serialized identity = [" + debugUint8Array(serIdentity) + "]"
|
||||
);
|
||||
|
||||
let deserIdentity;
|
||||
try {
|
||||
deserIdentity = rlnWasm.Identity.fromBytesLE(serIdentity);
|
||||
} catch (error) {
|
||||
console.error("Identity deserialization error:", error);
|
||||
return;
|
||||
}
|
||||
const deserIdentitySecret = deserIdentity.getSecretHash();
|
||||
const deserIdCommitment = deserIdentity.getCommitment();
|
||||
console.log(
|
||||
" - deserialized identity = [" +
|
||||
deserIdentitySecret.debug() +
|
||||
", " +
|
||||
deserIdCommitment.debug() +
|
||||
"]"
|
||||
);
|
||||
|
||||
console.log("\nBuilding Merkle path for stateless mode");
|
||||
const treeDepth = 20;
|
||||
const defaultLeaf = rlnWasm.WasmFr.zero();
|
||||
|
||||
const defaultHashes = [];
|
||||
try {
|
||||
defaultHashes[0] = rlnWasm.Hasher.poseidonHashPair(
|
||||
defaultLeaf,
|
||||
defaultLeaf
|
||||
);
|
||||
for (let i = 1; i < treeDepth - 1; i++) {
|
||||
defaultHashes[i] = rlnWasm.Hasher.poseidonHashPair(
|
||||
defaultHashes[i - 1],
|
||||
defaultHashes[i - 1]
|
||||
);
|
||||
}
|
||||
} catch (error) {
|
||||
console.error("Poseidon hash error:", error);
|
||||
return;
|
||||
}
|
||||
|
||||
const pathElements = new rlnWasm.VecWasmFr();
|
||||
pathElements.push(defaultLeaf);
|
||||
for (let i = 1; i < treeDepth; i++) {
|
||||
pathElements.push(defaultHashes[i - 1]);
|
||||
}
|
||||
const identityPathIndex = new Uint8Array(treeDepth);
|
||||
|
||||
console.log("\nVecWasmFr serialization: VecWasmFr <-> bytes");
|
||||
const serPathElements = pathElements.toBytesLE();
|
||||
console.log(
|
||||
" - serialized path_elements = [" + debugUint8Array(serPathElements) + "]"
|
||||
);
|
||||
|
||||
let deserPathElements;
|
||||
try {
|
||||
deserPathElements = rlnWasm.VecWasmFr.fromBytesLE(serPathElements);
|
||||
} catch (error) {
|
||||
console.error("Path elements deserialization error:", error);
|
||||
return;
|
||||
}
|
||||
console.log(" - deserialized path_elements = ", deserPathElements.debug());
|
||||
|
||||
console.log("\nUint8Array serialization: Uint8Array <-> bytes");
|
||||
const serPathIndex = rlnWasm.Uint8ArrayUtils.toBytesLE(identityPathIndex);
|
||||
console.log(
|
||||
" - serialized path_index = [" + debugUint8Array(serPathIndex) + "]"
|
||||
);
|
||||
|
||||
let deserPathIndex;
|
||||
try {
|
||||
deserPathIndex = rlnWasm.Uint8ArrayUtils.fromBytesLE(serPathIndex);
|
||||
} catch (error) {
|
||||
console.error("Path index deserialization error:", error);
|
||||
return;
|
||||
}
|
||||
console.log(" - deserialized path_index =", deserPathIndex);
|
||||
|
||||
console.log("\nComputing Merkle root for stateless mode");
|
||||
console.log(" - computing root for index 0 with rate_commitment");
|
||||
|
||||
let computedRoot;
|
||||
try {
|
||||
computedRoot = rlnWasm.Hasher.poseidonHashPair(rateCommitment, defaultLeaf);
|
||||
for (let i = 1; i < treeDepth; i++) {
|
||||
computedRoot = rlnWasm.Hasher.poseidonHashPair(
|
||||
computedRoot,
|
||||
defaultHashes[i - 1]
|
||||
);
|
||||
}
|
||||
} catch (error) {
|
||||
console.error("Poseidon hash error:", error);
|
||||
return;
|
||||
}
|
||||
console.log(" - computed_root = " + computedRoot.debug());
|
||||
|
||||
console.log("\nHashing signal");
|
||||
const signal = new Uint8Array([
|
||||
1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
|
||||
0, 0, 0, 0, 0, 0, 0,
|
||||
]);
|
||||
let x;
|
||||
try {
|
||||
x = rlnWasm.Hasher.hashToFieldLE(signal);
|
||||
} catch (error) {
|
||||
console.error("Hash signal error:", error);
|
||||
return;
|
||||
}
|
||||
console.log(" - x = " + x.debug());
|
||||
|
||||
console.log("\nHashing epoch");
|
||||
const epochStr = "test-epoch";
|
||||
let epoch;
|
||||
try {
|
||||
epoch = rlnWasm.Hasher.hashToFieldLE(new TextEncoder().encode(epochStr));
|
||||
} catch (error) {
|
||||
console.error("Hash epoch error:", error);
|
||||
return;
|
||||
}
|
||||
console.log(" - epoch = " + epoch.debug());
|
||||
|
||||
console.log("\nHashing RLN identifier");
|
||||
const rlnIdStr = "test-rln-identifier";
|
||||
let rlnIdentifier;
|
||||
try {
|
||||
rlnIdentifier = rlnWasm.Hasher.hashToFieldLE(
|
||||
new TextEncoder().encode(rlnIdStr)
|
||||
);
|
||||
} catch (error) {
|
||||
console.error("Hash RLN identifier error:", error);
|
||||
return;
|
||||
}
|
||||
console.log(" - rln_identifier = " + rlnIdentifier.debug());
|
||||
|
||||
console.log("\nComputing Poseidon hash for external nullifier");
|
||||
let externalNullifier;
|
||||
try {
|
||||
externalNullifier = rlnWasm.Hasher.poseidonHashPair(epoch, rlnIdentifier);
|
||||
} catch (error) {
|
||||
console.error("External nullifier hash error:", error);
|
||||
return;
|
||||
}
|
||||
console.log(" - external_nullifier = " + externalNullifier.debug());
|
||||
|
||||
console.log("\nCreating message_id");
|
||||
const messageId = rlnWasm.WasmFr.fromUint(0);
|
||||
console.log(" - message_id = " + messageId.debug());
|
||||
|
||||
console.log("\nCreating RLN Witness");
|
||||
const witness = new rlnWasm.WasmRLNWitnessInput(
|
||||
identitySecret,
|
||||
userMessageLimit,
|
||||
messageId,
|
||||
pathElements,
|
||||
identityPathIndex,
|
||||
x,
|
||||
externalNullifier
|
||||
);
|
||||
console.log("RLN Witness created successfully");
|
||||
|
||||
console.log(
|
||||
"\nWasmRLNWitnessInput serialization: WasmRLNWitnessInput <-> bytes"
|
||||
);
|
||||
let serWitness;
|
||||
try {
|
||||
serWitness = witness.toBytesLE();
|
||||
} catch (error) {
|
||||
console.error("Witness serialization error:", error);
|
||||
return;
|
||||
}
|
||||
console.log(
|
||||
" - serialized witness = [" + debugUint8Array(serWitness) + " ]"
|
||||
);
|
||||
|
||||
let deserWitness;
|
||||
try {
|
||||
deserWitness = rlnWasm.WasmRLNWitnessInput.fromBytesLE(serWitness);
|
||||
} catch (error) {
|
||||
console.error("Witness deserialization error:", error);
|
||||
return;
|
||||
}
|
||||
console.log(" - witness deserialized successfully");
|
||||
|
||||
console.log("\nCalculating witness");
|
||||
let witnessJson;
|
||||
try {
|
||||
witnessJson = witness.toBigIntJson();
|
||||
} catch (error) {
|
||||
console.error("Witness to BigInt JSON error:", error);
|
||||
return;
|
||||
}
|
||||
const calculatedWitness = await calculateWitness(
|
||||
circomPath,
|
||||
witnessJson,
|
||||
witnessCalculatorFile
|
||||
);
|
||||
console.log("Witness calculated successfully");
|
||||
|
||||
console.log("\nGenerating RLN Proof");
|
||||
let rln_proof;
|
||||
try {
|
||||
rln_proof = rlnInstance.generateRLNProofWithWitness(
|
||||
calculatedWitness,
|
||||
witness
|
||||
);
|
||||
} catch (error) {
|
||||
console.error("Proof generation error:", error);
|
||||
return;
|
||||
}
|
||||
console.log("Proof generated successfully");
|
||||
|
||||
console.log("\nGetting proof values");
|
||||
const proofValues = rln_proof.getValues();
|
||||
console.log(" - y = " + proofValues.y.debug());
|
||||
console.log(" - nullifier = " + proofValues.nullifier.debug());
|
||||
console.log(" - root = " + proofValues.root.debug());
|
||||
console.log(" - x = " + proofValues.x.debug());
|
||||
console.log(
|
||||
" - external_nullifier = " + proofValues.externalNullifier.debug()
|
||||
);
|
||||
|
||||
console.log("\nRLNProof serialization: RLNProof <-> bytes");
|
||||
let serProof;
|
||||
try {
|
||||
serProof = rln_proof.toBytesLE();
|
||||
} catch (error) {
|
||||
console.error("Proof serialization error:", error);
|
||||
return;
|
||||
}
|
||||
console.log(" - serialized proof = [" + debugUint8Array(serProof) + " ]");
|
||||
|
||||
let deserProof;
|
||||
try {
|
||||
deserProof = rlnWasm.WasmRLNProof.fromBytesLE(serProof);
|
||||
} catch (error) {
|
||||
console.error("Proof deserialization error:", error);
|
||||
return;
|
||||
}
|
||||
console.log(" - proof deserialized successfully");
|
||||
|
||||
console.log("\nRLNProofValues serialization: RLNProofValues <-> bytes");
|
||||
const serProofValues = proofValues.toBytesLE();
|
||||
console.log(
|
||||
" - serialized proof_values = [" + debugUint8Array(serProofValues) + " ]"
|
||||
);
|
||||
|
||||
let deserProofValues2;
|
||||
try {
|
||||
deserProofValues2 = rlnWasm.WasmRLNProofValues.fromBytesLE(serProofValues);
|
||||
} catch (error) {
|
||||
console.error("Proof values deserialization error:", error);
|
||||
return;
|
||||
}
|
||||
console.log(" - proof_values deserialized successfully");
|
||||
console.log(
|
||||
" - deserialized external_nullifier = " +
|
||||
deserProofValues2.externalNullifier.debug()
|
||||
);
|
||||
|
||||
console.log("\nVerifying Proof");
|
||||
const roots = new rlnWasm.VecWasmFr();
|
||||
roots.push(computedRoot);
|
||||
let isValid;
|
||||
try {
|
||||
isValid = rlnInstance.verifyWithRoots(rln_proof, roots, x);
|
||||
} catch (error) {
|
||||
console.error("Proof verification error:", error);
|
||||
return;
|
||||
}
|
||||
if (isValid) {
|
||||
console.log("Proof verified successfully");
|
||||
} else {
|
||||
console.log("Proof verification failed");
|
||||
return;
|
||||
}
|
||||
|
||||
console.log(
|
||||
"\nSimulating double-signaling attack (same epoch, different message)"
|
||||
);
|
||||
|
||||
console.log("\nHashing second signal");
|
||||
const signal2 = new Uint8Array([
|
||||
11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
|
||||
0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
|
||||
]);
|
||||
let x2;
|
||||
try {
|
||||
x2 = rlnWasm.Hasher.hashToFieldLE(signal2);
|
||||
} catch (error) {
|
||||
console.error("Hash second signal error:", error);
|
||||
return;
|
||||
}
|
||||
console.log(" - x2 = " + x2.debug());
|
||||
|
||||
console.log("\nCreating second message with the same id");
|
||||
const messageId2 = rlnWasm.WasmFr.fromUint(0);
|
||||
console.log(" - message_id2 = " + messageId2.debug());
|
||||
|
||||
console.log("\nCreating second RLN Witness");
|
||||
const witness2 = new rlnWasm.WasmRLNWitnessInput(
|
||||
identitySecret,
|
||||
userMessageLimit,
|
||||
messageId2,
|
||||
pathElements,
|
||||
identityPathIndex,
|
||||
x2,
|
||||
externalNullifier
|
||||
);
|
||||
console.log("Second RLN Witness created successfully");
|
||||
|
||||
console.log("\nCalculating second witness");
|
||||
let witnessJson2;
|
||||
try {
|
||||
witnessJson2 = witness2.toBigIntJson();
|
||||
} catch (error) {
|
||||
console.error("Second witness to BigInt JSON error:", error);
|
||||
return;
|
||||
}
|
||||
const calculatedWitness2 = await calculateWitness(
|
||||
circomPath,
|
||||
witnessJson2,
|
||||
witnessCalculatorFile
|
||||
);
|
||||
console.log("Second witness calculated successfully");
|
||||
|
||||
console.log("\nGenerating second RLN Proof");
|
||||
let rln_proof2;
|
||||
try {
|
||||
rln_proof2 = rlnInstance.generateRLNProofWithWitness(
|
||||
calculatedWitness2,
|
||||
witness2
|
||||
);
|
||||
} catch (error) {
|
||||
console.error("Second proof generation error:", error);
|
||||
return;
|
||||
}
|
||||
console.log("Second proof generated successfully");
|
||||
|
||||
console.log("\nVerifying second proof");
|
||||
let isValid2;
|
||||
try {
|
||||
isValid2 = rlnInstance.verifyWithRoots(rln_proof2, roots, x2);
|
||||
} catch (error) {
|
||||
console.error("Proof verification error:", error);
|
||||
return;
|
||||
}
|
||||
if (isValid2) {
|
||||
console.log("Second proof verified successfully");
|
||||
|
||||
console.log("\nRecovering identity secret");
|
||||
const proofValues1 = rln_proof.getValues();
|
||||
const proofValues2 = rln_proof2.getValues();
|
||||
let recoveredSecret;
|
||||
try {
|
||||
recoveredSecret = rlnWasm.WasmRLNProofValues.recoverIdSecret(
|
||||
proofValues1,
|
||||
proofValues2
|
||||
);
|
||||
} catch (error) {
|
||||
console.error("Identity recovery error:", error);
|
||||
return;
|
||||
}
|
||||
console.log(" - recovered_secret = " + recoveredSecret.debug());
|
||||
console.log(" - original_secret = " + identitySecret.debug());
|
||||
console.log("Slashing successful: Identity is recovered!");
|
||||
} else {
|
||||
console.log("Second proof verification failed");
|
||||
}
|
||||
}
|
||||
|
||||
main().catch(console.error);
|
||||
13
rln-wasm/examples/package.json
Normal file
13
rln-wasm/examples/package.json
Normal file
@@ -0,0 +1,13 @@
|
||||
{
|
||||
"name": "rln-wasm-node-example",
|
||||
"version": "1.0.0",
|
||||
"description": "Node.js example for RLN WASM",
|
||||
"type": "module",
|
||||
"main": "index.js",
|
||||
"scripts": {
|
||||
"start": "node index.js"
|
||||
},
|
||||
"dependencies": {
|
||||
"@waku/zerokit-rln-wasm": "file:../../pkg"
|
||||
}
|
||||
}
|
||||
@@ -1,331 +1,328 @@
|
||||
module.exports = async function builder(code, options) {
|
||||
// File generated with https://github.com/iden3/circom
|
||||
// following the instructions from:
|
||||
// https://github.com/vacp2p/zerokit/tree/master/rln#advanced-custom-circuit-compilation
|
||||
|
||||
options = options || {};
|
||||
export async function builder(code, options) {
|
||||
options = options || {};
|
||||
|
||||
let wasmModule;
|
||||
try {
|
||||
wasmModule = await WebAssembly.compile(code);
|
||||
} catch (err) {
|
||||
console.log(err);
|
||||
console.log("\nTry to run circom --c in order to generate c++ code instead\n");
|
||||
throw new Error(err);
|
||||
}
|
||||
let wasmModule;
|
||||
try {
|
||||
wasmModule = await WebAssembly.compile(code);
|
||||
} catch (err) {
|
||||
console.log(err);
|
||||
console.log(
|
||||
"\nTry to run circom --c in order to generate c++ code instead\n"
|
||||
);
|
||||
throw new Error(err);
|
||||
}
|
||||
|
||||
let wc;
|
||||
let wc;
|
||||
|
||||
let errStr = "";
|
||||
let msgStr = "";
|
||||
|
||||
const instance = await WebAssembly.instantiate(wasmModule, {
|
||||
runtime: {
|
||||
exceptionHandler : function(code) {
|
||||
let err;
|
||||
if (code == 1) {
|
||||
err = "Signal not found.\n";
|
||||
} else if (code == 2) {
|
||||
err = "Too many signals set.\n";
|
||||
} else if (code == 3) {
|
||||
err = "Signal already set.\n";
|
||||
} else if (code == 4) {
|
||||
err = "Assert Failed.\n";
|
||||
} else if (code == 5) {
|
||||
err = "Not enough memory.\n";
|
||||
} else if (code == 6) {
|
||||
err = "Input signal array access exceeds the size.\n";
|
||||
} else {
|
||||
err = "Unknown error.\n";
|
||||
}
|
||||
throw new Error(err + errStr);
|
||||
},
|
||||
printErrorMessage : function() {
|
||||
errStr += getMessage() + "\n";
|
||||
// console.error(getMessage());
|
||||
},
|
||||
writeBufferMessage : function() {
|
||||
const msg = getMessage();
|
||||
// Any calls to `log()` will always end with a `\n`, so that's when we print and reset
|
||||
if (msg === "\n") {
|
||||
console.log(msgStr);
|
||||
msgStr = "";
|
||||
} else {
|
||||
// If we've buffered other content, put a space in between the items
|
||||
if (msgStr !== "") {
|
||||
msgStr += " "
|
||||
}
|
||||
// Then append the message to the message we are creating
|
||||
msgStr += msg;
|
||||
}
|
||||
},
|
||||
showSharedRWMemory : function() {
|
||||
printSharedRWMemory ();
|
||||
}
|
||||
let errStr = "";
|
||||
let msgStr = "";
|
||||
|
||||
const instance = await WebAssembly.instantiate(wasmModule, {
|
||||
runtime: {
|
||||
exceptionHandler: function (code) {
|
||||
let err;
|
||||
if (code == 1) {
|
||||
err = "Signal not found.\n";
|
||||
} else if (code == 2) {
|
||||
err = "Too many signals set.\n";
|
||||
} else if (code == 3) {
|
||||
err = "Signal already set.\n";
|
||||
} else if (code == 4) {
|
||||
err = "Assert Failed.\n";
|
||||
} else if (code == 5) {
|
||||
err = "Not enough memory.\n";
|
||||
} else if (code == 6) {
|
||||
err = "Input signal array access exceeds the size.\n";
|
||||
} else {
|
||||
err = "Unknown error.\n";
|
||||
}
|
||||
});
|
||||
throw new Error(err + errStr);
|
||||
},
|
||||
printErrorMessage: function () {
|
||||
errStr += getMessage() + "\n";
|
||||
// console.error(getMessage());
|
||||
},
|
||||
writeBufferMessage: function () {
|
||||
const msg = getMessage();
|
||||
// Any calls to `log()` will always end with a `\n`, so that's when we print and reset
|
||||
if (msg === "\n") {
|
||||
console.log(msgStr);
|
||||
msgStr = "";
|
||||
} else {
|
||||
// If we've buffered other content, put a space in between the items
|
||||
if (msgStr !== "") {
|
||||
msgStr += " ";
|
||||
}
|
||||
// Then append the message to the message we are creating
|
||||
msgStr += msg;
|
||||
}
|
||||
},
|
||||
showSharedRWMemory: function () {
|
||||
printSharedRWMemory();
|
||||
},
|
||||
},
|
||||
});
|
||||
|
||||
const sanityCheck =
|
||||
options
|
||||
// options &&
|
||||
// (
|
||||
// options.sanityCheck ||
|
||||
// options.logGetSignal ||
|
||||
// options.logSetSignal ||
|
||||
// options.logStartComponent ||
|
||||
// options.logFinishComponent
|
||||
// );
|
||||
const sanityCheck = options;
|
||||
// options &&
|
||||
// (
|
||||
// options.sanityCheck ||
|
||||
// options.logGetSignal ||
|
||||
// options.logSetSignal ||
|
||||
// options.logStartComponent ||
|
||||
// options.logFinishComponent
|
||||
// );
|
||||
|
||||
|
||||
wc = new WitnessCalculator(instance, sanityCheck);
|
||||
return wc;
|
||||
wc = new WitnessCalculator(instance, sanityCheck);
|
||||
return wc;
|
||||
|
||||
function getMessage() {
|
||||
var message = "";
|
||||
var c = instance.exports.getMessageChar();
|
||||
while ( c != 0 ) {
|
||||
message += String.fromCharCode(c);
|
||||
c = instance.exports.getMessageChar();
|
||||
}
|
||||
return message;
|
||||
function getMessage() {
|
||||
var message = "";
|
||||
var c = instance.exports.getMessageChar();
|
||||
while (c != 0) {
|
||||
message += String.fromCharCode(c);
|
||||
c = instance.exports.getMessageChar();
|
||||
}
|
||||
|
||||
function printSharedRWMemory () {
|
||||
const shared_rw_memory_size = instance.exports.getFieldNumLen32();
|
||||
const arr = new Uint32Array(shared_rw_memory_size);
|
||||
for (let j=0; j<shared_rw_memory_size; j++) {
|
||||
arr[shared_rw_memory_size-1-j] = instance.exports.readSharedRWMemory(j);
|
||||
}
|
||||
return message;
|
||||
}
|
||||
|
||||
// If we've buffered other content, put a space in between the items
|
||||
if (msgStr !== "") {
|
||||
msgStr += " "
|
||||
}
|
||||
// Then append the value to the message we are creating
|
||||
msgStr += (fromArray32(arr).toString());
|
||||
}
|
||||
function printSharedRWMemory() {
|
||||
const shared_rw_memory_size = instance.exports.getFieldNumLen32();
|
||||
const arr = new Uint32Array(shared_rw_memory_size);
|
||||
for (let j = 0; j < shared_rw_memory_size; j++) {
|
||||
arr[shared_rw_memory_size - 1 - j] =
|
||||
instance.exports.readSharedRWMemory(j);
|
||||
}
|
||||
|
||||
};
|
||||
// If we've buffered other content, put a space in between the items
|
||||
if (msgStr !== "") {
|
||||
msgStr += " ";
|
||||
}
|
||||
// Then append the value to the message we are creating
|
||||
msgStr += fromArray32(arr).toString();
|
||||
}
|
||||
}
|
||||
|
||||
class WitnessCalculator {
|
||||
constructor(instance, sanityCheck) {
|
||||
this.instance = instance;
|
||||
constructor(instance, sanityCheck) {
|
||||
this.instance = instance;
|
||||
|
||||
this.version = this.instance.exports.getVersion();
|
||||
this.n32 = this.instance.exports.getFieldNumLen32();
|
||||
this.version = this.instance.exports.getVersion();
|
||||
this.n32 = this.instance.exports.getFieldNumLen32();
|
||||
|
||||
this.instance.exports.getRawPrime();
|
||||
const arr = new Uint32Array(this.n32);
|
||||
for (let i=0; i<this.n32; i++) {
|
||||
arr[this.n32-1-i] = this.instance.exports.readSharedRWMemory(i);
|
||||
this.instance.exports.getRawPrime();
|
||||
const arr = new Uint32Array(this.n32);
|
||||
for (let i = 0; i < this.n32; i++) {
|
||||
arr[this.n32 - 1 - i] = this.instance.exports.readSharedRWMemory(i);
|
||||
}
|
||||
this.prime = fromArray32(arr);
|
||||
|
||||
this.witnessSize = this.instance.exports.getWitnessSize();
|
||||
|
||||
this.sanityCheck = sanityCheck;
|
||||
}
|
||||
|
||||
circom_version() {
|
||||
return this.instance.exports.getVersion();
|
||||
}
|
||||
|
||||
async _doCalculateWitness(input, sanityCheck) {
|
||||
//input is assumed to be a map from signals to arrays of bigints
|
||||
this.instance.exports.init(this.sanityCheck || sanityCheck ? 1 : 0);
|
||||
const keys = Object.keys(input);
|
||||
var input_counter = 0;
|
||||
keys.forEach((k) => {
|
||||
const h = fnvHash(k);
|
||||
const hMSB = parseInt(h.slice(0, 8), 16);
|
||||
const hLSB = parseInt(h.slice(8, 16), 16);
|
||||
const fArr = flatArray(input[k]);
|
||||
let signalSize = this.instance.exports.getInputSignalSize(hMSB, hLSB);
|
||||
if (signalSize < 0) {
|
||||
throw new Error(`Signal ${k} not found\n`);
|
||||
}
|
||||
if (fArr.length < signalSize) {
|
||||
throw new Error(`Not enough values for input signal ${k}\n`);
|
||||
}
|
||||
if (fArr.length > signalSize) {
|
||||
throw new Error(`Too many values for input signal ${k}\n`);
|
||||
}
|
||||
for (let i = 0; i < fArr.length; i++) {
|
||||
const arrFr = toArray32(BigInt(fArr[i]) % this.prime, this.n32);
|
||||
for (let j = 0; j < this.n32; j++) {
|
||||
this.instance.exports.writeSharedRWMemory(j, arrFr[this.n32 - 1 - j]);
|
||||
}
|
||||
this.prime = fromArray32(arr);
|
||||
|
||||
this.witnessSize = this.instance.exports.getWitnessSize();
|
||||
|
||||
this.sanityCheck = sanityCheck;
|
||||
}
|
||||
|
||||
circom_version() {
|
||||
return this.instance.exports.getVersion();
|
||||
}
|
||||
|
||||
async _doCalculateWitness(input, sanityCheck) {
|
||||
//input is assumed to be a map from signals to arrays of bigints
|
||||
this.instance.exports.init((this.sanityCheck || sanityCheck) ? 1 : 0);
|
||||
const keys = Object.keys(input);
|
||||
var input_counter = 0;
|
||||
keys.forEach( (k) => {
|
||||
const h = fnvHash(k);
|
||||
const hMSB = parseInt(h.slice(0,8), 16);
|
||||
const hLSB = parseInt(h.slice(8,16), 16);
|
||||
const fArr = flatArray(input[k]);
|
||||
let signalSize = this.instance.exports.getInputSignalSize(hMSB, hLSB);
|
||||
if (signalSize < 0){
|
||||
throw new Error(`Signal ${k} not found\n`);
|
||||
}
|
||||
if (fArr.length < signalSize) {
|
||||
throw new Error(`Not enough values for input signal ${k}\n`);
|
||||
}
|
||||
if (fArr.length > signalSize) {
|
||||
throw new Error(`Too many values for input signal ${k}\n`);
|
||||
}
|
||||
for (let i=0; i<fArr.length; i++) {
|
||||
const arrFr = toArray32(BigInt(fArr[i])%this.prime,this.n32)
|
||||
for (let j=0; j<this.n32; j++) {
|
||||
this.instance.exports.writeSharedRWMemory(j,arrFr[this.n32-1-j]);
|
||||
}
|
||||
try {
|
||||
this.instance.exports.setInputSignal(hMSB, hLSB,i);
|
||||
input_counter++;
|
||||
} catch (err) {
|
||||
// console.log(`After adding signal ${i} of ${k}`)
|
||||
throw new Error(err);
|
||||
}
|
||||
}
|
||||
|
||||
});
|
||||
if (input_counter < this.instance.exports.getInputSize()) {
|
||||
throw new Error(`Not all inputs have been set. Only ${input_counter} out of ${this.instance.exports.getInputSize()}`);
|
||||
}
|
||||
}
|
||||
|
||||
async calculateWitness(input, sanityCheck) {
|
||||
|
||||
const w = [];
|
||||
|
||||
await this._doCalculateWitness(input, sanityCheck);
|
||||
|
||||
for (let i=0; i<this.witnessSize; i++) {
|
||||
this.instance.exports.getWitness(i);
|
||||
const arr = new Uint32Array(this.n32);
|
||||
for (let j=0; j<this.n32; j++) {
|
||||
arr[this.n32-1-j] = this.instance.exports.readSharedRWMemory(j);
|
||||
}
|
||||
w.push(fromArray32(arr));
|
||||
try {
|
||||
this.instance.exports.setInputSignal(hMSB, hLSB, i);
|
||||
input_counter++;
|
||||
} catch (err) {
|
||||
// console.log(`After adding signal ${i} of ${k}`)
|
||||
throw new Error(err);
|
||||
}
|
||||
|
||||
return w;
|
||||
}
|
||||
});
|
||||
if (input_counter < this.instance.exports.getInputSize()) {
|
||||
throw new Error(
|
||||
`Not all inputs have been set. Only ${input_counter} out of ${this.instance.exports.getInputSize()}`
|
||||
);
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
async calculateBinWitness(input, sanityCheck) {
|
||||
async calculateWitness(input, sanityCheck) {
|
||||
const w = [];
|
||||
|
||||
const buff32 = new Uint32Array(this.witnessSize*this.n32);
|
||||
const buff = new Uint8Array( buff32.buffer);
|
||||
await this._doCalculateWitness(input, sanityCheck);
|
||||
await this._doCalculateWitness(input, sanityCheck);
|
||||
|
||||
for (let i=0; i<this.witnessSize; i++) {
|
||||
this.instance.exports.getWitness(i);
|
||||
const pos = i*this.n32;
|
||||
for (let j=0; j<this.n32; j++) {
|
||||
buff32[pos+j] = this.instance.exports.readSharedRWMemory(j);
|
||||
}
|
||||
}
|
||||
|
||||
return buff;
|
||||
}
|
||||
|
||||
|
||||
async calculateWTNSBin(input, sanityCheck) {
|
||||
|
||||
const buff32 = new Uint32Array(this.witnessSize*this.n32+this.n32+11);
|
||||
const buff = new Uint8Array( buff32.buffer);
|
||||
await this._doCalculateWitness(input, sanityCheck);
|
||||
|
||||
//"wtns"
|
||||
buff[0] = "w".charCodeAt(0)
|
||||
buff[1] = "t".charCodeAt(0)
|
||||
buff[2] = "n".charCodeAt(0)
|
||||
buff[3] = "s".charCodeAt(0)
|
||||
|
||||
//version 2
|
||||
buff32[1] = 2;
|
||||
|
||||
//number of sections: 2
|
||||
buff32[2] = 2;
|
||||
|
||||
//id section 1
|
||||
buff32[3] = 1;
|
||||
|
||||
const n8 = this.n32*4;
|
||||
//id section 1 length in 64bytes
|
||||
const idSection1length = 8 + n8;
|
||||
const idSection1lengthHex = idSection1length.toString(16);
|
||||
buff32[4] = parseInt(idSection1lengthHex.slice(0,8), 16);
|
||||
buff32[5] = parseInt(idSection1lengthHex.slice(8,16), 16);
|
||||
|
||||
//this.n32
|
||||
buff32[6] = n8;
|
||||
|
||||
//prime number
|
||||
this.instance.exports.getRawPrime();
|
||||
|
||||
var pos = 7;
|
||||
for (let j=0; j<this.n32; j++) {
|
||||
buff32[pos+j] = this.instance.exports.readSharedRWMemory(j);
|
||||
}
|
||||
pos += this.n32;
|
||||
|
||||
// witness size
|
||||
buff32[pos] = this.witnessSize;
|
||||
pos++;
|
||||
|
||||
//id section 2
|
||||
buff32[pos] = 2;
|
||||
pos++;
|
||||
|
||||
// section 2 length
|
||||
const idSection2length = n8*this.witnessSize;
|
||||
const idSection2lengthHex = idSection2length.toString(16);
|
||||
buff32[pos] = parseInt(idSection2lengthHex.slice(0,8), 16);
|
||||
buff32[pos+1] = parseInt(idSection2lengthHex.slice(8,16), 16);
|
||||
|
||||
pos += 2;
|
||||
for (let i=0; i<this.witnessSize; i++) {
|
||||
this.instance.exports.getWitness(i);
|
||||
for (let j=0; j<this.n32; j++) {
|
||||
buff32[pos+j] = this.instance.exports.readSharedRWMemory(j);
|
||||
}
|
||||
pos += this.n32;
|
||||
}
|
||||
|
||||
return buff;
|
||||
for (let i = 0; i < this.witnessSize; i++) {
|
||||
this.instance.exports.getWitness(i);
|
||||
const arr = new Uint32Array(this.n32);
|
||||
for (let j = 0; j < this.n32; j++) {
|
||||
arr[this.n32 - 1 - j] = this.instance.exports.readSharedRWMemory(j);
|
||||
}
|
||||
w.push(fromArray32(arr));
|
||||
}
|
||||
|
||||
return w;
|
||||
}
|
||||
|
||||
async calculateBinWitness(input, sanityCheck) {
|
||||
const buff32 = new Uint32Array(this.witnessSize * this.n32);
|
||||
const buff = new Uint8Array(buff32.buffer);
|
||||
await this._doCalculateWitness(input, sanityCheck);
|
||||
|
||||
for (let i = 0; i < this.witnessSize; i++) {
|
||||
this.instance.exports.getWitness(i);
|
||||
const pos = i * this.n32;
|
||||
for (let j = 0; j < this.n32; j++) {
|
||||
buff32[pos + j] = this.instance.exports.readSharedRWMemory(j);
|
||||
}
|
||||
}
|
||||
|
||||
return buff;
|
||||
}
|
||||
|
||||
async calculateWTNSBin(input, sanityCheck) {
|
||||
const buff32 = new Uint32Array(this.witnessSize * this.n32 + this.n32 + 11);
|
||||
const buff = new Uint8Array(buff32.buffer);
|
||||
await this._doCalculateWitness(input, sanityCheck);
|
||||
|
||||
//"wtns"
|
||||
buff[0] = "w".charCodeAt(0);
|
||||
buff[1] = "t".charCodeAt(0);
|
||||
buff[2] = "n".charCodeAt(0);
|
||||
buff[3] = "s".charCodeAt(0);
|
||||
|
||||
//version 2
|
||||
buff32[1] = 2;
|
||||
|
||||
//number of sections: 2
|
||||
buff32[2] = 2;
|
||||
|
||||
//id section 1
|
||||
buff32[3] = 1;
|
||||
|
||||
const n8 = this.n32 * 4;
|
||||
//id section 1 length in 64bytes
|
||||
const idSection1length = 8 + n8;
|
||||
const idSection1lengthHex = idSection1length.toString(16);
|
||||
buff32[4] = parseInt(idSection1lengthHex.slice(0, 8), 16);
|
||||
buff32[5] = parseInt(idSection1lengthHex.slice(8, 16), 16);
|
||||
|
||||
//this.n32
|
||||
buff32[6] = n8;
|
||||
|
||||
//prime number
|
||||
this.instance.exports.getRawPrime();
|
||||
|
||||
var pos = 7;
|
||||
for (let j = 0; j < this.n32; j++) {
|
||||
buff32[pos + j] = this.instance.exports.readSharedRWMemory(j);
|
||||
}
|
||||
pos += this.n32;
|
||||
|
||||
// witness size
|
||||
buff32[pos] = this.witnessSize;
|
||||
pos++;
|
||||
|
||||
//id section 2
|
||||
buff32[pos] = 2;
|
||||
pos++;
|
||||
|
||||
// section 2 length
|
||||
const idSection2length = n8 * this.witnessSize;
|
||||
const idSection2lengthHex = idSection2length.toString(16);
|
||||
buff32[pos] = parseInt(idSection2lengthHex.slice(0, 8), 16);
|
||||
buff32[pos + 1] = parseInt(idSection2lengthHex.slice(8, 16), 16);
|
||||
|
||||
pos += 2;
|
||||
for (let i = 0; i < this.witnessSize; i++) {
|
||||
this.instance.exports.getWitness(i);
|
||||
for (let j = 0; j < this.n32; j++) {
|
||||
buff32[pos + j] = this.instance.exports.readSharedRWMemory(j);
|
||||
}
|
||||
pos += this.n32;
|
||||
}
|
||||
|
||||
return buff;
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
function toArray32(rem,size) {
|
||||
const res = []; //new Uint32Array(size); //has no unshift
|
||||
const radix = BigInt(0x100000000);
|
||||
while (rem) {
|
||||
res.unshift( Number(rem % radix));
|
||||
rem = rem / radix;
|
||||
function toArray32(rem, size) {
|
||||
const res = []; //new Uint32Array(size); //has no unshift
|
||||
const radix = BigInt(0x100000000);
|
||||
while (rem) {
|
||||
res.unshift(Number(rem % radix));
|
||||
rem = rem / radix;
|
||||
}
|
||||
if (size) {
|
||||
var i = size - res.length;
|
||||
while (i > 0) {
|
||||
res.unshift(0);
|
||||
i--;
|
||||
}
|
||||
if (size) {
|
||||
var i = size - res.length;
|
||||
while (i>0) {
|
||||
res.unshift(0);
|
||||
i--;
|
||||
}
|
||||
}
|
||||
return res;
|
||||
}
|
||||
return res;
|
||||
}
|
||||
|
||||
function fromArray32(arr) { //returns a BigInt
|
||||
var res = BigInt(0);
|
||||
const radix = BigInt(0x100000000);
|
||||
for (let i = 0; i<arr.length; i++) {
|
||||
res = res*radix + BigInt(arr[i]);
|
||||
}
|
||||
return res;
|
||||
function fromArray32(arr) {
|
||||
//returns a BigInt
|
||||
var res = BigInt(0);
|
||||
const radix = BigInt(0x100000000);
|
||||
for (let i = 0; i < arr.length; i++) {
|
||||
res = res * radix + BigInt(arr[i]);
|
||||
}
|
||||
return res;
|
||||
}
|
||||
|
||||
function flatArray(a) {
|
||||
var res = [];
|
||||
fillArray(res, a);
|
||||
return res;
|
||||
var res = [];
|
||||
fillArray(res, a);
|
||||
return res;
|
||||
|
||||
function fillArray(res, a) {
|
||||
if (Array.isArray(a)) {
|
||||
for (let i=0; i<a.length; i++) {
|
||||
fillArray(res, a[i]);
|
||||
}
|
||||
} else {
|
||||
res.push(a);
|
||||
}
|
||||
function fillArray(res, a) {
|
||||
if (Array.isArray(a)) {
|
||||
for (let i = 0; i < a.length; i++) {
|
||||
fillArray(res, a[i]);
|
||||
}
|
||||
} else {
|
||||
res.push(a);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
function fnvHash(str) {
|
||||
const uint64_max = BigInt(2) ** BigInt(64);
|
||||
let hash = BigInt("0xCBF29CE484222325");
|
||||
for (var i = 0; i < str.length; i++) {
|
||||
hash ^= BigInt(str[i].charCodeAt());
|
||||
hash *= BigInt(0x100000001B3);
|
||||
hash %= uint64_max;
|
||||
}
|
||||
let shash = hash.toString(16);
|
||||
let n = 16 - shash.length;
|
||||
shash = '0'.repeat(n).concat(shash);
|
||||
return shash;
|
||||
const uint64_max = BigInt(2) ** BigInt(64);
|
||||
let hash = BigInt("0xCBF29CE484222325");
|
||||
for (var i = 0; i < str.length; i++) {
|
||||
hash ^= BigInt(str[i].charCodeAt());
|
||||
hash *= BigInt(0x100000001b3);
|
||||
hash %= uint64_max;
|
||||
}
|
||||
let shash = hash.toString(16);
|
||||
let n = 16 - shash.length;
|
||||
shash = "0".repeat(n).concat(shash);
|
||||
return shash;
|
||||
}
|
||||
|
||||
@@ -1,320 +1,16 @@
|
||||
#![cfg(target_arch = "wasm32")]
|
||||
|
||||
extern crate wasm_bindgen;
|
||||
extern crate web_sys;
|
||||
pub mod wasm_rln;
|
||||
pub mod wasm_utils;
|
||||
|
||||
use std::vec::Vec;
|
||||
#[cfg(all(feature = "parallel", not(feature = "utils")))]
|
||||
pub use wasm_bindgen_rayon::init_thread_pool;
|
||||
#[cfg(not(feature = "utils"))]
|
||||
pub use wasm_rln::{WasmRLN, WasmRLNProof, WasmRLNProofValues, WasmRLNWitnessInput};
|
||||
pub use wasm_utils::{ExtendedIdentity, Hasher, Identity, VecWasmFr, WasmFr};
|
||||
|
||||
use js_sys::{BigInt as JsBigInt, Object, Uint8Array};
|
||||
use num_bigint::BigInt;
|
||||
use rln::public::{hash, poseidon_hash, RLN};
|
||||
use wasm_bindgen::prelude::*;
|
||||
|
||||
#[wasm_bindgen]
|
||||
#[cfg(feature = "panic_hook")]
|
||||
#[wasm_bindgen(js_name = initPanicHook)]
|
||||
pub fn init_panic_hook() {
|
||||
console_error_panic_hook::set_once();
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = RLN)]
|
||||
pub struct RLNWrapper {
|
||||
// The purpose of this wrapper is to hold a RLN instance with the 'static lifetime
|
||||
// because wasm_bindgen does not allow returning elements with lifetimes
|
||||
instance: RLN,
|
||||
}
|
||||
|
||||
// Macro to call methods with arbitrary amount of arguments,
|
||||
// which have the last argument is output buffer pointer
|
||||
// First argument to the macro is context,
|
||||
// second is the actual method on `RLN`
|
||||
// third is the aforementioned output buffer argument
|
||||
// rest are all other arguments to the method
|
||||
macro_rules! call_with_output_and_error_msg {
|
||||
// this variant is needed for the case when
|
||||
// there are zero other arguments
|
||||
($instance:expr, $method:ident, $error_msg:expr) => {
|
||||
{
|
||||
let mut output_data: Vec<u8> = Vec::new();
|
||||
let new_instance = $instance.process();
|
||||
if let Err(err) = new_instance.instance.$method(&mut output_data) {
|
||||
std::mem::forget(output_data);
|
||||
Err(format!("Msg: {:#?}, Error: {:#?}", $error_msg, err))
|
||||
} else {
|
||||
let result = Uint8Array::from(&output_data[..]);
|
||||
std::mem::forget(output_data);
|
||||
Ok(result)
|
||||
}
|
||||
}
|
||||
};
|
||||
($instance:expr, $method:ident, $error_msg:expr, $( $arg:expr ),* ) => {
|
||||
{
|
||||
let mut output_data: Vec<u8> = Vec::new();
|
||||
let new_instance = $instance.process();
|
||||
if let Err(err) = new_instance.instance.$method($($arg.process()),*, &mut output_data) {
|
||||
std::mem::forget(output_data);
|
||||
Err(format!("Msg: {:#?}, Error: {:#?}", $error_msg, err))
|
||||
} else {
|
||||
let result = Uint8Array::from(&output_data[..]);
|
||||
std::mem::forget(output_data);
|
||||
Ok(result)
|
||||
}
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
// Macro to call_with_error_msg methods with arbitrary amount of arguments,
|
||||
// First argument to the macro is context,
|
||||
// second is the actual method on `RLNWrapper`
|
||||
// rest are all other arguments to the method
|
||||
macro_rules! call_with_error_msg {
|
||||
($instance:expr, $method:ident, $error_msg:expr $(, $arg:expr)*) => {
|
||||
{
|
||||
let new_instance: &mut RLNWrapper = $instance.process();
|
||||
if let Err(err) = new_instance.instance.$method($($arg.process()),*) {
|
||||
Err(format!("Msg: {:#?}, Error: {:#?}", $error_msg, err))
|
||||
} else {
|
||||
Ok(())
|
||||
}
|
||||
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
macro_rules! call {
|
||||
($instance:expr, $method:ident $(, $arg:expr)*) => {
|
||||
{
|
||||
let new_instance: &mut RLNWrapper = $instance.process();
|
||||
new_instance.instance.$method($($arg.process()),*)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
macro_rules! call_bool_method_with_error_msg {
|
||||
($instance:expr, $method:ident, $error_msg:expr $(, $arg:expr)*) => {
|
||||
{
|
||||
let new_instance: &RLNWrapper = $instance.process();
|
||||
new_instance.instance.$method($($arg.process()),*).map_err(|err| format!("Msg: {:#?}, Error: {:#?}", $error_msg, err))
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Macro to execute a function with arbitrary amount of arguments,
|
||||
// First argument is the function to execute
|
||||
// Rest are all other arguments to the method
|
||||
macro_rules! fn_call_with_output_and_error_msg {
|
||||
// this variant is needed for the case when
|
||||
// there are zero other arguments
|
||||
($func:ident, $error_msg:expr) => {
|
||||
{
|
||||
let mut output_data: Vec<u8> = Vec::new();
|
||||
if let Err(err) = $func(&mut output_data) {
|
||||
std::mem::forget(output_data);
|
||||
Err(format!("Msg: {:#?}, Error: {:#?}", $error_msg, err))
|
||||
} else {
|
||||
let result = Uint8Array::from(&output_data[..]);
|
||||
std::mem::forget(output_data);
|
||||
Ok(result)
|
||||
}
|
||||
}
|
||||
};
|
||||
($func:ident, $error_msg:expr, $( $arg:expr ),* ) => {
|
||||
{
|
||||
let mut output_data: Vec<u8> = Vec::new();
|
||||
if let Err(err) = $func($($arg.process()),*, &mut output_data) {
|
||||
std::mem::forget(output_data);
|
||||
Err(format!("Msg: {:#?}, Error: {:#?}", $error_msg, err))
|
||||
} else {
|
||||
let result = Uint8Array::from(&output_data[..]);
|
||||
std::mem::forget(output_data);
|
||||
Ok(result)
|
||||
}
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
trait ProcessArg {
|
||||
type ReturnType;
|
||||
fn process(self) -> Self::ReturnType;
|
||||
}
|
||||
|
||||
impl ProcessArg for usize {
|
||||
type ReturnType = usize;
|
||||
fn process(self) -> Self::ReturnType {
|
||||
self
|
||||
}
|
||||
}
|
||||
|
||||
impl<T> ProcessArg for Vec<T> {
|
||||
type ReturnType = Vec<T>;
|
||||
fn process(self) -> Self::ReturnType {
|
||||
self
|
||||
}
|
||||
}
|
||||
|
||||
impl ProcessArg for *const RLN {
|
||||
type ReturnType = &'static RLN;
|
||||
fn process(self) -> Self::ReturnType {
|
||||
unsafe { &*self }
|
||||
}
|
||||
}
|
||||
|
||||
impl ProcessArg for *const RLNWrapper {
|
||||
type ReturnType = &'static RLNWrapper;
|
||||
fn process(self) -> Self::ReturnType {
|
||||
unsafe { &*self }
|
||||
}
|
||||
}
|
||||
|
||||
impl ProcessArg for *mut RLNWrapper {
|
||||
type ReturnType = &'static mut RLNWrapper;
|
||||
fn process(self) -> Self::ReturnType {
|
||||
unsafe { &mut *self }
|
||||
}
|
||||
}
|
||||
|
||||
impl<'a> ProcessArg for &'a [u8] {
|
||||
type ReturnType = &'a [u8];
|
||||
|
||||
fn process(self) -> Self::ReturnType {
|
||||
self
|
||||
}
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[wasm_bindgen(js_name = newRLN)]
|
||||
pub fn wasm_new(
|
||||
zkey: Uint8Array,
|
||||
vk: Uint8Array,
|
||||
) -> Result<*mut RLNWrapper, String> {
|
||||
let instance = RLN::new_with_params(zkey.to_vec(), vk.to_vec())
|
||||
.map_err(|err| format!("{:#?}", err))?;
|
||||
let wrapper = RLNWrapper { instance };
|
||||
Ok(Box::into_raw(Box::new(wrapper)))
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[wasm_bindgen(js_name = RLNWitnessToJson)]
|
||||
pub fn rln_witness_to_json(
|
||||
ctx: *mut RLNWrapper,
|
||||
serialized_witness: Uint8Array,
|
||||
) -> Result<Object, String> {
|
||||
let inputs = call!(
|
||||
ctx,
|
||||
get_rln_witness_bigint_json,
|
||||
&serialized_witness.to_vec()[..]
|
||||
)
|
||||
.map_err(|err| err.to_string())?;
|
||||
let js_value = serde_wasm_bindgen::to_value(&inputs).map_err(|err| err.to_string())?;
|
||||
Object::from_entries(&js_value).map_err(|err| format!("{:#?}", err))
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[wasm_bindgen]
|
||||
pub fn generate_rln_proof_with_witness(
|
||||
ctx: *mut RLNWrapper,
|
||||
calculated_witness: Vec<JsBigInt>,
|
||||
serialized_witness: Uint8Array,
|
||||
) -> Result<Uint8Array, String> {
|
||||
let mut witness_vec: Vec<BigInt> = vec![];
|
||||
|
||||
for v in calculated_witness {
|
||||
witness_vec.push(
|
||||
v.to_string(10)
|
||||
.map_err(|err| format!("{:#?}", err))?
|
||||
.as_string()
|
||||
.ok_or("not a string error")?
|
||||
.parse::<BigInt>()
|
||||
.map_err(|err| format!("{:#?}", err))?,
|
||||
);
|
||||
}
|
||||
|
||||
call_with_output_and_error_msg!(
|
||||
ctx,
|
||||
generate_rln_proof_with_witness,
|
||||
"could not generate proof",
|
||||
witness_vec,
|
||||
serialized_witness.to_vec()
|
||||
)
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[wasm_bindgen(js_name = generateMembershipKey)]
|
||||
pub fn wasm_key_gen(ctx: *const RLNWrapper) -> Result<Uint8Array, String> {
|
||||
call_with_output_and_error_msg!(ctx, key_gen, "could not generate membership keys")
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[wasm_bindgen(js_name = generateExtendedMembershipKey)]
|
||||
pub fn wasm_extended_key_gen(ctx: *const RLNWrapper) -> Result<Uint8Array, String> {
|
||||
call_with_output_and_error_msg!(ctx, extended_key_gen, "could not generate membership keys")
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[wasm_bindgen(js_name = generateSeededMembershipKey)]
|
||||
pub fn wasm_seeded_key_gen(ctx: *const RLNWrapper, seed: Uint8Array) -> Result<Uint8Array, String> {
|
||||
call_with_output_and_error_msg!(
|
||||
ctx,
|
||||
seeded_key_gen,
|
||||
"could not generate membership key",
|
||||
&seed.to_vec()[..]
|
||||
)
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[wasm_bindgen(js_name = generateSeededExtendedMembershipKey)]
|
||||
pub fn wasm_seeded_extended_key_gen(
|
||||
ctx: *const RLNWrapper,
|
||||
seed: Uint8Array,
|
||||
) -> Result<Uint8Array, String> {
|
||||
call_with_output_and_error_msg!(
|
||||
ctx,
|
||||
seeded_extended_key_gen,
|
||||
"could not generate membership key",
|
||||
&seed.to_vec()[..]
|
||||
)
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[wasm_bindgen(js_name = recovedIDSecret)]
|
||||
pub fn wasm_recover_id_secret(
|
||||
ctx: *const RLNWrapper,
|
||||
input_proof_data_1: Uint8Array,
|
||||
input_proof_data_2: Uint8Array,
|
||||
) -> Result<Uint8Array, String> {
|
||||
call_with_output_and_error_msg!(
|
||||
ctx,
|
||||
recover_id_secret,
|
||||
"could not recover id secret",
|
||||
&input_proof_data_1.to_vec()[..],
|
||||
&input_proof_data_2.to_vec()[..]
|
||||
)
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[wasm_bindgen(js_name = verifyWithRoots)]
|
||||
pub fn wasm_verify_with_roots(
|
||||
ctx: *const RLNWrapper,
|
||||
proof: Uint8Array,
|
||||
roots: Uint8Array,
|
||||
) -> Result<bool, String> {
|
||||
call_bool_method_with_error_msg!(
|
||||
ctx,
|
||||
verify_with_roots,
|
||||
"error while verifying proof with roots".to_string(),
|
||||
&proof.to_vec()[..],
|
||||
&roots.to_vec()[..]
|
||||
)
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = hash)]
|
||||
pub fn wasm_hash(input: Uint8Array) -> Result<Uint8Array, String> {
|
||||
fn_call_with_output_and_error_msg!(hash, "could not generate hash", &input.to_vec()[..])
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = poseidonHash)]
|
||||
pub fn wasm_poseidon_hash(input: Uint8Array) -> Result<Uint8Array, String> {
|
||||
fn_call_with_output_and_error_msg!(
|
||||
poseidon_hash,
|
||||
"could not generate poseidon hash",
|
||||
&input.to_vec()[..]
|
||||
)
|
||||
}
|
||||
|
||||
@@ -1,18 +0,0 @@
|
||||
const fs = require("fs");
|
||||
|
||||
// Utils functions for loading circom witness calculator and reading files from test
|
||||
|
||||
module.exports = {
|
||||
read_file: function (path) {
|
||||
return fs.readFileSync(path);
|
||||
},
|
||||
|
||||
calculateWitness: async function(circom_path, inputs){
|
||||
const wc = require("resources/witness_calculator.js");
|
||||
const wasmFile = fs.readFileSync(circom_path);
|
||||
const wasmFileBuffer = wasmFile.slice(wasmFile.byteOffset, wasmFile.byteOffset + wasmFile.byteLength);
|
||||
const witnessCalculator = await wc(wasmFileBuffer);
|
||||
const calculatedWitness = await witnessCalculator.calculateWitness(inputs, false);
|
||||
return JSON.stringify(calculatedWitness, (key, value) => typeof value === "bigint" ? value.toString() : value);
|
||||
}
|
||||
}
|
||||
253
rln-wasm/src/wasm_rln.rs
Normal file
253
rln-wasm/src/wasm_rln.rs
Normal file
@@ -0,0 +1,253 @@
|
||||
#![cfg(target_arch = "wasm32")]
|
||||
#![cfg(not(feature = "utils"))]
|
||||
|
||||
use js_sys::{BigInt as JsBigInt, Object, Uint8Array};
|
||||
use num_bigint::BigInt;
|
||||
use rln::prelude::*;
|
||||
use serde::Serialize;
|
||||
use wasm_bindgen::prelude::*;
|
||||
|
||||
use crate::wasm_utils::{VecWasmFr, WasmFr};
|
||||
|
||||
#[wasm_bindgen]
|
||||
pub struct WasmRLN(RLN);
|
||||
|
||||
#[wasm_bindgen]
|
||||
impl WasmRLN {
|
||||
#[wasm_bindgen(constructor)]
|
||||
pub fn new(zkey_data: &Uint8Array) -> Result<WasmRLN, String> {
|
||||
let rln = RLN::new_with_params(zkey_data.to_vec()).map_err(|err| err.to_string())?;
|
||||
Ok(WasmRLN(rln))
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = generateRLNProofWithWitness)]
|
||||
pub fn generate_rln_proof_with_witness(
|
||||
&self,
|
||||
calculated_witness: Vec<JsBigInt>,
|
||||
witness: &WasmRLNWitnessInput,
|
||||
) -> Result<WasmRLNProof, String> {
|
||||
let calculated_witness_bigint: Vec<BigInt> = calculated_witness
|
||||
.iter()
|
||||
.map(|js_bigint| {
|
||||
js_bigint
|
||||
.to_string(10)
|
||||
.ok()
|
||||
.and_then(|js_str| js_str.as_string())
|
||||
.ok_or_else(|| "Failed to convert JsBigInt to string".to_string())
|
||||
.and_then(|str_val| {
|
||||
str_val
|
||||
.parse::<BigInt>()
|
||||
.map_err(|err| format!("Failed to parse BigInt: {}", err))
|
||||
})
|
||||
})
|
||||
.collect::<Result<Vec<_>, _>>()?;
|
||||
|
||||
let (proof, proof_values) = self
|
||||
.0
|
||||
.generate_rln_proof_with_witness(calculated_witness_bigint, &witness.0)
|
||||
.map_err(|err| err.to_string())?;
|
||||
|
||||
let rln_proof = RLNProof {
|
||||
proof_values,
|
||||
proof,
|
||||
};
|
||||
|
||||
Ok(WasmRLNProof(rln_proof))
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = verifyWithRoots)]
|
||||
pub fn verify_with_roots(
|
||||
&self,
|
||||
rln_proof: &WasmRLNProof,
|
||||
roots: &VecWasmFr,
|
||||
x: &WasmFr,
|
||||
) -> Result<bool, String> {
|
||||
let roots_fr: Vec<Fr> = (0..roots.length())
|
||||
.filter_map(|i| roots.get(i))
|
||||
.map(|root| *root)
|
||||
.collect();
|
||||
|
||||
self.0
|
||||
.verify_with_roots(&rln_proof.0.proof, &rln_proof.0.proof_values, x, &roots_fr)
|
||||
.map_err(|err| err.to_string())
|
||||
}
|
||||
}
|
||||
|
||||
#[wasm_bindgen]
|
||||
pub struct WasmRLNProof(RLNProof);
|
||||
|
||||
#[wasm_bindgen]
|
||||
impl WasmRLNProof {
|
||||
#[wasm_bindgen(js_name = getValues)]
|
||||
pub fn get_values(&self) -> WasmRLNProofValues {
|
||||
WasmRLNProofValues(self.0.proof_values)
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = toBytesLE)]
|
||||
pub fn to_bytes_le(&self) -> Result<Uint8Array, String> {
|
||||
let bytes = rln_proof_to_bytes_le(&self.0).map_err(|err| err.to_string())?;
|
||||
Ok(Uint8Array::from(&bytes[..]))
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = toBytesBE)]
|
||||
pub fn to_bytes_be(&self) -> Result<Uint8Array, String> {
|
||||
let bytes = rln_proof_to_bytes_be(&self.0).map_err(|err| err.to_string())?;
|
||||
Ok(Uint8Array::from(&bytes[..]))
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = fromBytesLE)]
|
||||
pub fn from_bytes_le(bytes: &Uint8Array) -> Result<WasmRLNProof, String> {
|
||||
let bytes_vec = bytes.to_vec();
|
||||
let (proof, _) = bytes_le_to_rln_proof(&bytes_vec).map_err(|err| err.to_string())?;
|
||||
Ok(WasmRLNProof(proof))
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = fromBytesBE)]
|
||||
pub fn from_bytes_be(bytes: &Uint8Array) -> Result<WasmRLNProof, String> {
|
||||
let bytes_vec = bytes.to_vec();
|
||||
let (proof, _) = bytes_be_to_rln_proof(&bytes_vec).map_err(|err| err.to_string())?;
|
||||
Ok(WasmRLNProof(proof))
|
||||
}
|
||||
}
|
||||
|
||||
#[wasm_bindgen]
|
||||
pub struct WasmRLNProofValues(RLNProofValues);
|
||||
|
||||
#[wasm_bindgen]
|
||||
impl WasmRLNProofValues {
|
||||
#[wasm_bindgen(getter)]
|
||||
pub fn y(&self) -> WasmFr {
|
||||
WasmFr::from(self.0.y)
|
||||
}
|
||||
|
||||
#[wasm_bindgen(getter)]
|
||||
pub fn nullifier(&self) -> WasmFr {
|
||||
WasmFr::from(self.0.nullifier)
|
||||
}
|
||||
|
||||
#[wasm_bindgen(getter)]
|
||||
pub fn root(&self) -> WasmFr {
|
||||
WasmFr::from(self.0.root)
|
||||
}
|
||||
|
||||
#[wasm_bindgen(getter)]
|
||||
pub fn x(&self) -> WasmFr {
|
||||
WasmFr::from(self.0.x)
|
||||
}
|
||||
|
||||
#[wasm_bindgen(getter, js_name = externalNullifier)]
|
||||
pub fn external_nullifier(&self) -> WasmFr {
|
||||
WasmFr::from(self.0.external_nullifier)
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = toBytesLE)]
|
||||
pub fn to_bytes_le(&self) -> Uint8Array {
|
||||
Uint8Array::from(&rln_proof_values_to_bytes_le(&self.0)[..])
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = toBytesBE)]
|
||||
pub fn to_bytes_be(&self) -> Uint8Array {
|
||||
Uint8Array::from(&rln_proof_values_to_bytes_be(&self.0)[..])
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = fromBytesLE)]
|
||||
pub fn from_bytes_le(bytes: &Uint8Array) -> Result<WasmRLNProofValues, String> {
|
||||
let bytes_vec = bytes.to_vec();
|
||||
let (proof_values, _) =
|
||||
bytes_le_to_rln_proof_values(&bytes_vec).map_err(|err| err.to_string())?;
|
||||
Ok(WasmRLNProofValues(proof_values))
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = fromBytesBE)]
|
||||
pub fn from_bytes_be(bytes: &Uint8Array) -> Result<WasmRLNProofValues, String> {
|
||||
let bytes_vec = bytes.to_vec();
|
||||
let (proof_values, _) =
|
||||
bytes_be_to_rln_proof_values(&bytes_vec).map_err(|err| err.to_string())?;
|
||||
Ok(WasmRLNProofValues(proof_values))
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = recoverIdSecret)]
|
||||
pub fn recover_id_secret(
|
||||
proof_values_1: &WasmRLNProofValues,
|
||||
proof_values_2: &WasmRLNProofValues,
|
||||
) -> Result<WasmFr, String> {
|
||||
let recovered_identity_secret = recover_id_secret(&proof_values_1.0, &proof_values_2.0)
|
||||
.map_err(|err| err.to_string())?;
|
||||
|
||||
Ok(WasmFr::from(*recovered_identity_secret))
|
||||
}
|
||||
}
|
||||
|
||||
#[wasm_bindgen]
|
||||
pub struct WasmRLNWitnessInput(RLNWitnessInput);
|
||||
|
||||
#[wasm_bindgen]
|
||||
impl WasmRLNWitnessInput {
|
||||
#[wasm_bindgen(constructor)]
|
||||
pub fn new(
|
||||
identity_secret: &WasmFr,
|
||||
user_message_limit: &WasmFr,
|
||||
message_id: &WasmFr,
|
||||
path_elements: &VecWasmFr,
|
||||
identity_path_index: &Uint8Array,
|
||||
x: &WasmFr,
|
||||
external_nullifier: &WasmFr,
|
||||
) -> Result<WasmRLNWitnessInput, String> {
|
||||
let mut identity_secret_fr = identity_secret.inner();
|
||||
let path_elements: Vec<Fr> = path_elements.inner();
|
||||
let identity_path_index: Vec<u8> = identity_path_index.to_vec();
|
||||
|
||||
let witness = RLNWitnessInput::new(
|
||||
IdSecret::from(&mut identity_secret_fr),
|
||||
user_message_limit.inner(),
|
||||
message_id.inner(),
|
||||
path_elements,
|
||||
identity_path_index,
|
||||
x.inner(),
|
||||
external_nullifier.inner(),
|
||||
)
|
||||
.map_err(|err| err.to_string())?;
|
||||
|
||||
Ok(WasmRLNWitnessInput(witness))
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = toBigIntJson)]
|
||||
pub fn to_bigint_json(&self) -> Result<Object, String> {
|
||||
let bigint_json = rln_witness_to_bigint_json(&self.0).map_err(|err| err.to_string())?;
|
||||
|
||||
let serializer = serde_wasm_bindgen::Serializer::json_compatible();
|
||||
let js_value = bigint_json
|
||||
.serialize(&serializer)
|
||||
.map_err(|err| err.to_string())?;
|
||||
|
||||
js_value
|
||||
.dyn_into::<Object>()
|
||||
.map_err(|err| format!("{:#?}", err))
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = toBytesLE)]
|
||||
pub fn to_bytes_le(&self) -> Result<Uint8Array, String> {
|
||||
let bytes = rln_witness_to_bytes_le(&self.0).map_err(|err| err.to_string())?;
|
||||
Ok(Uint8Array::from(&bytes[..]))
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = toBytesBE)]
|
||||
pub fn to_bytes_be(&self) -> Result<Uint8Array, String> {
|
||||
let bytes = rln_witness_to_bytes_be(&self.0).map_err(|err| err.to_string())?;
|
||||
Ok(Uint8Array::from(&bytes[..]))
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = fromBytesLE)]
|
||||
pub fn from_bytes_le(bytes: &Uint8Array) -> Result<WasmRLNWitnessInput, String> {
|
||||
let bytes_vec = bytes.to_vec();
|
||||
let (witness, _) = bytes_le_to_rln_witness(&bytes_vec).map_err(|err| err.to_string())?;
|
||||
Ok(WasmRLNWitnessInput(witness))
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = fromBytesBE)]
|
||||
pub fn from_bytes_be(bytes: &Uint8Array) -> Result<WasmRLNWitnessInput, String> {
|
||||
let bytes_vec = bytes.to_vec();
|
||||
let (witness, _) = bytes_be_to_rln_witness(&bytes_vec).map_err(|err| err.to_string())?;
|
||||
Ok(WasmRLNWitnessInput(witness))
|
||||
}
|
||||
}
|
||||
420
rln-wasm/src/wasm_utils.rs
Normal file
420
rln-wasm/src/wasm_utils.rs
Normal file
@@ -0,0 +1,420 @@
|
||||
#![cfg(target_arch = "wasm32")]
|
||||
|
||||
use std::ops::Deref;
|
||||
|
||||
use js_sys::Uint8Array;
|
||||
use rln::prelude::*;
|
||||
use wasm_bindgen::prelude::*;
|
||||
|
||||
// WasmFr
|
||||
|
||||
#[wasm_bindgen]
|
||||
#[derive(Debug, Clone, Copy, PartialEq, Default)]
|
||||
pub struct WasmFr(Fr);
|
||||
|
||||
impl From<Fr> for WasmFr {
|
||||
fn from(fr: Fr) -> Self {
|
||||
Self(fr)
|
||||
}
|
||||
}
|
||||
|
||||
impl Deref for WasmFr {
|
||||
type Target = Fr;
|
||||
fn deref(&self) -> &Self::Target {
|
||||
&self.0
|
||||
}
|
||||
}
|
||||
|
||||
#[wasm_bindgen]
|
||||
impl WasmFr {
|
||||
#[wasm_bindgen(js_name = zero)]
|
||||
pub fn zero() -> Self {
|
||||
Self(Fr::from(0u32))
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = one)]
|
||||
pub fn one() -> Self {
|
||||
Self(Fr::from(1u32))
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = fromUint)]
|
||||
pub fn from_uint(value: u32) -> Self {
|
||||
Self(Fr::from(value))
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = fromBytesLE)]
|
||||
pub fn from_bytes_le(bytes: &Uint8Array) -> Result<Self, String> {
|
||||
let bytes_vec = bytes.to_vec();
|
||||
let (fr, _) = bytes_le_to_fr(&bytes_vec).map_err(|err| err.to_string())?;
|
||||
Ok(Self(fr))
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = fromBytesBE)]
|
||||
pub fn from_bytes_be(bytes: &Uint8Array) -> Result<Self, String> {
|
||||
let bytes_vec = bytes.to_vec();
|
||||
let (fr, _) = bytes_be_to_fr(&bytes_vec).map_err(|err| err.to_string())?;
|
||||
Ok(Self(fr))
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = toBytesLE)]
|
||||
pub fn to_bytes_le(&self) -> Uint8Array {
|
||||
let bytes = fr_to_bytes_le(&self.0);
|
||||
Uint8Array::from(&bytes[..])
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = toBytesBE)]
|
||||
pub fn to_bytes_be(&self) -> Uint8Array {
|
||||
let bytes = fr_to_bytes_be(&self.0);
|
||||
Uint8Array::from(&bytes[..])
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = debug)]
|
||||
pub fn debug(&self) -> String {
|
||||
format!("{:?}", self.0)
|
||||
}
|
||||
}
|
||||
|
||||
impl WasmFr {
|
||||
pub fn inner(&self) -> Fr {
|
||||
self.0
|
||||
}
|
||||
}
|
||||
|
||||
// VecWasmFr
|
||||
|
||||
#[wasm_bindgen]
|
||||
#[derive(Debug, Clone, PartialEq, Default)]
|
||||
pub struct VecWasmFr(Vec<Fr>);
|
||||
|
||||
#[wasm_bindgen]
|
||||
impl VecWasmFr {
|
||||
#[wasm_bindgen(constructor)]
|
||||
pub fn new() -> Self {
|
||||
Self(Vec::new())
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = fromBytesLE)]
|
||||
pub fn from_bytes_le(bytes: &Uint8Array) -> Result<VecWasmFr, String> {
|
||||
let bytes_vec = bytes.to_vec();
|
||||
bytes_le_to_vec_fr(&bytes_vec)
|
||||
.map(|(vec_fr, _)| VecWasmFr(vec_fr))
|
||||
.map_err(|err| err.to_string())
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = fromBytesBE)]
|
||||
pub fn from_bytes_be(bytes: &Uint8Array) -> Result<VecWasmFr, String> {
|
||||
let bytes_vec = bytes.to_vec();
|
||||
bytes_be_to_vec_fr(&bytes_vec)
|
||||
.map(|(vec_fr, _)| VecWasmFr(vec_fr))
|
||||
.map_err(|err| err.to_string())
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = toBytesLE)]
|
||||
pub fn to_bytes_le(&self) -> Uint8Array {
|
||||
let bytes = vec_fr_to_bytes_le(&self.0);
|
||||
Uint8Array::from(&bytes[..])
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = toBytesBE)]
|
||||
pub fn to_bytes_be(&self) -> Uint8Array {
|
||||
let bytes = vec_fr_to_bytes_be(&self.0);
|
||||
Uint8Array::from(&bytes[..])
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = get)]
|
||||
pub fn get(&self, index: usize) -> Option<WasmFr> {
|
||||
self.0.get(index).map(|&fr| WasmFr(fr))
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = length)]
|
||||
pub fn length(&self) -> usize {
|
||||
self.0.len()
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = push)]
|
||||
pub fn push(&mut self, element: &WasmFr) {
|
||||
self.0.push(element.0);
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = debug)]
|
||||
pub fn debug(&self) -> String {
|
||||
format!("{:?}", self.0)
|
||||
}
|
||||
}
|
||||
|
||||
impl VecWasmFr {
|
||||
pub fn inner(&self) -> Vec<Fr> {
|
||||
self.0.clone()
|
||||
}
|
||||
}
|
||||
|
||||
// Uint8Array
|
||||
|
||||
#[wasm_bindgen]
|
||||
pub struct Uint8ArrayUtils;
|
||||
|
||||
#[wasm_bindgen]
|
||||
impl Uint8ArrayUtils {
|
||||
#[wasm_bindgen(js_name = toBytesLE)]
|
||||
pub fn to_bytes_le(input: &Uint8Array) -> Uint8Array {
|
||||
let input_vec = input.to_vec();
|
||||
let bytes = vec_u8_to_bytes_le(&input_vec);
|
||||
Uint8Array::from(&bytes[..])
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = toBytesBE)]
|
||||
pub fn to_bytes_be(input: &Uint8Array) -> Uint8Array {
|
||||
let input_vec = input.to_vec();
|
||||
let bytes = vec_u8_to_bytes_be(&input_vec);
|
||||
Uint8Array::from(&bytes[..])
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = fromBytesLE)]
|
||||
pub fn from_bytes_le(bytes: &Uint8Array) -> Result<Uint8Array, String> {
|
||||
let bytes_vec = bytes.to_vec();
|
||||
bytes_le_to_vec_u8(&bytes_vec)
|
||||
.map(|(vec_u8, _)| Uint8Array::from(&vec_u8[..]))
|
||||
.map_err(|err| err.to_string())
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = fromBytesBE)]
|
||||
pub fn from_bytes_be(bytes: &Uint8Array) -> Result<Uint8Array, String> {
|
||||
let bytes_vec = bytes.to_vec();
|
||||
bytes_be_to_vec_u8(&bytes_vec)
|
||||
.map(|(vec_u8, _)| Uint8Array::from(&vec_u8[..]))
|
||||
.map_err(|err| err.to_string())
|
||||
}
|
||||
}
|
||||
|
||||
// Utility APIs
|
||||
|
||||
#[wasm_bindgen]
|
||||
pub struct Hasher;
|
||||
|
||||
#[wasm_bindgen]
|
||||
impl Hasher {
|
||||
#[wasm_bindgen(js_name = hashToFieldLE)]
|
||||
pub fn hash_to_field_le(input: &Uint8Array) -> Result<WasmFr, String> {
|
||||
hash_to_field_le(&input.to_vec())
|
||||
.map(WasmFr)
|
||||
.map_err(|err| err.to_string())
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = hashToFieldBE)]
|
||||
pub fn hash_to_field_be(input: &Uint8Array) -> Result<WasmFr, String> {
|
||||
hash_to_field_be(&input.to_vec())
|
||||
.map(WasmFr)
|
||||
.map_err(|err| err.to_string())
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = poseidonHashPair)]
|
||||
pub fn poseidon_hash_pair(a: &WasmFr, b: &WasmFr) -> Result<WasmFr, String> {
|
||||
poseidon_hash(&[a.0, b.0])
|
||||
.map(WasmFr)
|
||||
.map_err(|err| err.to_string())
|
||||
}
|
||||
}
|
||||
|
||||
#[wasm_bindgen]
|
||||
pub struct Identity {
|
||||
identity_secret: Fr,
|
||||
id_commitment: Fr,
|
||||
}
|
||||
|
||||
#[wasm_bindgen]
|
||||
impl Identity {
|
||||
#[wasm_bindgen(js_name = generate)]
|
||||
pub fn generate() -> Result<Identity, String> {
|
||||
let (identity_secret, id_commitment) = keygen().map_err(|err| err.to_string())?;
|
||||
Ok(Identity {
|
||||
identity_secret: *identity_secret,
|
||||
id_commitment,
|
||||
})
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = generateSeeded)]
|
||||
pub fn generate_seeded(seed: &Uint8Array) -> Result<Identity, String> {
|
||||
let seed_vec = seed.to_vec();
|
||||
let (identity_secret, id_commitment) =
|
||||
seeded_keygen(&seed_vec).map_err(|err| err.to_string())?;
|
||||
Ok(Identity {
|
||||
identity_secret,
|
||||
id_commitment,
|
||||
})
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = getSecretHash)]
|
||||
pub fn get_secret_hash(&self) -> WasmFr {
|
||||
WasmFr(self.identity_secret)
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = getCommitment)]
|
||||
pub fn get_commitment(&self) -> WasmFr {
|
||||
WasmFr(self.id_commitment)
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = toArray)]
|
||||
pub fn to_array(&self) -> VecWasmFr {
|
||||
VecWasmFr(vec![self.identity_secret, self.id_commitment])
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = toBytesLE)]
|
||||
pub fn to_bytes_le(&self) -> Uint8Array {
|
||||
let vec_fr = vec![self.identity_secret, self.id_commitment];
|
||||
let bytes = vec_fr_to_bytes_le(&vec_fr);
|
||||
Uint8Array::from(&bytes[..])
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = toBytesBE)]
|
||||
pub fn to_bytes_be(&self) -> Uint8Array {
|
||||
let vec_fr = vec![self.identity_secret, self.id_commitment];
|
||||
let bytes = vec_fr_to_bytes_be(&vec_fr);
|
||||
Uint8Array::from(&bytes[..])
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = fromBytesLE)]
|
||||
pub fn from_bytes_le(bytes: &Uint8Array) -> Result<Identity, String> {
|
||||
let bytes_vec = bytes.to_vec();
|
||||
let (vec_fr, _) = bytes_le_to_vec_fr(&bytes_vec).map_err(|err| err.to_string())?;
|
||||
if vec_fr.len() != 2 {
|
||||
return Err(format!("Expected 2 elements, got {}", vec_fr.len()));
|
||||
}
|
||||
Ok(Identity {
|
||||
identity_secret: vec_fr[0],
|
||||
id_commitment: vec_fr[1],
|
||||
})
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = fromBytesBE)]
|
||||
pub fn from_bytes_be(bytes: &Uint8Array) -> Result<Identity, String> {
|
||||
let bytes_vec = bytes.to_vec();
|
||||
let (vec_fr, _) = bytes_be_to_vec_fr(&bytes_vec).map_err(|err| err.to_string())?;
|
||||
if vec_fr.len() != 2 {
|
||||
return Err(format!("Expected 2 elements, got {}", vec_fr.len()));
|
||||
}
|
||||
Ok(Identity {
|
||||
identity_secret: vec_fr[0],
|
||||
id_commitment: vec_fr[1],
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
#[wasm_bindgen]
|
||||
pub struct ExtendedIdentity {
|
||||
identity_trapdoor: Fr,
|
||||
identity_nullifier: Fr,
|
||||
identity_secret: Fr,
|
||||
id_commitment: Fr,
|
||||
}
|
||||
|
||||
#[wasm_bindgen]
|
||||
impl ExtendedIdentity {
|
||||
#[wasm_bindgen(js_name = generate)]
|
||||
pub fn generate() -> Result<ExtendedIdentity, String> {
|
||||
let (identity_trapdoor, identity_nullifier, identity_secret, id_commitment) =
|
||||
extended_keygen().map_err(|err| err.to_string())?;
|
||||
Ok(ExtendedIdentity {
|
||||
identity_trapdoor,
|
||||
identity_nullifier,
|
||||
identity_secret,
|
||||
id_commitment,
|
||||
})
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = generateSeeded)]
|
||||
pub fn generate_seeded(seed: &Uint8Array) -> Result<ExtendedIdentity, String> {
|
||||
let seed_vec = seed.to_vec();
|
||||
let (identity_trapdoor, identity_nullifier, identity_secret, id_commitment) =
|
||||
extended_seeded_keygen(&seed_vec).map_err(|err| err.to_string())?;
|
||||
Ok(ExtendedIdentity {
|
||||
identity_trapdoor,
|
||||
identity_nullifier,
|
||||
identity_secret,
|
||||
id_commitment,
|
||||
})
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = getTrapdoor)]
|
||||
pub fn get_trapdoor(&self) -> WasmFr {
|
||||
WasmFr(self.identity_trapdoor)
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = getNullifier)]
|
||||
pub fn get_nullifier(&self) -> WasmFr {
|
||||
WasmFr(self.identity_nullifier)
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = getSecretHash)]
|
||||
pub fn get_secret_hash(&self) -> WasmFr {
|
||||
WasmFr(self.identity_secret)
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = getCommitment)]
|
||||
pub fn get_commitment(&self) -> WasmFr {
|
||||
WasmFr(self.id_commitment)
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = toArray)]
|
||||
pub fn to_array(&self) -> VecWasmFr {
|
||||
VecWasmFr(vec![
|
||||
self.identity_trapdoor,
|
||||
self.identity_nullifier,
|
||||
self.identity_secret,
|
||||
self.id_commitment,
|
||||
])
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = toBytesLE)]
|
||||
pub fn to_bytes_le(&self) -> Uint8Array {
|
||||
let vec_fr = vec![
|
||||
self.identity_trapdoor,
|
||||
self.identity_nullifier,
|
||||
self.identity_secret,
|
||||
self.id_commitment,
|
||||
];
|
||||
let bytes = vec_fr_to_bytes_le(&vec_fr);
|
||||
Uint8Array::from(&bytes[..])
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = toBytesBE)]
|
||||
pub fn to_bytes_be(&self) -> Uint8Array {
|
||||
let vec_fr = vec![
|
||||
self.identity_trapdoor,
|
||||
self.identity_nullifier,
|
||||
self.identity_secret,
|
||||
self.id_commitment,
|
||||
];
|
||||
let bytes = vec_fr_to_bytes_be(&vec_fr);
|
||||
Uint8Array::from(&bytes[..])
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = fromBytesLE)]
|
||||
pub fn from_bytes_le(bytes: &Uint8Array) -> Result<ExtendedIdentity, String> {
|
||||
let bytes_vec = bytes.to_vec();
|
||||
let (vec_fr, _) = bytes_le_to_vec_fr(&bytes_vec).map_err(|err| err.to_string())?;
|
||||
if vec_fr.len() != 4 {
|
||||
return Err(format!("Expected 4 elements, got {}", vec_fr.len()));
|
||||
}
|
||||
Ok(ExtendedIdentity {
|
||||
identity_trapdoor: vec_fr[0],
|
||||
identity_nullifier: vec_fr[1],
|
||||
identity_secret: vec_fr[2],
|
||||
id_commitment: vec_fr[3],
|
||||
})
|
||||
}
|
||||
|
||||
#[wasm_bindgen(js_name = fromBytesBE)]
|
||||
pub fn from_bytes_be(bytes: &Uint8Array) -> Result<ExtendedIdentity, String> {
|
||||
let bytes_vec = bytes.to_vec();
|
||||
let (vec_fr, _) = bytes_be_to_vec_fr(&bytes_vec).map_err(|err| err.to_string())?;
|
||||
if vec_fr.len() != 4 {
|
||||
return Err(format!("Expected 4 elements, got {}", vec_fr.len()));
|
||||
}
|
||||
Ok(ExtendedIdentity {
|
||||
identity_trapdoor: vec_fr[0],
|
||||
identity_nullifier: vec_fr[1],
|
||||
identity_secret: vec_fr[2],
|
||||
id_commitment: vec_fr[3],
|
||||
})
|
||||
}
|
||||
}
|
||||
247
rln-wasm/tests/browser.rs
Normal file
247
rln-wasm/tests/browser.rs
Normal file
@@ -0,0 +1,247 @@
|
||||
#![cfg(target_arch = "wasm32")]
|
||||
#![cfg(not(feature = "utils"))]
|
||||
|
||||
#[cfg(test)]
|
||||
mod test {
|
||||
use js_sys::{BigInt as JsBigInt, Date, Object, Uint8Array};
|
||||
use rln::prelude::*;
|
||||
use rln_wasm::{
|
||||
Hasher, Identity, VecWasmFr, WasmFr, WasmRLN, WasmRLNProof, WasmRLNWitnessInput,
|
||||
};
|
||||
use wasm_bindgen::{prelude::wasm_bindgen, JsValue};
|
||||
use wasm_bindgen_test::{console_log, wasm_bindgen_test, wasm_bindgen_test_configure};
|
||||
use zerokit_utils::merkle_tree::{
|
||||
OptimalMerkleProof, OptimalMerkleTree, ZerokitMerkleProof, ZerokitMerkleTree,
|
||||
};
|
||||
#[cfg(feature = "parallel")]
|
||||
use {rln_wasm::init_thread_pool, wasm_bindgen_futures::JsFuture, web_sys::window};
|
||||
|
||||
#[wasm_bindgen(inline_js = r#"
|
||||
export function isThreadpoolSupported() {
|
||||
return typeof SharedArrayBuffer !== 'undefined' &&
|
||||
typeof Atomics !== 'undefined' &&
|
||||
typeof crossOriginIsolated !== 'undefined' &&
|
||||
crossOriginIsolated;
|
||||
}
|
||||
|
||||
export function initWitnessCalculator(jsCode) {
|
||||
const processedCode = jsCode
|
||||
.replace(/export\s+async\s+function\s+builder/, 'async function builder')
|
||||
.replace(/export\s*\{\s*builder\s*\};?/g, '');
|
||||
|
||||
const moduleFunc = new Function(processedCode + '\nreturn { builder };');
|
||||
const witnessCalculatorModule = moduleFunc();
|
||||
|
||||
window.witnessCalculatorBuilder = witnessCalculatorModule.builder;
|
||||
|
||||
if (typeof window.witnessCalculatorBuilder !== 'function') {
|
||||
return false;
|
||||
}
|
||||
return true;
|
||||
}
|
||||
|
||||
export function readFile(data) {
|
||||
return new Uint8Array(data);
|
||||
}
|
||||
|
||||
export async function calculateWitness(circom_data, inputs) {
|
||||
const wasmBuffer = circom_data instanceof Uint8Array ? circom_data : new Uint8Array(circom_data);
|
||||
const witnessCalculator = await window.witnessCalculatorBuilder(wasmBuffer);
|
||||
const calculatedWitness = await witnessCalculator.calculateWitness(inputs, false);
|
||||
return JSON.stringify(calculatedWitness, (key, value) =>
|
||||
typeof value === "bigint" ? value.toString() : value
|
||||
);
|
||||
}
|
||||
"#)]
|
||||
extern "C" {
|
||||
#[wasm_bindgen(catch)]
|
||||
fn isThreadpoolSupported() -> Result<bool, JsValue>;
|
||||
|
||||
#[wasm_bindgen(catch)]
|
||||
fn initWitnessCalculator(js: &str) -> Result<bool, JsValue>;
|
||||
|
||||
#[wasm_bindgen(catch)]
|
||||
fn readFile(data: &[u8]) -> Result<Uint8Array, JsValue>;
|
||||
|
||||
#[wasm_bindgen(catch)]
|
||||
async fn calculateWitness(circom_data: &[u8], inputs: Object) -> Result<JsValue, JsValue>;
|
||||
}
|
||||
|
||||
const WITNESS_CALCULATOR_JS: &str = include_str!("../resources/witness_calculator.js");
|
||||
|
||||
const ARKZKEY_BYTES: &[u8] =
|
||||
include_bytes!("../../rln/resources/tree_depth_20/rln_final.arkzkey");
|
||||
|
||||
const CIRCOM_BYTES: &[u8] = include_bytes!("../../rln/resources/tree_depth_20/rln.wasm");
|
||||
|
||||
wasm_bindgen_test_configure!(run_in_browser);
|
||||
|
||||
#[wasm_bindgen_test]
|
||||
pub async fn rln_wasm_benchmark() {
|
||||
// Check if thread pool is supported
|
||||
#[cfg(feature = "parallel")]
|
||||
if !isThreadpoolSupported().unwrap() {
|
||||
panic!("Thread pool is NOT supported");
|
||||
} else {
|
||||
// Initialize thread pool
|
||||
let cpu_count = window().unwrap().navigator().hardware_concurrency() as usize;
|
||||
JsFuture::from(init_thread_pool(cpu_count)).await.unwrap();
|
||||
}
|
||||
|
||||
// Initialize witness calculator
|
||||
initWitnessCalculator(WITNESS_CALCULATOR_JS).unwrap();
|
||||
|
||||
let mut results = String::from("\nbenchmarks:\n");
|
||||
let iterations = 10;
|
||||
|
||||
let zkey = readFile(ARKZKEY_BYTES).unwrap();
|
||||
|
||||
// Benchmark RLN instance creation
|
||||
let start_rln_new = Date::now();
|
||||
for _ in 0..iterations {
|
||||
let _ = WasmRLN::new(&zkey).unwrap();
|
||||
}
|
||||
let rln_new_result = Date::now() - start_rln_new;
|
||||
|
||||
// Create RLN instance for other benchmarks
|
||||
let rln_instance = WasmRLN::new(&zkey).unwrap();
|
||||
let mut tree: OptimalMerkleTree<PoseidonHash> =
|
||||
OptimalMerkleTree::default(DEFAULT_TREE_DEPTH).unwrap();
|
||||
|
||||
// Benchmark generate identity
|
||||
let start_identity_gen = Date::now();
|
||||
for _ in 0..iterations {
|
||||
let _ = Identity::generate().unwrap();
|
||||
}
|
||||
let identity_gen_result = Date::now() - start_identity_gen;
|
||||
|
||||
// Generate identity for other benchmarks
|
||||
let identity_pair = Identity::generate().unwrap();
|
||||
let identity_secret = identity_pair.get_secret_hash();
|
||||
let id_commitment = identity_pair.get_commitment();
|
||||
|
||||
let epoch = Hasher::hash_to_field_le(&Uint8Array::from(b"test-epoch" as &[u8])).unwrap();
|
||||
let rln_identifier =
|
||||
Hasher::hash_to_field_le(&Uint8Array::from(b"test-rln-identifier" as &[u8])).unwrap();
|
||||
let external_nullifier = Hasher::poseidon_hash_pair(&epoch, &rln_identifier).unwrap();
|
||||
|
||||
let identity_index = tree.leaves_set();
|
||||
|
||||
let user_message_limit = WasmFr::from_uint(100);
|
||||
|
||||
let rate_commitment =
|
||||
Hasher::poseidon_hash_pair(&id_commitment, &user_message_limit).unwrap();
|
||||
tree.update_next(*rate_commitment).unwrap();
|
||||
|
||||
let message_id = WasmFr::from_uint(0);
|
||||
let signal: [u8; 32] = [0; 32];
|
||||
let x = Hasher::hash_to_field_le(&Uint8Array::from(&signal[..])).unwrap();
|
||||
|
||||
let merkle_proof: OptimalMerkleProof<PoseidonHash> = tree.proof(identity_index).unwrap();
|
||||
|
||||
let mut path_elements = VecWasmFr::new();
|
||||
for path_element in merkle_proof.get_path_elements() {
|
||||
path_elements.push(&WasmFr::from(path_element));
|
||||
}
|
||||
let path_index = Uint8Array::from(&merkle_proof.get_path_index()[..]);
|
||||
|
||||
let witness = WasmRLNWitnessInput::new(
|
||||
&identity_secret,
|
||||
&user_message_limit,
|
||||
&message_id,
|
||||
&path_elements,
|
||||
&path_index,
|
||||
&x,
|
||||
&external_nullifier,
|
||||
)
|
||||
.unwrap();
|
||||
|
||||
let bigint_json = witness.to_bigint_json().unwrap();
|
||||
|
||||
// Benchmark witness calculation
|
||||
let start_calculate_witness = Date::now();
|
||||
for _ in 0..iterations {
|
||||
let _ = calculateWitness(CIRCOM_BYTES, bigint_json.clone())
|
||||
.await
|
||||
.unwrap();
|
||||
}
|
||||
let calculate_witness_result = Date::now() - start_calculate_witness;
|
||||
|
||||
// Calculate witness for other benchmarks
|
||||
let calculated_witness_str = calculateWitness(CIRCOM_BYTES, bigint_json.clone())
|
||||
.await
|
||||
.unwrap()
|
||||
.as_string()
|
||||
.unwrap();
|
||||
let calculated_witness_vec_str: Vec<String> =
|
||||
serde_json::from_str(&calculated_witness_str).unwrap();
|
||||
let calculated_witness: Vec<JsBigInt> = calculated_witness_vec_str
|
||||
.iter()
|
||||
.map(|x| JsBigInt::new(&x.into()).unwrap())
|
||||
.collect();
|
||||
|
||||
// Benchmark proof generation with witness
|
||||
let start_generate_rln_proof_with_witness = Date::now();
|
||||
for _ in 0..iterations {
|
||||
let _ = rln_instance
|
||||
.generate_rln_proof_with_witness(calculated_witness.clone(), &witness)
|
||||
.unwrap();
|
||||
}
|
||||
let generate_rln_proof_with_witness_result =
|
||||
Date::now() - start_generate_rln_proof_with_witness;
|
||||
|
||||
// Generate proof with witness for other benchmarks
|
||||
let proof: WasmRLNProof = rln_instance
|
||||
.generate_rln_proof_with_witness(calculated_witness, &witness)
|
||||
.unwrap();
|
||||
|
||||
let root = WasmFr::from(tree.root());
|
||||
let mut roots = VecWasmFr::new();
|
||||
roots.push(&root);
|
||||
|
||||
// Benchmark proof verification with the root
|
||||
let start_verify_with_roots = Date::now();
|
||||
for _ in 0..iterations {
|
||||
let _ = rln_instance.verify_with_roots(&proof, &roots, &x).unwrap();
|
||||
}
|
||||
let verify_with_roots_result = Date::now() - start_verify_with_roots;
|
||||
|
||||
// Verify proof with the root for other benchmarks
|
||||
let is_proof_valid = rln_instance.verify_with_roots(&proof, &roots, &x).unwrap();
|
||||
assert!(is_proof_valid, "verification failed");
|
||||
|
||||
// Format and display the benchmark results
|
||||
let format_duration = |duration_ms: f64| -> String {
|
||||
let avg_ms = duration_ms / (iterations as f64);
|
||||
if avg_ms >= 1000.0 {
|
||||
format!("{:.3} s", avg_ms / 1000.0)
|
||||
} else {
|
||||
format!("{:.3} ms", avg_ms)
|
||||
}
|
||||
};
|
||||
|
||||
results.push_str(&format!(
|
||||
"RLN instance creation: {}\n",
|
||||
format_duration(rln_new_result)
|
||||
));
|
||||
results.push_str(&format!(
|
||||
"Identity generation: {}\n",
|
||||
format_duration(identity_gen_result)
|
||||
));
|
||||
results.push_str(&format!(
|
||||
"Witness calculation: {}\n",
|
||||
format_duration(calculate_witness_result)
|
||||
));
|
||||
results.push_str(&format!(
|
||||
"Proof generation with witness: {}\n",
|
||||
format_duration(generate_rln_proof_with_witness_result)
|
||||
));
|
||||
results.push_str(&format!(
|
||||
"Proof verification with roots: {}\n",
|
||||
format_duration(verify_with_roots_result)
|
||||
));
|
||||
|
||||
// Log the results
|
||||
console_log!("{results}");
|
||||
}
|
||||
}
|
||||
233
rln-wasm/tests/node.rs
Normal file
233
rln-wasm/tests/node.rs
Normal file
@@ -0,0 +1,233 @@
|
||||
#![cfg(target_arch = "wasm32")]
|
||||
#![cfg(not(feature = "utils"))]
|
||||
|
||||
#[cfg(test)]
|
||||
mod test {
|
||||
use js_sys::{BigInt as JsBigInt, Date, Object, Uint8Array};
|
||||
use rln::prelude::*;
|
||||
use rln_wasm::{
|
||||
Hasher, Identity, VecWasmFr, WasmFr, WasmRLN, WasmRLNProof, WasmRLNWitnessInput,
|
||||
};
|
||||
use wasm_bindgen::{prelude::wasm_bindgen, JsValue};
|
||||
use wasm_bindgen_test::{console_log, wasm_bindgen_test};
|
||||
use zerokit_utils::merkle_tree::{
|
||||
OptimalMerkleProof, OptimalMerkleTree, ZerokitMerkleProof, ZerokitMerkleTree,
|
||||
};
|
||||
|
||||
#[wasm_bindgen(inline_js = r#"
|
||||
const fs = require("fs");
|
||||
|
||||
let witnessCalculatorModule = null;
|
||||
|
||||
module.exports = {
|
||||
initWitnessCalculator: function(code) {
|
||||
const processedCode = code
|
||||
.replace(/export\s+async\s+function\s+builder/, 'async function builder')
|
||||
.replace(/export\s*\{\s*builder\s*\};?/g, '');
|
||||
|
||||
const moduleFunc = new Function(processedCode + '\nreturn { builder };');
|
||||
witnessCalculatorModule = moduleFunc();
|
||||
|
||||
if (typeof witnessCalculatorModule.builder !== 'function') {
|
||||
return false;
|
||||
}
|
||||
return true;
|
||||
},
|
||||
|
||||
readFile: function (path) {
|
||||
return fs.readFileSync(path);
|
||||
},
|
||||
|
||||
calculateWitness: async function (circom_path, inputs) {
|
||||
const wasmFile = fs.readFileSync(circom_path);
|
||||
const wasmFileBuffer = wasmFile.buffer.slice(
|
||||
wasmFile.byteOffset,
|
||||
wasmFile.byteOffset + wasmFile.byteLength
|
||||
);
|
||||
const witnessCalculator = await witnessCalculatorModule.builder(wasmFileBuffer);
|
||||
const calculatedWitness = await witnessCalculator.calculateWitness(
|
||||
inputs,
|
||||
false
|
||||
);
|
||||
return JSON.stringify(calculatedWitness, (key, value) =>
|
||||
typeof value === "bigint" ? value.toString() : value
|
||||
);
|
||||
},
|
||||
};
|
||||
"#)]
|
||||
extern "C" {
|
||||
#[wasm_bindgen(catch)]
|
||||
fn initWitnessCalculator(code: &str) -> Result<bool, JsValue>;
|
||||
|
||||
#[wasm_bindgen(catch)]
|
||||
fn readFile(path: &str) -> Result<Uint8Array, JsValue>;
|
||||
|
||||
#[wasm_bindgen(catch)]
|
||||
async fn calculateWitness(circom_path: &str, input: Object) -> Result<JsValue, JsValue>;
|
||||
}
|
||||
|
||||
const WITNESS_CALCULATOR_JS: &str = include_str!("../resources/witness_calculator.js");
|
||||
|
||||
const ARKZKEY_PATH: &str = "../rln/resources/tree_depth_20/rln_final.arkzkey";
|
||||
|
||||
const CIRCOM_PATH: &str = "../rln/resources/tree_depth_20/rln.wasm";
|
||||
|
||||
#[wasm_bindgen_test]
|
||||
pub async fn rln_wasm_benchmark() {
|
||||
// Initialize witness calculator
|
||||
initWitnessCalculator(WITNESS_CALCULATOR_JS).unwrap();
|
||||
|
||||
let mut results = String::from("\nbenchmarks:\n");
|
||||
let iterations = 10;
|
||||
|
||||
let zkey = readFile(ARKZKEY_PATH).unwrap();
|
||||
|
||||
// Benchmark RLN instance creation
|
||||
let start_rln_new = Date::now();
|
||||
for _ in 0..iterations {
|
||||
let _ = WasmRLN::new(&zkey).unwrap();
|
||||
}
|
||||
let rln_new_result = Date::now() - start_rln_new;
|
||||
|
||||
// Create RLN instance for other benchmarks
|
||||
let rln_instance = WasmRLN::new(&zkey).unwrap();
|
||||
let mut tree: OptimalMerkleTree<PoseidonHash> =
|
||||
OptimalMerkleTree::default(DEFAULT_TREE_DEPTH).unwrap();
|
||||
|
||||
// Benchmark generate identity
|
||||
let start_identity_gen = Date::now();
|
||||
for _ in 0..iterations {
|
||||
let _ = Identity::generate().unwrap();
|
||||
}
|
||||
let identity_gen_result = Date::now() - start_identity_gen;
|
||||
|
||||
// Generate identity for other benchmarks
|
||||
let identity_pair = Identity::generate().unwrap();
|
||||
let identity_secret = identity_pair.get_secret_hash();
|
||||
let id_commitment = identity_pair.get_commitment();
|
||||
|
||||
let epoch = Hasher::hash_to_field_le(&Uint8Array::from(b"test-epoch" as &[u8])).unwrap();
|
||||
let rln_identifier =
|
||||
Hasher::hash_to_field_le(&Uint8Array::from(b"test-rln-identifier" as &[u8])).unwrap();
|
||||
let external_nullifier = Hasher::poseidon_hash_pair(&epoch, &rln_identifier).unwrap();
|
||||
|
||||
let identity_index = tree.leaves_set();
|
||||
|
||||
let user_message_limit = WasmFr::from_uint(100);
|
||||
|
||||
let rate_commitment =
|
||||
Hasher::poseidon_hash_pair(&id_commitment, &user_message_limit).unwrap();
|
||||
tree.update_next(*rate_commitment).unwrap();
|
||||
|
||||
let message_id = WasmFr::from_uint(0);
|
||||
let signal: [u8; 32] = [0; 32];
|
||||
let x = Hasher::hash_to_field_le(&Uint8Array::from(&signal[..])).unwrap();
|
||||
|
||||
let merkle_proof: OptimalMerkleProof<PoseidonHash> = tree.proof(identity_index).unwrap();
|
||||
|
||||
let mut path_elements = VecWasmFr::new();
|
||||
for path_element in merkle_proof.get_path_elements() {
|
||||
path_elements.push(&WasmFr::from(path_element));
|
||||
}
|
||||
let path_index = Uint8Array::from(&merkle_proof.get_path_index()[..]);
|
||||
|
||||
let witness = WasmRLNWitnessInput::new(
|
||||
&identity_secret,
|
||||
&user_message_limit,
|
||||
&message_id,
|
||||
&path_elements,
|
||||
&path_index,
|
||||
&x,
|
||||
&external_nullifier,
|
||||
)
|
||||
.unwrap();
|
||||
|
||||
let bigint_json = witness.to_bigint_json().unwrap();
|
||||
|
||||
// Benchmark witness calculation
|
||||
let start_calculate_witness = Date::now();
|
||||
for _ in 0..iterations {
|
||||
let _ = calculateWitness(CIRCOM_PATH, bigint_json.clone())
|
||||
.await
|
||||
.unwrap();
|
||||
}
|
||||
let calculate_witness_result = Date::now() - start_calculate_witness;
|
||||
|
||||
// Calculate witness for other benchmarks
|
||||
let calculated_witness_str = calculateWitness(CIRCOM_PATH, bigint_json.clone())
|
||||
.await
|
||||
.unwrap()
|
||||
.as_string()
|
||||
.unwrap();
|
||||
let calculated_witness_vec_str: Vec<String> =
|
||||
serde_json::from_str(&calculated_witness_str).unwrap();
|
||||
let calculated_witness: Vec<JsBigInt> = calculated_witness_vec_str
|
||||
.iter()
|
||||
.map(|x| JsBigInt::new(&x.into()).unwrap())
|
||||
.collect();
|
||||
|
||||
// Benchmark proof generation with witness
|
||||
let start_generate_rln_proof_with_witness = Date::now();
|
||||
for _ in 0..iterations {
|
||||
let _ = rln_instance
|
||||
.generate_rln_proof_with_witness(calculated_witness.clone(), &witness)
|
||||
.unwrap();
|
||||
}
|
||||
let generate_rln_proof_with_witness_result =
|
||||
Date::now() - start_generate_rln_proof_with_witness;
|
||||
|
||||
// Generate proof with witness for other benchmarks
|
||||
let proof: WasmRLNProof = rln_instance
|
||||
.generate_rln_proof_with_witness(calculated_witness, &witness)
|
||||
.unwrap();
|
||||
|
||||
let root = WasmFr::from(tree.root());
|
||||
let mut roots = VecWasmFr::new();
|
||||
roots.push(&root);
|
||||
|
||||
// Benchmark proof verification with the root
|
||||
let start_verify_with_roots = Date::now();
|
||||
for _ in 0..iterations {
|
||||
let _ = rln_instance.verify_with_roots(&proof, &roots, &x).unwrap();
|
||||
}
|
||||
let verify_with_roots_result = Date::now() - start_verify_with_roots;
|
||||
|
||||
// Verify proof with the root for other benchmarks
|
||||
let is_proof_valid = rln_instance.verify_with_roots(&proof, &roots, &x).unwrap();
|
||||
assert!(is_proof_valid, "verification failed");
|
||||
|
||||
// Format and display the benchmark results
|
||||
let format_duration = |duration_ms: f64| -> String {
|
||||
let avg_ms = duration_ms / (iterations as f64);
|
||||
if avg_ms >= 1000.0 {
|
||||
format!("{:.3} s", avg_ms / 1000.0)
|
||||
} else {
|
||||
format!("{:.3} ms", avg_ms)
|
||||
}
|
||||
};
|
||||
|
||||
results.push_str(&format!(
|
||||
"RLN instance creation: {}\n",
|
||||
format_duration(rln_new_result)
|
||||
));
|
||||
results.push_str(&format!(
|
||||
"Identity generation: {}\n",
|
||||
format_duration(identity_gen_result)
|
||||
));
|
||||
results.push_str(&format!(
|
||||
"Witness calculation: {}\n",
|
||||
format_duration(calculate_witness_result)
|
||||
));
|
||||
results.push_str(&format!(
|
||||
"Proof generation with witness: {}\n",
|
||||
format_duration(generate_rln_proof_with_witness_result)
|
||||
));
|
||||
results.push_str(&format!(
|
||||
"Proof verification with roots: {}\n",
|
||||
format_duration(verify_with_roots_result)
|
||||
));
|
||||
|
||||
// Log the results
|
||||
console_log!("{results}");
|
||||
}
|
||||
}
|
||||
@@ -1,123 +0,0 @@
|
||||
#![cfg(target_arch = "wasm32")]
|
||||
|
||||
#[cfg(test)]
|
||||
mod tests {
|
||||
use js_sys::{BigInt as JsBigInt, Object, Uint8Array};
|
||||
use rln::circuit::{Fr, TEST_TREE_HEIGHT};
|
||||
use rln::hashers::{hash_to_field, poseidon_hash};
|
||||
use rln::utils::{bytes_le_to_fr, fr_to_bytes_le, normalize_usize};
|
||||
use rln_wasm::*;
|
||||
use rln::poseidon_tree::PoseidonTree;
|
||||
use rln::utils::vec_fr_to_bytes_le;
|
||||
use wasm_bindgen::{prelude::*, JsValue};
|
||||
use wasm_bindgen_test::wasm_bindgen_test;
|
||||
use zerokit_utils::merkle_tree::merkle_tree::ZerokitMerkleTree;
|
||||
use zerokit_utils::ZerokitMerkleProof;
|
||||
use rln::utils::vec_u8_to_bytes_le;
|
||||
|
||||
#[wasm_bindgen(module = "src/utils.js")]
|
||||
extern "C" {
|
||||
#[wasm_bindgen(catch)]
|
||||
fn read_file(path: &str) -> Result<Uint8Array, JsValue>;
|
||||
|
||||
#[wasm_bindgen(catch)]
|
||||
async fn calculateWitness(circom_path: &str, input: Object) -> Result<JsValue, JsValue>;
|
||||
}
|
||||
|
||||
#[wasm_bindgen_test]
|
||||
pub async fn test_basic_flow() {
|
||||
let tree_height = TEST_TREE_HEIGHT;
|
||||
let circom_path = format!("../rln/resources/tree_height_{TEST_TREE_HEIGHT}/rln.wasm");
|
||||
let zkey_path = format!("../rln/resources/tree_height_{TEST_TREE_HEIGHT}/rln_final.zkey");
|
||||
let vk_path =
|
||||
format!("../rln/resources/tree_height_{TEST_TREE_HEIGHT}/verification_key.arkvkey");
|
||||
let zkey = read_file(&zkey_path).unwrap();
|
||||
let vk = read_file(&vk_path).unwrap();
|
||||
|
||||
// Creating an instance of RLN
|
||||
let rln_instance = wasm_new(zkey, vk).unwrap();
|
||||
|
||||
let mut tree = PoseidonTree::default(TEST_TREE_HEIGHT).unwrap();
|
||||
|
||||
// Creating membership key
|
||||
let mem_keys = wasm_key_gen(rln_instance).unwrap();
|
||||
let id_key = mem_keys.subarray(0, 32);
|
||||
let id_commitment = mem_keys.subarray(32, 64);
|
||||
|
||||
// Prepare the message
|
||||
let signal = b"Hello World";
|
||||
|
||||
let identity_index = tree.leaves_set();
|
||||
// Setting up the epoch and rln_identifier
|
||||
let epoch = hash_to_field(b"test-epoch");
|
||||
let rln_identifier = hash_to_field(b"test-rln-identifier");
|
||||
|
||||
let external_nullifier = poseidon_hash(&[epoch, rln_identifier]);
|
||||
let external_nullifier = fr_to_bytes_le(&external_nullifier);
|
||||
|
||||
let user_message_limit = Fr::from(100);
|
||||
let message_id = fr_to_bytes_le(&Fr::from(0));
|
||||
|
||||
let (id_commitment_fr, _) = bytes_le_to_fr(&id_commitment.to_vec()[..]);
|
||||
let rate_commitment = poseidon_hash(&[id_commitment_fr, user_message_limit]);
|
||||
tree.update_next(rate_commitment).unwrap();
|
||||
|
||||
let x = hash_to_field(signal);
|
||||
let merkle_proof = tree.proof(identity_index).expect("proof should exist");
|
||||
let path_elements = merkle_proof.get_path_elements();
|
||||
let identity_path_index = merkle_proof.get_path_index();
|
||||
|
||||
|
||||
// Serializing the message
|
||||
let mut serialized_vec: Vec<u8> = Vec::new();
|
||||
serialized_vec.append(&mut id_key.to_vec());
|
||||
serialized_vec.append(&mut fr_to_bytes_le(&user_message_limit).to_vec());
|
||||
serialized_vec.append(&mut message_id.to_vec());
|
||||
serialized_vec.append(&mut vec_fr_to_bytes_le(&path_elements).unwrap());
|
||||
serialized_vec.append(&mut vec_u8_to_bytes_le(&identity_path_index).unwrap());
|
||||
serialized_vec.append(&mut fr_to_bytes_le(&x));
|
||||
serialized_vec.append(&mut external_nullifier.to_vec());
|
||||
let serialized_message = Uint8Array::from(&serialized_vec[..]);
|
||||
|
||||
// Obtaining inputs that should be sent to circom witness calculator
|
||||
let json_inputs =
|
||||
rln_witness_to_json(rln_instance, serialized_message.clone()).unwrap();
|
||||
|
||||
// Calculating witness with JS
|
||||
// (Using a JSON since wasm_bindgen does not like Result<Vec<JsBigInt>,JsValue>)
|
||||
let calculated_witness_json = calculateWitness(&circom_path, json_inputs)
|
||||
.await
|
||||
.unwrap()
|
||||
.as_string()
|
||||
.unwrap();
|
||||
let calculated_witness_vec_str: Vec<String> =
|
||||
serde_json::from_str(&calculated_witness_json).unwrap();
|
||||
let calculated_witness: Vec<JsBigInt> = calculated_witness_vec_str
|
||||
.iter()
|
||||
.map(|x| JsBigInt::new(&x.into()).unwrap())
|
||||
.collect();
|
||||
|
||||
// Generating proof
|
||||
let proof = generate_rln_proof_with_witness(
|
||||
rln_instance,
|
||||
calculated_witness.into(),
|
||||
serialized_message,
|
||||
)
|
||||
.unwrap();
|
||||
|
||||
// Add signal_len | signal
|
||||
let mut proof_bytes = proof.to_vec();
|
||||
proof_bytes.append(&mut normalize_usize(signal.len()));
|
||||
proof_bytes.append(&mut signal.to_vec());
|
||||
let proof_with_signal = Uint8Array::from(&proof_bytes[..]);
|
||||
|
||||
// Validating Proof with Roots
|
||||
let root = tree.root();
|
||||
let root_le = fr_to_bytes_le(&root);
|
||||
let roots = Uint8Array::from(&root_le[..]);
|
||||
let proof_with_signal = Uint8Array::from(&proof_bytes[..]);
|
||||
|
||||
let is_proof_valid = wasm_verify_with_roots(rln_instance, proof_with_signal, roots);
|
||||
assert!(is_proof_valid.unwrap(), "verifying proof with roots failed");
|
||||
}
|
||||
}
|
||||
222
rln-wasm/tests/utils.rs
Normal file
222
rln-wasm/tests/utils.rs
Normal file
@@ -0,0 +1,222 @@
|
||||
#![cfg(target_arch = "wasm32")]
|
||||
|
||||
#[cfg(test)]
|
||||
mod test {
|
||||
use std::assert_eq;
|
||||
|
||||
use ark_std::rand::thread_rng;
|
||||
use js_sys::Uint8Array;
|
||||
use rand::Rng;
|
||||
use rln::prelude::*;
|
||||
use rln_wasm::{ExtendedIdentity, Hasher, Identity, VecWasmFr, WasmFr};
|
||||
use wasm_bindgen_test::wasm_bindgen_test;
|
||||
|
||||
#[wasm_bindgen_test]
|
||||
fn test_keygen_wasm() {
|
||||
let identity = Identity::generate().unwrap();
|
||||
let identity_secret = *identity.get_secret_hash();
|
||||
let id_commitment = *identity.get_commitment();
|
||||
|
||||
assert_ne!(identity_secret, Fr::from(0u8));
|
||||
assert_ne!(id_commitment, Fr::from(0u8));
|
||||
|
||||
let arr = identity.to_array();
|
||||
assert_eq!(arr.length(), 2);
|
||||
assert_eq!(*arr.get(0).unwrap(), identity_secret);
|
||||
assert_eq!(*arr.get(1).unwrap(), id_commitment);
|
||||
}
|
||||
|
||||
#[wasm_bindgen_test]
|
||||
fn test_extended_keygen_wasm() {
|
||||
let identity = ExtendedIdentity::generate().unwrap();
|
||||
|
||||
let identity_trapdoor = *identity.get_trapdoor();
|
||||
let identity_nullifier = *identity.get_nullifier();
|
||||
let identity_secret = *identity.get_secret_hash();
|
||||
let id_commitment = *identity.get_commitment();
|
||||
|
||||
assert_ne!(identity_trapdoor, Fr::from(0u8));
|
||||
assert_ne!(identity_nullifier, Fr::from(0u8));
|
||||
assert_ne!(identity_secret, Fr::from(0u8));
|
||||
assert_ne!(id_commitment, Fr::from(0u8));
|
||||
|
||||
let arr = identity.to_array();
|
||||
assert_eq!(arr.length(), 4);
|
||||
assert_eq!(*arr.get(0).unwrap(), identity_trapdoor);
|
||||
assert_eq!(*arr.get(1).unwrap(), identity_nullifier);
|
||||
assert_eq!(*arr.get(2).unwrap(), identity_secret);
|
||||
assert_eq!(*arr.get(3).unwrap(), id_commitment);
|
||||
}
|
||||
|
||||
#[wasm_bindgen_test]
|
||||
fn test_seeded_keygen_wasm() {
|
||||
let seed_bytes: Vec<u8> = vec![0, 1, 2, 3, 4, 5, 6, 7, 8, 9];
|
||||
let seed = Uint8Array::from(&seed_bytes[..]);
|
||||
|
||||
let identity = Identity::generate_seeded(&seed).unwrap();
|
||||
let identity_secret = *identity.get_secret_hash();
|
||||
let id_commitment = *identity.get_commitment();
|
||||
|
||||
let expected_identity_secret_seed_bytes = str_to_fr(
|
||||
"0x766ce6c7e7a01bdf5b3f257616f603918c30946fa23480f2859c597817e6716",
|
||||
16,
|
||||
)
|
||||
.unwrap();
|
||||
let expected_id_commitment_seed_bytes = str_to_fr(
|
||||
"0xbf16d2b5c0d6f9d9d561e05bfca16a81b4b873bb063508fae360d8c74cef51f",
|
||||
16,
|
||||
)
|
||||
.unwrap();
|
||||
|
||||
assert_eq!(identity_secret, expected_identity_secret_seed_bytes);
|
||||
assert_eq!(id_commitment, expected_id_commitment_seed_bytes);
|
||||
}
|
||||
|
||||
#[wasm_bindgen_test]
|
||||
fn test_seeded_extended_keygen_wasm() {
|
||||
let seed_bytes: Vec<u8> = vec![0, 1, 2, 3, 4, 5, 6, 7, 8, 9];
|
||||
let seed = Uint8Array::from(&seed_bytes[..]);
|
||||
|
||||
let identity = ExtendedIdentity::generate_seeded(&seed).unwrap();
|
||||
|
||||
let identity_trapdoor = *identity.get_trapdoor();
|
||||
let identity_nullifier = *identity.get_nullifier();
|
||||
let identity_secret = *identity.get_secret_hash();
|
||||
let id_commitment = *identity.get_commitment();
|
||||
|
||||
let expected_identity_trapdoor_seed_bytes = str_to_fr(
|
||||
"0x766ce6c7e7a01bdf5b3f257616f603918c30946fa23480f2859c597817e6716",
|
||||
16,
|
||||
)
|
||||
.unwrap();
|
||||
let expected_identity_nullifier_seed_bytes = str_to_fr(
|
||||
"0x1f18714c7bc83b5bca9e89d404cf6f2f585bc4c0f7ed8b53742b7e2b298f50b4",
|
||||
16,
|
||||
)
|
||||
.unwrap();
|
||||
let expected_identity_secret_seed_bytes = str_to_fr(
|
||||
"0x2aca62aaa7abaf3686fff2caf00f55ab9462dc12db5b5d4bcf3994e671f8e521",
|
||||
16,
|
||||
)
|
||||
.unwrap();
|
||||
let expected_id_commitment_seed_bytes = str_to_fr(
|
||||
"0x68b66aa0a8320d2e56842581553285393188714c48f9b17acd198b4f1734c5c",
|
||||
16,
|
||||
)
|
||||
.unwrap();
|
||||
|
||||
assert_eq!(identity_trapdoor, expected_identity_trapdoor_seed_bytes);
|
||||
assert_eq!(identity_nullifier, expected_identity_nullifier_seed_bytes);
|
||||
assert_eq!(identity_secret, expected_identity_secret_seed_bytes);
|
||||
assert_eq!(id_commitment, expected_id_commitment_seed_bytes);
|
||||
}
|
||||
|
||||
#[wasm_bindgen_test]
|
||||
fn test_wasmfr() {
|
||||
let wasmfr_zero = WasmFr::zero();
|
||||
let fr_zero = Fr::from(0u8);
|
||||
assert_eq!(*wasmfr_zero, fr_zero);
|
||||
|
||||
let wasmfr_one = WasmFr::one();
|
||||
let fr_one = Fr::from(1u8);
|
||||
assert_eq!(*wasmfr_one, fr_one);
|
||||
|
||||
let wasmfr_int = WasmFr::from_uint(42);
|
||||
let fr_int = Fr::from(42u8);
|
||||
assert_eq!(*wasmfr_int, fr_int);
|
||||
|
||||
let wasmfr_debug_str = wasmfr_int.debug();
|
||||
assert_eq!(wasmfr_debug_str.to_string(), "42");
|
||||
|
||||
let identity = Identity::generate().unwrap();
|
||||
let mut id_secret_fr = *identity.get_secret_hash();
|
||||
let id_secret_hash = IdSecret::from(&mut id_secret_fr);
|
||||
let id_commitment = *identity.get_commitment();
|
||||
let wasmfr_id_secret_hash = *identity.get_secret_hash();
|
||||
assert_eq!(wasmfr_id_secret_hash, *id_secret_hash);
|
||||
let wasmfr_id_commitment = *identity.get_commitment();
|
||||
assert_eq!(wasmfr_id_commitment, id_commitment);
|
||||
}
|
||||
|
||||
#[wasm_bindgen_test]
|
||||
fn test_vec_wasmfr() {
|
||||
let vec_fr = vec![Fr::from(1u8), Fr::from(2u8), Fr::from(3u8), Fr::from(4u8)];
|
||||
let mut vec_wasmfr = VecWasmFr::new();
|
||||
for fr in &vec_fr {
|
||||
vec_wasmfr.push(&WasmFr::from(*fr));
|
||||
}
|
||||
|
||||
let bytes_le = vec_wasmfr.to_bytes_le();
|
||||
let expected_le = rln::utils::vec_fr_to_bytes_le(&vec_fr);
|
||||
assert_eq!(bytes_le.to_vec(), expected_le);
|
||||
|
||||
let bytes_be = vec_wasmfr.to_bytes_be();
|
||||
let expected_be = rln::utils::vec_fr_to_bytes_be(&vec_fr);
|
||||
assert_eq!(bytes_be.to_vec(), expected_be);
|
||||
|
||||
let vec_wasmfr_from_le = match VecWasmFr::from_bytes_le(&bytes_le) {
|
||||
Ok(v) => v,
|
||||
Err(err) => panic!("VecWasmFr::from_bytes_le call failed: {}", err),
|
||||
};
|
||||
assert_eq!(vec_wasmfr_from_le.length(), vec_wasmfr.length());
|
||||
for i in 0..vec_wasmfr.length() {
|
||||
assert_eq!(
|
||||
*vec_wasmfr_from_le.get(i).unwrap(),
|
||||
*vec_wasmfr.get(i).unwrap()
|
||||
);
|
||||
}
|
||||
|
||||
let vec_wasmfr_from_be = match VecWasmFr::from_bytes_be(&bytes_be) {
|
||||
Ok(v) => v,
|
||||
Err(err) => panic!("VecWasmFr::from_bytes_be call failed: {}", err),
|
||||
};
|
||||
for i in 0..vec_wasmfr.length() {
|
||||
assert_eq!(
|
||||
*vec_wasmfr_from_be.get(i).unwrap(),
|
||||
*vec_wasmfr.get(i).unwrap()
|
||||
);
|
||||
}
|
||||
}
|
||||
|
||||
#[wasm_bindgen_test]
|
||||
fn test_hash_to_field_wasm() {
|
||||
let mut rng = thread_rng();
|
||||
let signal_gen: [u8; 32] = rng.gen();
|
||||
let signal = Uint8Array::from(&signal_gen[..]);
|
||||
|
||||
let wasmfr_le_1 = Hasher::hash_to_field_le(&signal).unwrap();
|
||||
let fr_le_2 = hash_to_field_le(&signal_gen).unwrap();
|
||||
assert_eq!(*wasmfr_le_1, fr_le_2);
|
||||
|
||||
let wasmfr_be_1 = Hasher::hash_to_field_be(&signal).unwrap();
|
||||
let fr_be_2 = hash_to_field_be(&signal_gen).unwrap();
|
||||
assert_eq!(*wasmfr_be_1, fr_be_2);
|
||||
|
||||
assert_eq!(*wasmfr_le_1, *wasmfr_be_1);
|
||||
assert_eq!(fr_le_2, fr_be_2);
|
||||
|
||||
let hash_wasmfr_le_1 = wasmfr_le_1.to_bytes_le();
|
||||
let hash_fr_le_2 = fr_to_bytes_le(&fr_le_2);
|
||||
assert_eq!(hash_wasmfr_le_1.to_vec(), hash_fr_le_2);
|
||||
|
||||
let hash_wasmfr_be_1 = wasmfr_be_1.to_bytes_be();
|
||||
let hash_fr_be_2 = fr_to_bytes_be(&fr_be_2);
|
||||
assert_eq!(hash_wasmfr_be_1.to_vec(), hash_fr_be_2);
|
||||
|
||||
assert_ne!(hash_wasmfr_le_1.to_vec(), hash_wasmfr_be_1.to_vec());
|
||||
assert_ne!(hash_fr_le_2, hash_fr_be_2);
|
||||
}
|
||||
|
||||
#[wasm_bindgen_test]
|
||||
fn test_poseidon_hash_pair_wasm() {
|
||||
let input_1 = Fr::from(42u8);
|
||||
let input_2 = Fr::from(99u8);
|
||||
|
||||
let expected_hash = poseidon_hash(&[input_1, input_2]).unwrap();
|
||||
let wasmfr_1 = WasmFr::from_uint(42);
|
||||
let wasmfr_2 = WasmFr::from_uint(99);
|
||||
let received_hash = Hasher::poseidon_hash_pair(&wasmfr_1, &wasmfr_2).unwrap();
|
||||
|
||||
assert_eq!(*received_hash, expected_hash);
|
||||
}
|
||||
}
|
||||
126
rln/Cargo.toml
126
rln/Cargo.toml
@@ -1,6 +1,6 @@
|
||||
[package]
|
||||
name = "rln"
|
||||
version = "0.5.1"
|
||||
version = "1.0.0"
|
||||
edition = "2021"
|
||||
license = "MIT OR Apache-2.0"
|
||||
description = "APIs to manage, compute and verify zkSNARK proofs and RLN primitives"
|
||||
@@ -9,95 +9,81 @@ homepage = "https://vac.dev"
|
||||
repository = "https://github.com/vacp2p/zerokit"
|
||||
|
||||
[lib]
|
||||
crate-type = ["rlib", "staticlib"]
|
||||
crate-type = ["rlib", "staticlib", "cdylib"]
|
||||
bench = false
|
||||
|
||||
# This flag disable cargo doctests, i.e. testing example code-snippets in documentation
|
||||
# This flag disables cargo doctests, i.e. testing example code-snippets in documentation
|
||||
doctest = false
|
||||
|
||||
|
||||
[dependencies]
|
||||
# ZKP Generation
|
||||
ark-ec = { version = "=0.4.1", default-features = false }
|
||||
ark-ff = { version = "=0.4.1", default-features = false, features = ["asm"] }
|
||||
ark-std = { version = "=0.4.0", default-features = false }
|
||||
ark-bn254 = { version = "=0.4.0" }
|
||||
ark-groth16 = { version = "=0.4.0", features = [
|
||||
"parallel",
|
||||
], default-features = false }
|
||||
ark-relations = { version = "=0.4.0", default-features = false, features = [
|
||||
"std",
|
||||
] }
|
||||
ark-serialize = { version = "=0.4.1", default-features = false }
|
||||
ark-circom = { version = "=0.1.0", default-features = false, features = [
|
||||
"circom-2",
|
||||
] }
|
||||
ark-zkey = { version = "0.1.0", optional = true, default-features = false }
|
||||
ark-bn254 = { version = "0.5.0", features = ["std"] }
|
||||
ark-relations = { version = "0.5.1", features = ["std"] }
|
||||
ark-ff = { version = "0.5.0", default-features = false }
|
||||
ark-ec = { version = "0.5.0", default-features = false }
|
||||
ark-std = { version = "0.5.0", default-features = false }
|
||||
ark-poly = { version = "0.5.0", default-features = false }
|
||||
ark-groth16 = { version = "0.5.0", default-features = false }
|
||||
ark-serialize = { version = "0.5.0", default-features = false }
|
||||
|
||||
# WASM
|
||||
wasmer = { version = "=2.3.0", default-features = false }
|
||||
# Error Handling
|
||||
thiserror = "2.0.17"
|
||||
|
||||
# error handling
|
||||
color-eyre = "=0.6.2"
|
||||
thiserror = "=1.0.39"
|
||||
|
||||
# utilities
|
||||
cfg-if = "=1.0"
|
||||
num-bigint = { version = "=0.4.3", default-features = false, features = [
|
||||
# Utilities
|
||||
rayon = { version = "1.11.0", optional = true }
|
||||
byteorder = "1.5.0"
|
||||
cfg-if = "1.0.4"
|
||||
num-bigint = { version = "0.4.6", default-features = false, features = ["std"] }
|
||||
num-traits = "0.2.19"
|
||||
once_cell = "1.21.3"
|
||||
rand = "0.8.5"
|
||||
rand_chacha = "0.3.1"
|
||||
ruint = { version = "1.17.0", default-features = false, features = [
|
||||
"rand",
|
||||
"serde",
|
||||
"ark-ff-05",
|
||||
] }
|
||||
num-traits = "=0.2.15"
|
||||
once_cell = "=1.17.1"
|
||||
lazy_static = "=1.4.0"
|
||||
rand = "=0.8.5"
|
||||
rand_chacha = "=0.3.1"
|
||||
tiny-keccak = { version = "=2.0.2", features = ["keccak"] }
|
||||
utils = { package = "zerokit_utils", version = "=0.5.1", path = "../utils/", default-features = false }
|
||||
tiny-keccak = { version = "2.0.2", features = ["keccak"] }
|
||||
zeroize = "1.8.2"
|
||||
tempfile = "3.23.0"
|
||||
zerokit_utils = { version = "1.0.0", path = "../utils", default-features = false }
|
||||
|
||||
# FFI
|
||||
safer-ffi.version = "0.1"
|
||||
|
||||
# serialization
|
||||
serde_json = "=1.0.96"
|
||||
serde = { version = "=1.0.163", features = ["derive"] }
|
||||
# Serialization
|
||||
prost = "0.14.1"
|
||||
serde_json = "1.0.145"
|
||||
serde = { version = "1.0.228", features = ["derive"] }
|
||||
|
||||
document-features = { version = "=0.2.10", optional = true }
|
||||
# Documentation
|
||||
document-features = { version = "0.2.12", optional = true }
|
||||
|
||||
[dev-dependencies]
|
||||
sled = "=0.34.7"
|
||||
criterion = { version = "=0.4.0", features = ["html_reports"] }
|
||||
criterion = { version = "0.8.0", features = ["html_reports"] }
|
||||
|
||||
[features]
|
||||
default = ["parallel", "wasmer/sys-default", "pmtree-ft"]
|
||||
parallel = [
|
||||
"ark-ec/parallel",
|
||||
"ark-ff/parallel",
|
||||
"ark-std/parallel",
|
||||
"ark-groth16/parallel",
|
||||
"utils/parallel",
|
||||
]
|
||||
wasm = ["wasmer/js", "wasmer/std"]
|
||||
fullmerkletree = ["default"]
|
||||
arkzkey = ["ark-zkey"]
|
||||
default = ["parallel", "pmtree-ft"]
|
||||
stateless = []
|
||||
|
||||
# Note: pmtree feature is still experimental
|
||||
pmtree-ft = ["utils/pmtree-ft"]
|
||||
parallel = [
|
||||
"rayon",
|
||||
"ark-ff/parallel",
|
||||
"ark-ec/parallel",
|
||||
"ark-std/parallel",
|
||||
"ark-poly/parallel",
|
||||
"ark-groth16/parallel",
|
||||
"ark-serialize/parallel",
|
||||
"zerokit_utils/parallel",
|
||||
]
|
||||
fullmerkletree = [] # Pre-allocated tree, fastest access
|
||||
optimalmerkletree = [] # Sparse storage, memory efficient
|
||||
pmtree-ft = ["zerokit_utils/pmtree-ft"] # Persistent storage, disk-based
|
||||
headers = ["safer-ffi/headers"] # Generate C header file with safer-ffi
|
||||
|
||||
[[bench]]
|
||||
name = "pmtree_benchmark"
|
||||
harness = false
|
||||
|
||||
[[bench]]
|
||||
name = "circuit_loading_benchmark"
|
||||
harness = false
|
||||
|
||||
[[bench]]
|
||||
name = "circuit_loading_arkzkey_benchmark"
|
||||
harness = false
|
||||
required-features = ["arkzkey"]
|
||||
|
||||
[[bench]]
|
||||
name = "circuit_deser_benchmark"
|
||||
harness = false
|
||||
required-features = ["pmtree-ft"]
|
||||
|
||||
[[bench]]
|
||||
name = "poseidon_tree_benchmark"
|
||||
@@ -105,3 +91,7 @@ harness = false
|
||||
|
||||
[package.metadata.docs.rs]
|
||||
all-features = true
|
||||
|
||||
[[bin]]
|
||||
name = "generate-headers"
|
||||
required-features = ["headers"] # Do not build unless generating headers.
|
||||
|
||||
@@ -2,17 +2,21 @@
|
||||
command = "cargo"
|
||||
args = ["build", "--release"]
|
||||
|
||||
[tasks.test_default]
|
||||
[tasks.test]
|
||||
command = "cargo"
|
||||
args = ["test", "--release"]
|
||||
args = ["test", "--release", "--", "--nocapture"]
|
||||
|
||||
[tasks.test_stateless]
|
||||
command = "cargo"
|
||||
args = ["test", "--release", "--features", "stateless"]
|
||||
|
||||
[tasks.test_arkzkey]
|
||||
command = "cargo"
|
||||
args = ["test", "--release", "--features", "arkzkey"]
|
||||
args = [
|
||||
"test",
|
||||
"--release",
|
||||
"--no-default-features",
|
||||
"--features",
|
||||
"stateless",
|
||||
"--",
|
||||
"--nocapture",
|
||||
]
|
||||
|
||||
[tasks.bench]
|
||||
command = "cargo"
|
||||
|
||||
415
rln/README.md
415
rln/README.md
@@ -1,65 +1,22 @@
|
||||
# Zerokit RLN Module
|
||||
|
||||
This module provides APIs to manage, compute and verify [RLN](https://rfc.vac.dev/spec/32/) zkSNARK proofs and RLN primitives.
|
||||
[](https://crates.io/crates/rln)
|
||||
[](https://opensource.org/licenses/MIT)
|
||||
[](https://opensource.org/licenses/Apache-2.0)
|
||||
|
||||
## Pre-requisites
|
||||
The Zerokit RLN Module provides a Rust implementation for working with
|
||||
Rate-Limiting Nullifier [RLN](https://rfc.vac.dev/vac/raw/rln-v2) zkSNARK proofs and primitives.
|
||||
This module allows you to:
|
||||
|
||||
### Install dependencies and clone repo
|
||||
- Generate and verify RLN proofs
|
||||
- Work with Merkle trees for commitment storage
|
||||
- Implement rate-limiting mechanisms for distributed systems
|
||||
|
||||
```sh
|
||||
make installdeps
|
||||
git clone https://github.com/vacp2p/zerokit.git
|
||||
cd zerokit/rln
|
||||
```
|
||||
## Quick Start
|
||||
|
||||
### Build and Test
|
||||
|
||||
To build and test, run the following commands within the module folder
|
||||
|
||||
```bash
|
||||
cargo make build
|
||||
cargo make test
|
||||
```
|
||||
|
||||
### Compile ZK circuits
|
||||
|
||||
The `rln` (https://github.com/rate-limiting-nullifier/circom-rln) repository, which contains the RLN circuit implementation is a submodule of zerokit RLN.
|
||||
|
||||
To compile the RLN circuit
|
||||
|
||||
```sh
|
||||
# Update submodules
|
||||
git submodule update --init --recursive
|
||||
|
||||
# Install rln dependencies
|
||||
cd vendor/rln/ && npm install
|
||||
|
||||
# Build circuits
|
||||
./scripts/build-circuits.sh rln
|
||||
|
||||
# Copy over assets
|
||||
cp build/zkeyFiles/rln-final.zkey ../../resources/tree_height_15
|
||||
cp build/zkeyFiles/rln.wasm ../../resources/tree_height_15
|
||||
```
|
||||
|
||||
Note that the above code snippet will compile a RLN circuit with a Merkle tree of height equal `15` based on the default value set in `vendor/rln/circuit/rln.circom`.
|
||||
|
||||
In order to compile a RLN circuit with Merkle tree height `N`, it suffices to change `vendor/rln/circuit/rln.circom` to
|
||||
|
||||
```
|
||||
pragma circom 2.0.0;
|
||||
|
||||
include "./rln-base.circom";
|
||||
|
||||
component main {public [x, epoch, rln_identifier ]} = RLN(N);
|
||||
```
|
||||
|
||||
However, if `N` is too big, this might require a bigger Powers of Tau ceremony than the one hardcoded in `./scripts/build-circuits.sh`, which is `2^14`.
|
||||
In such case we refer to the official [Circom documentation](https://docs.circom.io/getting-started/proving-circuits/#powers-of-tau) for instructions on how to run an appropriate Powers of Tau ceremony and Phase 2 in order to compile the desired circuit.
|
||||
|
||||
Currently, the `rln` module comes with 2 [pre-compiled](https://github.com/vacp2p/zerokit/tree/master/rln/resources) RLN circuits having Merkle tree of height `20` and `32`, respectively.
|
||||
|
||||
## Getting started
|
||||
> [!IMPORTANT]
|
||||
> Version 0.7.0 is the only version that does not support WASM and x32 architecture.
|
||||
> WASM support is available in version 0.8.0 and above.
|
||||
|
||||
### Add RLN as dependency
|
||||
|
||||
@@ -67,146 +24,308 @@ We start by adding zerokit RLN to our `Cargo.toml`
|
||||
|
||||
```toml
|
||||
[dependencies]
|
||||
rln = { git = "https://github.com/vacp2p/zerokit" }
|
||||
rln = "1.0.0"
|
||||
```
|
||||
|
||||
### Create a RLN object
|
||||
## Basic Usage Example
|
||||
|
||||
First, we need to create a RLN object for a chosen input Merkle tree size.
|
||||
The RLN object constructor requires the following files:
|
||||
|
||||
Note that we need to pass to RLN object constructor the path where the circuit (`rln.wasm`, built for the input tree size), the corresponding proving key (`rln_final.zkey`) or (`rln_final.arkzkey`) and verification key (`verification_key.arkvkey`, optional) are found.
|
||||
- `rln_final.arkzkey`: The proving key in arkzkey format.
|
||||
- `graph.bin`: The graph file built for the input tree size
|
||||
|
||||
In the following we will use [cursors](https://doc.rust-lang.org/std/io/struct.Cursor.html) as readers/writers for interfacing with RLN public APIs.
|
||||
Additionally, `rln.wasm` is used for testing in the rln-wasm module.
|
||||
|
||||
```rust
|
||||
use rln::protocol::*;
|
||||
use rln::public::*;
|
||||
use std::io::Cursor;
|
||||
use rln::prelude::{keygen, poseidon_hash, hash_to_field_le, RLN, RLNWitnessInput, Fr, IdSecret};
|
||||
|
||||
// We set the RLN parameters:
|
||||
// - the tree height;
|
||||
// - the tree config, if it is not defined, the default value will be set
|
||||
let tree_height = 20;
|
||||
let input = Cursor::new(json!({}).to_string());
|
||||
fn main() {
|
||||
// 1. Initialize RLN with parameters:
|
||||
// - the tree depth;
|
||||
// - the tree config, if it is not defined, the default value will be set
|
||||
let tree_depth = 20;
|
||||
let mut rln = RLN::new(tree_depth, "").unwrap();
|
||||
|
||||
// We create a new RLN instance
|
||||
let mut rln = RLN::new(tree_height, input);
|
||||
// 2. Generate an identity keypair
|
||||
let (identity_secret, id_commitment) = keygen();
|
||||
|
||||
// 3. Add a rate commitment to the Merkle tree
|
||||
let leaf_index = 10;
|
||||
let user_message_limit = Fr::from(10);
|
||||
let rate_commitment = poseidon_hash(&[id_commitment, user_message_limit]);
|
||||
rln.set_leaf(leaf_index, rate_commitment).unwrap();
|
||||
|
||||
// 4. Get the Merkle proof for the added commitment
|
||||
let (path_elements, identity_path_index) = rln.get_merkle_proof(leaf_index).unwrap();
|
||||
|
||||
// 5. Set up external nullifier (epoch + app identifier)
|
||||
// We generate epoch from a date seed and we ensure is
|
||||
// mapped to a field element by hashing-to-field its content
|
||||
let epoch = hash_to_field_le(b"Today at noon, this year");
|
||||
// We generate rln_identifier from an application identifier and
|
||||
// we ensure is mapped to a field element by hashing-to-field its content
|
||||
let rln_identifier = hash_to_field_le(b"test-rln-identifier");
|
||||
// We generate a external nullifier
|
||||
let external_nullifier = poseidon_hash(&[epoch, rln_identifier]);
|
||||
// We choose a message_id satisfy 0 <= message_id < user_message_limit
|
||||
let message_id = Fr::from(1);
|
||||
|
||||
// 6. Define the message signal
|
||||
let signal = b"RLN is awesome";
|
||||
|
||||
// 7. Compute x from the signal
|
||||
let x = hash_to_field_le(signal);
|
||||
|
||||
// 8. Create witness input for RLN proof generation
|
||||
let witness = RLNWitnessInput::new(
|
||||
identity_secret,
|
||||
user_message_limit,
|
||||
message_id,
|
||||
path_elements,
|
||||
identity_path_index,
|
||||
x,
|
||||
external_nullifier,
|
||||
)
|
||||
.unwrap();
|
||||
|
||||
// 9. Generate a RLN proof
|
||||
// We generate proof and proof values from the witness
|
||||
let (proof, proof_values) = rln.generate_rln_proof(&witness).unwrap();
|
||||
|
||||
// 10. Verify the RLN proof
|
||||
// We verify the proof using the proof and proof values and the hashed signal x
|
||||
let verified = rln.verify_rln_proof(&proof, &proof_values, &x).unwrap();
|
||||
assert!(verified);
|
||||
}
|
||||
```
|
||||
|
||||
### Generate an identity keypair
|
||||
|
||||
We generate an identity keypair
|
||||
|
||||
```rust
|
||||
// We generate an identity pair
|
||||
let mut buffer = Cursor::new(Vec::<u8>::new());
|
||||
rln.key_gen(&mut buffer).unwrap();
|
||||
|
||||
// We deserialize the keygen output to obtain
|
||||
// the identity_secret and id_commitment
|
||||
let (identity_secret_hash, id_commitment) = deserialize_identity_pair(buffer.into_inner());
|
||||
```
|
||||
|
||||
### Add Rate commitment to the RLN Merkle tree
|
||||
|
||||
```rust
|
||||
// We define the tree index where id_commitment will be added
|
||||
let id_index = 10;
|
||||
let user_message_limit = 10;
|
||||
|
||||
// We serialize id_commitment and pass it to set_leaf
|
||||
let rate_commitment = poseidon_hash(&[id_commitment, user_message_limit]);
|
||||
let mut buffer = Cursor::new(serialize_field_element(rate_commitment));
|
||||
rln.set_leaf(id_index, &mut buffer).unwrap();
|
||||
```
|
||||
|
||||
Note that when tree leaves are not explicitly set by the user (in this example, all those with index less and greater than `10`), their values is set to an hardcoded default (all-`0` bytes in current implementation).
|
||||
|
||||
### Set external nullifier
|
||||
### Comments for the code above for point 5
|
||||
|
||||
The `external nullifier` includes two parameters.
|
||||
|
||||
The first one is `epoch` and it's used to identify messages received in a certain time frame. It usually corresponds to the current UNIX time but can also be set to a random value or generated by a seed, provided that it corresponds to a field element.
|
||||
The first one is `epoch` and it's used to identify messages received in a certain time frame.
|
||||
It usually corresponds to the current UNIX time but can also be set to a random value or generated by a seed,
|
||||
provided that it corresponds to a field element.
|
||||
|
||||
The second one is `rln_identifier` and it's used to prevent a RLN ZK proof generated for one application to be re-used in another one.
|
||||
The second one is `rln_identifier` and it's used to prevent a RLN ZK proof generated
|
||||
for one application to be re-used in another one.
|
||||
|
||||
```rust
|
||||
// We generate epoch from a date seed and we ensure is
|
||||
// mapped to a field element by hashing-to-field its content
|
||||
let epoch = hash_to_field(b"Today at noon, this year");
|
||||
// We generate rln_identifier from a date seed and we ensure is
|
||||
// mapped to a field element by hashing-to-field its content
|
||||
let rln_identifier = hash_to_field(b"test-rln-identifier");
|
||||
### Features
|
||||
|
||||
let external_nullifier = poseidon_hash(&[epoch, rln_identifier]);
|
||||
- **Stateless Mode**: Allows the use of RLN without maintaining state of the Merkle tree.
|
||||
- **Pre-compiled Circuits**: Ready-to-use circuits with Merkle tree depth of 20
|
||||
- **Wasm Support**: WebAssembly bindings via rln-wasm crate with features like:
|
||||
- Browser and Node.js compatibility
|
||||
- Optional parallel feature support using [wasm-bindgen-rayon](https://github.com/RReverser/wasm-bindgen-rayon)
|
||||
- Headless browser testing capabilities
|
||||
- **Merkle Tree Implementations**: Multiple tree variants optimized for different use cases:
|
||||
- **Full Merkle Tree**: Fastest access with complete pre-allocated tree in memory. Best for frequent random access (enable with `fullmerkletree` feature).
|
||||
- **Optimal Merkle Tree**: Memory-efficient sparse storage using HashMap. Ideal for partially populated trees (enable with `optimalmerkletree` feature).
|
||||
- **Persistent Merkle Tree**: Disk-based storage with [sled](https://github.com/spacejam/sled) for persistence across application restarts and large datasets (enable with `pmtree-ft` feature).
|
||||
|
||||
## Building and Testing
|
||||
|
||||
### Prerequisites
|
||||
|
||||
```sh
|
||||
git clone https://github.com/vacp2p/zerokit.git
|
||||
make installdeps
|
||||
cd zerokit/rln
|
||||
```
|
||||
|
||||
### Set signal
|
||||
### Build Commands
|
||||
|
||||
The signal is the message for which we are computing a RLN proof.
|
||||
```sh
|
||||
# Build with default features
|
||||
cargo make build
|
||||
|
||||
```rust
|
||||
// We set our signal
|
||||
let signal = b"RLN is awesome";
|
||||
# Test with default features
|
||||
cargo make test
|
||||
|
||||
# Test with stateless features
|
||||
cargo make test_stateless
|
||||
```
|
||||
|
||||
### Generate a RLN proof
|
||||
## Advanced: Custom Circuit Compilation
|
||||
|
||||
We prepare the input to the proof generation routine.
|
||||
The `circom-rln` (<https://github.com/rate-limiting-nullifier/circom-rln>) repository,
|
||||
which contains the RLN circuit implementation used for pre-compiled RLN circuit for zerokit RLN.
|
||||
If you want to compile your own RLN circuit, you can follow the instructions below.
|
||||
|
||||
Input buffer is serialized as `[ identity_key | id_index | external_nullifier | user_message_limit | message_id | signal_len | signal ]`.
|
||||
### 1. Compile ZK Circuits for getting the zkey file
|
||||
|
||||
```rust
|
||||
// We prepare input to the proof generation routine
|
||||
let proof_input = prepare_prove_input(identity_secret_hash, id_index, external_nullifier, signal);
|
||||
This script actually generates not only the zkey file for the RLN circuit,
|
||||
but also the execution wasm file used for witness calculation.
|
||||
However, the wasm file is not needed for the `rln` module,
|
||||
because current implementation uses the iden3 graph file for witness calculation.
|
||||
This graph file is generated by the `circom-witnesscalc` tool in [step 2](#2-generate-witness-calculation-graph).
|
||||
|
||||
To customize the circuit parameters, modify `circom-rln/circuits/rln.circom`:
|
||||
|
||||
```circom
|
||||
pragma circom 2.1.0;
|
||||
include "./rln.circom";
|
||||
component main { public [x, externalNullifier] } = RLN(N, M);
|
||||
```
|
||||
|
||||
We are now ready to generate a RLN ZK proof along with the _public outputs_ of the ZK circuit evaluation.
|
||||
Where:
|
||||
|
||||
```rust
|
||||
- `N`: Merkle tree depth, determining the maximum membership capacity (2^N members).
|
||||
|
||||
// We generate a RLN proof for proof_input
|
||||
let mut in_buffer = Cursor::new(proof_input);
|
||||
let mut out_buffer = Cursor::new(Vec::<u8>::new());
|
||||
rln.generate_rln_proof(&mut in_buffer, &mut out_buffer)
|
||||
.unwrap();
|
||||
- `M`: Bit size for range checks, setting an upper bound for the number of messages per epoch (2^M messages).
|
||||
|
||||
// We get the public outputs returned by the circuit evaluation
|
||||
let proof_data = out_buffer.into_inner();
|
||||
> [!NOTE]
|
||||
> However, if `N` is too big, this might require a larger Powers of Tau ceremony
|
||||
> than the one hardcoded in `./scripts/build-circuits.sh`, which is `2^14`.
|
||||
> In such case, we refer to the official
|
||||
> [Circom documentation](https://docs.circom.io/getting-started/proving-circuits/#powers-of-tau)
|
||||
> for instructions on how to run an appropriate Powers of Tau ceremony and Phase 2 in order to compile the desired circuit. \
|
||||
> Additionally, while `M` sets an upper bound on the number of messages per epoch (`2^M`),
|
||||
> you can configure lower message limit for your use case, as long as it satisfies `user_message_limit ≤ 2^M`. \
|
||||
> Currently, the `rln` module comes with a [pre-compiled](https://github.com/vacp2p/zerokit/tree/master/rln/resources/tree_depth_20)
|
||||
> RLN circuit with a Merkle tree of depth `20` and a bit size of `16`,
|
||||
> allowing up to `2^20` registered members and a `2^16` message limit per epoch.
|
||||
|
||||
#### Install circom compiler
|
||||
|
||||
You can follow the instructions below or refer to the
|
||||
[installing Circom](https://docs.circom.io/getting-started/installation/#installing-circom) guide for more details.
|
||||
Make sure to use the specific version `v2.1.0`.
|
||||
|
||||
```sh
|
||||
# Clone the circom repository
|
||||
git clone https://github.com/iden3/circom.git
|
||||
|
||||
# Checkout the specific version
|
||||
cd circom && git checkout v2.1.0
|
||||
|
||||
# Build the circom compiler
|
||||
cargo build --release
|
||||
|
||||
# Install the circom binary globally
|
||||
cargo install --path circom
|
||||
|
||||
# Check the circom version to ensure it's v2.1.0
|
||||
circom --version
|
||||
```
|
||||
|
||||
The byte vector `proof_data` is serialized as `[ zk-proof | tree_root | external_nullifier | share_x | share_y | nullifier ]`.
|
||||
#### Generate the zkey and verification key files example
|
||||
|
||||
### Verify a RLN proof
|
||||
```sh
|
||||
# Clone the circom-rln repository
|
||||
git clone https://github.com/rate-limiting-nullifier/circom-rln
|
||||
|
||||
We prepare the input to the proof verification routine.
|
||||
# Install dependencies
|
||||
cd circom-rln && npm install
|
||||
|
||||
Input buffer is serialized as `[proof_data | signal_len | signal ]`, where `proof_data` is (computed as) the output obtained by `generate_rln_proof`.
|
||||
# Build circuits
|
||||
./scripts/build-circuits.sh rln
|
||||
|
||||
```rust
|
||||
// We prepare input to the proof verification routine
|
||||
let verify_data = prepare_verify_input(proof_data, signal);
|
||||
|
||||
// We verify the zk-proof against the provided proof values
|
||||
let mut in_buffer = Cursor::new(verify_data);
|
||||
let verified = rln.verify(&mut in_buffer).unwrap();
|
||||
# Use the generated zkey file in subsequent steps
|
||||
cp zkeyFiles/rln/final.zkey <path_to_rln_final.zkey>
|
||||
```
|
||||
|
||||
We check if the proof verification was successful:
|
||||
### 2. Generate Witness Calculation Graph
|
||||
|
||||
```rust
|
||||
// We ensure the proof is valid
|
||||
assert!(verified);
|
||||
The execution graph file used for witness calculation can be compiled following instructions
|
||||
in the [circom-witnesscalc](https://github.com/iden3/circom-witnesscalc) repository.
|
||||
As mentioned in step 1, we should use `rln.circom` file from `circom-rln` repository.
|
||||
|
||||
```sh
|
||||
# Clone the circom-witnesscalc repository
|
||||
git clone https://github.com/iden3/circom-witnesscalc
|
||||
|
||||
# Load the submodules
|
||||
cd circom-witnesscalc && git submodule update --init --recursive
|
||||
|
||||
# Build the circom-witnesscalc tool
|
||||
cargo build
|
||||
|
||||
# Generate the witness calculation graph
|
||||
cargo run --package circom_witnesscalc --bin build-circuit ../circom-rln/circuits/rln.circom <path_to_graph.bin>
|
||||
```
|
||||
|
||||
## Get involved!
|
||||
The `rln` module comes with [pre-compiled](https://github.com/vacp2p/zerokit/tree/master/rln/resources/tree_depth_20)
|
||||
execution graph files for the RLN circuit.
|
||||
|
||||
### 3. Generate Arkzkey Representation for zkey file
|
||||
|
||||
For faster loading, compile the zkey file into the arkzkey format using
|
||||
[ark-zkey](https://github.com/seemenkina/ark-zkey).
|
||||
This is fork of the [original](https://github.com/zkmopro/ark-zkey) repository with the uncompressed arkzkey support.
|
||||
|
||||
```sh
|
||||
# Clone the ark-zkey repository
|
||||
git clone https://github.com/seemenkina/ark-zkey.git
|
||||
|
||||
# Build the ark-zkey tool
|
||||
cd ark-zkey && cargo build
|
||||
|
||||
# Generate the arkzkey representation for the zkey file
|
||||
cargo run --bin arkzkey-util <path_to_rln_final.zkey>
|
||||
```
|
||||
|
||||
This will generate the `rln_final.arkzkey` file, which is used by the `rln` module.
|
||||
|
||||
Currently, the `rln` module comes with
|
||||
[pre-compiled](https://github.com/vacp2p/zerokit/tree/master/rln/resources/tree_depth_20) arkzkey keys for the RLN circuit.
|
||||
|
||||
> [!NOTE]
|
||||
> You can use this [convert_zkey.sh](./convert_zkey.sh) script
|
||||
> to automate the process of generating the arkzkey file from any zkey file
|
||||
|
||||
Run the script as follows:
|
||||
|
||||
```sh
|
||||
chmod +x ./convert_zkey.sh
|
||||
./convert_zkey.sh <path_to_rln_final.zkey>
|
||||
```
|
||||
|
||||
## FFI Interface
|
||||
|
||||
RLN provides C-compatible bindings for integration with C, C++, Nim, and other languages through [safer_ffi](https://getditto.github.io/safer_ffi/).
|
||||
|
||||
The FFI layer is organized into several modules:
|
||||
|
||||
- [`ffi_rln.rs`](./src/ffi/ffi_rln.rs) – Implements core RLN functionality, including initialization functions, proof generation, and proof verification.
|
||||
- [`ffi_tree.rs`](./src/ffi/ffi_tree.rs) – Provides all tree-related operations and helper functions for Merkle tree management.
|
||||
- [`ffi_utils.rs`](./src/ffi/ffi_utils.rs) – Contains all utility functions and structure definitions used across the FFI layer.
|
||||
|
||||
### Examples
|
||||
|
||||
Working examples demonstrating proof generation, proof verification and slashing in C and Nim:
|
||||
|
||||
- [C example](./ffi_c_examples/main.c) and [README](./ffi_c_examples/Readme.md)
|
||||
- [Nim example](./ffi_nim_examples/main.nim) and [README](./ffi_nim_examples/Readme.md)
|
||||
|
||||
### Memory Management
|
||||
|
||||
- All **heap-allocated** objects returned from Rust FFI **must** be freed using their corresponding FFI `_free` functions.
|
||||
|
||||
## Detailed Protocol Flow
|
||||
|
||||
1. **Identity Creation**: Generate a secret key and commitment
|
||||
2. **Rate Commitment**: Add commitment to a Merkle tree
|
||||
3. **External Nullifier Setup**: Combine epoch and application identifier
|
||||
4. **Proof Generation**: Create a zkSNARK proof that:
|
||||
- Proves membership in the Merkle tree
|
||||
- Ensures rate-limiting constraints are satisfied
|
||||
- Generates a nullifier to prevent double-usage
|
||||
5. **Proof Verification**: Verify the proof without revealing the prover's identity
|
||||
6. **Slashing Mechanism**: Detect and penalize double-usage attempts
|
||||
|
||||
## Getting Involved
|
||||
|
||||
Zerokit RLN public and FFI APIs allow interaction with many more features than what briefly showcased above.
|
||||
|
||||
We invite you to check our API documentation by running
|
||||
|
||||
```rust
|
||||
```bash
|
||||
cargo doc --no-deps
|
||||
```
|
||||
|
||||
and look at unit tests to have an hint on how to interface and use them.
|
||||
|
||||
- Check the [unit tests](https://github.com/vacp2p/zerokit/tree/master/rln/tests) for more usage examples
|
||||
- [RFC specification](https://rfc.vac.dev/vac/raw/rln-v2) for the Rate-Limiting Nullifier protocol
|
||||
- [GitHub repository](https://github.com/vacp2p/zerokit) for the latest updates
|
||||
|
||||
@@ -1,22 +0,0 @@
|
||||
use criterion::{criterion_group, criterion_main, Criterion};
|
||||
use rln::circuit::{vk_from_ark_serialized, VK_BYTES};
|
||||
|
||||
// Here we benchmark how long the deserialization of the
|
||||
// verifying_key takes, only testing the json => verifying_key conversion,
|
||||
// and skipping conversion from bytes => string => serde_json::Value
|
||||
pub fn vk_deserialize_benchmark(c: &mut Criterion) {
|
||||
let vk = VK_BYTES;
|
||||
|
||||
c.bench_function("vk::vk_from_ark_serialized", |b| {
|
||||
b.iter(|| {
|
||||
let _ = vk_from_ark_serialized(vk);
|
||||
})
|
||||
});
|
||||
}
|
||||
|
||||
criterion_group! {
|
||||
name = benches;
|
||||
config = Criterion::default().measurement_time(std::time::Duration::from_secs(10));
|
||||
targets = vk_deserialize_benchmark
|
||||
}
|
||||
criterion_main!(benches);
|
||||
@@ -1,43 +0,0 @@
|
||||
use criterion::{criterion_group, criterion_main, Criterion};
|
||||
use rln::circuit::{
|
||||
read_arkzkey_from_bytes_compressed, read_arkzkey_from_bytes_uncompressed, ARKZKEY_BYTES,
|
||||
ARKZKEY_BYTES_UNCOMPR,
|
||||
};
|
||||
|
||||
pub fn uncompressed_bench(c: &mut Criterion) {
|
||||
let arkzkey = ARKZKEY_BYTES_UNCOMPR.to_vec();
|
||||
let size = arkzkey.len() as f32;
|
||||
println!(
|
||||
"Size of uncompressed arkzkey: {:.2?} MB",
|
||||
size / 1024.0 / 1024.0
|
||||
);
|
||||
|
||||
c.bench_function("arkzkey::arkzkey_from_raw_uncompressed", |b| {
|
||||
b.iter(|| {
|
||||
let r = read_arkzkey_from_bytes_uncompressed(&arkzkey);
|
||||
assert_eq!(r.is_ok(), true);
|
||||
})
|
||||
});
|
||||
}
|
||||
pub fn compressed_bench(c: &mut Criterion) {
|
||||
let arkzkey = ARKZKEY_BYTES.to_vec();
|
||||
let size = arkzkey.len() as f32;
|
||||
println!(
|
||||
"Size of compressed arkzkey: {:.2?} MB",
|
||||
size / 1024.0 / 1024.0
|
||||
);
|
||||
|
||||
c.bench_function("arkzkey::arkzkey_from_raw_compressed", |b| {
|
||||
b.iter(|| {
|
||||
let r = read_arkzkey_from_bytes_compressed(&arkzkey);
|
||||
assert_eq!(r.is_ok(), true);
|
||||
})
|
||||
});
|
||||
}
|
||||
|
||||
criterion_group! {
|
||||
name = benches;
|
||||
config = Criterion::default().measurement_time(std::time::Duration::from_secs(250));
|
||||
targets = uncompressed_bench, compressed_bench
|
||||
}
|
||||
criterion_main!(benches);
|
||||
@@ -1,24 +0,0 @@
|
||||
use ark_circom::read_zkey;
|
||||
use criterion::{criterion_group, criterion_main, Criterion};
|
||||
use std::io::Cursor;
|
||||
|
||||
pub fn zkey_load_benchmark(c: &mut Criterion) {
|
||||
let zkey = rln::circuit::ZKEY_BYTES.to_vec();
|
||||
let size = zkey.len() as f32;
|
||||
println!("Size of zkey: {:.2?} MB", size / 1024.0 / 1024.0);
|
||||
|
||||
c.bench_function("zkey::zkey_from_raw", |b| {
|
||||
b.iter(|| {
|
||||
let mut reader = Cursor::new(zkey.clone());
|
||||
let r = read_zkey(&mut reader);
|
||||
assert_eq!(r.is_ok(), true);
|
||||
})
|
||||
});
|
||||
}
|
||||
|
||||
criterion_group! {
|
||||
name = benches;
|
||||
config = Criterion::default().measurement_time(std::time::Duration::from_secs(250));
|
||||
targets = zkey_load_benchmark
|
||||
}
|
||||
criterion_main!(benches);
|
||||
@@ -1,11 +1,11 @@
|
||||
use criterion::{criterion_group, criterion_main, Criterion};
|
||||
use rln::{circuit::Fr, pm_tree_adapter::PmTree};
|
||||
use utils::ZerokitMerkleTree;
|
||||
use rln::prelude::*;
|
||||
use zerokit_utils::merkle_tree::ZerokitMerkleTree;
|
||||
|
||||
pub fn pmtree_benchmark(c: &mut Criterion) {
|
||||
let mut tree = PmTree::default(2).unwrap();
|
||||
|
||||
let leaves: Vec<Fr> = (0..4).map(|s| Fr::from(s)).collect();
|
||||
let leaves: Vec<Fr> = (0..4).map(Fr::from).collect();
|
||||
|
||||
c.bench_function("Pmtree::set", |b| {
|
||||
b.iter(|| {
|
||||
@@ -13,7 +13,7 @@ pub fn pmtree_benchmark(c: &mut Criterion) {
|
||||
})
|
||||
});
|
||||
|
||||
c.bench_function("Pmtree:delete", |b| {
|
||||
c.bench_function("Pmtree::delete", |b| {
|
||||
b.iter(|| {
|
||||
tree.delete(0).unwrap();
|
||||
})
|
||||
@@ -21,17 +21,11 @@ pub fn pmtree_benchmark(c: &mut Criterion) {
|
||||
|
||||
c.bench_function("Pmtree::override_range", |b| {
|
||||
b.iter(|| {
|
||||
tree.override_range(0, leaves.clone(), [0, 1, 2, 3])
|
||||
tree.override_range(0, leaves.clone().into_iter(), [0, 1, 2, 3].into_iter())
|
||||
.unwrap();
|
||||
})
|
||||
});
|
||||
|
||||
c.bench_function("Pmtree::compute_root", |b| {
|
||||
b.iter(|| {
|
||||
tree.compute_root().unwrap();
|
||||
})
|
||||
});
|
||||
|
||||
c.bench_function("Pmtree::get", |b| {
|
||||
b.iter(|| {
|
||||
tree.get(0).unwrap();
|
||||
|
||||
@@ -1,18 +1,15 @@
|
||||
use criterion::{criterion_group, criterion_main, BenchmarkId, Criterion};
|
||||
use rln::{
|
||||
circuit::{Fr, TEST_TREE_HEIGHT},
|
||||
hashers::PoseidonHash,
|
||||
};
|
||||
use utils::{FullMerkleTree, OptimalMerkleTree, ZerokitMerkleTree};
|
||||
use rln::prelude::*;
|
||||
use zerokit_utils::merkle_tree::{FullMerkleTree, OptimalMerkleTree, ZerokitMerkleTree};
|
||||
|
||||
pub fn get_leaves(n: u32) -> Vec<Fr> {
|
||||
(0..n).map(|s| Fr::from(s)).collect()
|
||||
(0..n).map(Fr::from).collect()
|
||||
}
|
||||
|
||||
pub fn optimal_merkle_tree_poseidon_benchmark(c: &mut Criterion) {
|
||||
c.bench_function("OptimalMerkleTree::<Poseidon>::full_height_gen", |b| {
|
||||
c.bench_function("OptimalMerkleTree::<Poseidon>::full_depth_gen", |b| {
|
||||
b.iter(|| {
|
||||
OptimalMerkleTree::<PoseidonHash>::default(TEST_TREE_HEIGHT).unwrap();
|
||||
OptimalMerkleTree::<PoseidonHash>::default(DEFAULT_TREE_DEPTH).unwrap();
|
||||
})
|
||||
});
|
||||
|
||||
@@ -20,7 +17,7 @@ pub fn optimal_merkle_tree_poseidon_benchmark(c: &mut Criterion) {
|
||||
for &n in [1u32, 10, 100].iter() {
|
||||
let leaves = get_leaves(n);
|
||||
|
||||
let mut tree = OptimalMerkleTree::<PoseidonHash>::default(TEST_TREE_HEIGHT).unwrap();
|
||||
let mut tree = OptimalMerkleTree::<PoseidonHash>::default(DEFAULT_TREE_DEPTH).unwrap();
|
||||
group.bench_function(
|
||||
BenchmarkId::new("OptimalMerkleTree::<Poseidon>::set", n),
|
||||
|b| {
|
||||
@@ -41,9 +38,9 @@ pub fn optimal_merkle_tree_poseidon_benchmark(c: &mut Criterion) {
|
||||
}
|
||||
|
||||
pub fn full_merkle_tree_poseidon_benchmark(c: &mut Criterion) {
|
||||
c.bench_function("FullMerkleTree::<Poseidon>::full_height_gen", |b| {
|
||||
c.bench_function("FullMerkleTree::<Poseidon>::full_depth_gen", |b| {
|
||||
b.iter(|| {
|
||||
FullMerkleTree::<PoseidonHash>::default(TEST_TREE_HEIGHT).unwrap();
|
||||
FullMerkleTree::<PoseidonHash>::default(DEFAULT_TREE_DEPTH).unwrap();
|
||||
})
|
||||
});
|
||||
|
||||
@@ -51,7 +48,7 @@ pub fn full_merkle_tree_poseidon_benchmark(c: &mut Criterion) {
|
||||
for &n in [1u32, 10, 100].iter() {
|
||||
let leaves = get_leaves(n);
|
||||
|
||||
let mut tree = FullMerkleTree::<PoseidonHash>::default(TEST_TREE_HEIGHT).unwrap();
|
||||
let mut tree = FullMerkleTree::<PoseidonHash>::default(DEFAULT_TREE_DEPTH).unwrap();
|
||||
group.bench_function(
|
||||
BenchmarkId::new("FullMerkleTree::<Poseidon>::set", n),
|
||||
|b| {
|
||||
|
||||
56
rln/convert_zkey.sh
Executable file
56
rln/convert_zkey.sh
Executable file
@@ -0,0 +1,56 @@
|
||||
#!/bin/bash
|
||||
|
||||
# Convert zkey to arkzkey using /tmp directory
|
||||
# Usage: ./convert_zkey.sh <path_to_zkey_file>
|
||||
|
||||
set -e
|
||||
|
||||
# Check input
|
||||
if [ $# -eq 0 ]; then
|
||||
echo "Usage: $0 <path_to_zkey_file>"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
ZKEY_FILE="$1"
|
||||
|
||||
if [ ! -f "$ZKEY_FILE" ]; then
|
||||
echo "Error: File '$ZKEY_FILE' does not exist"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
# Get absolute path before changing directories
|
||||
ZKEY_ABSOLUTE_PATH=$(realpath "$ZKEY_FILE")
|
||||
|
||||
# Create temp directory in /tmp
|
||||
TEMP_DIR="/tmp/ark-zkey-$$"
|
||||
echo "Using temp directory: $TEMP_DIR"
|
||||
|
||||
# Cleanup function
|
||||
cleanup() {
|
||||
echo "Cleaning up temp directory: $TEMP_DIR"
|
||||
rm -rf "$TEMP_DIR"
|
||||
}
|
||||
|
||||
# Setup cleanup trap
|
||||
trap cleanup EXIT
|
||||
|
||||
# Create temp directory and clone ark-zkey
|
||||
mkdir -p "$TEMP_DIR"
|
||||
cd "$TEMP_DIR"
|
||||
git clone https://github.com/seemenkina/ark-zkey.git
|
||||
cd ark-zkey
|
||||
cargo build
|
||||
|
||||
# Convert
|
||||
cargo run --bin arkzkey-util "$ZKEY_ABSOLUTE_PATH"
|
||||
|
||||
# Check if arkzkey file was created (tool creates it in same directory as input)
|
||||
ARKZKEY_FILE="${ZKEY_ABSOLUTE_PATH%.zkey}.arkzkey"
|
||||
|
||||
if [ ! -f "$ARKZKEY_FILE" ]; then
|
||||
echo "Could not find generated .arkzkey file at $ARKZKEY_FILE"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
echo "Conversion successful!"
|
||||
echo "Output file: $ARKZKEY_FILE"
|
||||
47
rln/ffi_c_examples/Readme.md
Normal file
47
rln/ffi_c_examples/Readme.md
Normal file
@@ -0,0 +1,47 @@
|
||||
# RLN FFI C example
|
||||
|
||||
This example demonstrates how to use the RLN C FFI in both stateless and non-stateless modes.
|
||||
|
||||
## Non-stateless mode
|
||||
|
||||
### Compile lib non-stateless
|
||||
|
||||
```bash
|
||||
cargo build -p rln
|
||||
cargo run --features headers --bin generate-headers
|
||||
mv -v rln.h rln/ffi_c_examples/
|
||||
```
|
||||
|
||||
### Compile and run example non-stateless
|
||||
|
||||
```bash
|
||||
cd rln/ffi_c_examples/
|
||||
gcc -Wall main.c -o main -lrln -L../../target/debug
|
||||
./main
|
||||
```
|
||||
|
||||
## Stateless mode
|
||||
|
||||
### Compile lib stateless
|
||||
|
||||
```bash
|
||||
cargo build -p rln --no-default-features --features stateless
|
||||
cargo run --no-default-features --features stateless,headers --bin generate-headers
|
||||
mv -v rln.h rln/ffi_c_examples/
|
||||
```
|
||||
|
||||
### Compile example stateless
|
||||
|
||||
```bash
|
||||
cd rln/ffi_c_examples/
|
||||
gcc -Wall -DSTATELESS main.c -o main -lrln -L../../target/debug
|
||||
./main
|
||||
```
|
||||
|
||||
## Note
|
||||
|
||||
### Find C lib used by Rust
|
||||
|
||||
```bash
|
||||
cargo +nightly rustc --release -p rln -- -Z unstable-options --print native-static-libs
|
||||
```
|
||||
668
rln/ffi_c_examples/main.c
Normal file
668
rln/ffi_c_examples/main.c
Normal file
@@ -0,0 +1,668 @@
|
||||
#include <stdlib.h>
|
||||
#include <stdio.h>
|
||||
#include <string.h>
|
||||
|
||||
#include "rln.h"
|
||||
|
||||
int main(int argc, char const *const argv[])
|
||||
{
|
||||
printf("Creating RLN instance\n");
|
||||
|
||||
#ifdef STATELESS
|
||||
CResult_FFI_RLN_ptr_Vec_uint8_t ffi_rln_new_result = ffi_rln_new();
|
||||
#else
|
||||
const char *config_path = "../resources/tree_depth_20/config.json";
|
||||
CResult_FFI_RLN_ptr_Vec_uint8_t ffi_rln_new_result = ffi_rln_new(20, config_path);
|
||||
#endif
|
||||
|
||||
if (!ffi_rln_new_result.ok)
|
||||
{
|
||||
fprintf(stderr, "Initial RLN instance creation error: %s\n", ffi_rln_new_result.err.ptr);
|
||||
ffi_c_string_free(ffi_rln_new_result.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
|
||||
FFI_RLN_t *rln = ffi_rln_new_result.ok;
|
||||
printf("RLN instance created successfully\n");
|
||||
|
||||
printf("\nGenerating identity keys\n");
|
||||
CResult_Vec_CFr_Vec_uint8_t keys_result = ffi_key_gen();
|
||||
if (keys_result.err.ptr)
|
||||
{
|
||||
fprintf(stderr, "Key generation error: %s\n", keys_result.err.ptr);
|
||||
ffi_c_string_free(keys_result.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
Vec_CFr_t keys = keys_result.ok;
|
||||
const CFr_t *identity_secret = ffi_vec_cfr_get(&keys, 0);
|
||||
const CFr_t *id_commitment = ffi_vec_cfr_get(&keys, 1);
|
||||
printf("Identity generated\n");
|
||||
|
||||
Vec_uint8_t debug = ffi_cfr_debug(identity_secret);
|
||||
printf(" - identity_secret = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
|
||||
debug = ffi_cfr_debug(id_commitment);
|
||||
printf(" - id_commitment = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
|
||||
printf("\nCreating message limit\n");
|
||||
CFr_t *user_message_limit = ffi_uint_to_cfr(1);
|
||||
|
||||
debug = ffi_cfr_debug(user_message_limit);
|
||||
printf(" - user_message_limit = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
|
||||
printf("\nComputing rate commitment\n");
|
||||
CResult_CFr_ptr_Vec_uint8_t rate_commitment_result = ffi_poseidon_hash_pair(id_commitment, user_message_limit);
|
||||
if (!rate_commitment_result.ok)
|
||||
{
|
||||
fprintf(stderr, "Rate commitment hash error: %s\n", rate_commitment_result.err.ptr);
|
||||
ffi_c_string_free(rate_commitment_result.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
CFr_t *rate_commitment = rate_commitment_result.ok;
|
||||
|
||||
debug = ffi_cfr_debug(rate_commitment);
|
||||
printf(" - rate_commitment = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
|
||||
printf("\nCFr serialization: CFr <-> bytes\n");
|
||||
Vec_uint8_t ser_rate_commitment = ffi_cfr_to_bytes_le(rate_commitment);
|
||||
|
||||
debug = ffi_vec_u8_debug(&ser_rate_commitment);
|
||||
printf(" - serialized rate_commitment = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
|
||||
CResult_CFr_ptr_Vec_uint8_t deser_rate_commitment_result = ffi_bytes_le_to_cfr(&ser_rate_commitment);
|
||||
if (!deser_rate_commitment_result.ok)
|
||||
{
|
||||
fprintf(stderr, "Rate commitment deserialization error: %s\n", deser_rate_commitment_result.err.ptr);
|
||||
ffi_c_string_free(deser_rate_commitment_result.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
CFr_t *deser_rate_commitment = deser_rate_commitment_result.ok;
|
||||
|
||||
debug = ffi_cfr_debug(deser_rate_commitment);
|
||||
printf(" - deserialized rate_commitment = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
|
||||
ffi_vec_u8_free(ser_rate_commitment);
|
||||
ffi_cfr_free(deser_rate_commitment);
|
||||
|
||||
printf("\nVec<CFr> serialization: Vec<CFr> <-> bytes\n");
|
||||
Vec_uint8_t ser_keys = ffi_vec_cfr_to_bytes_le(&keys);
|
||||
|
||||
debug = ffi_vec_u8_debug(&ser_keys);
|
||||
printf(" - serialized keys = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
|
||||
CResult_Vec_CFr_Vec_uint8_t deser_keys_result = ffi_bytes_le_to_vec_cfr(&ser_keys);
|
||||
if (deser_keys_result.err.ptr)
|
||||
{
|
||||
fprintf(stderr, "Keys deserialization error: %s\n", deser_keys_result.err.ptr);
|
||||
ffi_c_string_free(deser_keys_result.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
|
||||
debug = ffi_vec_cfr_debug(&deser_keys_result.ok);
|
||||
printf(" - deserialized keys = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
|
||||
Vec_CFr_t deser_keys = deser_keys_result.ok;
|
||||
ffi_vec_cfr_free(deser_keys);
|
||||
|
||||
ffi_vec_u8_free(ser_keys);
|
||||
|
||||
#ifdef STATELESS
|
||||
#define TREE_DEPTH 20
|
||||
#define CFR_SIZE 32
|
||||
|
||||
printf("\nBuilding Merkle path for stateless mode\n");
|
||||
CFr_t *default_leaf = ffi_cfr_zero();
|
||||
|
||||
CFr_t *default_hashes[TREE_DEPTH - 1];
|
||||
CResult_CFr_ptr_Vec_uint8_t hash_result = ffi_poseidon_hash_pair(default_leaf, default_leaf);
|
||||
if (!hash_result.ok)
|
||||
{
|
||||
fprintf(stderr, "Poseidon hash error: %s\n", hash_result.err.ptr);
|
||||
ffi_c_string_free(hash_result.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
default_hashes[0] = hash_result.ok;
|
||||
for (size_t i = 1; i < TREE_DEPTH - 1; i++)
|
||||
{
|
||||
hash_result = ffi_poseidon_hash_pair(default_hashes[i - 1], default_hashes[i - 1]);
|
||||
if (!hash_result.ok)
|
||||
{
|
||||
fprintf(stderr, "Poseidon hash error: %s\n", hash_result.err.ptr);
|
||||
ffi_c_string_free(hash_result.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
default_hashes[i] = hash_result.ok;
|
||||
}
|
||||
|
||||
Vec_CFr_t path_elements = ffi_vec_cfr_new(TREE_DEPTH);
|
||||
ffi_vec_cfr_push(&path_elements, default_leaf);
|
||||
for (size_t i = 0; i < TREE_DEPTH - 1; i++)
|
||||
{
|
||||
ffi_vec_cfr_push(&path_elements, default_hashes[i]);
|
||||
}
|
||||
|
||||
printf("\nVec<CFr> serialization: Vec<CFr> <-> bytes\n");
|
||||
Vec_uint8_t ser_path_elements = ffi_vec_cfr_to_bytes_le(&path_elements);
|
||||
|
||||
debug = ffi_vec_u8_debug(&ser_path_elements);
|
||||
printf(" - serialized path_elements = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
|
||||
CResult_Vec_CFr_Vec_uint8_t deser_path_elements_result = ffi_bytes_le_to_vec_cfr(&ser_path_elements);
|
||||
if (deser_path_elements_result.err.ptr)
|
||||
{
|
||||
fprintf(stderr, "Path elements deserialization error: %s\n", deser_path_elements_result.err.ptr);
|
||||
ffi_c_string_free(deser_path_elements_result.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
|
||||
debug = ffi_vec_cfr_debug(&deser_path_elements_result.ok);
|
||||
printf(" - deserialized path_elements = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
|
||||
Vec_CFr_t deser_path_elements = deser_path_elements_result.ok;
|
||||
ffi_vec_cfr_free(deser_path_elements);
|
||||
|
||||
ffi_vec_u8_free(ser_path_elements);
|
||||
|
||||
uint8_t path_index_arr[TREE_DEPTH] = {0};
|
||||
Vec_uint8_t identity_path_index = {
|
||||
.ptr = path_index_arr,
|
||||
.len = TREE_DEPTH,
|
||||
.cap = TREE_DEPTH};
|
||||
|
||||
printf("\nVec<uint8> serialization: Vec<uint8> <-> bytes\n");
|
||||
Vec_uint8_t ser_path_index = ffi_vec_u8_to_bytes_le(&identity_path_index);
|
||||
|
||||
debug = ffi_vec_u8_debug(&ser_path_index);
|
||||
printf(" - serialized path_index = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
|
||||
CResult_Vec_uint8_Vec_uint8_t deser_path_index_result = ffi_bytes_le_to_vec_u8(&ser_path_index);
|
||||
if (deser_path_index_result.err.ptr)
|
||||
{
|
||||
fprintf(stderr, "Path index deserialization error: %s\n", deser_path_index_result.err.ptr);
|
||||
ffi_c_string_free(deser_path_index_result.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
|
||||
debug = ffi_vec_u8_debug(&deser_path_index_result.ok);
|
||||
printf(" - deserialized path_index = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
|
||||
Vec_uint8_t deser_path_index = deser_path_index_result.ok;
|
||||
ffi_vec_u8_free(deser_path_index);
|
||||
|
||||
ffi_vec_u8_free(ser_path_index);
|
||||
|
||||
printf("\nComputing Merkle root for stateless mode\n");
|
||||
printf(" - computing root for index 0 with rate_commitment\n");
|
||||
CResult_CFr_ptr_Vec_uint8_t root_result = ffi_poseidon_hash_pair(rate_commitment, default_leaf);
|
||||
if (!root_result.ok)
|
||||
{
|
||||
fprintf(stderr, "Poseidon hash error: %s\n", root_result.err.ptr);
|
||||
ffi_c_string_free(root_result.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
CFr_t *computed_root = root_result.ok;
|
||||
for (size_t i = 1; i < TREE_DEPTH; i++)
|
||||
{
|
||||
root_result = ffi_poseidon_hash_pair(computed_root, default_hashes[i - 1]);
|
||||
if (!root_result.ok)
|
||||
{
|
||||
fprintf(stderr, "Poseidon hash error: %s\n", root_result.err.ptr);
|
||||
ffi_c_string_free(root_result.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
CFr_t *next_root = root_result.ok;
|
||||
ffi_cfr_free(computed_root);
|
||||
computed_root = next_root;
|
||||
}
|
||||
|
||||
debug = ffi_cfr_debug(computed_root);
|
||||
printf(" - computed_root = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
#else
|
||||
printf("\nAdding rate_commitment to tree\n");
|
||||
CBoolResult_t set_err = ffi_set_next_leaf(&rln, rate_commitment);
|
||||
if (!set_err.ok)
|
||||
{
|
||||
fprintf(stderr, "Set next leaf error: %s\n", set_err.err.ptr);
|
||||
ffi_c_string_free(set_err.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
|
||||
size_t leaf_index = ffi_leaves_set(&rln) - 1;
|
||||
printf(" - added to tree at index %zu\n", leaf_index);
|
||||
|
||||
printf("\nGetting Merkle proof\n");
|
||||
CResult_FFI_MerkleProof_ptr_Vec_uint8_t proof_result = ffi_get_merkle_proof(&rln, leaf_index);
|
||||
if (!proof_result.ok)
|
||||
{
|
||||
fprintf(stderr, "Get proof error: %s\n", proof_result.err.ptr);
|
||||
ffi_c_string_free(proof_result.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
FFI_MerkleProof_t *merkle_proof = proof_result.ok;
|
||||
printf(" - proof obtained (depth: %zu)\n", merkle_proof->path_elements.len);
|
||||
#endif
|
||||
|
||||
printf("\nHashing signal\n");
|
||||
uint8_t signal[32] = {1, 2, 3, 4, 5, 6, 7, 8, 9, 10};
|
||||
Vec_uint8_t signal_vec = {signal, 32, 32};
|
||||
CResult_CFr_ptr_Vec_uint8_t x_result = ffi_hash_to_field_le(&signal_vec);
|
||||
if (!x_result.ok)
|
||||
{
|
||||
fprintf(stderr, "Hash signal error: %s\n", x_result.err.ptr);
|
||||
ffi_c_string_free(x_result.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
CFr_t *x = x_result.ok;
|
||||
|
||||
debug = ffi_cfr_debug(x);
|
||||
printf(" - x = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
|
||||
printf("\nHashing epoch\n");
|
||||
const char *epoch_str = "test-epoch";
|
||||
Vec_uint8_t epoch_vec = {(uint8_t *)epoch_str, strlen(epoch_str), strlen(epoch_str)};
|
||||
CResult_CFr_ptr_Vec_uint8_t epoch_result = ffi_hash_to_field_le(&epoch_vec);
|
||||
if (!epoch_result.ok)
|
||||
{
|
||||
fprintf(stderr, "Hash epoch error: %s\n", epoch_result.err.ptr);
|
||||
ffi_c_string_free(epoch_result.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
CFr_t *epoch = epoch_result.ok;
|
||||
|
||||
debug = ffi_cfr_debug(epoch);
|
||||
printf(" - epoch = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
|
||||
printf("\nHashing RLN identifier\n");
|
||||
const char *rln_id_str = "test-rln-identifier";
|
||||
Vec_uint8_t rln_id_vec = {(uint8_t *)rln_id_str, strlen(rln_id_str), strlen(rln_id_str)};
|
||||
CResult_CFr_ptr_Vec_uint8_t rln_identifier_result = ffi_hash_to_field_le(&rln_id_vec);
|
||||
if (!rln_identifier_result.ok)
|
||||
{
|
||||
fprintf(stderr, "Hash RLN identifier error: %s\n", rln_identifier_result.err.ptr);
|
||||
ffi_c_string_free(rln_identifier_result.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
CFr_t *rln_identifier = rln_identifier_result.ok;
|
||||
|
||||
debug = ffi_cfr_debug(rln_identifier);
|
||||
printf(" - rln_identifier = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
|
||||
printf("\nComputing Poseidon hash for external nullifier\n");
|
||||
CResult_CFr_ptr_Vec_uint8_t external_nullifier_result = ffi_poseidon_hash_pair(epoch, rln_identifier);
|
||||
if (!external_nullifier_result.ok)
|
||||
{
|
||||
fprintf(stderr, "External nullifier hash error: %s\n", external_nullifier_result.err.ptr);
|
||||
ffi_c_string_free(external_nullifier_result.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
CFr_t *external_nullifier = external_nullifier_result.ok;
|
||||
|
||||
debug = ffi_cfr_debug(external_nullifier);
|
||||
printf(" - external_nullifier = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
|
||||
printf("\nCreating message_id\n");
|
||||
CFr_t *message_id = ffi_uint_to_cfr(0);
|
||||
|
||||
debug = ffi_cfr_debug(message_id);
|
||||
printf(" - message_id = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
|
||||
printf("\nCreating RLN Witness\n");
|
||||
#ifdef STATELESS
|
||||
CResult_FFI_RLNWitnessInput_ptr_Vec_uint8_t witness_result = ffi_rln_witness_input_new(
|
||||
identity_secret,
|
||||
user_message_limit,
|
||||
message_id,
|
||||
&path_elements,
|
||||
&identity_path_index,
|
||||
x,
|
||||
external_nullifier);
|
||||
|
||||
if (!witness_result.ok)
|
||||
{
|
||||
fprintf(stderr, "RLN Witness creation error: %s\n", witness_result.err.ptr);
|
||||
ffi_c_string_free(witness_result.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
FFI_RLNWitnessInput_t *witness = witness_result.ok;
|
||||
printf("RLN Witness created successfully\n");
|
||||
#else
|
||||
CResult_FFI_RLNWitnessInput_ptr_Vec_uint8_t witness_result = ffi_rln_witness_input_new(
|
||||
identity_secret,
|
||||
user_message_limit,
|
||||
message_id,
|
||||
&merkle_proof->path_elements,
|
||||
&merkle_proof->path_index,
|
||||
x,
|
||||
external_nullifier);
|
||||
|
||||
if (!witness_result.ok)
|
||||
{
|
||||
fprintf(stderr, "RLN Witness creation error: %s\n", witness_result.err.ptr);
|
||||
ffi_c_string_free(witness_result.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
FFI_RLNWitnessInput_t *witness = witness_result.ok;
|
||||
printf("RLN Witness created successfully\n");
|
||||
#endif
|
||||
|
||||
printf("\nRLNWitnessInput serialization: RLNWitnessInput <-> bytes\n");
|
||||
CResult_Vec_uint8_Vec_uint8_t ser_witness_result = ffi_rln_witness_to_bytes_le(&witness);
|
||||
if (ser_witness_result.err.ptr)
|
||||
{
|
||||
fprintf(stderr, "Witness serialization error: %s\n", ser_witness_result.err.ptr);
|
||||
ffi_c_string_free(ser_witness_result.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
Vec_uint8_t ser_witness = ser_witness_result.ok;
|
||||
|
||||
debug = ffi_vec_u8_debug(&ser_witness);
|
||||
printf(" - serialized witness = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
|
||||
CResult_FFI_RLNWitnessInput_ptr_Vec_uint8_t deser_witness_result = ffi_bytes_le_to_rln_witness(&ser_witness);
|
||||
if (!deser_witness_result.ok)
|
||||
{
|
||||
fprintf(stderr, "Witness deserialization error: %s\n", deser_witness_result.err.ptr);
|
||||
ffi_c_string_free(deser_witness_result.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
|
||||
FFI_RLNWitnessInput_t *deser_witness = deser_witness_result.ok;
|
||||
printf(" - witness deserialized successfully\n");
|
||||
|
||||
ffi_rln_witness_input_free(deser_witness);
|
||||
ffi_vec_u8_free(ser_witness);
|
||||
|
||||
printf("\nGenerating RLN Proof\n");
|
||||
CResult_FFI_RLNProof_ptr_Vec_uint8_t proof_gen_result = ffi_generate_rln_proof(
|
||||
&rln,
|
||||
&witness);
|
||||
|
||||
if (!proof_gen_result.ok)
|
||||
{
|
||||
fprintf(stderr, "Proof generation error: %s\n", proof_gen_result.err.ptr);
|
||||
ffi_c_string_free(proof_gen_result.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
|
||||
FFI_RLNProof_t *rln_proof = proof_gen_result.ok;
|
||||
printf("Proof generated successfully\n");
|
||||
|
||||
printf("\nGetting proof values\n");
|
||||
FFI_RLNProofValues_t *proof_values = ffi_rln_proof_get_values(&rln_proof);
|
||||
|
||||
CFr_t *y = ffi_rln_proof_values_get_y(&proof_values);
|
||||
debug = ffi_cfr_debug(y);
|
||||
printf(" - y = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
ffi_cfr_free(y);
|
||||
|
||||
CFr_t *nullifier = ffi_rln_proof_values_get_nullifier(&proof_values);
|
||||
debug = ffi_cfr_debug(nullifier);
|
||||
printf(" - nullifier = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
ffi_cfr_free(nullifier);
|
||||
|
||||
CFr_t *root = ffi_rln_proof_values_get_root(&proof_values);
|
||||
debug = ffi_cfr_debug(root);
|
||||
printf(" - root = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
ffi_cfr_free(root);
|
||||
|
||||
CFr_t *x_val = ffi_rln_proof_values_get_x(&proof_values);
|
||||
debug = ffi_cfr_debug(x_val);
|
||||
printf(" - x = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
ffi_cfr_free(x_val);
|
||||
|
||||
CFr_t *ext_nullifier = ffi_rln_proof_values_get_external_nullifier(&proof_values);
|
||||
debug = ffi_cfr_debug(ext_nullifier);
|
||||
printf(" - external_nullifier = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
ffi_cfr_free(ext_nullifier);
|
||||
|
||||
printf("\nRLNProof serialization: RLNProof <-> bytes\n");
|
||||
CResult_Vec_uint8_Vec_uint8_t ser_proof_result = ffi_rln_proof_to_bytes_le(&rln_proof);
|
||||
if (ser_proof_result.err.ptr)
|
||||
{
|
||||
fprintf(stderr, "Proof serialization error: %s\n", ser_proof_result.err.ptr);
|
||||
ffi_c_string_free(ser_proof_result.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
Vec_uint8_t ser_proof = ser_proof_result.ok;
|
||||
|
||||
debug = ffi_vec_u8_debug(&ser_proof);
|
||||
printf(" - serialized proof = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
|
||||
CResult_FFI_RLNProof_ptr_Vec_uint8_t deser_proof_result = ffi_bytes_le_to_rln_proof(&ser_proof);
|
||||
if (!deser_proof_result.ok)
|
||||
{
|
||||
fprintf(stderr, "Proof deserialization error: %s\n", deser_proof_result.err.ptr);
|
||||
ffi_c_string_free(deser_proof_result.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
|
||||
FFI_RLNProof_t *deser_proof = deser_proof_result.ok;
|
||||
printf(" - proof deserialized successfully\n");
|
||||
|
||||
printf("\nRLNProofValues serialization: RLNProofValues <-> bytes\n");
|
||||
Vec_uint8_t ser_proof_values = ffi_rln_proof_values_to_bytes_le(&proof_values);
|
||||
|
||||
debug = ffi_vec_u8_debug(&ser_proof_values);
|
||||
printf(" - serialized proof_values = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
|
||||
CResult_FFI_RLNProofValues_ptr_Vec_uint8_t deser_proof_values_result = ffi_bytes_le_to_rln_proof_values(&ser_proof_values);
|
||||
if (!deser_proof_values_result.ok)
|
||||
{
|
||||
fprintf(stderr, "Proof values deserialization error: %s\n", deser_proof_values_result.err.ptr);
|
||||
ffi_c_string_free(deser_proof_values_result.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
FFI_RLNProofValues_t *deser_proof_values = deser_proof_values_result.ok;
|
||||
printf(" - proof_values deserialized successfully\n");
|
||||
|
||||
CFr_t *deser_external_nullifier = ffi_rln_proof_values_get_external_nullifier(&deser_proof_values);
|
||||
debug = ffi_cfr_debug(deser_external_nullifier);
|
||||
printf(" - deserialized external_nullifier = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
ffi_cfr_free(deser_external_nullifier);
|
||||
|
||||
ffi_rln_proof_values_free(deser_proof_values);
|
||||
ffi_vec_u8_free(ser_proof_values);
|
||||
ffi_rln_proof_free(deser_proof);
|
||||
ffi_vec_u8_free(ser_proof);
|
||||
|
||||
printf("\nVerifying Proof\n");
|
||||
#ifdef STATELESS
|
||||
Vec_CFr_t roots = ffi_vec_cfr_from_cfr(computed_root);
|
||||
CBoolResult_t verify_err = ffi_verify_with_roots(&rln, &rln_proof, &roots, x);
|
||||
#else
|
||||
CBoolResult_t verify_err = ffi_verify_rln_proof(&rln, &rln_proof, x);
|
||||
#endif
|
||||
|
||||
if (!verify_err.ok)
|
||||
{
|
||||
fprintf(stderr, "Proof verification error: %s\n", verify_err.err.ptr);
|
||||
ffi_c_string_free(verify_err.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
|
||||
printf("Proof verified successfully\n");
|
||||
|
||||
ffi_rln_proof_free(rln_proof);
|
||||
|
||||
printf("\nSimulating double-signaling attack (same epoch, different message)\n");
|
||||
|
||||
printf("\nHashing second signal\n");
|
||||
uint8_t signal2[32] = {11, 12, 13, 14, 15, 16, 17, 18, 19, 20};
|
||||
Vec_uint8_t signal2_vec = {signal2, 32, 32};
|
||||
CResult_CFr_ptr_Vec_uint8_t x2_result = ffi_hash_to_field_le(&signal2_vec);
|
||||
if (!x2_result.ok)
|
||||
{
|
||||
fprintf(stderr, "Hash second signal error: %s\n", x2_result.err.ptr);
|
||||
ffi_c_string_free(x2_result.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
CFr_t *x2 = x2_result.ok;
|
||||
|
||||
debug = ffi_cfr_debug(x2);
|
||||
printf(" - x2 = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
|
||||
printf("\nCreating second message with the same id\n");
|
||||
CFr_t *message_id2 = ffi_uint_to_cfr(0);
|
||||
|
||||
debug = ffi_cfr_debug(message_id2);
|
||||
printf(" - message_id2 = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
|
||||
printf("\nCreating second RLN Witness\n");
|
||||
#ifdef STATELESS
|
||||
CResult_FFI_RLNWitnessInput_ptr_Vec_uint8_t witness_result2 = ffi_rln_witness_input_new(
|
||||
identity_secret,
|
||||
user_message_limit,
|
||||
message_id2,
|
||||
&path_elements,
|
||||
&identity_path_index,
|
||||
x2,
|
||||
external_nullifier);
|
||||
|
||||
if (!witness_result2.ok)
|
||||
{
|
||||
fprintf(stderr, "Second RLN Witness creation error: %s\n", witness_result2.err.ptr);
|
||||
ffi_c_string_free(witness_result2.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
FFI_RLNWitnessInput_t *witness2 = witness_result2.ok;
|
||||
printf("Second RLN Witness created successfully\n");
|
||||
#else
|
||||
CResult_FFI_RLNWitnessInput_ptr_Vec_uint8_t witness_result2 = ffi_rln_witness_input_new(
|
||||
identity_secret,
|
||||
user_message_limit,
|
||||
message_id2,
|
||||
&merkle_proof->path_elements,
|
||||
&merkle_proof->path_index,
|
||||
x2,
|
||||
external_nullifier);
|
||||
|
||||
if (!witness_result2.ok)
|
||||
{
|
||||
fprintf(stderr, "Second RLN Witness creation error: %s\n", witness_result2.err.ptr);
|
||||
ffi_c_string_free(witness_result2.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
FFI_RLNWitnessInput_t *witness2 = witness_result2.ok;
|
||||
printf("Second RLN Witness created successfully\n");
|
||||
#endif
|
||||
printf("\nGenerating second RLN Proof\n");
|
||||
CResult_FFI_RLNProof_ptr_Vec_uint8_t proof_gen_result2 = ffi_generate_rln_proof(
|
||||
&rln,
|
||||
&witness2);
|
||||
|
||||
if (!proof_gen_result2.ok)
|
||||
{
|
||||
fprintf(stderr, "Second proof generation error: %s\n", proof_gen_result2.err.ptr);
|
||||
ffi_c_string_free(proof_gen_result2.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
|
||||
FFI_RLNProof_t *rln_proof2 = proof_gen_result2.ok;
|
||||
printf("Second proof generated successfully\n");
|
||||
|
||||
FFI_RLNProofValues_t *proof_values2 = ffi_rln_proof_get_values(&rln_proof2);
|
||||
|
||||
printf("\nVerifying second proof\n");
|
||||
#ifdef STATELESS
|
||||
CBoolResult_t verify_err2 = ffi_verify_with_roots(&rln, &rln_proof2, &roots, x2);
|
||||
#else
|
||||
CBoolResult_t verify_err2 = ffi_verify_rln_proof(&rln, &rln_proof2, x2);
|
||||
#endif
|
||||
|
||||
if (!verify_err2.ok)
|
||||
{
|
||||
fprintf(stderr, "Proof verification error: %s\n", verify_err2.err.ptr);
|
||||
ffi_c_string_free(verify_err2.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
|
||||
printf("Second proof verified successfully\n");
|
||||
|
||||
ffi_rln_proof_free(rln_proof2);
|
||||
|
||||
printf("\nRecovering identity secret\n");
|
||||
CResult_CFr_ptr_Vec_uint8_t recover_result = ffi_recover_id_secret(&proof_values, &proof_values2);
|
||||
if (!recover_result.ok)
|
||||
{
|
||||
fprintf(stderr, "Identity recovery error: %s\n", recover_result.err.ptr);
|
||||
ffi_c_string_free(recover_result.err);
|
||||
return EXIT_FAILURE;
|
||||
}
|
||||
|
||||
CFr_t *recovered_secret = recover_result.ok;
|
||||
|
||||
debug = ffi_cfr_debug(recovered_secret);
|
||||
printf(" - recovered_secret = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
|
||||
debug = ffi_cfr_debug(identity_secret);
|
||||
printf(" - original_secret = %s\n", debug.ptr);
|
||||
ffi_c_string_free(debug);
|
||||
|
||||
printf("Slashing successful: Identity is recovered!\n");
|
||||
|
||||
ffi_cfr_free(recovered_secret);
|
||||
|
||||
ffi_rln_proof_values_free(proof_values2);
|
||||
ffi_rln_proof_values_free(proof_values);
|
||||
ffi_cfr_free(x2);
|
||||
ffi_cfr_free(message_id2);
|
||||
|
||||
#ifdef STATELESS
|
||||
ffi_rln_witness_input_free(witness2);
|
||||
ffi_rln_witness_input_free(witness);
|
||||
ffi_vec_cfr_free(roots);
|
||||
ffi_vec_cfr_free(path_elements);
|
||||
for (size_t i = 0; i < TREE_DEPTH - 1; i++)
|
||||
{
|
||||
ffi_cfr_free(default_hashes[i]);
|
||||
}
|
||||
ffi_cfr_free(default_leaf);
|
||||
ffi_cfr_free(computed_root);
|
||||
#else
|
||||
ffi_rln_witness_input_free(witness2);
|
||||
ffi_rln_witness_input_free(witness);
|
||||
ffi_merkle_proof_free(merkle_proof);
|
||||
#endif
|
||||
|
||||
ffi_cfr_free(rate_commitment);
|
||||
ffi_cfr_free(x);
|
||||
ffi_cfr_free(epoch);
|
||||
ffi_cfr_free(rln_identifier);
|
||||
ffi_cfr_free(external_nullifier);
|
||||
ffi_cfr_free(user_message_limit);
|
||||
ffi_cfr_free(message_id);
|
||||
ffi_vec_cfr_free(keys);
|
||||
ffi_rln_free(rln);
|
||||
|
||||
return EXIT_SUCCESS;
|
||||
}
|
||||
124
rln/ffi_nim_examples/README.md
Normal file
124
rln/ffi_nim_examples/README.md
Normal file
@@ -0,0 +1,124 @@
|
||||
# RLN FFI Nim example
|
||||
|
||||
This example demonstrates how to use the RLN C FFI from Nim in both stateless and non-stateless modes. It covers:
|
||||
|
||||
- Creating an RLN handle (stateless or with Merkle tree backend)
|
||||
- Generating identity keys and commitments
|
||||
- Building a witness (mock Merkle path in stateless mode, real Merkle proof in non-stateless mode)
|
||||
- Generating and verifying a proof
|
||||
- Serializing/deserializing FFI objects (CFr, Vec\<CFr>, RLNWitnessInput, RLNProof, RLNProofValues)
|
||||
- Simulating a double-signaling attack and recovering the identity secret
|
||||
|
||||
## Build the RLN library
|
||||
|
||||
From the repository root:
|
||||
|
||||
```bash
|
||||
# Stateless build (no tree APIs)
|
||||
cargo build -p rln --release --no-default-features --features stateless
|
||||
|
||||
# Non-stateless build (with tree APIs)
|
||||
cargo build -p rln --release
|
||||
```
|
||||
|
||||
This produces the shared library in `target/release`:
|
||||
|
||||
- macOS: `librln.dylib`
|
||||
- Linux: `librln.so`
|
||||
- Windows: `rln.dll`
|
||||
|
||||
## Build the Nim example (two modes)
|
||||
|
||||
From this directory:
|
||||
|
||||
```bash
|
||||
# Stateless mode (no tree APIs, uses mock Merkle path)
|
||||
nim c -d:release -d:ffiStateless main.nim
|
||||
|
||||
# Non-stateless mode (uses exported tree APIs to insert leaf and fetch proof)
|
||||
nim c -d:release main.nim
|
||||
```
|
||||
|
||||
Notes:
|
||||
|
||||
- The example links dynamically. If your OS linker cannot find the library at runtime,
|
||||
set an rpath or environment variable as shown below.
|
||||
- The example auto-picks a platform-specific default library name.
|
||||
You can override it with `-d:RLN_LIB:"/absolute/path/to/lib"` if needed.
|
||||
|
||||
## Run the example
|
||||
|
||||
Ensure the dynamic loader can find the RLN library, then run the binary.
|
||||
|
||||
macOS:
|
||||
|
||||
```bash
|
||||
DYLD_LIBRARY_PATH=../../target/release ./main
|
||||
```
|
||||
|
||||
Linux:
|
||||
|
||||
```bash
|
||||
LD_LIBRARY_PATH=../../target/release ./main
|
||||
```
|
||||
|
||||
Windows (PowerShell):
|
||||
|
||||
```powershell
|
||||
$env:PATH = "$PWD\..\..\target\release;$env:PATH"
|
||||
./main.exe
|
||||
```
|
||||
|
||||
You should see detailed output showing each step, for example:
|
||||
|
||||
```text
|
||||
Creating RLN instance
|
||||
RLN instance created successfully
|
||||
|
||||
Generating identity keys
|
||||
Identity generated
|
||||
- identity_secret = ...
|
||||
- id_commitment = ...
|
||||
|
||||
Creating message limit
|
||||
- user_message_limit = ...
|
||||
|
||||
Computing rate commitment
|
||||
- rate_commitment = ...
|
||||
|
||||
CFr serialization: CFr <-> bytes
|
||||
- serialized rate_commitment = ...
|
||||
- deserialized rate_commitment = ...
|
||||
|
||||
Vec<CFr> serialization: Vec<CFr> <-> bytes
|
||||
- serialized keys = ...
|
||||
- deserialized keys = ...
|
||||
|
||||
... (Merkle path, hashing, witness, proof, verification, and slashing steps) ...
|
||||
|
||||
Proof verified successfully
|
||||
Slashing successful: Identity is recovered!
|
||||
```
|
||||
|
||||
## What the example does
|
||||
|
||||
### Stateless mode
|
||||
|
||||
1. Creates an RLN handle via the stateless constructor.
|
||||
2. Generates identity keys, sets a `user_message_limit` and `message_id`.
|
||||
3. Hashes a signal, epoch, and RLN identifier to field elements.
|
||||
4. Computes `rateCommitment = Poseidon(id_commitment, user_message_limit)`.
|
||||
5. Builds a mock Merkle path for an empty depth-20 tree at index 0 (no exported tree APIs):
|
||||
- Path siblings: level 0 sibling is `0`, then each level uses precomputed default hashes `H(0,0)`, `H(H(0,0),H(0,0))`, ...
|
||||
- Path indices: all zeros (left at every level)
|
||||
- Root: folds the path upwards with `rateCommitment` at index 0
|
||||
6. Builds the witness, generates the proof, and verifies it with `ffi_verify_with_roots`, passing a one-element roots vector containing the computed root.
|
||||
7. Simulates a double-signaling attack and recovers the identity secret from two proofs.
|
||||
|
||||
### Non-stateless mode
|
||||
|
||||
1. Creates an RLN handle with a Merkle tree backend and configuration.
|
||||
2. Generates identity keys and computes `rateCommitment = Poseidon(id_commitment, user_message_limit)`.
|
||||
3. Inserts the leaf with `ffi_set_next_leaf` and fetches a real Merkle path for index 0 via `ffi_get_merkle_proof`.
|
||||
4. Builds the witness from the exported proof, generates the proof, and verifies with `ffi_verify_rln_proof` using the current tree root.
|
||||
5. Simulates a double-signaling attack and recovers the identity secret from two proofs.
|
||||
940
rln/ffi_nim_examples/main.nim
Normal file
940
rln/ffi_nim_examples/main.nim
Normal file
@@ -0,0 +1,940 @@
|
||||
# Embed rpaths to find Cargo's built library relative to the executable
|
||||
when defined(macosx):
|
||||
{.passL: "-Wl,-rpath,@executable_path/../../target/release".}
|
||||
when defined(linux):
|
||||
{.passL: "-Wl,-rpath,'$ORIGIN/../../target/release'".}
|
||||
|
||||
# Portable dynlib name with override capability (-d:RLN_LIB:"...")
|
||||
when defined(macosx):
|
||||
const RLN_LIB* {.strdefine.} = "librln.dylib"
|
||||
elif defined(linux):
|
||||
const RLN_LIB* {.strdefine.} = "librln.so"
|
||||
elif defined(windows):
|
||||
const RLN_LIB* {.strdefine.} = "rln.dll"
|
||||
else:
|
||||
const RLN_LIB* {.strdefine.} = "rln"
|
||||
|
||||
# FFI objects
|
||||
type
|
||||
CSize* = csize_t
|
||||
CFr* = object
|
||||
FFI_RLN* = object
|
||||
FFI_RLNProof* = object
|
||||
FFI_RLNWitnessInput* = object
|
||||
|
||||
Vec_CFr* = object
|
||||
dataPtr*: ptr CFr
|
||||
len*: CSize
|
||||
cap*: CSize
|
||||
|
||||
Vec_uint8* = object
|
||||
dataPtr*: ptr uint8
|
||||
len*: CSize
|
||||
cap*: CSize
|
||||
|
||||
SliceRefU8* = object
|
||||
dataPtr*: ptr uint8
|
||||
len*: CSize
|
||||
|
||||
FFI_MerkleProof* = object
|
||||
path_elements*: Vec_CFr
|
||||
path_index*: Vec_uint8
|
||||
|
||||
CResultRLNPtrVecU8* = object
|
||||
ok*: ptr FFI_RLN
|
||||
err*: Vec_uint8
|
||||
|
||||
CResultProofPtrVecU8* = object
|
||||
ok*: ptr FFI_RLNProof
|
||||
err*: Vec_uint8
|
||||
|
||||
CResultWitnessInputPtrVecU8* = object
|
||||
ok*: ptr FFI_RLNWitnessInput
|
||||
err*: Vec_uint8
|
||||
|
||||
FFI_RLNProofValues* = object
|
||||
|
||||
CResultCFrPtrVecU8* = object
|
||||
ok*: ptr CFr
|
||||
err*: Vec_uint8
|
||||
|
||||
CResultRLNProofValuesPtrVecU8* = object
|
||||
ok*: ptr FFI_RLNProofValues
|
||||
err*: Vec_uint8
|
||||
|
||||
CResultMerkleProofPtrVecU8* = object
|
||||
ok*: ptr FFI_MerkleProof
|
||||
err*: Vec_uint8
|
||||
|
||||
CResultVecCFrVecU8* = object
|
||||
ok*: Vec_CFr
|
||||
err*: Vec_uint8
|
||||
|
||||
CResultVecU8VecU8* = object
|
||||
ok*: Vec_uint8
|
||||
err*: Vec_uint8
|
||||
|
||||
CResultBigIntJsonVecU8* = object
|
||||
ok*: Vec_uint8
|
||||
err*: Vec_uint8
|
||||
|
||||
CBoolResult* = object
|
||||
ok*: bool
|
||||
err*: Vec_uint8
|
||||
|
||||
# CFr functions
|
||||
proc ffi_cfr_zero*(): ptr CFr {.importc: "ffi_cfr_zero", cdecl,
|
||||
dynlib: RLN_LIB.}
|
||||
proc ffi_cfr_one*(): ptr CFr {.importc: "ffi_cfr_one", cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_cfr_free*(x: ptr CFr) {.importc: "ffi_cfr_free", cdecl,
|
||||
dynlib: RLN_LIB.}
|
||||
proc ffi_uint_to_cfr*(value: uint32): ptr CFr {.importc: "ffi_uint_to_cfr",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_cfr_debug*(cfr: ptr CFr): Vec_uint8 {.importc: "ffi_cfr_debug", cdecl,
|
||||
dynlib: RLN_LIB.}
|
||||
proc ffi_cfr_to_bytes_le*(cfr: ptr CFr): Vec_uint8 {.importc: "ffi_cfr_to_bytes_le",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_cfr_to_bytes_be*(cfr: ptr CFr): Vec_uint8 {.importc: "ffi_cfr_to_bytes_be",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_bytes_le_to_cfr*(bytes: ptr Vec_uint8): CResultCFrPtrVecU8 {.importc: "ffi_bytes_le_to_cfr",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_bytes_be_to_cfr*(bytes: ptr Vec_uint8): CResultCFrPtrVecU8 {.importc: "ffi_bytes_be_to_cfr",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
|
||||
# Vec<CFr> functions
|
||||
proc ffi_vec_cfr_new*(capacity: CSize): Vec_CFr {.importc: "ffi_vec_cfr_new",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_vec_cfr_from_cfr*(cfr: ptr CFr): Vec_CFr {.importc: "ffi_vec_cfr_from_cfr",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_vec_cfr_push*(v: ptr Vec_CFr, cfr: ptr CFr) {.importc: "ffi_vec_cfr_push",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_vec_cfr_len*(v: ptr Vec_CFr): CSize {.importc: "ffi_vec_cfr_len",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_vec_cfr_get*(v: ptr Vec_CFr, i: CSize): ptr CFr {.importc: "ffi_vec_cfr_get",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_vec_cfr_to_bytes_le*(v: ptr Vec_CFr): Vec_uint8 {.importc: "ffi_vec_cfr_to_bytes_le",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_vec_cfr_to_bytes_be*(v: ptr Vec_CFr): Vec_uint8 {.importc: "ffi_vec_cfr_to_bytes_be",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_bytes_le_to_vec_cfr*(bytes: ptr Vec_uint8): CResultVecCFrVecU8 {.importc: "ffi_bytes_le_to_vec_cfr",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_bytes_be_to_vec_cfr*(bytes: ptr Vec_uint8): CResultVecCFrVecU8 {.importc: "ffi_bytes_be_to_vec_cfr",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_vec_cfr_debug*(v: ptr Vec_CFr): Vec_uint8 {.importc: "ffi_vec_cfr_debug",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_vec_cfr_free*(v: Vec_CFr) {.importc: "ffi_vec_cfr_free", cdecl,
|
||||
dynlib: RLN_LIB.}
|
||||
|
||||
# Vec<u8> functions
|
||||
proc ffi_vec_u8_to_bytes_le*(v: ptr Vec_uint8): Vec_uint8 {.importc: "ffi_vec_u8_to_bytes_le",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_vec_u8_to_bytes_be*(v: ptr Vec_uint8): Vec_uint8 {.importc: "ffi_vec_u8_to_bytes_be",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_bytes_le_to_vec_u8*(bytes: ptr Vec_uint8): CResultVecU8VecU8 {.importc: "ffi_bytes_le_to_vec_u8",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_bytes_be_to_vec_u8*(bytes: ptr Vec_uint8): CResultVecU8VecU8 {.importc: "ffi_bytes_be_to_vec_u8",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_vec_u8_debug*(v: ptr Vec_uint8): Vec_uint8 {.importc: "ffi_vec_u8_debug",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_vec_u8_free*(v: Vec_uint8) {.importc: "ffi_vec_u8_free", cdecl,
|
||||
dynlib: RLN_LIB.}
|
||||
|
||||
# Hashing functions
|
||||
proc ffi_hash_to_field_le*(input: ptr Vec_uint8): CResultCFrPtrVecU8 {.importc: "ffi_hash_to_field_le",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_hash_to_field_be*(input: ptr Vec_uint8): CResultCFrPtrVecU8 {.importc: "ffi_hash_to_field_be",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_poseidon_hash_pair*(a: ptr CFr,
|
||||
b: ptr CFr): CResultCFrPtrVecU8 {.importc: "ffi_poseidon_hash_pair", cdecl,
|
||||
dynlib: RLN_LIB.}
|
||||
|
||||
# Keygen function
|
||||
proc ffi_key_gen*(): CResultVecCFrVecU8 {.importc: "ffi_key_gen", cdecl,
|
||||
dynlib: RLN_LIB.}
|
||||
proc ffi_seeded_key_gen*(seed: ptr Vec_uint8): CResultVecCFrVecU8 {.importc: "ffi_seeded_key_gen",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_extended_key_gen*(): CResultVecCFrVecU8 {.importc: "ffi_extended_key_gen",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_seeded_extended_key_gen*(seed: ptr Vec_uint8): CResultVecCFrVecU8 {.importc: "ffi_seeded_extended_key_gen",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
|
||||
# RLN instance functions
|
||||
when defined(ffiStateless):
|
||||
proc ffi_rln_new*(): CResultRLNPtrVecU8 {.importc: "ffi_rln_new", cdecl,
|
||||
dynlib: RLN_LIB.}
|
||||
proc ffi_rln_new_with_params*(zkey_data: ptr Vec_uint8,
|
||||
graph_data: ptr Vec_uint8): CResultRLNPtrVecU8 {.importc: "ffi_rln_new_with_params",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
else:
|
||||
proc ffi_rln_new*(treeDepth: CSize, config: cstring): CResultRLNPtrVecU8 {.importc: "ffi_rln_new",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_rln_new_with_params*(treeDepth: CSize, zkey_data: ptr Vec_uint8,
|
||||
graph_data: ptr Vec_uint8, config: cstring): CResultRLNPtrVecU8 {.importc: "ffi_rln_new_with_params",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
|
||||
proc ffi_rln_free*(rln: ptr FFI_RLN) {.importc: "ffi_rln_free", cdecl,
|
||||
dynlib: RLN_LIB.}
|
||||
|
||||
# Witness input functions
|
||||
proc ffi_rln_witness_input_new*(
|
||||
identity_secret: ptr CFr,
|
||||
user_message_limit: ptr CFr,
|
||||
message_id: ptr CFr,
|
||||
path_elements: ptr Vec_CFr,
|
||||
identity_path_index: ptr Vec_uint8,
|
||||
x: ptr CFr,
|
||||
external_nullifier: ptr CFr
|
||||
): CResultWitnessInputPtrVecU8 {.importc: "ffi_rln_witness_input_new", cdecl,
|
||||
dynlib: RLN_LIB.}
|
||||
proc ffi_rln_witness_to_bytes_le*(witness: ptr ptr FFI_RLNWitnessInput): CResultVecU8VecU8 {.importc: "ffi_rln_witness_to_bytes_le",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_rln_witness_to_bytes_be*(witness: ptr ptr FFI_RLNWitnessInput): CResultVecU8VecU8 {.importc: "ffi_rln_witness_to_bytes_be",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_bytes_le_to_rln_witness*(bytes: ptr Vec_uint8): CResultWitnessInputPtrVecU8 {.importc: "ffi_bytes_le_to_rln_witness",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_bytes_be_to_rln_witness*(bytes: ptr Vec_uint8): CResultWitnessInputPtrVecU8 {.importc: "ffi_bytes_be_to_rln_witness",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_rln_witness_to_bigint_json*(witness: ptr ptr FFI_RLNWitnessInput): CResultBigIntJsonVecU8 {.importc: "ffi_rln_witness_to_bigint_json",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_rln_witness_input_free*(witness: ptr FFI_RLNWitnessInput) {.importc: "ffi_rln_witness_input_free",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
|
||||
# Proof generation/verification functions
|
||||
proc ffi_generate_rln_proof*(
|
||||
rln: ptr ptr FFI_RLN,
|
||||
witness: ptr ptr FFI_RLNWitnessInput
|
||||
): CResultProofPtrVecU8 {.importc: "ffi_generate_rln_proof", cdecl,
|
||||
dynlib: RLN_LIB.}
|
||||
|
||||
proc ffi_generate_rln_proof_with_witness*(
|
||||
rln: ptr ptr FFI_RLN,
|
||||
calculated_witness: ptr Vec_uint8,
|
||||
witness: ptr ptr FFI_RLNWitnessInput
|
||||
): CResultProofPtrVecU8 {.importc: "ffi_generate_rln_proof_with_witness",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
|
||||
when not defined(ffiStateless):
|
||||
proc ffi_verify_rln_proof*(
|
||||
rln: ptr ptr FFI_RLN,
|
||||
proof: ptr ptr FFI_RLNProof,
|
||||
x: ptr CFr
|
||||
): CBoolResult {.importc: "ffi_verify_rln_proof", cdecl,
|
||||
dynlib: RLN_LIB.}
|
||||
|
||||
proc ffi_verify_with_roots*(
|
||||
rln: ptr ptr FFI_RLN,
|
||||
proof: ptr ptr FFI_RLNProof,
|
||||
roots: ptr Vec_CFr,
|
||||
x: ptr CFr
|
||||
): CBoolResult {.importc: "ffi_verify_with_roots", cdecl,
|
||||
dynlib: RLN_LIB.}
|
||||
|
||||
proc ffi_rln_proof_free*(p: ptr FFI_RLNProof) {.importc: "ffi_rln_proof_free",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
|
||||
# Merkle tree operations (non-stateless mode)
|
||||
when not defined(ffiStateless):
|
||||
proc ffi_set_tree*(rln: ptr ptr FFI_RLN,
|
||||
tree_depth: CSize): CBoolResult {.importc: "ffi_set_tree",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_delete_leaf*(rln: ptr ptr FFI_RLN,
|
||||
index: CSize): CBoolResult {.importc: "ffi_delete_leaf",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_set_leaf*(rln: ptr ptr FFI_RLN, index: CSize,
|
||||
leaf: ptr CFr): CBoolResult {.importc: "ffi_set_leaf",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_get_leaf*(rln: ptr ptr FFI_RLN,
|
||||
index: CSize): CResultCFrPtrVecU8 {.importc: "ffi_get_leaf",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_set_next_leaf*(rln: ptr ptr FFI_RLN,
|
||||
leaf: ptr CFr): CBoolResult {.importc: "ffi_set_next_leaf",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_set_leaves_from*(rln: ptr ptr FFI_RLN, index: CSize,
|
||||
leaves: ptr Vec_CFr): CBoolResult {.importc: "ffi_set_leaves_from",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_init_tree_with_leaves*(rln: ptr ptr FFI_RLN,
|
||||
leaves: ptr Vec_CFr): CBoolResult {.importc: "ffi_init_tree_with_leaves",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_atomic_operation*(rln: ptr ptr FFI_RLN, index: CSize,
|
||||
leaves: ptr Vec_CFr,
|
||||
indices: ptr Vec_uint8): CBoolResult {.importc: "ffi_atomic_operation",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_seq_atomic_operation*(rln: ptr ptr FFI_RLN, leaves: ptr Vec_CFr,
|
||||
indices: ptr Vec_uint8): CBoolResult {.importc: "ffi_seq_atomic_operation",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_get_root*(rln: ptr ptr FFI_RLN): ptr CFr {.importc: "ffi_get_root",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_leaves_set*(rln: ptr ptr FFI_RLN): CSize {.importc: "ffi_leaves_set",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_get_merkle_proof*(rln: ptr ptr FFI_RLN,
|
||||
index: CSize): CResultMerkleProofPtrVecU8 {.importc: "ffi_get_merkle_proof",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_set_metadata*(rln: ptr ptr FFI_RLN,
|
||||
metadata: ptr Vec_uint8): CBoolResult {.importc: "ffi_set_metadata",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_get_metadata*(rln: ptr ptr FFI_RLN): CResultVecU8VecU8 {.importc: "ffi_get_metadata",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_flush*(rln: ptr ptr FFI_RLN): CBoolResult {.importc: "ffi_flush",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_merkle_proof_free*(p: ptr FFI_MerkleProof) {.importc: "ffi_merkle_proof_free",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
|
||||
# Identity secret recovery
|
||||
proc ffi_recover_id_secret*(proof_values_1: ptr ptr FFI_RLNProofValues,
|
||||
proof_values_2: ptr ptr FFI_RLNProofValues): CResultCFrPtrVecU8 {.importc: "ffi_recover_id_secret",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
|
||||
# RLNProof serialization
|
||||
proc ffi_rln_proof_to_bytes_le*(proof: ptr ptr FFI_RLNProof): CResultVecU8VecU8 {.importc: "ffi_rln_proof_to_bytes_le",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_rln_proof_to_bytes_be*(proof: ptr ptr FFI_RLNProof): CResultVecU8VecU8 {.importc: "ffi_rln_proof_to_bytes_be",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_bytes_le_to_rln_proof*(bytes: ptr Vec_uint8): CResultProofPtrVecU8 {.importc: "ffi_bytes_le_to_rln_proof",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_bytes_be_to_rln_proof*(bytes: ptr Vec_uint8): CResultProofPtrVecU8 {.importc: "ffi_bytes_be_to_rln_proof",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
|
||||
# RLNProofValues functions
|
||||
proc ffi_rln_proof_get_values*(proof: ptr ptr FFI_RLNProof): ptr FFI_RLNProofValues {.importc: "ffi_rln_proof_get_values",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_rln_proof_values_get_y*(pv: ptr ptr FFI_RLNProofValues): ptr CFr {.importc: "ffi_rln_proof_values_get_y",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_rln_proof_values_get_nullifier*(pv: ptr ptr FFI_RLNProofValues): ptr CFr {.importc: "ffi_rln_proof_values_get_nullifier",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_rln_proof_values_get_root*(pv: ptr ptr FFI_RLNProofValues): ptr CFr {.importc: "ffi_rln_proof_values_get_root",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_rln_proof_values_get_x*(pv: ptr ptr FFI_RLNProofValues): ptr CFr {.importc: "ffi_rln_proof_values_get_x",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_rln_proof_values_get_external_nullifier*(pv: ptr ptr FFI_RLNProofValues): ptr CFr {.importc: "ffi_rln_proof_values_get_external_nullifier",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_rln_proof_values_to_bytes_le*(pv: ptr ptr FFI_RLNProofValues): Vec_uint8 {.importc: "ffi_rln_proof_values_to_bytes_le",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_rln_proof_values_to_bytes_be*(pv: ptr ptr FFI_RLNProofValues): Vec_uint8 {.importc: "ffi_rln_proof_values_to_bytes_be",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_bytes_le_to_rln_proof_values*(bytes: ptr Vec_uint8): CResultRLNProofValuesPtrVecU8 {.importc: "ffi_bytes_le_to_rln_proof_values",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_bytes_be_to_rln_proof_values*(bytes: ptr Vec_uint8): CResultRLNProofValuesPtrVecU8 {.importc: "ffi_bytes_be_to_rln_proof_values",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
proc ffi_rln_proof_values_free*(pv: ptr FFI_RLNProofValues) {.importc: "ffi_rln_proof_values_free",
|
||||
cdecl, dynlib: RLN_LIB.}
|
||||
|
||||
# Helpers functions
|
||||
proc asVecU8*(buf: var seq[uint8]): Vec_uint8 =
|
||||
result.dataPtr = if buf.len == 0: nil else: addr buf[0]
|
||||
result.len = CSize(buf.len)
|
||||
result.cap = CSize(buf.len)
|
||||
|
||||
proc asString*(v: Vec_uint8): string =
|
||||
if v.dataPtr.isNil or v.len == 0: return ""
|
||||
result = newString(v.len.int)
|
||||
copyMem(addr result[0], v.dataPtr, v.len.int)
|
||||
|
||||
proc ffi_c_string_free*(s: Vec_uint8) {.importc: "ffi_c_string_free", cdecl,
|
||||
dynlib: RLN_LIB.}
|
||||
|
||||
when isMainModule:
|
||||
echo "Creating RLN instance"
|
||||
|
||||
var rlnRes: CResultRLNPtrVecU8
|
||||
when defined(ffiStateless):
|
||||
rlnRes = ffi_rln_new()
|
||||
else:
|
||||
let config_path = """../resources/tree_depth_20/config.json""".cstring
|
||||
rlnRes = ffi_rln_new(CSize(20), config_path)
|
||||
|
||||
if rlnRes.ok.isNil:
|
||||
stderr.writeLine "Initial RLN instance creation error: ", asString(rlnRes.err)
|
||||
ffi_c_string_free(rlnRes.err)
|
||||
quit 1
|
||||
|
||||
var rln = rlnRes.ok
|
||||
echo "RLN instance created successfully"
|
||||
|
||||
echo "\nGenerating identity keys"
|
||||
var keysResult = ffi_key_gen()
|
||||
if keysResult.err.dataPtr != nil:
|
||||
let errMsg = asString(keysResult.err)
|
||||
ffi_c_string_free(keysResult.err)
|
||||
echo "Key generation error: ", errMsg
|
||||
quit 1
|
||||
var keys = keysResult.ok
|
||||
let identitySecret = ffi_vec_cfr_get(addr keys, CSize(0))
|
||||
let idCommitment = ffi_vec_cfr_get(addr keys, CSize(1))
|
||||
echo "Identity generated"
|
||||
|
||||
block:
|
||||
let debug = ffi_cfr_debug(identitySecret)
|
||||
echo " - identity_secret = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
|
||||
block:
|
||||
let debug = ffi_cfr_debug(idCommitment)
|
||||
echo " - id_commitment = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
|
||||
echo "\nCreating message limit"
|
||||
let userMessageLimit = ffi_uint_to_cfr(1'u32)
|
||||
|
||||
block:
|
||||
let debug = ffi_cfr_debug(userMessageLimit)
|
||||
echo " - user_message_limit = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
|
||||
echo "\nComputing rate commitment"
|
||||
let rateCommitmentResult = ffi_poseidon_hash_pair(idCommitment, userMessageLimit)
|
||||
if rateCommitmentResult.ok.isNil:
|
||||
let errMsg = asString(rateCommitmentResult.err)
|
||||
ffi_c_string_free(rateCommitmentResult.err)
|
||||
echo "Rate commitment hash error: ", errMsg
|
||||
quit 1
|
||||
let rateCommitment = rateCommitmentResult.ok
|
||||
|
||||
block:
|
||||
let debug = ffi_cfr_debug(rateCommitment)
|
||||
echo " - rate_commitment = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
|
||||
echo "\nCFr serialization: CFr <-> bytes"
|
||||
var serRateCommitment = ffi_cfr_to_bytes_be(rateCommitment)
|
||||
|
||||
block:
|
||||
let debug = ffi_vec_u8_debug(addr serRateCommitment)
|
||||
echo " - serialized rate_commitment = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
|
||||
let deserRateCommitmentResult = ffi_bytes_be_to_cfr(addr serRateCommitment)
|
||||
if deserRateCommitmentResult.ok.isNil:
|
||||
stderr.writeLine "Rate commitment deserialization error: ", asString(
|
||||
deserRateCommitmentResult.err)
|
||||
ffi_c_string_free(deserRateCommitmentResult.err)
|
||||
quit 1
|
||||
let deserRateCommitment = deserRateCommitmentResult.ok
|
||||
|
||||
block:
|
||||
let debug = ffi_cfr_debug(deserRateCommitment)
|
||||
echo " - deserialized rate_commitment = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
|
||||
ffi_vec_u8_free(serRateCommitment)
|
||||
ffi_cfr_free(deserRateCommitment)
|
||||
|
||||
echo "\nVec<CFr> serialization: Vec<CFr> <-> bytes"
|
||||
var serKeys = ffi_vec_cfr_to_bytes_be(addr keys)
|
||||
|
||||
block:
|
||||
let debug = ffi_vec_u8_debug(addr serKeys)
|
||||
echo " - serialized keys = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
|
||||
let deserKeysResult = ffi_bytes_be_to_vec_cfr(addr serKeys)
|
||||
if deserKeysResult.err.dataPtr != nil:
|
||||
stderr.writeLine "Keys deserialization error: ", asString(
|
||||
deserKeysResult.err)
|
||||
ffi_c_string_free(deserKeysResult.err)
|
||||
quit 1
|
||||
|
||||
block:
|
||||
var okKeys = deserKeysResult.ok
|
||||
let debug = ffi_vec_cfr_debug(addr okKeys)
|
||||
echo " - deserialized keys = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
|
||||
ffi_vec_cfr_free(deserKeysResult.ok)
|
||||
ffi_vec_u8_free(serKeys)
|
||||
|
||||
when defined(ffiStateless):
|
||||
const treeDepth = 20
|
||||
const CFR_SIZE = 32
|
||||
|
||||
echo "\nBuilding Merkle path for stateless mode"
|
||||
|
||||
let defaultLeaf = ffi_cfr_zero()
|
||||
var defaultHashes: array[treeDepth-1, ptr CFr]
|
||||
block:
|
||||
let hashResult = ffi_poseidon_hash_pair(defaultLeaf, defaultLeaf)
|
||||
if hashResult.ok.isNil:
|
||||
let errMsg = asString(hashResult.err)
|
||||
ffi_c_string_free(hashResult.err)
|
||||
echo "Poseidon hash error: ", errMsg
|
||||
quit 1
|
||||
defaultHashes[0] = hashResult.ok
|
||||
for i in 1..treeDepth-2:
|
||||
let hashResult = ffi_poseidon_hash_pair(defaultHashes[i-1], defaultHashes[i-1])
|
||||
if hashResult.ok.isNil:
|
||||
let errMsg = asString(hashResult.err)
|
||||
ffi_c_string_free(hashResult.err)
|
||||
echo "Poseidon hash error: ", errMsg
|
||||
quit 1
|
||||
defaultHashes[i] = hashResult.ok
|
||||
|
||||
var pathElements = ffi_vec_cfr_new(CSize(treeDepth))
|
||||
ffi_vec_cfr_push(addr pathElements, defaultLeaf)
|
||||
for i in 0..treeDepth-2:
|
||||
ffi_vec_cfr_push(addr pathElements, defaultHashes[i])
|
||||
|
||||
echo "\nVec<CFr> serialization: Vec<CFr> <-> bytes"
|
||||
var serPathElements = ffi_vec_cfr_to_bytes_be(addr pathElements)
|
||||
|
||||
block:
|
||||
let debug = ffi_vec_u8_debug(addr serPathElements)
|
||||
echo " - serialized path_elements = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
|
||||
let deserPathElements = ffi_bytes_be_to_vec_cfr(addr serPathElements)
|
||||
if deserPathElements.err.dataPtr != nil:
|
||||
stderr.writeLine "Path elements deserialization error: ", asString(
|
||||
deserPathElements.err)
|
||||
ffi_c_string_free(deserPathElements.err)
|
||||
quit 1
|
||||
|
||||
block:
|
||||
var okPathElems = deserPathElements.ok
|
||||
let debug = ffi_vec_cfr_debug(addr okPathElems)
|
||||
echo " - deserialized path_elements = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
|
||||
ffi_vec_cfr_free(deserPathElements.ok)
|
||||
ffi_vec_u8_free(serPathElements)
|
||||
|
||||
var pathIndexSeq = newSeq[uint8](treeDepth)
|
||||
var identityPathIndex = asVecU8(pathIndexSeq)
|
||||
|
||||
echo "\nVec<uint8> serialization: Vec<uint8> <-> bytes"
|
||||
var serPathIndex = ffi_vec_u8_to_bytes_be(addr identityPathIndex)
|
||||
|
||||
block:
|
||||
let debug = ffi_vec_u8_debug(addr serPathIndex)
|
||||
echo " - serialized path_index = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
|
||||
let deserPathIndex = ffi_bytes_be_to_vec_u8(addr serPathIndex)
|
||||
if deserPathIndex.err.dataPtr != nil:
|
||||
stderr.writeLine "Path index deserialization error: ", asString(
|
||||
deserPathIndex.err)
|
||||
ffi_c_string_free(deserPathIndex.err)
|
||||
quit 1
|
||||
|
||||
block:
|
||||
var okPathIdx = deserPathIndex.ok
|
||||
let debug = ffi_vec_u8_debug(addr okPathIdx)
|
||||
echo " - deserialized path_index = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
|
||||
ffi_vec_u8_free(deserPathIndex.ok)
|
||||
ffi_vec_u8_free(serPathIndex)
|
||||
|
||||
echo "\nComputing Merkle root for stateless mode"
|
||||
echo " - computing root for index 0 with rate_commitment"
|
||||
let rootResult = ffi_poseidon_hash_pair(rateCommitment, defaultLeaf)
|
||||
if rootResult.ok.isNil:
|
||||
let errMsg = asString(rootResult.err)
|
||||
ffi_c_string_free(rootResult.err)
|
||||
echo "Poseidon hash error: ", errMsg
|
||||
quit 1
|
||||
var computedRoot = rootResult.ok
|
||||
for i in 1..treeDepth-1:
|
||||
let nextResult = ffi_poseidon_hash_pair(computedRoot, defaultHashes[i-1])
|
||||
if nextResult.ok.isNil:
|
||||
let errMsg = asString(nextResult.err)
|
||||
ffi_c_string_free(nextResult.err)
|
||||
echo "Poseidon hash error: ", errMsg
|
||||
quit 1
|
||||
let next = nextResult.ok
|
||||
ffi_cfr_free(computedRoot)
|
||||
computedRoot = next
|
||||
|
||||
block:
|
||||
let debug = ffi_cfr_debug(computedRoot)
|
||||
echo " - computed_root = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
else:
|
||||
echo "\nAdding rate_commitment to tree"
|
||||
var rcPtr = rateCommitment
|
||||
let setErr = ffi_set_next_leaf(addr rln, rcPtr)
|
||||
if not setErr.ok:
|
||||
stderr.writeLine "Set next leaf error: ", asString(setErr.err)
|
||||
ffi_c_string_free(setErr.err)
|
||||
quit 1
|
||||
|
||||
let leafIndex = ffi_leaves_set(addr rln) - 1
|
||||
echo " - added to tree at index ", leafIndex
|
||||
|
||||
echo "\nGetting Merkle proof"
|
||||
let proofResult = ffi_get_merkle_proof(addr rln, leafIndex)
|
||||
if proofResult.ok.isNil:
|
||||
stderr.writeLine "Get proof error: ", asString(proofResult.err)
|
||||
ffi_c_string_free(proofResult.err)
|
||||
quit 1
|
||||
let merkleProof = proofResult.ok
|
||||
echo " - proof obtained (depth: ", merkleProof.path_elements.len, ")"
|
||||
|
||||
echo "\nHashing signal"
|
||||
var signal: array[32, uint8] = [1'u8, 2, 3, 4, 5, 6, 7, 8, 9, 10, 0, 0, 0, 0,
|
||||
0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]
|
||||
var signalVec = Vec_uint8(dataPtr: cast[ptr uint8](addr signal[0]),
|
||||
len: CSize(signal.len), cap: CSize(signal.len))
|
||||
let xResult = ffi_hash_to_field_be(addr signalVec)
|
||||
if xResult.ok.isNil:
|
||||
stderr.writeLine "Hash signal error: ", asString(xResult.err)
|
||||
ffi_c_string_free(xResult.err)
|
||||
quit 1
|
||||
let x = xResult.ok
|
||||
|
||||
block:
|
||||
let debug = ffi_cfr_debug(x)
|
||||
echo " - x = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
|
||||
echo "\nHashing epoch"
|
||||
let epochStr = "test-epoch"
|
||||
var epochBytes = newSeq[uint8](epochStr.len)
|
||||
for i in 0..<epochStr.len: epochBytes[i] = uint8(epochStr[i])
|
||||
var epochVec = asVecU8(epochBytes)
|
||||
let epochResult = ffi_hash_to_field_be(addr epochVec)
|
||||
if epochResult.ok.isNil:
|
||||
stderr.writeLine "Hash epoch error: ", asString(epochResult.err)
|
||||
ffi_c_string_free(epochResult.err)
|
||||
quit 1
|
||||
let epoch = epochResult.ok
|
||||
|
||||
block:
|
||||
let debug = ffi_cfr_debug(epoch)
|
||||
echo " - epoch = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
|
||||
echo "\nHashing RLN identifier"
|
||||
let rlnIdStr = "test-rln-identifier"
|
||||
var rlnIdBytes = newSeq[uint8](rlnIdStr.len)
|
||||
for i in 0..<rlnIdStr.len: rlnIdBytes[i] = uint8(rlnIdStr[i])
|
||||
var rlnIdVec = asVecU8(rlnIdBytes)
|
||||
let rlnIdentifierResult = ffi_hash_to_field_be(addr rlnIdVec)
|
||||
if rlnIdentifierResult.ok.isNil:
|
||||
stderr.writeLine "Hash RLN identifier error: ", asString(
|
||||
rlnIdentifierResult.err)
|
||||
ffi_c_string_free(rlnIdentifierResult.err)
|
||||
quit 1
|
||||
let rlnIdentifier = rlnIdentifierResult.ok
|
||||
|
||||
block:
|
||||
let debug = ffi_cfr_debug(rlnIdentifier)
|
||||
echo " - rln_identifier = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
|
||||
echo "\nComputing Poseidon hash for external nullifier"
|
||||
let externalNullifierResult = ffi_poseidon_hash_pair(epoch, rlnIdentifier)
|
||||
if externalNullifierResult.ok.isNil:
|
||||
let errMsg = asString(externalNullifierResult.err)
|
||||
ffi_c_string_free(externalNullifierResult.err)
|
||||
echo "External nullifier hash error: ", errMsg
|
||||
quit 1
|
||||
let externalNullifier = externalNullifierResult.ok
|
||||
|
||||
block:
|
||||
let debug = ffi_cfr_debug(externalNullifier)
|
||||
echo " - external_nullifier = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
|
||||
echo "\nCreating message_id"
|
||||
let messageId = ffi_uint_to_cfr(0'u32)
|
||||
|
||||
block:
|
||||
let debug = ffi_cfr_debug(messageId)
|
||||
echo " - message_id = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
|
||||
echo "\nCreating RLN Witness"
|
||||
when defined(ffiStateless):
|
||||
var witnessRes = ffi_rln_witness_input_new(identitySecret,
|
||||
userMessageLimit, messageId, addr pathElements, addr identityPathIndex,
|
||||
x, externalNullifier)
|
||||
if witnessRes.ok.isNil:
|
||||
stderr.writeLine "RLN Witness creation error: ", asString(witnessRes.err)
|
||||
ffi_c_string_free(witnessRes.err)
|
||||
quit 1
|
||||
var witness = witnessRes.ok
|
||||
echo "RLN Witness created successfully"
|
||||
else:
|
||||
var witnessRes = ffi_rln_witness_input_new(identitySecret,
|
||||
userMessageLimit, messageId, addr merkleProof.path_elements,
|
||||
addr merkleProof.path_index, x, externalNullifier)
|
||||
if witnessRes.ok.isNil:
|
||||
stderr.writeLine "RLN Witness creation error: ", asString(witnessRes.err)
|
||||
ffi_c_string_free(witnessRes.err)
|
||||
quit 1
|
||||
var witness = witnessRes.ok
|
||||
echo "RLN Witness created successfully"
|
||||
|
||||
echo "\nRLNWitnessInput serialization: RLNWitnessInput <-> bytes"
|
||||
let serWitnessResult = ffi_rln_witness_to_bytes_be(addr witness)
|
||||
if serWitnessResult.err.dataPtr != nil:
|
||||
stderr.writeLine "Witness serialization error: ", asString(
|
||||
serWitnessResult.err)
|
||||
ffi_c_string_free(serWitnessResult.err)
|
||||
quit 1
|
||||
var serWitness = serWitnessResult.ok
|
||||
|
||||
block:
|
||||
let debug = ffi_vec_u8_debug(addr serWitness)
|
||||
echo " - serialized witness = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
|
||||
let deserWitnessResult = ffi_bytes_be_to_rln_witness(addr serWitness)
|
||||
if deserWitnessResult.ok.isNil:
|
||||
stderr.writeLine "Witness deserialization error: ", asString(
|
||||
deserWitnessResult.err)
|
||||
ffi_c_string_free(deserWitnessResult.err)
|
||||
quit 1
|
||||
|
||||
echo " - witness deserialized successfully"
|
||||
ffi_rln_witness_input_free(deserWitnessResult.ok)
|
||||
ffi_vec_u8_free(serWitness)
|
||||
|
||||
echo "\nGenerating RLN Proof"
|
||||
var proofRes = ffi_generate_rln_proof(addr rln, addr witness)
|
||||
|
||||
if proofRes.ok.isNil:
|
||||
stderr.writeLine "Proof generation error: ", asString(proofRes.err)
|
||||
ffi_c_string_free(proofRes.err)
|
||||
quit 1
|
||||
|
||||
var proof = proofRes.ok
|
||||
echo "Proof generated successfully"
|
||||
|
||||
echo "\nGetting proof values"
|
||||
var proofValues = ffi_rln_proof_get_values(addr proof)
|
||||
|
||||
block:
|
||||
let y = ffi_rln_proof_values_get_y(addr proofValues)
|
||||
let debug = ffi_cfr_debug(y)
|
||||
echo " - y = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
ffi_cfr_free(y)
|
||||
|
||||
block:
|
||||
let nullifier = ffi_rln_proof_values_get_nullifier(addr proofValues)
|
||||
let debug = ffi_cfr_debug(nullifier)
|
||||
echo " - nullifier = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
ffi_cfr_free(nullifier)
|
||||
|
||||
block:
|
||||
let root = ffi_rln_proof_values_get_root(addr proofValues)
|
||||
let debug = ffi_cfr_debug(root)
|
||||
echo " - root = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
ffi_cfr_free(root)
|
||||
|
||||
block:
|
||||
let xVal = ffi_rln_proof_values_get_x(addr proofValues)
|
||||
let debug = ffi_cfr_debug(xVal)
|
||||
echo " - x = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
ffi_cfr_free(xVal)
|
||||
|
||||
block:
|
||||
let extNullifier = ffi_rln_proof_values_get_external_nullifier(
|
||||
addr proofValues)
|
||||
let debug = ffi_cfr_debug(extNullifier)
|
||||
echo " - external_nullifier = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
ffi_cfr_free(extNullifier)
|
||||
|
||||
echo "\nRLNProof serialization: RLNProof <-> bytes"
|
||||
let serProofResult = ffi_rln_proof_to_bytes_be(addr proof)
|
||||
if serProofResult.err.dataPtr != nil:
|
||||
stderr.writeLine "Proof serialization error: ", asString(serProofResult.err)
|
||||
ffi_c_string_free(serProofResult.err)
|
||||
quit 1
|
||||
var serProof = serProofResult.ok
|
||||
|
||||
block:
|
||||
let debug = ffi_vec_u8_debug(addr serProof)
|
||||
echo " - serialized proof = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
|
||||
let deserProofResult = ffi_bytes_be_to_rln_proof(addr serProof)
|
||||
if deserProofResult.ok.isNil:
|
||||
stderr.writeLine "Proof deserialization error: ", asString(
|
||||
deserProofResult.err)
|
||||
ffi_c_string_free(deserProofResult.err)
|
||||
quit 1
|
||||
|
||||
var deserProof = deserProofResult.ok
|
||||
echo " - proof deserialized successfully"
|
||||
|
||||
echo "\nRLNProofValues serialization: RLNProofValues <-> bytes"
|
||||
var serProofValues = ffi_rln_proof_values_to_bytes_be(addr proofValues)
|
||||
|
||||
block:
|
||||
let debug = ffi_vec_u8_debug(addr serProofValues)
|
||||
echo " - serialized proof_values = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
|
||||
let deserProofValuesResult = ffi_bytes_be_to_rln_proof_values(
|
||||
addr serProofValues)
|
||||
if deserProofValuesResult.ok.isNil:
|
||||
stderr.writeLine "Proof values deserialization error: ", asString(
|
||||
deserProofValuesResult.err)
|
||||
ffi_c_string_free(deserProofValuesResult.err)
|
||||
quit 1
|
||||
var deserProofValues = deserProofValuesResult.ok
|
||||
echo " - proof_values deserialized successfully"
|
||||
|
||||
block:
|
||||
let deserExternalNullifier = ffi_rln_proof_values_get_external_nullifier(
|
||||
addr deserProofValues)
|
||||
let debug = ffi_cfr_debug(deserExternalNullifier)
|
||||
echo " - deserialized external_nullifier = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
ffi_cfr_free(deserExternalNullifier)
|
||||
|
||||
ffi_rln_proof_values_free(deserProofValues)
|
||||
ffi_vec_u8_free(serProofValues)
|
||||
ffi_rln_proof_free(deserProof)
|
||||
ffi_vec_u8_free(serProof)
|
||||
|
||||
echo "\nVerifying Proof"
|
||||
when defined(ffiStateless):
|
||||
var roots = ffi_vec_cfr_from_cfr(computedRoot)
|
||||
let verifyErr = ffi_verify_with_roots(addr rln, addr proof, addr roots, x)
|
||||
else:
|
||||
let verifyErr = ffi_verify_rln_proof(addr rln, addr proof, x)
|
||||
|
||||
if not verifyErr.ok:
|
||||
stderr.writeLine "Proof verification error: ", asString(verifyErr.err)
|
||||
ffi_c_string_free(verifyErr.err)
|
||||
quit 1
|
||||
|
||||
echo "Proof verified successfully"
|
||||
|
||||
ffi_rln_proof_free(proof)
|
||||
|
||||
echo "\nSimulating double-signaling attack (same epoch, different message)"
|
||||
|
||||
echo "\nHashing second signal"
|
||||
var signal2: array[32, uint8] = [11'u8, 12, 13, 14, 15, 16, 17, 18, 19, 20, 0,
|
||||
0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]
|
||||
var signal2Vec = Vec_uint8(dataPtr: cast[ptr uint8](addr signal2[0]),
|
||||
len: CSize(signal2.len), cap: CSize(signal2.len))
|
||||
let x2Result = ffi_hash_to_field_be(addr signal2Vec)
|
||||
if x2Result.ok.isNil:
|
||||
stderr.writeLine "Hash second signal error: ", asString(x2Result.err)
|
||||
ffi_c_string_free(x2Result.err)
|
||||
quit 1
|
||||
let x2 = x2Result.ok
|
||||
|
||||
block:
|
||||
let debug = ffi_cfr_debug(x2)
|
||||
echo " - x2 = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
|
||||
echo "\nCreating second message with the same id"
|
||||
let messageId2 = ffi_uint_to_cfr(0'u32)
|
||||
|
||||
block:
|
||||
let debug = ffi_cfr_debug(messageId2)
|
||||
echo " - message_id2 = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
|
||||
echo "\nCreating second RLN Witness"
|
||||
when defined(ffiStateless):
|
||||
var witnessRes2 = ffi_rln_witness_input_new(identitySecret,
|
||||
userMessageLimit, messageId2, addr pathElements, addr identityPathIndex,
|
||||
x2, externalNullifier)
|
||||
if witnessRes2.ok.isNil:
|
||||
stderr.writeLine "Second RLN Witness creation error: ", asString(
|
||||
witnessRes2.err)
|
||||
ffi_c_string_free(witnessRes2.err)
|
||||
quit 1
|
||||
var witness2 = witnessRes2.ok
|
||||
echo "Second RLN Witness created successfully"
|
||||
else:
|
||||
var witnessRes2 = ffi_rln_witness_input_new(identitySecret,
|
||||
userMessageLimit, messageId2, addr merkleProof.path_elements,
|
||||
addr merkleProof.path_index, x2, externalNullifier)
|
||||
if witnessRes2.ok.isNil:
|
||||
stderr.writeLine "Second RLN Witness creation error: ", asString(
|
||||
witnessRes2.err)
|
||||
ffi_c_string_free(witnessRes2.err)
|
||||
quit 1
|
||||
var witness2 = witnessRes2.ok
|
||||
echo "Second RLN Witness created successfully"
|
||||
|
||||
echo "\nGenerating second RLN Proof"
|
||||
var proofRes2 = ffi_generate_rln_proof(addr rln, addr witness2)
|
||||
|
||||
if proofRes2.ok.isNil:
|
||||
stderr.writeLine "Second proof generation error: ", asString(proofRes2.err)
|
||||
ffi_c_string_free(proofRes2.err)
|
||||
quit 1
|
||||
|
||||
var proof2 = proofRes2.ok
|
||||
echo "Second proof generated successfully"
|
||||
|
||||
var proofValues2 = ffi_rln_proof_get_values(addr proof2)
|
||||
|
||||
echo "\nVerifying second proof"
|
||||
when defined(ffiStateless):
|
||||
let verifyErr2 = ffi_verify_with_roots(addr rln, addr proof2, addr roots, x2)
|
||||
else:
|
||||
let verifyErr2 = ffi_verify_rln_proof(addr rln, addr proof2, x2)
|
||||
|
||||
if not verifyErr2.ok:
|
||||
stderr.writeLine "Proof verification error: ", asString(
|
||||
verifyErr2.err)
|
||||
ffi_c_string_free(verifyErr2.err)
|
||||
quit 1
|
||||
|
||||
echo "Second proof verified successfully"
|
||||
|
||||
echo "\nRecovering identity secret"
|
||||
let recoverRes = ffi_recover_id_secret(addr proofValues, addr proofValues2)
|
||||
if recoverRes.ok.isNil:
|
||||
stderr.writeLine "Identity recovery error: ", asString(recoverRes.err)
|
||||
ffi_c_string_free(recoverRes.err)
|
||||
quit 1
|
||||
|
||||
let recoveredSecret = recoverRes.ok
|
||||
|
||||
block:
|
||||
let debug = ffi_cfr_debug(recoveredSecret)
|
||||
echo " - recovered_secret = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
|
||||
block:
|
||||
let debug = ffi_cfr_debug(identitySecret)
|
||||
echo " - original_secret = ", asString(debug)
|
||||
ffi_c_string_free(debug)
|
||||
|
||||
echo "Slashing successful: Identity is recovered!"
|
||||
ffi_cfr_free(recoveredSecret)
|
||||
|
||||
ffi_rln_proof_values_free(proofValues2)
|
||||
ffi_rln_proof_values_free(proofValues)
|
||||
ffi_rln_proof_free(proof2)
|
||||
ffi_cfr_free(x2)
|
||||
ffi_cfr_free(messageId2)
|
||||
|
||||
when defined(ffiStateless):
|
||||
ffi_rln_witness_input_free(witness2)
|
||||
ffi_rln_witness_input_free(witness)
|
||||
ffi_vec_cfr_free(roots)
|
||||
ffi_vec_cfr_free(pathElements)
|
||||
for i in 0..treeDepth-2:
|
||||
ffi_cfr_free(defaultHashes[i])
|
||||
ffi_cfr_free(defaultLeaf)
|
||||
ffi_cfr_free(computedRoot)
|
||||
else:
|
||||
ffi_rln_witness_input_free(witness2)
|
||||
ffi_rln_witness_input_free(witness)
|
||||
ffi_merkle_proof_free(merkleProof)
|
||||
|
||||
ffi_cfr_free(rateCommitment)
|
||||
ffi_cfr_free(x)
|
||||
ffi_cfr_free(epoch)
|
||||
ffi_cfr_free(rlnIdentifier)
|
||||
ffi_cfr_free(externalNullifier)
|
||||
ffi_cfr_free(userMessageLimit)
|
||||
ffi_cfr_free(messageId)
|
||||
ffi_vec_cfr_free(keys)
|
||||
ffi_rln_free(rln)
|
||||
9
rln/resources/tree_depth_20/config.json
Normal file
9
rln/resources/tree_depth_20/config.json
Normal file
@@ -0,0 +1,9 @@
|
||||
{
|
||||
"path": "./database",
|
||||
"temporary": false,
|
||||
"cache_capacity": 1073741824,
|
||||
"flush_every_ms": 500,
|
||||
"mode": "HighThroughput",
|
||||
"use_compression": false,
|
||||
"tree_depth": 20
|
||||
}
|
||||
BIN
rln/resources/tree_depth_20/graph.bin
Normal file
BIN
rln/resources/tree_depth_20/graph.bin
Normal file
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
5
rln/src/bin/generate-headers.rs
Normal file
5
rln/src/bin/generate-headers.rs
Normal file
@@ -0,0 +1,5 @@
|
||||
use rln::ffi;
|
||||
|
||||
fn main() -> std::io::Result<()> {
|
||||
ffi::generate_headers()
|
||||
}
|
||||
@@ -1,225 +0,0 @@
|
||||
// This crate provides interfaces for the zero-knowledge circuit and keys
|
||||
|
||||
use ark_bn254::{
|
||||
Bn254, Fq as ArkFq, Fq2 as ArkFq2, Fr as ArkFr, G1Affine as ArkG1Affine,
|
||||
G1Projective as ArkG1Projective, G2Affine as ArkG2Affine, G2Projective as ArkG2Projective,
|
||||
};
|
||||
use ark_groth16::{ProvingKey, VerifyingKey};
|
||||
use ark_relations::r1cs::ConstraintMatrices;
|
||||
use ark_serialize::CanonicalDeserialize;
|
||||
use cfg_if::cfg_if;
|
||||
use color_eyre::{Report, Result};
|
||||
|
||||
#[cfg(not(target_arch = "wasm32"))]
|
||||
use {
|
||||
ark_circom::WitnessCalculator,
|
||||
lazy_static::lazy_static,
|
||||
std::sync::{Arc, Mutex},
|
||||
wasmer::{Module, Store},
|
||||
};
|
||||
|
||||
#[cfg(feature = "arkzkey")]
|
||||
use {
|
||||
ark_zkey::{read_arkzkey_from_bytes, SerializableConstraintMatrices, SerializableProvingKey},
|
||||
color_eyre::eyre::WrapErr,
|
||||
};
|
||||
|
||||
#[cfg(not(feature = "arkzkey"))]
|
||||
use {ark_circom::read_zkey, std::io::Cursor};
|
||||
|
||||
#[cfg(feature = "arkzkey")]
|
||||
pub const ARKZKEY_BYTES: &[u8] = include_bytes!("../resources/tree_height_20/rln_final.arkzkey");
|
||||
#[cfg(feature = "arkzkey")]
|
||||
pub const ARKZKEY_BYTES_UNCOMPR: &[u8] =
|
||||
include_bytes!("../resources/tree_height_20/rln_final_uncompr.arkzkey");
|
||||
|
||||
pub const ZKEY_BYTES: &[u8] = include_bytes!("../resources/tree_height_20/rln_final.zkey");
|
||||
pub const VK_BYTES: &[u8] = include_bytes!("../resources/tree_height_20/verification_key.arkvkey");
|
||||
const WASM_BYTES: &[u8] = include_bytes!("../resources/tree_height_20/rln.wasm");
|
||||
|
||||
#[cfg(not(target_arch = "wasm32"))]
|
||||
lazy_static! {
|
||||
#[cfg(not(target_arch = "wasm32"))]
|
||||
static ref ZKEY: (ProvingKey<Curve>, ConstraintMatrices<Fr>) = {
|
||||
cfg_if! {
|
||||
if #[cfg(feature = "arkzkey")] {
|
||||
read_arkzkey_from_bytes_uncompressed(ARKZKEY_BYTES_UNCOMPR).expect("Failed to read arkzkey")
|
||||
} else {
|
||||
let mut reader = Cursor::new(ZKEY_BYTES);
|
||||
read_zkey(&mut reader).expect("Failed to read zkey")
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
#[cfg(not(target_arch = "wasm32"))]
|
||||
static ref VK: VerifyingKey<Curve> = vk_from_ark_serialized(VK_BYTES).expect("Failed to read vk");
|
||||
|
||||
#[cfg(not(target_arch = "wasm32"))]
|
||||
static ref WITNESS_CALCULATOR: Arc<Mutex<WitnessCalculator>> = {
|
||||
circom_from_raw(WASM_BYTES).expect("Failed to create witness calculator")
|
||||
};
|
||||
}
|
||||
|
||||
pub const TEST_TREE_HEIGHT: usize = 20;
|
||||
|
||||
// The following types define the pairing friendly elliptic curve, the underlying finite fields and groups default to this module
|
||||
// Note that proofs are serialized assuming Fr to be 4x8 = 32 bytes in size. Hence, changing to a curve with different encoding will make proof verification to fail
|
||||
pub type Curve = Bn254;
|
||||
pub type Fr = ArkFr;
|
||||
pub type Fq = ArkFq;
|
||||
pub type Fq2 = ArkFq2;
|
||||
pub type G1Affine = ArkG1Affine;
|
||||
pub type G1Projective = ArkG1Projective;
|
||||
pub type G2Affine = ArkG2Affine;
|
||||
pub type G2Projective = ArkG2Projective;
|
||||
|
||||
// Loads the proving key using a bytes vector
|
||||
pub fn zkey_from_raw(zkey_data: &[u8]) -> Result<(ProvingKey<Curve>, ConstraintMatrices<Fr>)> {
|
||||
if zkey_data.is_empty() {
|
||||
return Err(Report::msg("No proving key found!"));
|
||||
}
|
||||
|
||||
let proving_key_and_matrices = match () {
|
||||
#[cfg(feature = "arkzkey")]
|
||||
() => read_arkzkey_from_bytes(zkey_data)?,
|
||||
#[cfg(not(feature = "arkzkey"))]
|
||||
() => {
|
||||
let mut reader = Cursor::new(zkey_data);
|
||||
read_zkey(&mut reader)?
|
||||
}
|
||||
};
|
||||
|
||||
Ok(proving_key_and_matrices)
|
||||
}
|
||||
|
||||
// Loads the proving key
|
||||
#[cfg(not(target_arch = "wasm32"))]
|
||||
pub fn zkey_from_folder() -> &'static (ProvingKey<Curve>, ConstraintMatrices<Fr>) {
|
||||
&ZKEY
|
||||
}
|
||||
|
||||
// Loads the verification key from a bytes vector
|
||||
pub fn vk_from_raw(vk_data: &[u8], zkey_data: &[u8]) -> Result<VerifyingKey<Curve>> {
|
||||
if !vk_data.is_empty() {
|
||||
return vk_from_ark_serialized(vk_data);
|
||||
}
|
||||
|
||||
if !zkey_data.is_empty() {
|
||||
let (proving_key, _matrices) = zkey_from_raw(zkey_data)?;
|
||||
return Ok(proving_key.vk);
|
||||
}
|
||||
|
||||
Err(Report::msg("No proving/verification key found!"))
|
||||
}
|
||||
|
||||
// Loads the verification key
|
||||
#[cfg(not(target_arch = "wasm32"))]
|
||||
pub fn vk_from_folder() -> &'static VerifyingKey<Curve> {
|
||||
&VK
|
||||
}
|
||||
|
||||
// Initializes the witness calculator using a bytes vector
|
||||
#[cfg(not(target_arch = "wasm32"))]
|
||||
pub fn circom_from_raw(wasm_buffer: &[u8]) -> Result<Arc<Mutex<WitnessCalculator>>> {
|
||||
let module = Module::new(&Store::default(), wasm_buffer)?;
|
||||
let result = WitnessCalculator::from_module(module)?;
|
||||
Ok(Arc::new(Mutex::new(result)))
|
||||
}
|
||||
|
||||
// Initializes the witness calculator
|
||||
#[cfg(not(target_arch = "wasm32"))]
|
||||
pub fn circom_from_folder() -> &'static Arc<Mutex<WitnessCalculator>> {
|
||||
&WITNESS_CALCULATOR
|
||||
}
|
||||
|
||||
// Computes the verification key from a bytes vector containing pre-processed ark-serialized verification key
|
||||
// uncompressed, unchecked
|
||||
pub fn vk_from_ark_serialized(data: &[u8]) -> Result<VerifyingKey<Curve>> {
|
||||
let vk = VerifyingKey::<Curve>::deserialize_uncompressed_unchecked(data)?;
|
||||
Ok(vk)
|
||||
}
|
||||
|
||||
// Checks verification key to be correct with respect to proving key
|
||||
#[cfg(not(target_arch = "wasm32"))]
|
||||
pub fn check_vk_from_zkey(verifying_key: VerifyingKey<Curve>) -> Result<()> {
|
||||
let (proving_key, _matrices) = zkey_from_folder();
|
||||
if proving_key.vk == verifying_key {
|
||||
Ok(())
|
||||
} else {
|
||||
Err(Report::msg("verifying_keys are not equal"))
|
||||
}
|
||||
}
|
||||
|
||||
////////////////////////////////////////////////////////
|
||||
// Functions from [arkz-key](https://github.com/zkmopro/ark-zkey/blob/main/src/lib.rs#L106)
|
||||
// without print and allow to choose between compressed and uncompressed arkzkey
|
||||
////////////////////////////////////////////////////////
|
||||
#[cfg(feature = "arkzkey")]
|
||||
pub fn read_arkzkey_from_bytes_uncompressed(
|
||||
arkzkey_data: &[u8],
|
||||
) -> Result<(ProvingKey<Curve>, ConstraintMatrices<Fr>)> {
|
||||
if arkzkey_data.is_empty() {
|
||||
return Err(Report::msg("No proving key found!"));
|
||||
}
|
||||
|
||||
let mut cursor = std::io::Cursor::new(arkzkey_data);
|
||||
|
||||
let serialized_proving_key =
|
||||
SerializableProvingKey::deserialize_uncompressed_unchecked(&mut cursor)
|
||||
.wrap_err("Failed to deserialize proving key")?;
|
||||
|
||||
let serialized_constraint_matrices =
|
||||
SerializableConstraintMatrices::deserialize_uncompressed_unchecked(&mut cursor)
|
||||
.wrap_err("Failed to deserialize constraint matrices")?;
|
||||
|
||||
// Get on right form for API
|
||||
let proving_key: ProvingKey<Bn254> = serialized_proving_key.0;
|
||||
let constraint_matrices: ConstraintMatrices<ark_bn254::Fr> = ConstraintMatrices {
|
||||
num_instance_variables: serialized_constraint_matrices.num_instance_variables,
|
||||
num_witness_variables: serialized_constraint_matrices.num_witness_variables,
|
||||
num_constraints: serialized_constraint_matrices.num_constraints,
|
||||
a_num_non_zero: serialized_constraint_matrices.a_num_non_zero,
|
||||
b_num_non_zero: serialized_constraint_matrices.b_num_non_zero,
|
||||
c_num_non_zero: serialized_constraint_matrices.c_num_non_zero,
|
||||
a: serialized_constraint_matrices.a.data,
|
||||
b: serialized_constraint_matrices.b.data,
|
||||
c: serialized_constraint_matrices.c.data,
|
||||
};
|
||||
|
||||
Ok((proving_key, constraint_matrices))
|
||||
}
|
||||
|
||||
#[cfg(feature = "arkzkey")]
|
||||
pub fn read_arkzkey_from_bytes_compressed(
|
||||
arkzkey_data: &[u8],
|
||||
) -> Result<(ProvingKey<Curve>, ConstraintMatrices<Fr>)> {
|
||||
if arkzkey_data.is_empty() {
|
||||
return Err(Report::msg("No proving key found!"));
|
||||
}
|
||||
|
||||
let mut cursor = std::io::Cursor::new(arkzkey_data);
|
||||
|
||||
let serialized_proving_key =
|
||||
SerializableProvingKey::deserialize_compressed_unchecked(&mut cursor)
|
||||
.wrap_err("Failed to deserialize proving key")?;
|
||||
|
||||
let serialized_constraint_matrices =
|
||||
SerializableConstraintMatrices::deserialize_compressed_unchecked(&mut cursor)
|
||||
.wrap_err("Failed to deserialize constraint matrices")?;
|
||||
|
||||
// Get on right form for API
|
||||
let proving_key: ProvingKey<Bn254> = serialized_proving_key.0;
|
||||
let constraint_matrices: ConstraintMatrices<ark_bn254::Fr> = ConstraintMatrices {
|
||||
num_instance_variables: serialized_constraint_matrices.num_instance_variables,
|
||||
num_witness_variables: serialized_constraint_matrices.num_witness_variables,
|
||||
num_constraints: serialized_constraint_matrices.num_constraints,
|
||||
a_num_non_zero: serialized_constraint_matrices.a_num_non_zero,
|
||||
b_num_non_zero: serialized_constraint_matrices.b_num_non_zero,
|
||||
c_num_non_zero: serialized_constraint_matrices.c_num_non_zero,
|
||||
a: serialized_constraint_matrices.a.data,
|
||||
b: serialized_constraint_matrices.b.data,
|
||||
c: serialized_constraint_matrices.c.data,
|
||||
};
|
||||
|
||||
Ok((proving_key, constraint_matrices))
|
||||
}
|
||||
25
rln/src/circuit/error.rs
Normal file
25
rln/src/circuit/error.rs
Normal file
@@ -0,0 +1,25 @@
|
||||
/// Errors that can occur during zkey reading operations
|
||||
#[derive(Debug, thiserror::Error)]
|
||||
pub enum ZKeyReadError {
|
||||
#[error("Empty zkey bytes provided")]
|
||||
EmptyBytes,
|
||||
#[error("{0}")]
|
||||
SerializationError(#[from] ark_serialize::SerializationError),
|
||||
}
|
||||
|
||||
/// Errors that can occur during witness calculation
|
||||
#[derive(Debug, thiserror::Error)]
|
||||
pub enum WitnessCalcError {
|
||||
#[error("Failed to deserialize witness calculation graph: {0}")]
|
||||
GraphDeserialization(#[from] std::io::Error),
|
||||
#[error("Failed to evaluate witness calculation graph: {0}")]
|
||||
GraphEvaluation(String),
|
||||
#[error("Invalid input length for '{name}': expected {expected}, got {actual}")]
|
||||
InvalidInputLength {
|
||||
name: String,
|
||||
expected: usize,
|
||||
actual: usize,
|
||||
},
|
||||
#[error("Missing required input: {0}")]
|
||||
MissingInput(String),
|
||||
}
|
||||
113
rln/src/circuit/iden3calc.rs
Normal file
113
rln/src/circuit/iden3calc.rs
Normal file
@@ -0,0 +1,113 @@
|
||||
// This crate is based on the code by iden3. Its preimage can be found here:
|
||||
// https://github.com/iden3/circom-witnesscalc/blob/5cb365b6e4d9052ecc69d4567fcf5bc061c20e94/src/lib.rs
|
||||
|
||||
mod graph;
|
||||
mod proto;
|
||||
mod storage;
|
||||
|
||||
use std::collections::HashMap;
|
||||
|
||||
use graph::Node;
|
||||
use ruint::aliases::U256;
|
||||
use storage::deserialize_witnesscalc_graph;
|
||||
use zeroize::zeroize_flat_type;
|
||||
|
||||
use self::graph::fr_to_u256;
|
||||
use super::{error::WitnessCalcError, Fr};
|
||||
use crate::utils::FrOrSecret;
|
||||
|
||||
pub(crate) type InputSignalsInfo = HashMap<String, (usize, usize)>;
|
||||
|
||||
pub(crate) fn calc_witness<I: IntoIterator<Item = (String, Vec<FrOrSecret>)>>(
|
||||
inputs: I,
|
||||
graph_data: &[u8],
|
||||
) -> Result<Vec<Fr>, WitnessCalcError> {
|
||||
let mut inputs: HashMap<String, Vec<U256>> = inputs
|
||||
.into_iter()
|
||||
.map(|(key, value)| {
|
||||
(
|
||||
key,
|
||||
value
|
||||
.iter()
|
||||
.map(|f_| match f_ {
|
||||
FrOrSecret::IdSecret(s) => s.to_u256(),
|
||||
FrOrSecret::Fr(f) => fr_to_u256(f),
|
||||
})
|
||||
.collect(),
|
||||
)
|
||||
})
|
||||
.collect();
|
||||
|
||||
let (nodes, signals, input_mapping): (Vec<Node>, Vec<usize>, InputSignalsInfo) =
|
||||
deserialize_witnesscalc_graph(std::io::Cursor::new(graph_data))?;
|
||||
|
||||
let mut inputs_buffer = get_inputs_buffer(get_inputs_size(&nodes));
|
||||
|
||||
populate_inputs(&inputs, &input_mapping, &mut inputs_buffer)?;
|
||||
|
||||
if let Some(v) = inputs.get_mut("identitySecret") {
|
||||
// DO NOT USE: unsafe { zeroize_flat_type(v) } only clears the Vec pointer, not the data—can cause memory leaks
|
||||
|
||||
for val in v.iter_mut() {
|
||||
unsafe { zeroize_flat_type(val) };
|
||||
}
|
||||
}
|
||||
|
||||
let res = graph::evaluate(&nodes, inputs_buffer.as_slice(), &signals)
|
||||
.map_err(WitnessCalcError::GraphEvaluation)?;
|
||||
|
||||
for val in inputs_buffer.iter_mut() {
|
||||
unsafe { zeroize_flat_type(val) };
|
||||
}
|
||||
|
||||
Ok(res)
|
||||
}
|
||||
|
||||
fn get_inputs_size(nodes: &[Node]) -> usize {
|
||||
let mut start = false;
|
||||
let mut max_index = 0usize;
|
||||
for &node in nodes.iter() {
|
||||
if let Node::Input(i) = node {
|
||||
if i > max_index {
|
||||
max_index = i;
|
||||
}
|
||||
start = true
|
||||
} else if start {
|
||||
break;
|
||||
}
|
||||
}
|
||||
max_index + 1
|
||||
}
|
||||
|
||||
fn populate_inputs(
|
||||
input_list: &HashMap<String, Vec<U256>>,
|
||||
inputs_info: &InputSignalsInfo,
|
||||
input_buffer: &mut [U256],
|
||||
) -> Result<(), WitnessCalcError> {
|
||||
for (key, value) in input_list {
|
||||
let (offset, len) = inputs_info
|
||||
.get(key)
|
||||
.ok_or_else(|| WitnessCalcError::MissingInput(key.clone()))?;
|
||||
|
||||
if *len != value.len() {
|
||||
return Err(WitnessCalcError::InvalidInputLength {
|
||||
name: key.clone(),
|
||||
expected: *len,
|
||||
actual: value.len(),
|
||||
});
|
||||
}
|
||||
|
||||
for (i, v) in value.iter().enumerate() {
|
||||
input_buffer[offset + i] = *v;
|
||||
}
|
||||
}
|
||||
|
||||
Ok(())
|
||||
}
|
||||
|
||||
/// Allocates inputs vec with position 0 set to 1
|
||||
fn get_inputs_buffer(size: usize) -> Vec<U256> {
|
||||
let mut inputs = vec![U256::ZERO; size];
|
||||
inputs[0] = U256::from(1);
|
||||
inputs
|
||||
}
|
||||
584
rln/src/circuit/iden3calc/graph.rs
Normal file
584
rln/src/circuit/iden3calc/graph.rs
Normal file
@@ -0,0 +1,584 @@
|
||||
// This crate is based on the code by iden3. Its preimage can be found here:
|
||||
// https://github.com/iden3/circom-witnesscalc/blob/5cb365b6e4d9052ecc69d4567fcf5bc061c20e94/src/graph.rs
|
||||
|
||||
use std::cmp::Ordering;
|
||||
|
||||
use ark_ff::{BigInt, BigInteger, One, PrimeField, Zero};
|
||||
use ark_serialize::{CanonicalDeserialize, CanonicalSerialize, Compress, Validate};
|
||||
use ruint::{aliases::U256, uint};
|
||||
use serde::{Deserialize, Serialize};
|
||||
|
||||
use super::proto;
|
||||
use crate::circuit::Fr;
|
||||
|
||||
const M: U256 =
|
||||
uint!(21888242871839275222246405745257275088548364400416034343698204186575808495617_U256);
|
||||
|
||||
fn ark_se<S, A: CanonicalSerialize>(a: &A, s: S) -> Result<S::Ok, S::Error>
|
||||
where
|
||||
S: serde::Serializer,
|
||||
{
|
||||
let mut bytes = vec![];
|
||||
a.serialize_with_mode(&mut bytes, Compress::Yes)
|
||||
.map_err(serde::ser::Error::custom)?;
|
||||
s.serialize_bytes(&bytes)
|
||||
}
|
||||
|
||||
fn ark_de<'de, D, A: CanonicalDeserialize>(data: D) -> Result<A, D::Error>
|
||||
where
|
||||
D: serde::de::Deserializer<'de>,
|
||||
{
|
||||
let s: Vec<u8> = serde::de::Deserialize::deserialize(data)?;
|
||||
let a = A::deserialize_with_mode(s.as_slice(), Compress::Yes, Validate::Yes);
|
||||
a.map_err(serde::de::Error::custom)
|
||||
}
|
||||
|
||||
#[inline(always)]
|
||||
pub(crate) fn fr_to_u256(x: &Fr) -> U256 {
|
||||
U256::from_limbs(x.into_bigint().0)
|
||||
}
|
||||
|
||||
#[inline(always)]
|
||||
pub(crate) fn u256_to_fr(x: &U256) -> Result<Fr, String> {
|
||||
Fr::from_bigint(BigInt::new(x.into_limbs()))
|
||||
.ok_or_else(|| "Failed to convert U256 to Fr".to_string())
|
||||
}
|
||||
|
||||
#[derive(Hash, PartialEq, Eq, Debug, Clone, Copy, Serialize, Deserialize)]
|
||||
pub(crate) enum Operation {
|
||||
Mul,
|
||||
Div,
|
||||
Add,
|
||||
Sub,
|
||||
Pow,
|
||||
Idiv,
|
||||
Mod,
|
||||
Eq,
|
||||
Neq,
|
||||
Lt,
|
||||
Gt,
|
||||
Leq,
|
||||
Geq,
|
||||
Land,
|
||||
Lor,
|
||||
Shl,
|
||||
Shr,
|
||||
Bor,
|
||||
Band,
|
||||
Bxor,
|
||||
}
|
||||
|
||||
impl Operation {
|
||||
fn eval_fr(&self, a: Fr, b: Fr) -> Result<Fr, String> {
|
||||
use Operation::*;
|
||||
match self {
|
||||
Mul => Ok(a * b),
|
||||
// We always should return something on the circuit execution.
|
||||
// So in case of division by 0 we would return 0. And the proof
|
||||
// should be invalid in the end.
|
||||
Div => {
|
||||
if b.is_zero() {
|
||||
Ok(Fr::zero())
|
||||
} else {
|
||||
Ok(a / b)
|
||||
}
|
||||
}
|
||||
Add => Ok(a + b),
|
||||
Sub => Ok(a - b),
|
||||
// Modular exponentiation to prevent overflow and keep result in field
|
||||
Pow => {
|
||||
let a_u256 = fr_to_u256(&a);
|
||||
let b_u256 = fr_to_u256(&b);
|
||||
let result = a_u256.pow_mod(b_u256, M);
|
||||
u256_to_fr(&result)
|
||||
}
|
||||
// Integer division (not field division)
|
||||
Idiv => {
|
||||
if b.is_zero() {
|
||||
Ok(Fr::zero())
|
||||
} else {
|
||||
let a_u256 = fr_to_u256(&a);
|
||||
let b_u256 = fr_to_u256(&b);
|
||||
u256_to_fr(&(a_u256 / b_u256))
|
||||
}
|
||||
}
|
||||
// Integer modulo (not field arithmetic)
|
||||
Mod => {
|
||||
if b.is_zero() {
|
||||
Ok(Fr::zero())
|
||||
} else {
|
||||
let a_u256 = fr_to_u256(&a);
|
||||
let b_u256 = fr_to_u256(&b);
|
||||
u256_to_fr(&(a_u256 % b_u256))
|
||||
}
|
||||
}
|
||||
Eq => Ok(match a.cmp(&b) {
|
||||
Ordering::Equal => Fr::one(),
|
||||
_ => Fr::zero(),
|
||||
}),
|
||||
Neq => Ok(match a.cmp(&b) {
|
||||
Ordering::Equal => Fr::zero(),
|
||||
_ => Fr::one(),
|
||||
}),
|
||||
Lt => u256_to_fr(&u_lt(&fr_to_u256(&a), &fr_to_u256(&b))),
|
||||
Gt => u256_to_fr(&u_gt(&fr_to_u256(&a), &fr_to_u256(&b))),
|
||||
Leq => u256_to_fr(&u_lte(&fr_to_u256(&a), &fr_to_u256(&b))),
|
||||
Geq => u256_to_fr(&u_gte(&fr_to_u256(&a), &fr_to_u256(&b))),
|
||||
Land => Ok(if a.is_zero() || b.is_zero() {
|
||||
Fr::zero()
|
||||
} else {
|
||||
Fr::one()
|
||||
}),
|
||||
Lor => Ok(if a.is_zero() && b.is_zero() {
|
||||
Fr::zero()
|
||||
} else {
|
||||
Fr::one()
|
||||
}),
|
||||
Shl => shl(a, b),
|
||||
Shr => shr(a, b),
|
||||
Bor => bit_or(a, b),
|
||||
Band => bit_and(a, b),
|
||||
Bxor => bit_xor(a, b),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl From<&Operation> for proto::DuoOp {
|
||||
fn from(v: &Operation) -> Self {
|
||||
match v {
|
||||
Operation::Mul => proto::DuoOp::Mul,
|
||||
Operation::Div => proto::DuoOp::Div,
|
||||
Operation::Add => proto::DuoOp::Add,
|
||||
Operation::Sub => proto::DuoOp::Sub,
|
||||
Operation::Pow => proto::DuoOp::Pow,
|
||||
Operation::Idiv => proto::DuoOp::Idiv,
|
||||
Operation::Mod => proto::DuoOp::Mod,
|
||||
Operation::Eq => proto::DuoOp::Eq,
|
||||
Operation::Neq => proto::DuoOp::Neq,
|
||||
Operation::Lt => proto::DuoOp::Lt,
|
||||
Operation::Gt => proto::DuoOp::Gt,
|
||||
Operation::Leq => proto::DuoOp::Leq,
|
||||
Operation::Geq => proto::DuoOp::Geq,
|
||||
Operation::Land => proto::DuoOp::Land,
|
||||
Operation::Lor => proto::DuoOp::Lor,
|
||||
Operation::Shl => proto::DuoOp::Shl,
|
||||
Operation::Shr => proto::DuoOp::Shr,
|
||||
Operation::Bor => proto::DuoOp::Bor,
|
||||
Operation::Band => proto::DuoOp::Band,
|
||||
Operation::Bxor => proto::DuoOp::Bxor,
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
#[derive(Hash, PartialEq, Eq, Debug, Clone, Copy, Serialize, Deserialize)]
|
||||
pub(crate) enum UnoOperation {
|
||||
Neg,
|
||||
Id, // identity - just return self
|
||||
}
|
||||
|
||||
impl UnoOperation {
|
||||
fn eval_fr(&self, a: Fr) -> Result<Fr, String> {
|
||||
match self {
|
||||
UnoOperation::Neg => {
|
||||
if a.is_zero() {
|
||||
Ok(Fr::zero())
|
||||
} else {
|
||||
let mut x = Fr::MODULUS;
|
||||
x.sub_with_borrow(&a.into_bigint());
|
||||
Fr::from_bigint(x).ok_or_else(|| "Failed to compute negation".to_string())
|
||||
}
|
||||
}
|
||||
_ => Err(format!(
|
||||
"uno operator {:?} not implemented for Montgomery",
|
||||
self
|
||||
)),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl From<&UnoOperation> for proto::UnoOp {
|
||||
fn from(v: &UnoOperation) -> Self {
|
||||
match v {
|
||||
UnoOperation::Neg => proto::UnoOp::Neg,
|
||||
UnoOperation::Id => proto::UnoOp::Id,
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
#[derive(Hash, PartialEq, Eq, Debug, Clone, Copy, Serialize, Deserialize)]
|
||||
pub(crate) enum TresOperation {
|
||||
TernCond,
|
||||
}
|
||||
|
||||
impl TresOperation {
|
||||
fn eval_fr(&self, a: Fr, b: Fr, c: Fr) -> Result<Fr, String> {
|
||||
match self {
|
||||
TresOperation::TernCond => {
|
||||
if a.is_zero() {
|
||||
Ok(c)
|
||||
} else {
|
||||
Ok(b)
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl From<&TresOperation> for proto::TresOp {
|
||||
fn from(v: &TresOperation) -> Self {
|
||||
match v {
|
||||
TresOperation::TernCond => proto::TresOp::TernCond,
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
#[derive(Debug, Clone, Copy, PartialEq, Eq, Serialize, Deserialize)]
|
||||
pub(crate) enum Node {
|
||||
Input(usize),
|
||||
Constant(U256),
|
||||
#[serde(serialize_with = "ark_se", deserialize_with = "ark_de")]
|
||||
MontConstant(Fr),
|
||||
UnoOp(UnoOperation, usize),
|
||||
Op(Operation, usize, usize),
|
||||
TresOp(TresOperation, usize, usize, usize),
|
||||
}
|
||||
|
||||
pub(crate) fn evaluate(
|
||||
nodes: &[Node],
|
||||
inputs: &[U256],
|
||||
outputs: &[usize],
|
||||
) -> Result<Vec<Fr>, String> {
|
||||
// Evaluate the graph.
|
||||
let mut values = Vec::with_capacity(nodes.len());
|
||||
for &node in nodes.iter() {
|
||||
let value = match node {
|
||||
Node::Constant(c) => u256_to_fr(&c)?,
|
||||
Node::MontConstant(c) => c,
|
||||
Node::Input(i) => u256_to_fr(&inputs[i])?,
|
||||
Node::Op(op, a, b) => op.eval_fr(values[a], values[b])?,
|
||||
Node::UnoOp(op, a) => op.eval_fr(values[a])?,
|
||||
Node::TresOp(op, a, b, c) => op.eval_fr(values[a], values[b], values[c])?,
|
||||
};
|
||||
values.push(value);
|
||||
}
|
||||
|
||||
// Convert from Montgomery form and return the outputs.
|
||||
let mut out = vec![Fr::from(0); outputs.len()];
|
||||
for i in 0..outputs.len() {
|
||||
out[i] = values[outputs[i]];
|
||||
}
|
||||
|
||||
Ok(out)
|
||||
}
|
||||
|
||||
fn shl(a: Fr, b: Fr) -> Result<Fr, String> {
|
||||
if b.is_zero() {
|
||||
return Ok(a);
|
||||
}
|
||||
|
||||
if b.cmp(&Fr::from(Fr::MODULUS_BIT_SIZE)).is_ge() {
|
||||
return Ok(Fr::zero());
|
||||
}
|
||||
|
||||
let n = b.into_bigint().0[0] as u32;
|
||||
let a = a.into_bigint();
|
||||
Fr::from_bigint(a << n).ok_or_else(|| "Failed to compute left shift".to_string())
|
||||
}
|
||||
|
||||
fn shr(a: Fr, b: Fr) -> Result<Fr, String> {
|
||||
if b.is_zero() {
|
||||
return Ok(a);
|
||||
}
|
||||
|
||||
match b.cmp(&Fr::from(254u64)) {
|
||||
Ordering::Equal => return Ok(Fr::zero()),
|
||||
Ordering::Greater => return Ok(Fr::zero()),
|
||||
_ => (),
|
||||
};
|
||||
|
||||
let mut n = b.into_bigint().to_bytes_le()[0];
|
||||
let mut result = a.into_bigint();
|
||||
let c = result.as_mut();
|
||||
while n >= 64 {
|
||||
for i in 0..3 {
|
||||
c[i as usize] = c[(i + 1) as usize];
|
||||
}
|
||||
c[3] = 0;
|
||||
n -= 64;
|
||||
}
|
||||
|
||||
if n == 0 {
|
||||
return Fr::from_bigint(result).ok_or_else(|| "Failed to compute right shift".to_string());
|
||||
}
|
||||
|
||||
let mask: u64 = (1 << n) - 1;
|
||||
let mut carrier: u64 = c[3] & mask;
|
||||
c[3] >>= n;
|
||||
for i in (0..3).rev() {
|
||||
let new_carrier = c[i] & mask;
|
||||
c[i] = (c[i] >> n) | (carrier << (64 - n));
|
||||
carrier = new_carrier;
|
||||
}
|
||||
Fr::from_bigint(result).ok_or_else(|| "Failed to compute right shift".to_string())
|
||||
}
|
||||
|
||||
fn bit_and(a: Fr, b: Fr) -> Result<Fr, String> {
|
||||
let a = a.into_bigint();
|
||||
let b = b.into_bigint();
|
||||
let c: [u64; 4] = [
|
||||
a.0[0] & b.0[0],
|
||||
a.0[1] & b.0[1],
|
||||
a.0[2] & b.0[2],
|
||||
a.0[3] & b.0[3],
|
||||
];
|
||||
let mut d: BigInt<4> = BigInt::new(c);
|
||||
if d > Fr::MODULUS {
|
||||
d.sub_with_borrow(&Fr::MODULUS);
|
||||
}
|
||||
|
||||
Fr::from_bigint(d).ok_or_else(|| "Failed to compute bitwise AND".to_string())
|
||||
}
|
||||
|
||||
fn bit_or(a: Fr, b: Fr) -> Result<Fr, String> {
|
||||
let a = a.into_bigint();
|
||||
let b = b.into_bigint();
|
||||
let c: [u64; 4] = [
|
||||
a.0[0] | b.0[0],
|
||||
a.0[1] | b.0[1],
|
||||
a.0[2] | b.0[2],
|
||||
a.0[3] | b.0[3],
|
||||
];
|
||||
let mut d: BigInt<4> = BigInt::new(c);
|
||||
if d > Fr::MODULUS {
|
||||
d.sub_with_borrow(&Fr::MODULUS);
|
||||
}
|
||||
|
||||
Fr::from_bigint(d).ok_or_else(|| "Failed to compute bitwise OR".to_string())
|
||||
}
|
||||
|
||||
fn bit_xor(a: Fr, b: Fr) -> Result<Fr, String> {
|
||||
let a = a.into_bigint();
|
||||
let b = b.into_bigint();
|
||||
let c: [u64; 4] = [
|
||||
a.0[0] ^ b.0[0],
|
||||
a.0[1] ^ b.0[1],
|
||||
a.0[2] ^ b.0[2],
|
||||
a.0[3] ^ b.0[3],
|
||||
];
|
||||
let mut d: BigInt<4> = BigInt::new(c);
|
||||
if d > Fr::MODULUS {
|
||||
d.sub_with_borrow(&Fr::MODULUS);
|
||||
}
|
||||
|
||||
Fr::from_bigint(d).ok_or_else(|| "Failed to compute bitwise XOR".to_string())
|
||||
}
|
||||
|
||||
// M / 2
|
||||
const HALF_M: U256 =
|
||||
uint!(10944121435919637611123202872628637544274182200208017171849102093287904247808_U256);
|
||||
|
||||
fn u_gte(a: &U256, b: &U256) -> U256 {
|
||||
let a_neg = &HALF_M < a;
|
||||
let b_neg = &HALF_M < b;
|
||||
|
||||
match (a_neg, b_neg) {
|
||||
(false, false) => U256::from(a >= b),
|
||||
(true, false) => uint!(0_U256),
|
||||
(false, true) => uint!(1_U256),
|
||||
(true, true) => U256::from(a >= b),
|
||||
}
|
||||
}
|
||||
|
||||
fn u_lte(a: &U256, b: &U256) -> U256 {
|
||||
let a_neg = &HALF_M < a;
|
||||
let b_neg = &HALF_M < b;
|
||||
|
||||
match (a_neg, b_neg) {
|
||||
(false, false) => U256::from(a <= b),
|
||||
(true, false) => uint!(1_U256),
|
||||
(false, true) => uint!(0_U256),
|
||||
(true, true) => U256::from(a <= b),
|
||||
}
|
||||
}
|
||||
|
||||
fn u_gt(a: &U256, b: &U256) -> U256 {
|
||||
let a_neg = &HALF_M < a;
|
||||
let b_neg = &HALF_M < b;
|
||||
|
||||
match (a_neg, b_neg) {
|
||||
(false, false) => U256::from(a > b),
|
||||
(true, false) => uint!(0_U256),
|
||||
(false, true) => uint!(1_U256),
|
||||
(true, true) => U256::from(a > b),
|
||||
}
|
||||
}
|
||||
|
||||
fn u_lt(a: &U256, b: &U256) -> U256 {
|
||||
let a_neg = &HALF_M < a;
|
||||
let b_neg = &HALF_M < b;
|
||||
|
||||
match (a_neg, b_neg) {
|
||||
(false, false) => U256::from(a < b),
|
||||
(true, false) => uint!(1_U256),
|
||||
(false, true) => uint!(0_U256),
|
||||
(true, true) => U256::from(a < b),
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
mod test {
|
||||
use std::{ops::Div, str::FromStr};
|
||||
|
||||
use ruint::uint;
|
||||
|
||||
use super::*;
|
||||
|
||||
#[test]
|
||||
fn test_ok() {
|
||||
let a = Fr::from(4u64);
|
||||
let b = Fr::from(2u64);
|
||||
let c = shl(a, b).unwrap();
|
||||
assert_eq!(c.cmp(&Fr::from(16u64)), Ordering::Equal)
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_div() {
|
||||
assert_eq!(
|
||||
Operation::Div
|
||||
.eval_fr(Fr::from(2u64), Fr::from(3u64))
|
||||
.unwrap(),
|
||||
Fr::from_str(
|
||||
"7296080957279758407415468581752425029516121466805344781232734728858602831873"
|
||||
)
|
||||
.unwrap()
|
||||
);
|
||||
|
||||
assert_eq!(
|
||||
Operation::Div
|
||||
.eval_fr(Fr::from(6u64), Fr::from(2u64))
|
||||
.unwrap(),
|
||||
Fr::from_str("3").unwrap()
|
||||
);
|
||||
|
||||
assert_eq!(
|
||||
Operation::Div
|
||||
.eval_fr(Fr::from(7u64), Fr::from(2u64))
|
||||
.unwrap(),
|
||||
Fr::from_str(
|
||||
"10944121435919637611123202872628637544274182200208017171849102093287904247812"
|
||||
)
|
||||
.unwrap()
|
||||
);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_idiv() {
|
||||
assert_eq!(
|
||||
Operation::Idiv
|
||||
.eval_fr(Fr::from(2u64), Fr::from(3u64))
|
||||
.unwrap(),
|
||||
Fr::from_str("0").unwrap()
|
||||
);
|
||||
|
||||
assert_eq!(
|
||||
Operation::Idiv
|
||||
.eval_fr(Fr::from(6u64), Fr::from(2u64))
|
||||
.unwrap(),
|
||||
Fr::from_str("3").unwrap()
|
||||
);
|
||||
|
||||
assert_eq!(
|
||||
Operation::Idiv
|
||||
.eval_fr(Fr::from(7u64), Fr::from(2u64))
|
||||
.unwrap(),
|
||||
Fr::from_str("3").unwrap()
|
||||
);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_fr_mod() {
|
||||
assert_eq!(
|
||||
Operation::Mod
|
||||
.eval_fr(Fr::from(7u64), Fr::from(2u64))
|
||||
.unwrap(),
|
||||
Fr::from_str("1").unwrap()
|
||||
);
|
||||
|
||||
assert_eq!(
|
||||
Operation::Mod
|
||||
.eval_fr(Fr::from(7u64), Fr::from(9u64))
|
||||
.unwrap(),
|
||||
Fr::from_str("7").unwrap()
|
||||
);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_u_gte() {
|
||||
let result = u_gte(&uint!(10_U256), &uint!(3_U256));
|
||||
assert_eq!(result, uint!(1_U256));
|
||||
|
||||
let result = u_gte(&uint!(3_U256), &uint!(3_U256));
|
||||
assert_eq!(result, uint!(1_U256));
|
||||
|
||||
let result = u_gte(&uint!(2_U256), &uint!(3_U256));
|
||||
assert_eq!(result, uint!(0_U256));
|
||||
|
||||
// -1 >= 3 => 0
|
||||
let result = u_gte(
|
||||
&uint!(
|
||||
21888242871839275222246405745257275088548364400416034343698204186575808495616_U256
|
||||
),
|
||||
&uint!(3_U256),
|
||||
);
|
||||
assert_eq!(result, uint!(0_U256));
|
||||
|
||||
// -1 >= -2 => 1
|
||||
let result = u_gte(
|
||||
&uint!(
|
||||
21888242871839275222246405745257275088548364400416034343698204186575808495616_U256
|
||||
),
|
||||
&uint!(
|
||||
21888242871839275222246405745257275088548364400416034343698204186575808495615_U256
|
||||
),
|
||||
);
|
||||
assert_eq!(result, uint!(1_U256));
|
||||
|
||||
// -2 >= -1 => 0
|
||||
let result = u_gte(
|
||||
&uint!(
|
||||
21888242871839275222246405745257275088548364400416034343698204186575808495615_U256
|
||||
),
|
||||
&uint!(
|
||||
21888242871839275222246405745257275088548364400416034343698204186575808495616_U256
|
||||
),
|
||||
);
|
||||
assert_eq!(result, uint!(0_U256));
|
||||
|
||||
// -2 == -2 => 1
|
||||
let result = u_gte(
|
||||
&uint!(
|
||||
21888242871839275222246405745257275088548364400416034343698204186575808495615_U256
|
||||
),
|
||||
&uint!(
|
||||
21888242871839275222246405745257275088548364400416034343698204186575808495615_U256
|
||||
),
|
||||
);
|
||||
assert_eq!(result, uint!(1_U256));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_x() {
|
||||
let x = M.div(uint!(2_U256));
|
||||
|
||||
println!("x: {:?}", x.as_limbs());
|
||||
println!("x: {M}");
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_2() {
|
||||
let nodes: Vec<Node> = vec![];
|
||||
// let node = nodes[0];
|
||||
let node = nodes.first();
|
||||
println!("{node:?}");
|
||||
}
|
||||
}
|
||||
117
rln/src/circuit/iden3calc/proto.rs
Normal file
117
rln/src/circuit/iden3calc/proto.rs
Normal file
@@ -0,0 +1,117 @@
|
||||
// This crate has been generated by prost-build during compilation of the code by iden3
|
||||
// and modified manually. The *.proto file used to generate this on can be found here:
|
||||
// https://github.com/iden3/circom-witnesscalc/blob/5cb365b6e4d9052ecc69d4567fcf5bc061c20e94/protos/messages.proto
|
||||
|
||||
use std::collections::HashMap;
|
||||
|
||||
#[derive(Clone, PartialEq, prost::Message)]
|
||||
pub(crate) struct BigUInt {
|
||||
#[prost(bytes = "vec", tag = "1")]
|
||||
pub value_le: Vec<u8>,
|
||||
}
|
||||
#[derive(Clone, Copy, PartialEq, prost::Message)]
|
||||
pub(crate) struct InputNode {
|
||||
#[prost(uint32, tag = "1")]
|
||||
pub idx: u32,
|
||||
}
|
||||
#[derive(Clone, PartialEq, prost::Message)]
|
||||
pub(crate) struct ConstantNode {
|
||||
#[prost(message, optional, tag = "1")]
|
||||
pub value: Option<BigUInt>,
|
||||
}
|
||||
#[derive(Clone, Copy, PartialEq, prost::Message)]
|
||||
pub(crate) struct UnoOpNode {
|
||||
#[prost(enumeration = "UnoOp", tag = "1")]
|
||||
pub op: i32,
|
||||
#[prost(uint32, tag = "2")]
|
||||
pub a_idx: u32,
|
||||
}
|
||||
#[derive(Clone, Copy, PartialEq, prost::Message)]
|
||||
pub(crate) struct DuoOpNode {
|
||||
#[prost(enumeration = "DuoOp", tag = "1")]
|
||||
pub op: i32,
|
||||
#[prost(uint32, tag = "2")]
|
||||
pub a_idx: u32,
|
||||
#[prost(uint32, tag = "3")]
|
||||
pub b_idx: u32,
|
||||
}
|
||||
#[derive(Clone, Copy, PartialEq, prost::Message)]
|
||||
pub(crate) struct TresOpNode {
|
||||
#[prost(enumeration = "TresOp", tag = "1")]
|
||||
pub op: i32,
|
||||
#[prost(uint32, tag = "2")]
|
||||
pub a_idx: u32,
|
||||
#[prost(uint32, tag = "3")]
|
||||
pub b_idx: u32,
|
||||
#[prost(uint32, tag = "4")]
|
||||
pub c_idx: u32,
|
||||
}
|
||||
#[derive(Clone, PartialEq, prost::Message)]
|
||||
pub(crate) struct Node {
|
||||
#[prost(oneof = "node::Node", tags = "1, 2, 3, 4, 5")]
|
||||
pub node: Option<node::Node>,
|
||||
}
|
||||
/// Nested message and enum types in `Node`.
|
||||
pub(crate) mod node {
|
||||
#[derive(Clone, PartialEq, prost::Oneof)]
|
||||
pub(crate) enum Node {
|
||||
#[prost(message, tag = "1")]
|
||||
Input(super::InputNode),
|
||||
#[prost(message, tag = "2")]
|
||||
Constant(super::ConstantNode),
|
||||
#[prost(message, tag = "3")]
|
||||
UnoOp(super::UnoOpNode),
|
||||
#[prost(message, tag = "4")]
|
||||
DuoOp(super::DuoOpNode),
|
||||
#[prost(message, tag = "5")]
|
||||
TresOp(super::TresOpNode),
|
||||
}
|
||||
}
|
||||
#[derive(Clone, Copy, PartialEq, prost::Message)]
|
||||
pub(crate) struct SignalDescription {
|
||||
#[prost(uint32, tag = "1")]
|
||||
pub offset: u32,
|
||||
#[prost(uint32, tag = "2")]
|
||||
pub len: u32,
|
||||
}
|
||||
#[derive(Clone, PartialEq, prost::Message)]
|
||||
pub(crate) struct GraphMetadata {
|
||||
#[prost(uint32, repeated, tag = "1")]
|
||||
pub witness_signals: Vec<u32>,
|
||||
#[prost(map = "string, message", tag = "2")]
|
||||
pub inputs: HashMap<String, SignalDescription>,
|
||||
}
|
||||
#[derive(Clone, Copy, Debug, PartialEq, prost::Enumeration)]
|
||||
pub(crate) enum DuoOp {
|
||||
Mul = 0,
|
||||
Div = 1,
|
||||
Add = 2,
|
||||
Sub = 3,
|
||||
Pow = 4,
|
||||
Idiv = 5,
|
||||
Mod = 6,
|
||||
Eq = 7,
|
||||
Neq = 8,
|
||||
Lt = 9,
|
||||
Gt = 10,
|
||||
Leq = 11,
|
||||
Geq = 12,
|
||||
Land = 13,
|
||||
Lor = 14,
|
||||
Shl = 15,
|
||||
Shr = 16,
|
||||
Bor = 17,
|
||||
Band = 18,
|
||||
Bxor = 19,
|
||||
}
|
||||
|
||||
#[derive(Clone, Copy, Debug, PartialEq, prost::Enumeration)]
|
||||
pub(crate) enum UnoOp {
|
||||
Neg = 0,
|
||||
Id = 1,
|
||||
}
|
||||
|
||||
#[derive(Clone, Copy, Debug, PartialEq, prost::Enumeration)]
|
||||
pub(crate) enum TresOp {
|
||||
TernCond = 0,
|
||||
}
|
||||
528
rln/src/circuit/iden3calc/storage.rs
Normal file
528
rln/src/circuit/iden3calc/storage.rs
Normal file
@@ -0,0 +1,528 @@
|
||||
// This crate is based on the code by iden3. Its preimage can be found here:
|
||||
// https://github.com/iden3/circom-witnesscalc/blob/5cb365b6e4d9052ecc69d4567fcf5bc061c20e94/src/storage.rs
|
||||
|
||||
use std::io::{Read, Write};
|
||||
|
||||
use ark_ff::PrimeField;
|
||||
use byteorder::{LittleEndian, ReadBytesExt, WriteBytesExt};
|
||||
use prost::Message;
|
||||
|
||||
use super::{
|
||||
graph::{self, Operation, TresOperation, UnoOperation},
|
||||
proto, InputSignalsInfo,
|
||||
};
|
||||
use crate::circuit::Fr;
|
||||
|
||||
/// Format of the wtns.graph file:
|
||||
/// + magic line: wtns.graph.001
|
||||
/// + 4 bytes unsigned LE 32-bit integer: number of nodes
|
||||
/// + series of protobuf serialized nodes. Each node prefixed by varint length
|
||||
/// + protobuf serialized GraphMetadata
|
||||
/// + 8 bytes unsigned LE 64-bit integer: offset of GraphMetadata message
|
||||
const WITNESSCALC_GRAPH_MAGIC: &[u8] = b"wtns.graph.001";
|
||||
|
||||
const MAX_VARINT_LENGTH: usize = 10;
|
||||
|
||||
impl TryFrom<proto::Node> for graph::Node {
|
||||
type Error = std::io::Error;
|
||||
|
||||
fn try_from(value: proto::Node) -> Result<Self, Self::Error> {
|
||||
let node = value.node.ok_or_else(|| {
|
||||
std::io::Error::new(
|
||||
std::io::ErrorKind::InvalidData,
|
||||
"Proto::Node must have a node field",
|
||||
)
|
||||
})?;
|
||||
match node {
|
||||
proto::node::Node::Input(input_node) => Ok(graph::Node::Input(input_node.idx as usize)),
|
||||
proto::node::Node::Constant(constant_node) => {
|
||||
let i = constant_node.value.ok_or_else(|| {
|
||||
std::io::Error::new(
|
||||
std::io::ErrorKind::InvalidData,
|
||||
"Constant node must have a value",
|
||||
)
|
||||
})?;
|
||||
Ok(graph::Node::MontConstant(Fr::from_le_bytes_mod_order(
|
||||
i.value_le.as_slice(),
|
||||
)))
|
||||
}
|
||||
proto::node::Node::UnoOp(uno_op_node) => {
|
||||
let op = proto::UnoOp::try_from(uno_op_node.op).map_err(|_| {
|
||||
std::io::Error::new(
|
||||
std::io::ErrorKind::InvalidData,
|
||||
"UnoOp must be valid enum value",
|
||||
)
|
||||
})?;
|
||||
Ok(graph::Node::UnoOp(op.into(), uno_op_node.a_idx as usize))
|
||||
}
|
||||
proto::node::Node::DuoOp(duo_op_node) => {
|
||||
let op = proto::DuoOp::try_from(duo_op_node.op).map_err(|_| {
|
||||
std::io::Error::new(
|
||||
std::io::ErrorKind::InvalidData,
|
||||
"DuoOp must be valid enum value",
|
||||
)
|
||||
})?;
|
||||
Ok(graph::Node::Op(
|
||||
op.into(),
|
||||
duo_op_node.a_idx as usize,
|
||||
duo_op_node.b_idx as usize,
|
||||
))
|
||||
}
|
||||
proto::node::Node::TresOp(tres_op_node) => {
|
||||
let op = proto::TresOp::try_from(tres_op_node.op).map_err(|_| {
|
||||
std::io::Error::new(
|
||||
std::io::ErrorKind::InvalidData,
|
||||
"TresOp must be valid enum value",
|
||||
)
|
||||
})?;
|
||||
Ok(graph::Node::TresOp(
|
||||
op.into(),
|
||||
tres_op_node.a_idx as usize,
|
||||
tres_op_node.b_idx as usize,
|
||||
tres_op_node.c_idx as usize,
|
||||
))
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl From<&graph::Node> for proto::node::Node {
|
||||
fn from(node: &graph::Node) -> Self {
|
||||
match node {
|
||||
graph::Node::Input(i) => proto::node::Node::Input(proto::InputNode { idx: *i as u32 }),
|
||||
graph::Node::Constant(_) => {
|
||||
panic!("We are not supposed to write Constant to the witnesscalc graph. All Constant should be converted to MontConstant.");
|
||||
}
|
||||
graph::Node::UnoOp(op, a) => {
|
||||
let op = proto::UnoOp::from(op);
|
||||
proto::node::Node::UnoOp(proto::UnoOpNode {
|
||||
op: op as i32,
|
||||
a_idx: *a as u32,
|
||||
})
|
||||
}
|
||||
graph::Node::Op(op, a, b) => proto::node::Node::DuoOp(proto::DuoOpNode {
|
||||
op: proto::DuoOp::from(op) as i32,
|
||||
a_idx: *a as u32,
|
||||
b_idx: *b as u32,
|
||||
}),
|
||||
graph::Node::TresOp(op, a, b, c) => proto::node::Node::TresOp(proto::TresOpNode {
|
||||
op: proto::TresOp::from(op) as i32,
|
||||
a_idx: *a as u32,
|
||||
b_idx: *b as u32,
|
||||
c_idx: *c as u32,
|
||||
}),
|
||||
graph::Node::MontConstant(c) => {
|
||||
let bi = Into::<num_bigint::BigUint>::into(*c);
|
||||
let i = proto::BigUInt {
|
||||
value_le: bi.to_bytes_le(),
|
||||
};
|
||||
proto::node::Node::Constant(proto::ConstantNode { value: Some(i) })
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl From<proto::UnoOp> for UnoOperation {
|
||||
fn from(value: proto::UnoOp) -> Self {
|
||||
match value {
|
||||
proto::UnoOp::Neg => UnoOperation::Neg,
|
||||
proto::UnoOp::Id => UnoOperation::Id,
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl From<proto::DuoOp> for Operation {
|
||||
fn from(value: proto::DuoOp) -> Self {
|
||||
match value {
|
||||
proto::DuoOp::Mul => Operation::Mul,
|
||||
proto::DuoOp::Div => Operation::Div,
|
||||
proto::DuoOp::Add => Operation::Add,
|
||||
proto::DuoOp::Sub => Operation::Sub,
|
||||
proto::DuoOp::Pow => Operation::Pow,
|
||||
proto::DuoOp::Idiv => Operation::Idiv,
|
||||
proto::DuoOp::Mod => Operation::Mod,
|
||||
proto::DuoOp::Eq => Operation::Eq,
|
||||
proto::DuoOp::Neq => Operation::Neq,
|
||||
proto::DuoOp::Lt => Operation::Lt,
|
||||
proto::DuoOp::Gt => Operation::Gt,
|
||||
proto::DuoOp::Leq => Operation::Leq,
|
||||
proto::DuoOp::Geq => Operation::Geq,
|
||||
proto::DuoOp::Land => Operation::Land,
|
||||
proto::DuoOp::Lor => Operation::Lor,
|
||||
proto::DuoOp::Shl => Operation::Shl,
|
||||
proto::DuoOp::Shr => Operation::Shr,
|
||||
proto::DuoOp::Bor => Operation::Bor,
|
||||
proto::DuoOp::Band => Operation::Band,
|
||||
proto::DuoOp::Bxor => Operation::Bxor,
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl From<proto::TresOp> for graph::TresOperation {
|
||||
fn from(value: proto::TresOp) -> Self {
|
||||
match value {
|
||||
proto::TresOp::TernCond => TresOperation::TernCond,
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
#[allow(dead_code)]
|
||||
pub(crate) fn serialize_witnesscalc_graph<T: Write>(
|
||||
mut w: T,
|
||||
nodes: &Vec<graph::Node>,
|
||||
witness_signals: &[usize],
|
||||
input_signals: &InputSignalsInfo,
|
||||
) -> std::io::Result<()> {
|
||||
let mut ptr = 0usize;
|
||||
w.write_all(WITNESSCALC_GRAPH_MAGIC)?;
|
||||
ptr += WITNESSCALC_GRAPH_MAGIC.len();
|
||||
|
||||
w.write_u64::<LittleEndian>(nodes.len() as u64)?;
|
||||
ptr += 8;
|
||||
|
||||
let metadata = proto::GraphMetadata {
|
||||
witness_signals: witness_signals
|
||||
.iter()
|
||||
.map(|x| *x as u32)
|
||||
.collect::<Vec<u32>>(),
|
||||
inputs: input_signals
|
||||
.iter()
|
||||
.map(|(k, v)| {
|
||||
let sig = proto::SignalDescription {
|
||||
offset: v.0 as u32,
|
||||
len: v.1 as u32,
|
||||
};
|
||||
(k.clone(), sig)
|
||||
})
|
||||
.collect(),
|
||||
};
|
||||
|
||||
// capacity of buf should be enough to hold the largest message + 10 bytes
|
||||
// of varint length
|
||||
let mut buf = Vec::with_capacity(metadata.encoded_len() + MAX_VARINT_LENGTH);
|
||||
|
||||
for node in nodes {
|
||||
let node_pb = proto::Node {
|
||||
node: Some(proto::node::Node::from(node)),
|
||||
};
|
||||
|
||||
assert_eq!(buf.len(), 0);
|
||||
node_pb.encode_length_delimited(&mut buf)?;
|
||||
ptr += buf.len();
|
||||
|
||||
w.write_all(&buf)?;
|
||||
buf.clear();
|
||||
}
|
||||
|
||||
metadata.encode_length_delimited(&mut buf)?;
|
||||
w.write_all(&buf)?;
|
||||
buf.clear();
|
||||
|
||||
w.write_u64::<LittleEndian>(ptr as u64)?;
|
||||
|
||||
Ok(())
|
||||
}
|
||||
|
||||
fn read_message_length<R: Read>(rw: &mut WriteBackReader<R>) -> std::io::Result<usize> {
|
||||
let mut buf = [0u8; MAX_VARINT_LENGTH];
|
||||
let bytes_read = rw.read(&mut buf)?;
|
||||
if bytes_read == 0 {
|
||||
return Err(std::io::Error::new(
|
||||
std::io::ErrorKind::UnexpectedEof,
|
||||
"Unexpected EOF",
|
||||
));
|
||||
}
|
||||
|
||||
let len_delimiter = prost::decode_length_delimiter(buf.as_ref())?;
|
||||
|
||||
let lnln = prost::length_delimiter_len(len_delimiter);
|
||||
|
||||
if lnln < bytes_read {
|
||||
rw.write_all(&buf[lnln..bytes_read])?;
|
||||
}
|
||||
|
||||
Ok(len_delimiter)
|
||||
}
|
||||
|
||||
fn read_message<R: Read, M: Message + std::default::Default>(
|
||||
rw: &mut WriteBackReader<R>,
|
||||
) -> std::io::Result<M> {
|
||||
let ln = read_message_length(rw)?;
|
||||
let mut buf = vec![0u8; ln];
|
||||
let bytes_read = rw.read(&mut buf)?;
|
||||
if bytes_read != ln {
|
||||
return Err(std::io::Error::new(
|
||||
std::io::ErrorKind::UnexpectedEof,
|
||||
"Unexpected EOF",
|
||||
));
|
||||
}
|
||||
|
||||
let msg = prost::Message::decode(&buf[..])?;
|
||||
|
||||
Ok(msg)
|
||||
}
|
||||
|
||||
pub(crate) fn deserialize_witnesscalc_graph(
|
||||
r: impl Read,
|
||||
) -> std::io::Result<(Vec<graph::Node>, Vec<usize>, InputSignalsInfo)> {
|
||||
let mut br = WriteBackReader::new(r);
|
||||
let mut magic = [0u8; WITNESSCALC_GRAPH_MAGIC.len()];
|
||||
|
||||
br.read_exact(&mut magic)?;
|
||||
|
||||
if !magic.eq(WITNESSCALC_GRAPH_MAGIC) {
|
||||
return Err(std::io::Error::new(
|
||||
std::io::ErrorKind::InvalidData,
|
||||
"Invalid magic",
|
||||
));
|
||||
}
|
||||
|
||||
let nodes_num = br.read_u64::<LittleEndian>()?;
|
||||
let mut nodes = Vec::with_capacity(nodes_num as usize);
|
||||
for _ in 0..nodes_num {
|
||||
let n: proto::Node = read_message(&mut br)?;
|
||||
nodes.push(n.try_into()?);
|
||||
}
|
||||
|
||||
let md: proto::GraphMetadata = read_message(&mut br)?;
|
||||
|
||||
let witness_signals = md
|
||||
.witness_signals
|
||||
.iter()
|
||||
.map(|x| *x as usize)
|
||||
.collect::<Vec<usize>>();
|
||||
|
||||
let input_signals = md
|
||||
.inputs
|
||||
.iter()
|
||||
.map(|(k, v)| (k.clone(), (v.offset as usize, v.len as usize)))
|
||||
.collect::<InputSignalsInfo>();
|
||||
|
||||
Ok((nodes, witness_signals, input_signals))
|
||||
}
|
||||
|
||||
struct WriteBackReader<R: Read> {
|
||||
reader: R,
|
||||
buffer: Vec<u8>,
|
||||
}
|
||||
|
||||
impl<R: Read> WriteBackReader<R> {
|
||||
fn new(reader: R) -> Self {
|
||||
WriteBackReader {
|
||||
reader,
|
||||
buffer: Vec::new(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl<R: Read> Read for WriteBackReader<R> {
|
||||
fn read(&mut self, buf: &mut [u8]) -> std::io::Result<usize> {
|
||||
if buf.is_empty() {
|
||||
return Ok(0);
|
||||
}
|
||||
|
||||
let mut n = 0usize;
|
||||
|
||||
if !self.buffer.is_empty() {
|
||||
n = std::cmp::min(buf.len(), self.buffer.len());
|
||||
self.buffer[self.buffer.len() - n..]
|
||||
.iter()
|
||||
.rev()
|
||||
.enumerate()
|
||||
.for_each(|(i, x)| {
|
||||
buf[i] = *x;
|
||||
});
|
||||
self.buffer.truncate(self.buffer.len() - n);
|
||||
}
|
||||
|
||||
while n < buf.len() {
|
||||
let m = self.reader.read(&mut buf[n..])?;
|
||||
if m == 0 {
|
||||
break;
|
||||
}
|
||||
n += m;
|
||||
}
|
||||
|
||||
Ok(n)
|
||||
}
|
||||
}
|
||||
|
||||
impl<R: Read> Write for WriteBackReader<R> {
|
||||
fn write(&mut self, buf: &[u8]) -> std::io::Result<usize> {
|
||||
self.buffer.reserve(buf.len());
|
||||
self.buffer.extend(buf.iter().rev());
|
||||
Ok(buf.len())
|
||||
}
|
||||
|
||||
fn flush(&mut self) -> std::io::Result<()> {
|
||||
Ok(())
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
mod test {
|
||||
use core::str::FromStr;
|
||||
use std::collections::HashMap;
|
||||
|
||||
use byteorder::ByteOrder;
|
||||
use graph::{Operation, TresOperation, UnoOperation};
|
||||
|
||||
use super::*;
|
||||
|
||||
#[test]
|
||||
fn test_read_message() {
|
||||
let mut buf = Vec::new();
|
||||
let n1 = proto::Node {
|
||||
node: Some(proto::node::Node::Input(proto::InputNode { idx: 1 })),
|
||||
};
|
||||
n1.encode_length_delimited(&mut buf).unwrap();
|
||||
|
||||
let n2 = proto::Node {
|
||||
node: Some(proto::node::Node::Input(proto::InputNode { idx: 2 })),
|
||||
};
|
||||
n2.encode_length_delimited(&mut buf).unwrap();
|
||||
|
||||
let mut reader = std::io::Cursor::new(&buf);
|
||||
|
||||
let mut rw = WriteBackReader::new(&mut reader);
|
||||
|
||||
let got_n1: proto::Node = read_message(&mut rw).unwrap();
|
||||
assert!(n1.eq(&got_n1));
|
||||
|
||||
let got_n2: proto::Node = read_message(&mut rw).unwrap();
|
||||
assert!(n2.eq(&got_n2));
|
||||
|
||||
assert_eq!(reader.position(), buf.len() as u64);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_read_message_variant() {
|
||||
let nodes = vec![
|
||||
proto::Node {
|
||||
node: Some(proto::node::Node::from(&graph::Node::Input(0))),
|
||||
},
|
||||
proto::Node {
|
||||
node: Some(proto::node::Node::from(&graph::Node::MontConstant(
|
||||
Fr::from_str("1").unwrap(),
|
||||
))),
|
||||
},
|
||||
proto::Node {
|
||||
node: Some(proto::node::Node::from(&graph::Node::UnoOp(
|
||||
UnoOperation::Id,
|
||||
4,
|
||||
))),
|
||||
},
|
||||
proto::Node {
|
||||
node: Some(proto::node::Node::from(&graph::Node::Op(
|
||||
Operation::Mul,
|
||||
5,
|
||||
6,
|
||||
))),
|
||||
},
|
||||
proto::Node {
|
||||
node: Some(proto::node::Node::from(&graph::Node::TresOp(
|
||||
TresOperation::TernCond,
|
||||
7,
|
||||
8,
|
||||
9,
|
||||
))),
|
||||
},
|
||||
];
|
||||
|
||||
let mut buf = Vec::new();
|
||||
for n in &nodes {
|
||||
n.encode_length_delimited(&mut buf).unwrap();
|
||||
}
|
||||
|
||||
let mut nodes_got: Vec<proto::Node> = Vec::new();
|
||||
let mut reader = std::io::Cursor::new(&buf);
|
||||
let mut rw = WriteBackReader::new(&mut reader);
|
||||
for _ in 0..nodes.len() {
|
||||
nodes_got.push(read_message(&mut rw).unwrap());
|
||||
}
|
||||
|
||||
assert_eq!(nodes, nodes_got);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_write_back_reader() {
|
||||
let data = [1u8, 2, 3, 4, 5, 6];
|
||||
let mut r = WriteBackReader::new(std::io::Cursor::new(&data));
|
||||
|
||||
let buf = &mut [0u8; 5];
|
||||
r.read_exact(buf).unwrap();
|
||||
assert_eq!(buf, &[1, 2, 3, 4, 5]);
|
||||
|
||||
// return [4, 5] to reader
|
||||
r.write_all(&buf[3..]).unwrap();
|
||||
// return [2, 3] to reader
|
||||
r.write_all(&buf[1..3]).unwrap();
|
||||
|
||||
buf.fill(0);
|
||||
|
||||
// read 3 bytes, expect [2, 3, 4] after returns
|
||||
let mut n = r.read(&mut buf[..3]).unwrap();
|
||||
assert_eq!(n, 3);
|
||||
assert_eq!(buf, &[2, 3, 4, 0, 0]);
|
||||
|
||||
buf.fill(0);
|
||||
|
||||
// read everything left in reader
|
||||
n = r.read(buf).unwrap();
|
||||
assert_eq!(n, 2);
|
||||
assert_eq!(buf, &[5, 6, 0, 0, 0]);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_deserialize_inputs() {
|
||||
let nodes = vec![
|
||||
graph::Node::Input(0),
|
||||
graph::Node::MontConstant(Fr::from_str("1").unwrap()),
|
||||
graph::Node::UnoOp(UnoOperation::Id, 4),
|
||||
graph::Node::Op(Operation::Mul, 5, 6),
|
||||
graph::Node::TresOp(TresOperation::TernCond, 7, 8, 9),
|
||||
];
|
||||
|
||||
let witness_signals = vec![4, 1];
|
||||
|
||||
let mut input_signals: InputSignalsInfo = HashMap::new();
|
||||
input_signals.insert("sig1".to_string(), (1, 3));
|
||||
input_signals.insert("sig2".to_string(), (5, 1));
|
||||
|
||||
let mut tmp = Vec::new();
|
||||
serialize_witnesscalc_graph(&mut tmp, &nodes, &witness_signals, &input_signals).unwrap();
|
||||
|
||||
let mut reader = std::io::Cursor::new(&tmp);
|
||||
|
||||
let (nodes_res, witness_signals_res, input_signals_res) =
|
||||
deserialize_witnesscalc_graph(&mut reader).unwrap();
|
||||
|
||||
assert_eq!(nodes, nodes_res);
|
||||
assert_eq!(input_signals, input_signals_res);
|
||||
assert_eq!(witness_signals, witness_signals_res);
|
||||
|
||||
let metadata_start = LittleEndian::read_u64(&tmp[tmp.len() - 8..]);
|
||||
|
||||
let mt_reader = std::io::Cursor::new(&tmp[metadata_start as usize..]);
|
||||
let mut rw = WriteBackReader::new(mt_reader);
|
||||
let metadata: proto::GraphMetadata = read_message(&mut rw).unwrap();
|
||||
|
||||
let metadata_want = proto::GraphMetadata {
|
||||
witness_signals: vec![4, 1],
|
||||
inputs: input_signals
|
||||
.iter()
|
||||
.map(|(k, v)| {
|
||||
(
|
||||
k.clone(),
|
||||
proto::SignalDescription {
|
||||
offset: v.0 as u32,
|
||||
len: v.1 as u32,
|
||||
},
|
||||
)
|
||||
})
|
||||
.collect(),
|
||||
};
|
||||
|
||||
assert_eq!(metadata, metadata_want);
|
||||
}
|
||||
}
|
||||
151
rln/src/circuit/mod.rs
Normal file
151
rln/src/circuit/mod.rs
Normal file
@@ -0,0 +1,151 @@
|
||||
// This crate provides interfaces for the zero-knowledge circuit and keys
|
||||
|
||||
pub(crate) mod error;
|
||||
pub(crate) mod iden3calc;
|
||||
pub(crate) mod qap;
|
||||
|
||||
#[cfg(not(target_arch = "wasm32"))]
|
||||
use std::sync::LazyLock;
|
||||
|
||||
use ark_bn254::{
|
||||
Bn254, Fq as ArkFq, Fq2 as ArkFq2, Fr as ArkFr, G1Affine as ArkG1Affine,
|
||||
G1Projective as ArkG1Projective, G2Affine as ArkG2Affine, G2Projective as ArkG2Projective,
|
||||
};
|
||||
use ark_ff::Field;
|
||||
use ark_groth16::{
|
||||
Proof as ArkProof, ProvingKey as ArkProvingKey, VerifyingKey as ArkVerifyingKey,
|
||||
};
|
||||
use ark_relations::r1cs::ConstraintMatrices;
|
||||
use ark_serialize::{CanonicalDeserialize, CanonicalSerialize};
|
||||
|
||||
use self::error::ZKeyReadError;
|
||||
|
||||
#[cfg(not(target_arch = "wasm32"))]
|
||||
const GRAPH_BYTES: &[u8] = include_bytes!("../../resources/tree_depth_20/graph.bin");
|
||||
|
||||
#[cfg(not(target_arch = "wasm32"))]
|
||||
const ARKZKEY_BYTES: &[u8] = include_bytes!("../../resources/tree_depth_20/rln_final.arkzkey");
|
||||
|
||||
#[cfg(not(target_arch = "wasm32"))]
|
||||
static ARKZKEY: LazyLock<Zkey> = LazyLock::new(|| {
|
||||
read_arkzkey_from_bytes_uncompressed(ARKZKEY_BYTES).expect("Default zkey must be valid")
|
||||
});
|
||||
|
||||
pub const DEFAULT_TREE_DEPTH: usize = 20;
|
||||
pub const COMPRESS_PROOF_SIZE: usize = 128;
|
||||
|
||||
// The following types define the pairing friendly elliptic curve, the underlying finite fields and groups default to this module
|
||||
// Note that proofs are serialized assuming Fr to be 4x8 = 32 bytes in size. Hence, changing to a curve with different encoding will make proof verification to fail
|
||||
|
||||
/// BN254 pairing-friendly elliptic curve.
|
||||
pub type Curve = Bn254;
|
||||
|
||||
/// Scalar field Fr of the BN254 curve.
|
||||
pub type Fr = ArkFr;
|
||||
|
||||
/// Base field Fq of the BN254 curve.
|
||||
pub type Fq = ArkFq;
|
||||
|
||||
/// Quadratic extension field element for the BN254 curve.
|
||||
pub type Fq2 = ArkFq2;
|
||||
|
||||
/// Affine representation of a G1 group element on the BN254 curve.
|
||||
pub type G1Affine = ArkG1Affine;
|
||||
|
||||
/// Projective representation of a G1 group element on the BN254 curve.
|
||||
pub type G1Projective = ArkG1Projective;
|
||||
|
||||
/// Affine representation of a G2 group element on the BN254 curve.
|
||||
pub type G2Affine = ArkG2Affine;
|
||||
|
||||
/// Projective representation of a G2 group element on the BN254 curve.
|
||||
pub type G2Projective = ArkG2Projective;
|
||||
|
||||
/// Groth16 proof for the BN254 curve.
|
||||
pub type Proof = ArkProof<Curve>;
|
||||
|
||||
/// Proving key for the Groth16 proof system.
|
||||
pub type ProvingKey = ArkProvingKey<Curve>;
|
||||
|
||||
/// Combining the proving key and constraint matrices.
|
||||
pub type Zkey = (ArkProvingKey<Curve>, ConstraintMatrices<Fr>);
|
||||
|
||||
/// Verifying key for the Groth16 proof system.
|
||||
pub type VerifyingKey = ArkVerifyingKey<Curve>;
|
||||
|
||||
/// Loads the zkey from raw bytes
|
||||
pub fn zkey_from_raw(zkey_data: &[u8]) -> Result<Zkey, ZKeyReadError> {
|
||||
if zkey_data.is_empty() {
|
||||
return Err(ZKeyReadError::EmptyBytes);
|
||||
}
|
||||
|
||||
let proving_key_and_matrices = read_arkzkey_from_bytes_uncompressed(zkey_data)?;
|
||||
|
||||
Ok(proving_key_and_matrices)
|
||||
}
|
||||
|
||||
// Loads default zkey from folder
|
||||
#[cfg(not(target_arch = "wasm32"))]
|
||||
pub fn zkey_from_folder() -> &'static Zkey {
|
||||
&ARKZKEY
|
||||
}
|
||||
|
||||
// Loads default graph from folder
|
||||
#[cfg(not(target_arch = "wasm32"))]
|
||||
pub fn graph_from_folder() -> &'static [u8] {
|
||||
GRAPH_BYTES
|
||||
}
|
||||
|
||||
// The following functions and structs are based on code from ark-zkey:
|
||||
// https://github.com/zkmopro/ark-zkey/blob/main/src/lib.rs#L106
|
||||
|
||||
#[derive(CanonicalSerialize, CanonicalDeserialize, Clone, Debug, PartialEq)]
|
||||
struct SerializableProvingKey(ArkProvingKey<Curve>);
|
||||
|
||||
#[derive(CanonicalSerialize, CanonicalDeserialize, Clone, Debug, PartialEq)]
|
||||
struct SerializableConstraintMatrices<F: Field> {
|
||||
num_instance_variables: usize,
|
||||
num_witness_variables: usize,
|
||||
num_constraints: usize,
|
||||
a_num_non_zero: usize,
|
||||
b_num_non_zero: usize,
|
||||
c_num_non_zero: usize,
|
||||
a: SerializableMatrix<F>,
|
||||
b: SerializableMatrix<F>,
|
||||
c: SerializableMatrix<F>,
|
||||
}
|
||||
|
||||
#[derive(CanonicalSerialize, CanonicalDeserialize, Clone, Debug, PartialEq)]
|
||||
struct SerializableMatrix<F: Field> {
|
||||
pub data: Vec<Vec<(F, usize)>>,
|
||||
}
|
||||
|
||||
fn read_arkzkey_from_bytes_uncompressed(arkzkey_data: &[u8]) -> Result<Zkey, ZKeyReadError> {
|
||||
if arkzkey_data.is_empty() {
|
||||
return Err(ZKeyReadError::EmptyBytes);
|
||||
}
|
||||
|
||||
let mut cursor = std::io::Cursor::new(arkzkey_data);
|
||||
|
||||
let serialized_proving_key =
|
||||
SerializableProvingKey::deserialize_uncompressed_unchecked(&mut cursor)?;
|
||||
|
||||
let serialized_constraint_matrices =
|
||||
SerializableConstraintMatrices::deserialize_uncompressed_unchecked(&mut cursor)?;
|
||||
|
||||
let proving_key: ProvingKey = serialized_proving_key.0;
|
||||
let constraint_matrices: ConstraintMatrices<Fr> = ConstraintMatrices {
|
||||
num_instance_variables: serialized_constraint_matrices.num_instance_variables,
|
||||
num_witness_variables: serialized_constraint_matrices.num_witness_variables,
|
||||
num_constraints: serialized_constraint_matrices.num_constraints,
|
||||
a_num_non_zero: serialized_constraint_matrices.a_num_non_zero,
|
||||
b_num_non_zero: serialized_constraint_matrices.b_num_non_zero,
|
||||
c_num_non_zero: serialized_constraint_matrices.c_num_non_zero,
|
||||
a: serialized_constraint_matrices.a.data,
|
||||
b: serialized_constraint_matrices.b.data,
|
||||
c: serialized_constraint_matrices.c.data,
|
||||
};
|
||||
let zkey = (proving_key, constraint_matrices);
|
||||
|
||||
Ok(zkey)
|
||||
}
|
||||
119
rln/src/circuit/qap.rs
Normal file
119
rln/src/circuit/qap.rs
Normal file
@@ -0,0 +1,119 @@
|
||||
// This crate is based on the code by arkworks. Its preimage can be found here:
|
||||
// https://github.com/arkworks-rs/circom-compat/blob/3c95ed98e23a408b4d99a53e483a9bba39685a4e/src/circom/qap.rs
|
||||
|
||||
use ark_ff::PrimeField;
|
||||
use ark_groth16::r1cs_to_qap::{evaluate_constraint, LibsnarkReduction, R1CSToQAP};
|
||||
use ark_poly::EvaluationDomain;
|
||||
use ark_relations::r1cs::{ConstraintMatrices, ConstraintSystemRef, SynthesisError};
|
||||
use ark_std::{cfg_into_iter, cfg_iter, cfg_iter_mut, vec};
|
||||
#[cfg(feature = "parallel")]
|
||||
use rayon::iter::{
|
||||
IndexedParallelIterator, IntoParallelIterator, IntoParallelRefIterator,
|
||||
IntoParallelRefMutIterator, ParallelIterator,
|
||||
};
|
||||
|
||||
/// Implements the witness map used by snarkjs. The arkworks witness map calculates the
|
||||
/// coefficients of H through computing (AB-C)/Z in the evaluation domain and going back to the
|
||||
/// coefficients domain. snarkjs instead precomputes the Lagrange form of the powers of tau bases
|
||||
/// in a domain twice as large and the witness map is computed as the odd coefficients of (AB-C)
|
||||
/// in that domain. This serves as HZ when computing the C proof element.
|
||||
pub(crate) struct CircomReduction;
|
||||
|
||||
impl R1CSToQAP for CircomReduction {
|
||||
#[allow(clippy::type_complexity)]
|
||||
fn instance_map_with_evaluation<F: PrimeField, D: EvaluationDomain<F>>(
|
||||
cs: ConstraintSystemRef<F>,
|
||||
t: &F,
|
||||
) -> Result<(Vec<F>, Vec<F>, Vec<F>, F, usize, usize), SynthesisError> {
|
||||
LibsnarkReduction::instance_map_with_evaluation::<F, D>(cs, t)
|
||||
}
|
||||
|
||||
fn witness_map_from_matrices<F: PrimeField, D: EvaluationDomain<F>>(
|
||||
matrices: &ConstraintMatrices<F>,
|
||||
num_inputs: usize,
|
||||
num_constraints: usize,
|
||||
full_assignment: &[F],
|
||||
) -> Result<Vec<F>, SynthesisError> {
|
||||
let zero = F::zero();
|
||||
let domain =
|
||||
D::new(num_constraints + num_inputs).ok_or(SynthesisError::PolynomialDegreeTooLarge)?;
|
||||
let domain_size = domain.size();
|
||||
|
||||
let mut a = vec![zero; domain_size];
|
||||
let mut b = vec![zero; domain_size];
|
||||
|
||||
#[allow(unexpected_cfgs)]
|
||||
cfg_iter_mut!(a[..num_constraints])
|
||||
.zip(cfg_iter_mut!(b[..num_constraints]))
|
||||
.zip(cfg_iter!(&matrices.a))
|
||||
.zip(cfg_iter!(&matrices.b))
|
||||
.for_each(|(((a, b), at_i), bt_i)| {
|
||||
*a = evaluate_constraint(at_i, full_assignment);
|
||||
*b = evaluate_constraint(bt_i, full_assignment);
|
||||
});
|
||||
|
||||
{
|
||||
let start = num_constraints;
|
||||
let end = start + num_inputs;
|
||||
a[start..end].clone_from_slice(&full_assignment[..num_inputs]);
|
||||
}
|
||||
|
||||
let mut c = vec![zero; domain_size];
|
||||
#[allow(unexpected_cfgs)]
|
||||
cfg_iter_mut!(c[..num_constraints])
|
||||
.zip(&a)
|
||||
.zip(&b)
|
||||
.for_each(|((c_i, &a), &b)| {
|
||||
*c_i = a * b;
|
||||
});
|
||||
|
||||
domain.ifft_in_place(&mut a);
|
||||
domain.ifft_in_place(&mut b);
|
||||
|
||||
let root_of_unity = {
|
||||
let domain_size_double = 2 * domain_size;
|
||||
let domain_double =
|
||||
D::new(domain_size_double).ok_or(SynthesisError::PolynomialDegreeTooLarge)?;
|
||||
domain_double.element(1)
|
||||
};
|
||||
D::distribute_powers_and_mul_by_const(&mut a, root_of_unity, F::one());
|
||||
D::distribute_powers_and_mul_by_const(&mut b, root_of_unity, F::one());
|
||||
|
||||
domain.fft_in_place(&mut a);
|
||||
domain.fft_in_place(&mut b);
|
||||
|
||||
let mut ab = domain.mul_polynomials_in_evaluation_domain(&a, &b);
|
||||
drop(a);
|
||||
drop(b);
|
||||
|
||||
domain.ifft_in_place(&mut c);
|
||||
D::distribute_powers_and_mul_by_const(&mut c, root_of_unity, F::one());
|
||||
domain.fft_in_place(&mut c);
|
||||
|
||||
#[allow(unexpected_cfgs)]
|
||||
cfg_iter_mut!(ab)
|
||||
.zip(c)
|
||||
.for_each(|(ab_i, c_i)| *ab_i -= &c_i);
|
||||
|
||||
Ok(ab)
|
||||
}
|
||||
|
||||
fn h_query_scalars<F: PrimeField, D: EvaluationDomain<F>>(
|
||||
max_power: usize,
|
||||
t: F,
|
||||
_: F,
|
||||
delta_inverse: F,
|
||||
) -> Result<Vec<F>, SynthesisError> {
|
||||
// the usual H query has domain-1 powers. Z has domain powers. So HZ has 2*domain-1 powers.
|
||||
#[allow(unexpected_cfgs)]
|
||||
let mut scalars = cfg_into_iter!(0..2 * max_power + 1)
|
||||
.map(|i| delta_inverse * t.pow([i as u64]))
|
||||
.collect::<Vec<_>>();
|
||||
let domain_size = scalars.len();
|
||||
let domain = D::new(domain_size).ok_or(SynthesisError::PolynomialDegreeTooLarge)?;
|
||||
// generate the lagrange coefficients
|
||||
domain.ifft_in_place(&mut scalars);
|
||||
#[allow(unexpected_cfgs)]
|
||||
Ok(cfg_into_iter!(scalars).skip(1).step_by(2).collect())
|
||||
}
|
||||
}
|
||||
83
rln/src/error.rs
Normal file
83
rln/src/error.rs
Normal file
@@ -0,0 +1,83 @@
|
||||
use std::{array::TryFromSliceError, num::TryFromIntError};
|
||||
|
||||
use ark_relations::r1cs::SynthesisError;
|
||||
use num_bigint::{BigInt, ParseBigIntError};
|
||||
use thiserror::Error;
|
||||
use zerokit_utils::error::{FromConfigError, HashError, ZerokitMerkleTreeError};
|
||||
|
||||
use crate::circuit::{
|
||||
error::{WitnessCalcError, ZKeyReadError},
|
||||
Fr,
|
||||
};
|
||||
|
||||
/// Errors that can occur during RLN utility operations (conversions, parsing, etc.)
|
||||
#[derive(Debug, thiserror::Error)]
|
||||
pub enum UtilsError {
|
||||
#[error("Expected radix 10 or 16")]
|
||||
WrongRadix,
|
||||
#[error("Failed to parse big integer: {0}")]
|
||||
ParseBigInt(#[from] ParseBigIntError),
|
||||
#[error("Failed to convert to usize: {0}")]
|
||||
ToUsize(#[from] TryFromIntError),
|
||||
#[error("Failed to convert from slice: {0}")]
|
||||
FromSlice(#[from] TryFromSliceError),
|
||||
#[error("Input data too short: expected at least {expected} bytes, got {actual} bytes")]
|
||||
InsufficientData { expected: usize, actual: usize },
|
||||
}
|
||||
|
||||
/// Errors that can occur during RLN protocol operations (proof generation, verification, etc.)
|
||||
#[derive(Debug, thiserror::Error)]
|
||||
pub enum ProtocolError {
|
||||
#[error("Error producing proof: {0}")]
|
||||
Synthesis(#[from] SynthesisError),
|
||||
#[error("RLN utility error: {0}")]
|
||||
Utils(#[from] UtilsError),
|
||||
#[error("Error calculating witness: {0}")]
|
||||
WitnessCalc(#[from] WitnessCalcError),
|
||||
#[error("Expected to read {0} bytes but read only {1} bytes")]
|
||||
InvalidReadLen(usize, usize),
|
||||
#[error("Cannot convert bigint {0:?} to biguint")]
|
||||
BigUintConversion(BigInt),
|
||||
#[error("Message id ({0}) is not within user_message_limit ({1})")]
|
||||
InvalidMessageId(Fr, Fr),
|
||||
#[error("Merkle proof length mismatch: expected {0}, got {1}")]
|
||||
InvalidMerkleProofLength(usize, usize),
|
||||
#[error("External nullifiers mismatch: {0} != {1}")]
|
||||
ExternalNullifierMismatch(Fr, Fr),
|
||||
#[error("Cannot recover secret: division by zero")]
|
||||
DivisionByZero,
|
||||
#[error("Merkle tree operation error: {0}")]
|
||||
MerkleTree(#[from] ZerokitMerkleTreeError),
|
||||
#[error("Hash computation error: {0}")]
|
||||
Hash(#[from] HashError),
|
||||
#[error("Proof serialization error: {0}")]
|
||||
SerializationError(#[from] ark_serialize::SerializationError),
|
||||
}
|
||||
|
||||
/// Errors that can occur during proof verification
|
||||
#[derive(Error, Debug)]
|
||||
pub enum VerifyError {
|
||||
#[error("Invalid proof provided")]
|
||||
InvalidProof,
|
||||
#[error("Expected one of the provided roots")]
|
||||
InvalidRoot,
|
||||
#[error("Signal value does not match")]
|
||||
InvalidSignal,
|
||||
}
|
||||
|
||||
/// Top-level RLN error type encompassing all RLN operations
|
||||
#[derive(Debug, thiserror::Error)]
|
||||
pub enum RLNError {
|
||||
#[error("Configuration error: {0}")]
|
||||
Config(#[from] FromConfigError),
|
||||
#[error("Merkle tree error: {0}")]
|
||||
MerkleTree(#[from] ZerokitMerkleTreeError),
|
||||
#[error("Hash error: {0}")]
|
||||
Hash(#[from] HashError),
|
||||
#[error("ZKey error: {0}")]
|
||||
ZKey(#[from] ZKeyReadError),
|
||||
#[error("Protocol error: {0}")]
|
||||
Protocol(#[from] ProtocolError),
|
||||
#[error("Verification error: {0}")]
|
||||
Verify(#[from] VerifyError),
|
||||
}
|
||||
550
rln/src/ffi.rs
550
rln/src/ffi.rs
@@ -1,550 +0,0 @@
|
||||
// This crate implements the public Foreign Function Interface (FFI) for the RLN module
|
||||
|
||||
use std::slice;
|
||||
|
||||
use crate::public::{hash as public_hash, poseidon_hash as public_poseidon_hash, RLN};
|
||||
|
||||
// Macro to call methods with arbitrary amount of arguments,
|
||||
// First argument to the macro is context,
|
||||
// second is the actual method on `RLN`
|
||||
// rest are all other arguments to the method
|
||||
macro_rules! call {
|
||||
($instance:expr, $method:ident $(, $arg:expr)*) => {
|
||||
{
|
||||
let new_instance: &mut RLN = $instance.process();
|
||||
match new_instance.$method($($arg.process()),*) {
|
||||
Ok(()) => {
|
||||
true
|
||||
}
|
||||
Err(err) => {
|
||||
eprintln!("execution error: {err}");
|
||||
false
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Macro to call methods with arbitrary amount of arguments,
|
||||
// which have the last argument is output buffer pointer
|
||||
// First argument to the macro is context,
|
||||
// second is the actual method on `RLN`
|
||||
// third is the aforementioned output buffer argument
|
||||
// rest are all other arguments to the method
|
||||
macro_rules! call_with_output_arg {
|
||||
// this variant is needed for the case when
|
||||
// there are zero other arguments
|
||||
($instance:expr, $method:ident, $output_arg:expr) => {
|
||||
{
|
||||
let mut output_data: Vec<u8> = Vec::new();
|
||||
let new_instance = $instance.process();
|
||||
match new_instance.$method(&mut output_data) {
|
||||
Ok(()) => {
|
||||
unsafe { *$output_arg = Buffer::from(&output_data[..]) };
|
||||
std::mem::forget(output_data);
|
||||
true
|
||||
}
|
||||
Err(err) => {
|
||||
std::mem::forget(output_data);
|
||||
eprintln!("execution error: {err}");
|
||||
false
|
||||
}
|
||||
}
|
||||
}
|
||||
};
|
||||
($instance:expr, $method:ident, $output_arg:expr, $( $arg:expr ),* ) => {
|
||||
{
|
||||
let mut output_data: Vec<u8> = Vec::new();
|
||||
let new_instance = $instance.process();
|
||||
match new_instance.$method($($arg.process()),*, &mut output_data) {
|
||||
Ok(()) => {
|
||||
unsafe { *$output_arg = Buffer::from(&output_data[..]) };
|
||||
std::mem::forget(output_data);
|
||||
true
|
||||
}
|
||||
Err(err) => {
|
||||
std::mem::forget(output_data);
|
||||
eprintln!("execution error: {err}");
|
||||
false
|
||||
}
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
}
|
||||
|
||||
// Macro to call methods with arbitrary amount of arguments,
|
||||
// which are not implemented in a ctx RLN object
|
||||
// First argument is the method to call
|
||||
// Second argument is the output buffer argument
|
||||
// The remaining arguments are all other inputs to the method
|
||||
macro_rules! no_ctx_call_with_output_arg {
|
||||
($method:ident, $output_arg:expr, $( $arg:expr ),* ) => {
|
||||
{
|
||||
let mut output_data: Vec<u8> = Vec::new();
|
||||
match $method($($arg.process()),*, &mut output_data) {
|
||||
Ok(()) => {
|
||||
unsafe { *$output_arg = Buffer::from(&output_data[..]) };
|
||||
std::mem::forget(output_data);
|
||||
true
|
||||
}
|
||||
Err(err) => {
|
||||
std::mem::forget(output_data);
|
||||
eprintln!("execution error: {err}");
|
||||
false
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Macro to call methods with arbitrary amount of arguments,
|
||||
// which have the last argument as bool
|
||||
// First argument to the macro is context,
|
||||
// second is the actual method on `RLN`
|
||||
// third is the aforementioned bool argument
|
||||
// rest are all other arguments to the method
|
||||
macro_rules! call_with_bool_arg {
|
||||
($instance:expr, $method:ident, $bool_arg:expr, $( $arg:expr ),* ) => {
|
||||
{
|
||||
let new_instance = $instance.process();
|
||||
if match new_instance.$method($($arg.process()),*,) {
|
||||
Ok(result) => result,
|
||||
Err(err) => {
|
||||
eprintln!("execution error: {err}");
|
||||
return false
|
||||
},
|
||||
} {
|
||||
unsafe { *$bool_arg = true };
|
||||
} else {
|
||||
unsafe { *$bool_arg = false };
|
||||
};
|
||||
true
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
trait ProcessArg {
|
||||
type ReturnType;
|
||||
fn process(self) -> Self::ReturnType;
|
||||
}
|
||||
|
||||
impl ProcessArg for usize {
|
||||
type ReturnType = usize;
|
||||
fn process(self) -> Self::ReturnType {
|
||||
self
|
||||
}
|
||||
}
|
||||
|
||||
impl ProcessArg for *const Buffer {
|
||||
type ReturnType = &'static [u8];
|
||||
fn process(self) -> Self::ReturnType {
|
||||
<&[u8]>::from(unsafe { &*self })
|
||||
}
|
||||
}
|
||||
|
||||
impl ProcessArg for *const RLN {
|
||||
type ReturnType = &'static RLN;
|
||||
fn process(self) -> Self::ReturnType {
|
||||
unsafe { &*self }
|
||||
}
|
||||
}
|
||||
|
||||
impl ProcessArg for *mut RLN {
|
||||
type ReturnType = &'static mut RLN;
|
||||
fn process(self) -> Self::ReturnType {
|
||||
unsafe { &mut *self }
|
||||
}
|
||||
}
|
||||
|
||||
/// Buffer struct is taken from
|
||||
/// <https://github.com/celo-org/celo-threshold-bls-rs/blob/master/crates/threshold-bls-ffi/src/ffi.rs>
|
||||
///
|
||||
/// Also heavily inspired by <https://github.com/kilic/rln/blob/master/src/ffi.rs>
|
||||
|
||||
#[repr(C)]
|
||||
#[derive(Clone, Debug, PartialEq)]
|
||||
pub struct Buffer {
|
||||
pub ptr: *const u8,
|
||||
pub len: usize,
|
||||
}
|
||||
|
||||
impl From<&[u8]> for Buffer {
|
||||
fn from(src: &[u8]) -> Self {
|
||||
Self {
|
||||
ptr: src.as_ptr(),
|
||||
len: src.len(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl<'a> From<&Buffer> for &'a [u8] {
|
||||
fn from(src: &Buffer) -> &'a [u8] {
|
||||
unsafe { slice::from_raw_parts(src.ptr, src.len) }
|
||||
}
|
||||
}
|
||||
|
||||
// TODO: check if there are security implications by using this clippy
|
||||
// #[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
|
||||
////////////////////////////////////////////////////////
|
||||
// RLN APIs
|
||||
////////////////////////////////////////////////////////
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[cfg(not(feature = "stateless"))]
|
||||
#[no_mangle]
|
||||
pub extern "C" fn new(tree_height: usize, input_buffer: *const Buffer, ctx: *mut *mut RLN) -> bool {
|
||||
match RLN::new(tree_height, input_buffer.process()) {
|
||||
Ok(rln) => {
|
||||
unsafe { *ctx = Box::into_raw(Box::new(rln)) };
|
||||
true
|
||||
}
|
||||
Err(err) => {
|
||||
eprintln!("could not instantiate rln: {err}");
|
||||
false
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[cfg(feature = "stateless")]
|
||||
#[no_mangle]
|
||||
pub extern "C" fn new(ctx: *mut *mut RLN) -> bool {
|
||||
match RLN::new() {
|
||||
Ok(rln) => {
|
||||
unsafe { *ctx = Box::into_raw(Box::new(rln)) };
|
||||
true
|
||||
}
|
||||
Err(err) => {
|
||||
eprintln!("could not instantiate rln: {err}");
|
||||
false
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[cfg(not(feature = "stateless"))]
|
||||
#[no_mangle]
|
||||
pub extern "C" fn new_with_params(
|
||||
tree_height: usize,
|
||||
circom_buffer: *const Buffer,
|
||||
zkey_buffer: *const Buffer,
|
||||
vk_buffer: *const Buffer,
|
||||
tree_config: *const Buffer,
|
||||
ctx: *mut *mut RLN,
|
||||
) -> bool {
|
||||
match RLN::new_with_params(
|
||||
tree_height,
|
||||
circom_buffer.process().to_vec(),
|
||||
zkey_buffer.process().to_vec(),
|
||||
vk_buffer.process().to_vec(),
|
||||
tree_config.process(),
|
||||
) {
|
||||
Ok(rln) => {
|
||||
unsafe { *ctx = Box::into_raw(Box::new(rln)) };
|
||||
true
|
||||
}
|
||||
Err(err) => {
|
||||
eprintln!("could not instantiate rln: {err}");
|
||||
false
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[cfg(feature = "stateless")]
|
||||
#[no_mangle]
|
||||
pub extern "C" fn new_with_params(
|
||||
circom_buffer: *const Buffer,
|
||||
zkey_buffer: *const Buffer,
|
||||
vk_buffer: *const Buffer,
|
||||
ctx: *mut *mut RLN,
|
||||
) -> bool {
|
||||
match RLN::new_with_params(
|
||||
circom_buffer.process().to_vec(),
|
||||
zkey_buffer.process().to_vec(),
|
||||
vk_buffer.process().to_vec(),
|
||||
) {
|
||||
Ok(rln) => {
|
||||
unsafe { *ctx = Box::into_raw(Box::new(rln)) };
|
||||
true
|
||||
}
|
||||
Err(err) => {
|
||||
eprintln!("could not instantiate rln: {err}");
|
||||
false
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
////////////////////////////////////////////////////////
|
||||
// Merkle tree APIs
|
||||
////////////////////////////////////////////////////////
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[no_mangle]
|
||||
#[cfg(not(feature = "stateless"))]
|
||||
pub extern "C" fn set_tree(ctx: *mut RLN, tree_height: usize) -> bool {
|
||||
call!(ctx, set_tree, tree_height)
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[no_mangle]
|
||||
#[cfg(not(feature = "stateless"))]
|
||||
pub extern "C" fn delete_leaf(ctx: *mut RLN, index: usize) -> bool {
|
||||
call!(ctx, delete_leaf, index)
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[no_mangle]
|
||||
#[cfg(not(feature = "stateless"))]
|
||||
pub extern "C" fn set_leaf(ctx: *mut RLN, index: usize, input_buffer: *const Buffer) -> bool {
|
||||
call!(ctx, set_leaf, index, input_buffer)
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[no_mangle]
|
||||
#[cfg(not(feature = "stateless"))]
|
||||
pub extern "C" fn get_leaf(ctx: *mut RLN, index: usize, output_buffer: *mut Buffer) -> bool {
|
||||
call_with_output_arg!(ctx, get_leaf, output_buffer, index)
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[no_mangle]
|
||||
#[cfg(not(feature = "stateless"))]
|
||||
pub extern "C" fn leaves_set(ctx: *mut RLN) -> usize {
|
||||
ctx.process().leaves_set()
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[no_mangle]
|
||||
#[cfg(not(feature = "stateless"))]
|
||||
pub extern "C" fn set_next_leaf(ctx: *mut RLN, input_buffer: *const Buffer) -> bool {
|
||||
call!(ctx, set_next_leaf, input_buffer)
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[no_mangle]
|
||||
#[cfg(not(feature = "stateless"))]
|
||||
pub extern "C" fn set_leaves_from(
|
||||
ctx: *mut RLN,
|
||||
index: usize,
|
||||
input_buffer: *const Buffer,
|
||||
) -> bool {
|
||||
call!(ctx, set_leaves_from, index, input_buffer)
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[no_mangle]
|
||||
#[cfg(not(feature = "stateless"))]
|
||||
pub extern "C" fn init_tree_with_leaves(ctx: *mut RLN, input_buffer: *const Buffer) -> bool {
|
||||
call!(ctx, init_tree_with_leaves, input_buffer)
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[no_mangle]
|
||||
#[cfg(not(feature = "stateless"))]
|
||||
pub extern "C" fn atomic_operation(
|
||||
ctx: *mut RLN,
|
||||
index: usize,
|
||||
leaves_buffer: *const Buffer,
|
||||
indices_buffer: *const Buffer,
|
||||
) -> bool {
|
||||
call!(ctx, atomic_operation, index, leaves_buffer, indices_buffer)
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[no_mangle]
|
||||
#[cfg(not(feature = "stateless"))]
|
||||
pub extern "C" fn seq_atomic_operation(
|
||||
ctx: *mut RLN,
|
||||
leaves_buffer: *const Buffer,
|
||||
indices_buffer: *const Buffer,
|
||||
) -> bool {
|
||||
call!(
|
||||
ctx,
|
||||
atomic_operation,
|
||||
ctx.process().leaves_set(),
|
||||
leaves_buffer,
|
||||
indices_buffer
|
||||
)
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[no_mangle]
|
||||
#[cfg(not(feature = "stateless"))]
|
||||
pub extern "C" fn get_root(ctx: *const RLN, output_buffer: *mut Buffer) -> bool {
|
||||
call_with_output_arg!(ctx, get_root, output_buffer)
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[no_mangle]
|
||||
#[cfg(not(feature = "stateless"))]
|
||||
pub extern "C" fn get_proof(ctx: *const RLN, index: usize, output_buffer: *mut Buffer) -> bool {
|
||||
call_with_output_arg!(ctx, get_proof, output_buffer, index)
|
||||
}
|
||||
|
||||
////////////////////////////////////////////////////////
|
||||
// zkSNARKs APIs
|
||||
////////////////////////////////////////////////////////
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[no_mangle]
|
||||
pub extern "C" fn prove(
|
||||
ctx: *mut RLN,
|
||||
input_buffer: *const Buffer,
|
||||
output_buffer: *mut Buffer,
|
||||
) -> bool {
|
||||
call_with_output_arg!(ctx, prove, output_buffer, input_buffer)
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[no_mangle]
|
||||
pub extern "C" fn verify(
|
||||
ctx: *const RLN,
|
||||
proof_buffer: *const Buffer,
|
||||
proof_is_valid_ptr: *mut bool,
|
||||
) -> bool {
|
||||
call_with_bool_arg!(ctx, verify, proof_is_valid_ptr, proof_buffer)
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[no_mangle]
|
||||
#[cfg(not(feature = "stateless"))]
|
||||
pub extern "C" fn generate_rln_proof(
|
||||
ctx: *mut RLN,
|
||||
input_buffer: *const Buffer,
|
||||
output_buffer: *mut Buffer,
|
||||
) -> bool {
|
||||
call_with_output_arg!(ctx, generate_rln_proof, output_buffer, input_buffer)
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[no_mangle]
|
||||
pub extern "C" fn generate_rln_proof_with_witness(
|
||||
ctx: *mut RLN,
|
||||
input_buffer: *const Buffer,
|
||||
output_buffer: *mut Buffer,
|
||||
) -> bool {
|
||||
call_with_output_arg!(
|
||||
ctx,
|
||||
generate_rln_proof_with_witness,
|
||||
output_buffer,
|
||||
input_buffer
|
||||
)
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[no_mangle]
|
||||
#[cfg(not(feature = "stateless"))]
|
||||
pub extern "C" fn verify_rln_proof(
|
||||
ctx: *const RLN,
|
||||
proof_buffer: *const Buffer,
|
||||
proof_is_valid_ptr: *mut bool,
|
||||
) -> bool {
|
||||
call_with_bool_arg!(ctx, verify_rln_proof, proof_is_valid_ptr, proof_buffer)
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[no_mangle]
|
||||
pub extern "C" fn verify_with_roots(
|
||||
ctx: *const RLN,
|
||||
proof_buffer: *const Buffer,
|
||||
roots_buffer: *const Buffer,
|
||||
proof_is_valid_ptr: *mut bool,
|
||||
) -> bool {
|
||||
call_with_bool_arg!(
|
||||
ctx,
|
||||
verify_with_roots,
|
||||
proof_is_valid_ptr,
|
||||
proof_buffer,
|
||||
roots_buffer
|
||||
)
|
||||
}
|
||||
|
||||
////////////////////////////////////////////////////////
|
||||
// Utils
|
||||
////////////////////////////////////////////////////////
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[no_mangle]
|
||||
pub extern "C" fn key_gen(ctx: *const RLN, output_buffer: *mut Buffer) -> bool {
|
||||
call_with_output_arg!(ctx, key_gen, output_buffer)
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[no_mangle]
|
||||
pub extern "C" fn seeded_key_gen(
|
||||
ctx: *const RLN,
|
||||
input_buffer: *const Buffer,
|
||||
output_buffer: *mut Buffer,
|
||||
) -> bool {
|
||||
call_with_output_arg!(ctx, seeded_key_gen, output_buffer, input_buffer)
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[no_mangle]
|
||||
pub extern "C" fn extended_key_gen(ctx: *const RLN, output_buffer: *mut Buffer) -> bool {
|
||||
call_with_output_arg!(ctx, extended_key_gen, output_buffer)
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[no_mangle]
|
||||
pub extern "C" fn seeded_extended_key_gen(
|
||||
ctx: *const RLN,
|
||||
input_buffer: *const Buffer,
|
||||
output_buffer: *mut Buffer,
|
||||
) -> bool {
|
||||
call_with_output_arg!(ctx, seeded_extended_key_gen, output_buffer, input_buffer)
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[no_mangle]
|
||||
pub extern "C" fn recover_id_secret(
|
||||
ctx: *const RLN,
|
||||
input_proof_buffer_1: *const Buffer,
|
||||
input_proof_buffer_2: *const Buffer,
|
||||
output_buffer: *mut Buffer,
|
||||
) -> bool {
|
||||
call_with_output_arg!(
|
||||
ctx,
|
||||
recover_id_secret,
|
||||
output_buffer,
|
||||
input_proof_buffer_1,
|
||||
input_proof_buffer_2
|
||||
)
|
||||
}
|
||||
|
||||
////////////////////////////////////////////////////////
|
||||
// Persistent metadata APIs
|
||||
////////////////////////////////////////////////////////
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[no_mangle]
|
||||
#[cfg(not(feature = "stateless"))]
|
||||
pub extern "C" fn set_metadata(ctx: *mut RLN, input_buffer: *const Buffer) -> bool {
|
||||
call!(ctx, set_metadata, input_buffer)
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[no_mangle]
|
||||
#[cfg(not(feature = "stateless"))]
|
||||
pub extern "C" fn get_metadata(ctx: *const RLN, output_buffer: *mut Buffer) -> bool {
|
||||
call_with_output_arg!(ctx, get_metadata, output_buffer)
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[no_mangle]
|
||||
#[cfg(not(feature = "stateless"))]
|
||||
pub extern "C" fn flush(ctx: *mut RLN) -> bool {
|
||||
call!(ctx, flush)
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[no_mangle]
|
||||
pub extern "C" fn hash(input_buffer: *const Buffer, output_buffer: *mut Buffer) -> bool {
|
||||
no_ctx_call_with_output_arg!(public_hash, output_buffer, input_buffer)
|
||||
}
|
||||
|
||||
#[allow(clippy::not_unsafe_ptr_arg_deref)]
|
||||
#[no_mangle]
|
||||
pub extern "C" fn poseidon_hash(input_buffer: *const Buffer, output_buffer: *mut Buffer) -> bool {
|
||||
no_ctx_call_with_output_arg!(public_poseidon_hash, output_buffer, input_buffer)
|
||||
}
|
||||
566
rln/src/ffi/ffi_rln.rs
Normal file
566
rln/src/ffi/ffi_rln.rs
Normal file
@@ -0,0 +1,566 @@
|
||||
#![allow(non_camel_case_types)]
|
||||
|
||||
use num_bigint::BigInt;
|
||||
use safer_ffi::{boxed::Box_, derive_ReprC, ffi_export, prelude::repr_c};
|
||||
#[cfg(not(feature = "stateless"))]
|
||||
use {safer_ffi::prelude::char_p, std::fs::File, std::io::Read};
|
||||
|
||||
use super::ffi_utils::{CBoolResult, CFr, CResult};
|
||||
use crate::prelude::*;
|
||||
|
||||
#[cfg(not(feature = "stateless"))]
|
||||
const MAX_CONFIG_SIZE: u64 = 1024 * 1024; // 1MB
|
||||
|
||||
// FFI_RLN
|
||||
|
||||
#[derive_ReprC]
|
||||
#[repr(opaque)]
|
||||
pub struct FFI_RLN(pub(crate) RLN);
|
||||
|
||||
// RLN initialization APIs
|
||||
|
||||
#[cfg(not(feature = "stateless"))]
|
||||
#[ffi_export]
|
||||
pub fn ffi_rln_new(
|
||||
tree_depth: usize,
|
||||
config_path: char_p::Ref<'_>,
|
||||
) -> CResult<repr_c::Box<FFI_RLN>, repr_c::String> {
|
||||
let config_str = File::open(config_path.to_str())
|
||||
.and_then(|mut file| {
|
||||
let metadata = file.metadata()?;
|
||||
if metadata.len() > MAX_CONFIG_SIZE {
|
||||
return Err(std::io::Error::new(
|
||||
std::io::ErrorKind::InvalidData,
|
||||
format!(
|
||||
"Config file too large: {} bytes (max {} bytes)",
|
||||
metadata.len(),
|
||||
MAX_CONFIG_SIZE
|
||||
),
|
||||
));
|
||||
}
|
||||
let mut s = String::new();
|
||||
file.read_to_string(&mut s)?;
|
||||
Ok(s)
|
||||
})
|
||||
.unwrap_or_default();
|
||||
|
||||
match RLN::new(tree_depth, config_str.as_str()) {
|
||||
Ok(rln) => CResult {
|
||||
ok: Some(Box_::new(FFI_RLN(rln))),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(feature = "stateless")]
|
||||
#[ffi_export]
|
||||
pub fn ffi_rln_new() -> CResult<repr_c::Box<FFI_RLN>, repr_c::String> {
|
||||
match RLN::new() {
|
||||
Ok(rln) => CResult {
|
||||
ok: Some(Box_::new(FFI_RLN(rln))),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(not(feature = "stateless"))]
|
||||
#[ffi_export]
|
||||
pub fn ffi_rln_new_with_params(
|
||||
tree_depth: usize,
|
||||
zkey_data: &repr_c::Vec<u8>,
|
||||
graph_data: &repr_c::Vec<u8>,
|
||||
config_path: char_p::Ref<'_>,
|
||||
) -> CResult<repr_c::Box<FFI_RLN>, repr_c::String> {
|
||||
let config_str = File::open(config_path.to_str())
|
||||
.and_then(|mut file| {
|
||||
let metadata = file.metadata()?;
|
||||
if metadata.len() > MAX_CONFIG_SIZE {
|
||||
return Err(std::io::Error::new(
|
||||
std::io::ErrorKind::InvalidData,
|
||||
format!(
|
||||
"Config file too large: {} bytes (max {} bytes)",
|
||||
metadata.len(),
|
||||
MAX_CONFIG_SIZE
|
||||
),
|
||||
));
|
||||
}
|
||||
let mut s = String::new();
|
||||
file.read_to_string(&mut s)?;
|
||||
Ok(s)
|
||||
})
|
||||
.unwrap_or_default();
|
||||
|
||||
match RLN::new_with_params(
|
||||
tree_depth,
|
||||
zkey_data.to_vec(),
|
||||
graph_data.to_vec(),
|
||||
config_str.as_str(),
|
||||
) {
|
||||
Ok(rln) => CResult {
|
||||
ok: Some(Box_::new(FFI_RLN(rln))),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(feature = "stateless")]
|
||||
#[ffi_export]
|
||||
pub fn ffi_rln_new_with_params(
|
||||
zkey_data: &repr_c::Vec<u8>,
|
||||
graph_data: &repr_c::Vec<u8>,
|
||||
) -> CResult<repr_c::Box<FFI_RLN>, repr_c::String> {
|
||||
match RLN::new_with_params(zkey_data.to_vec(), graph_data.to_vec()) {
|
||||
Ok(rln) => CResult {
|
||||
ok: Some(Box_::new(FFI_RLN(rln))),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_rln_free(rln: repr_c::Box<FFI_RLN>) {
|
||||
drop(rln);
|
||||
}
|
||||
|
||||
// RLNProof
|
||||
|
||||
#[derive_ReprC]
|
||||
#[repr(opaque)]
|
||||
pub struct FFI_RLNProof(pub(crate) RLNProof);
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_rln_proof_get_values(
|
||||
rln_proof: &repr_c::Box<FFI_RLNProof>,
|
||||
) -> repr_c::Box<FFI_RLNProofValues> {
|
||||
Box_::new(FFI_RLNProofValues(rln_proof.0.proof_values))
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_rln_proof_to_bytes_le(
|
||||
rln_proof: &repr_c::Box<FFI_RLNProof>,
|
||||
) -> CResult<repr_c::Vec<u8>, repr_c::String> {
|
||||
match rln_proof_to_bytes_le(&rln_proof.0) {
|
||||
Ok(bytes) => CResult {
|
||||
ok: Some(bytes.into()),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_rln_proof_to_bytes_be(
|
||||
rln_proof: &repr_c::Box<FFI_RLNProof>,
|
||||
) -> CResult<repr_c::Vec<u8>, repr_c::String> {
|
||||
match rln_proof_to_bytes_be(&rln_proof.0) {
|
||||
Ok(bytes) => CResult {
|
||||
ok: Some(bytes.into()),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_bytes_le_to_rln_proof(
|
||||
bytes: &repr_c::Vec<u8>,
|
||||
) -> CResult<repr_c::Box<FFI_RLNProof>, repr_c::String> {
|
||||
match bytes_le_to_rln_proof(bytes) {
|
||||
Ok((rln_proof, _)) => CResult {
|
||||
ok: Some(Box_::new(FFI_RLNProof(rln_proof))),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_bytes_be_to_rln_proof(
|
||||
bytes: &repr_c::Vec<u8>,
|
||||
) -> CResult<repr_c::Box<FFI_RLNProof>, repr_c::String> {
|
||||
match bytes_be_to_rln_proof(bytes) {
|
||||
Ok((rln_proof, _)) => CResult {
|
||||
ok: Some(Box_::new(FFI_RLNProof(rln_proof))),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_rln_proof_free(rln_proof: repr_c::Box<FFI_RLNProof>) {
|
||||
drop(rln_proof);
|
||||
}
|
||||
|
||||
// RLNWitnessInput
|
||||
|
||||
#[derive_ReprC]
|
||||
#[repr(opaque)]
|
||||
pub struct FFI_RLNWitnessInput(pub(crate) RLNWitnessInput);
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_rln_witness_input_new(
|
||||
identity_secret: &CFr,
|
||||
user_message_limit: &CFr,
|
||||
message_id: &CFr,
|
||||
path_elements: &repr_c::Vec<CFr>,
|
||||
identity_path_index: &repr_c::Vec<u8>,
|
||||
x: &CFr,
|
||||
external_nullifier: &CFr,
|
||||
) -> CResult<repr_c::Box<FFI_RLNWitnessInput>, repr_c::String> {
|
||||
let mut identity_secret_fr = identity_secret.0;
|
||||
let path_elements: Vec<Fr> = path_elements.iter().map(|cfr| cfr.0).collect();
|
||||
let identity_path_index: Vec<u8> = identity_path_index.iter().copied().collect();
|
||||
match RLNWitnessInput::new(
|
||||
IdSecret::from(&mut identity_secret_fr),
|
||||
user_message_limit.0,
|
||||
message_id.0,
|
||||
path_elements,
|
||||
identity_path_index,
|
||||
x.0,
|
||||
external_nullifier.0,
|
||||
) {
|
||||
Ok(witness) => CResult {
|
||||
ok: Some(Box_::new(FFI_RLNWitnessInput(witness))),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_rln_witness_to_bytes_le(
|
||||
witness: &repr_c::Box<FFI_RLNWitnessInput>,
|
||||
) -> CResult<repr_c::Vec<u8>, repr_c::String> {
|
||||
match rln_witness_to_bytes_le(&witness.0) {
|
||||
Ok(bytes) => CResult {
|
||||
ok: Some(bytes.into()),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_rln_witness_to_bytes_be(
|
||||
witness: &repr_c::Box<FFI_RLNWitnessInput>,
|
||||
) -> CResult<repr_c::Vec<u8>, repr_c::String> {
|
||||
match rln_witness_to_bytes_be(&witness.0) {
|
||||
Ok(bytes) => CResult {
|
||||
ok: Some(bytes.into()),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_bytes_le_to_rln_witness(
|
||||
bytes: &repr_c::Vec<u8>,
|
||||
) -> CResult<repr_c::Box<FFI_RLNWitnessInput>, repr_c::String> {
|
||||
match bytes_le_to_rln_witness(bytes) {
|
||||
Ok((witness, _)) => CResult {
|
||||
ok: Some(Box_::new(FFI_RLNWitnessInput(witness))),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_bytes_be_to_rln_witness(
|
||||
bytes: &repr_c::Vec<u8>,
|
||||
) -> CResult<repr_c::Box<FFI_RLNWitnessInput>, repr_c::String> {
|
||||
match bytes_be_to_rln_witness(bytes) {
|
||||
Ok((witness, _)) => CResult {
|
||||
ok: Some(Box_::new(FFI_RLNWitnessInput(witness))),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_rln_witness_to_bigint_json(
|
||||
witness: &repr_c::Box<FFI_RLNWitnessInput>,
|
||||
) -> CResult<repr_c::String, repr_c::String> {
|
||||
match rln_witness_to_bigint_json(&witness.0) {
|
||||
Ok(json) => CResult {
|
||||
ok: Some(json.to_string().into()),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_rln_witness_input_free(witness: repr_c::Box<FFI_RLNWitnessInput>) {
|
||||
drop(witness);
|
||||
}
|
||||
|
||||
// RLNProofValues
|
||||
|
||||
#[derive_ReprC]
|
||||
#[repr(opaque)]
|
||||
pub struct FFI_RLNProofValues(pub(crate) RLNProofValues);
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_rln_proof_values_get_y(pv: &repr_c::Box<FFI_RLNProofValues>) -> repr_c::Box<CFr> {
|
||||
CFr::from(pv.0.y).into()
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_rln_proof_values_get_nullifier(
|
||||
pv: &repr_c::Box<FFI_RLNProofValues>,
|
||||
) -> repr_c::Box<CFr> {
|
||||
CFr::from(pv.0.nullifier).into()
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_rln_proof_values_get_root(pv: &repr_c::Box<FFI_RLNProofValues>) -> repr_c::Box<CFr> {
|
||||
CFr::from(pv.0.root).into()
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_rln_proof_values_get_x(pv: &repr_c::Box<FFI_RLNProofValues>) -> repr_c::Box<CFr> {
|
||||
CFr::from(pv.0.x).into()
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_rln_proof_values_get_external_nullifier(
|
||||
pv: &repr_c::Box<FFI_RLNProofValues>,
|
||||
) -> repr_c::Box<CFr> {
|
||||
CFr::from(pv.0.external_nullifier).into()
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_rln_proof_values_to_bytes_le(pv: &repr_c::Box<FFI_RLNProofValues>) -> repr_c::Vec<u8> {
|
||||
rln_proof_values_to_bytes_le(&pv.0).into()
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_rln_proof_values_to_bytes_be(pv: &repr_c::Box<FFI_RLNProofValues>) -> repr_c::Vec<u8> {
|
||||
rln_proof_values_to_bytes_be(&pv.0).into()
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_bytes_le_to_rln_proof_values(
|
||||
bytes: &repr_c::Vec<u8>,
|
||||
) -> CResult<repr_c::Box<FFI_RLNProofValues>, repr_c::String> {
|
||||
match bytes_le_to_rln_proof_values(bytes) {
|
||||
Ok((pv, _)) => CResult {
|
||||
ok: Some(Box_::new(FFI_RLNProofValues(pv))),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(format!("{:?}", err).into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_bytes_be_to_rln_proof_values(
|
||||
bytes: &repr_c::Vec<u8>,
|
||||
) -> CResult<repr_c::Box<FFI_RLNProofValues>, repr_c::String> {
|
||||
match bytes_be_to_rln_proof_values(bytes) {
|
||||
Ok((pv, _)) => CResult {
|
||||
ok: Some(Box_::new(FFI_RLNProofValues(pv))),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(format!("{:?}", err).into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_rln_proof_values_free(proof_values: repr_c::Box<FFI_RLNProofValues>) {
|
||||
drop(proof_values);
|
||||
}
|
||||
|
||||
// Proof generation APIs
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_generate_rln_proof(
|
||||
rln: &repr_c::Box<FFI_RLN>,
|
||||
witness: &repr_c::Box<FFI_RLNWitnessInput>,
|
||||
) -> CResult<repr_c::Box<FFI_RLNProof>, repr_c::String> {
|
||||
match rln.0.generate_rln_proof(&witness.0) {
|
||||
Ok((proof, proof_values)) => {
|
||||
let rln_proof = RLNProof {
|
||||
proof_values,
|
||||
proof,
|
||||
};
|
||||
CResult {
|
||||
ok: Some(Box_::new(FFI_RLNProof(rln_proof))),
|
||||
err: None,
|
||||
}
|
||||
}
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_generate_rln_proof_with_witness(
|
||||
rln: &repr_c::Box<FFI_RLN>,
|
||||
calculated_witness: &repr_c::Vec<repr_c::String>,
|
||||
witness: &repr_c::Box<FFI_RLNWitnessInput>,
|
||||
) -> CResult<repr_c::Box<FFI_RLNProof>, repr_c::String> {
|
||||
let calculated_witness_bigint: Result<Vec<BigInt>, _> = calculated_witness
|
||||
.iter()
|
||||
.map(|s| {
|
||||
let s_str = unsafe { std::str::from_utf8_unchecked(s.as_bytes()) };
|
||||
s_str.parse::<BigInt>()
|
||||
})
|
||||
.collect();
|
||||
|
||||
let calculated_witness_bigint = match calculated_witness_bigint {
|
||||
Ok(w) => w,
|
||||
Err(err) => {
|
||||
return CResult {
|
||||
ok: None,
|
||||
err: Some(format!("Failed to parse witness: {}", err).into()),
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
match rln
|
||||
.0
|
||||
.generate_rln_proof_with_witness(calculated_witness_bigint, &witness.0)
|
||||
{
|
||||
Ok((proof, proof_values)) => {
|
||||
let rln_proof = RLNProof {
|
||||
proof_values,
|
||||
proof,
|
||||
};
|
||||
CResult {
|
||||
ok: Some(Box_::new(FFI_RLNProof(rln_proof))),
|
||||
err: None,
|
||||
}
|
||||
}
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
// Proof verification APIs
|
||||
|
||||
#[cfg(not(feature = "stateless"))]
|
||||
#[ffi_export]
|
||||
pub fn ffi_verify_rln_proof(
|
||||
rln: &repr_c::Box<FFI_RLN>,
|
||||
rln_proof: &repr_c::Box<FFI_RLNProof>,
|
||||
x: &CFr,
|
||||
) -> CBoolResult {
|
||||
match rln
|
||||
.0
|
||||
.verify_rln_proof(&rln_proof.0.proof, &rln_proof.0.proof_values, &x.0)
|
||||
{
|
||||
Ok(verified) => CBoolResult {
|
||||
ok: verified,
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CBoolResult {
|
||||
ok: false,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_verify_with_roots(
|
||||
rln: &repr_c::Box<FFI_RLN>,
|
||||
rln_proof: &repr_c::Box<FFI_RLNProof>,
|
||||
roots: &repr_c::Vec<CFr>,
|
||||
x: &CFr,
|
||||
) -> CBoolResult {
|
||||
let roots_fr: Vec<Fr> = roots.iter().map(|cfr| cfr.0).collect();
|
||||
|
||||
match rln.0.verify_with_roots(
|
||||
&rln_proof.0.proof,
|
||||
&rln_proof.0.proof_values,
|
||||
&x.0,
|
||||
&roots_fr,
|
||||
) {
|
||||
Ok(verified) => CBoolResult {
|
||||
ok: verified,
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CBoolResult {
|
||||
ok: false,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
// Identity secret recovery API
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_recover_id_secret(
|
||||
proof_values_1: &repr_c::Box<FFI_RLNProofValues>,
|
||||
proof_values_2: &repr_c::Box<FFI_RLNProofValues>,
|
||||
) -> CResult<repr_c::Box<CFr>, repr_c::String> {
|
||||
match recover_id_secret(&proof_values_1.0, &proof_values_2.0) {
|
||||
Ok(secret) => CResult {
|
||||
ok: Some(Box_::new(CFr::from(*secret))),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
269
rln/src/ffi/ffi_tree.rs
Normal file
269
rln/src/ffi/ffi_tree.rs
Normal file
@@ -0,0 +1,269 @@
|
||||
#![allow(non_camel_case_types)]
|
||||
#![cfg(not(feature = "stateless"))]
|
||||
|
||||
use safer_ffi::{boxed::Box_, derive_ReprC, ffi_export, prelude::repr_c};
|
||||
|
||||
use super::{
|
||||
ffi_rln::FFI_RLN,
|
||||
ffi_utils::{CBoolResult, CFr, CResult},
|
||||
};
|
||||
|
||||
// MerkleProof
|
||||
|
||||
#[derive_ReprC]
|
||||
#[repr(C)]
|
||||
pub struct FFI_MerkleProof {
|
||||
pub path_elements: repr_c::Vec<CFr>,
|
||||
pub path_index: repr_c::Vec<u8>,
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_merkle_proof_free(merkle_proof: repr_c::Box<FFI_MerkleProof>) {
|
||||
drop(merkle_proof);
|
||||
}
|
||||
|
||||
// Merkle tree management APIs
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_set_tree(rln: &mut repr_c::Box<FFI_RLN>, tree_depth: usize) -> CBoolResult {
|
||||
match rln.0.set_tree(tree_depth) {
|
||||
Ok(_) => CBoolResult {
|
||||
ok: true,
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CBoolResult {
|
||||
ok: false,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
// Merkle tree leaf operations
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_delete_leaf(rln: &mut repr_c::Box<FFI_RLN>, index: usize) -> CBoolResult {
|
||||
match rln.0.delete_leaf(index) {
|
||||
Ok(_) => CBoolResult {
|
||||
ok: true,
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CBoolResult {
|
||||
ok: false,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_set_leaf(rln: &mut repr_c::Box<FFI_RLN>, index: usize, leaf: &CFr) -> CBoolResult {
|
||||
match rln.0.set_leaf(index, leaf.0) {
|
||||
Ok(_) => CBoolResult {
|
||||
ok: true,
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CBoolResult {
|
||||
ok: false,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_get_leaf(
|
||||
rln: &repr_c::Box<FFI_RLN>,
|
||||
index: usize,
|
||||
) -> CResult<repr_c::Box<CFr>, repr_c::String> {
|
||||
match rln.0.get_leaf(index) {
|
||||
Ok(leaf) => CResult {
|
||||
ok: Some(CFr::from(leaf).into()),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_leaves_set(rln: &repr_c::Box<FFI_RLN>) -> usize {
|
||||
rln.0.leaves_set()
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_set_next_leaf(rln: &mut repr_c::Box<FFI_RLN>, leaf: &CFr) -> CBoolResult {
|
||||
match rln.0.set_next_leaf(leaf.0) {
|
||||
Ok(_) => CBoolResult {
|
||||
ok: true,
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CBoolResult {
|
||||
ok: false,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_set_leaves_from(
|
||||
rln: &mut repr_c::Box<FFI_RLN>,
|
||||
index: usize,
|
||||
leaves: &repr_c::Vec<CFr>,
|
||||
) -> CBoolResult {
|
||||
let leaves_vec: Vec<_> = leaves.iter().map(|cfr| cfr.0).collect();
|
||||
match rln.0.set_leaves_from(index, leaves_vec) {
|
||||
Ok(_) => CBoolResult {
|
||||
ok: true,
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CBoolResult {
|
||||
ok: false,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_init_tree_with_leaves(
|
||||
rln: &mut repr_c::Box<FFI_RLN>,
|
||||
leaves: &repr_c::Vec<CFr>,
|
||||
) -> CBoolResult {
|
||||
let leaves_vec: Vec<_> = leaves.iter().map(|cfr| cfr.0).collect();
|
||||
match rln.0.init_tree_with_leaves(leaves_vec) {
|
||||
Ok(_) => CBoolResult {
|
||||
ok: true,
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CBoolResult {
|
||||
ok: false,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
// Atomic operations
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_atomic_operation(
|
||||
rln: &mut repr_c::Box<FFI_RLN>,
|
||||
index: usize,
|
||||
leaves: &repr_c::Vec<CFr>,
|
||||
indices: &repr_c::Vec<usize>,
|
||||
) -> CBoolResult {
|
||||
let leaves_vec: Vec<_> = leaves.iter().map(|cfr| cfr.0).collect();
|
||||
let indices_vec: Vec<_> = indices.iter().copied().collect();
|
||||
match rln.0.atomic_operation(index, leaves_vec, indices_vec) {
|
||||
Ok(_) => CBoolResult {
|
||||
ok: true,
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CBoolResult {
|
||||
ok: false,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_seq_atomic_operation(
|
||||
rln: &mut repr_c::Box<FFI_RLN>,
|
||||
leaves: &repr_c::Vec<CFr>,
|
||||
indices: &repr_c::Vec<u8>,
|
||||
) -> CBoolResult {
|
||||
let index = rln.0.leaves_set();
|
||||
let leaves_vec: Vec<_> = leaves.iter().map(|cfr| cfr.0).collect();
|
||||
let indices_vec: Vec<_> = indices.iter().map(|x| *x as usize).collect();
|
||||
match rln.0.atomic_operation(index, leaves_vec, indices_vec) {
|
||||
Ok(_) => CBoolResult {
|
||||
ok: true,
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CBoolResult {
|
||||
ok: false,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
// Root and proof operations
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_get_root(rln: &repr_c::Box<FFI_RLN>) -> repr_c::Box<CFr> {
|
||||
CFr::from(rln.0.get_root()).into()
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_get_merkle_proof(
|
||||
rln: &repr_c::Box<FFI_RLN>,
|
||||
index: usize,
|
||||
) -> CResult<repr_c::Box<FFI_MerkleProof>, repr_c::String> {
|
||||
match rln.0.get_merkle_proof(index) {
|
||||
Ok((path_elements, path_index)) => {
|
||||
let path_elements: repr_c::Vec<CFr> = path_elements
|
||||
.iter()
|
||||
.map(|fr| CFr::from(*fr))
|
||||
.collect::<Vec<_>>()
|
||||
.into();
|
||||
|
||||
let path_index: repr_c::Vec<u8> = path_index.into();
|
||||
|
||||
let merkle_proof = FFI_MerkleProof {
|
||||
path_elements,
|
||||
path_index,
|
||||
};
|
||||
|
||||
CResult {
|
||||
ok: Some(Box_::new(merkle_proof)),
|
||||
err: None,
|
||||
}
|
||||
}
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
// Persistent metadata APIs
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_set_metadata(rln: &mut repr_c::Box<FFI_RLN>, metadata: &repr_c::Vec<u8>) -> CBoolResult {
|
||||
match rln.0.set_metadata(metadata) {
|
||||
Ok(_) => CBoolResult {
|
||||
ok: true,
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CBoolResult {
|
||||
ok: false,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_get_metadata(rln: &repr_c::Box<FFI_RLN>) -> CResult<repr_c::Vec<u8>, repr_c::String> {
|
||||
match rln.0.get_metadata() {
|
||||
Ok(metadata) => CResult {
|
||||
ok: Some(metadata.into()),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_flush(rln: &mut repr_c::Box<FFI_RLN>) -> CBoolResult {
|
||||
match rln.0.flush() {
|
||||
Ok(_) => CBoolResult {
|
||||
ok: true,
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CBoolResult {
|
||||
ok: false,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
407
rln/src/ffi/ffi_utils.rs
Normal file
407
rln/src/ffi/ffi_utils.rs
Normal file
@@ -0,0 +1,407 @@
|
||||
#![allow(non_camel_case_types)]
|
||||
|
||||
use std::ops::Deref;
|
||||
|
||||
use safer_ffi::{
|
||||
boxed::Box_,
|
||||
derive_ReprC, ffi_export,
|
||||
prelude::{repr_c, ReprC},
|
||||
};
|
||||
|
||||
use crate::prelude::*;
|
||||
|
||||
// CResult
|
||||
|
||||
#[derive_ReprC]
|
||||
#[repr(C)]
|
||||
pub struct CResult<T: ReprC, Err: ReprC> {
|
||||
pub ok: Option<T>,
|
||||
pub err: Option<Err>,
|
||||
}
|
||||
|
||||
// CBoolResult
|
||||
|
||||
#[derive_ReprC]
|
||||
#[repr(C)]
|
||||
pub struct CBoolResult {
|
||||
pub ok: bool,
|
||||
pub err: Option<repr_c::String>,
|
||||
}
|
||||
|
||||
// CFr
|
||||
|
||||
#[derive_ReprC]
|
||||
#[repr(opaque)]
|
||||
#[derive(Debug, Clone, Copy, PartialEq)]
|
||||
pub struct CFr(pub(crate) Fr);
|
||||
|
||||
impl Deref for CFr {
|
||||
type Target = Fr;
|
||||
fn deref(&self) -> &Self::Target {
|
||||
&self.0
|
||||
}
|
||||
}
|
||||
|
||||
impl From<Fr> for CFr {
|
||||
fn from(fr: Fr) -> Self {
|
||||
Self(fr)
|
||||
}
|
||||
}
|
||||
|
||||
impl From<CFr> for repr_c::Box<CFr> {
|
||||
fn from(cfr: CFr) -> Self {
|
||||
Box_::new(cfr)
|
||||
}
|
||||
}
|
||||
|
||||
impl From<&CFr> for repr_c::Box<CFr> {
|
||||
fn from(cfr: &CFr) -> Self {
|
||||
CFr(cfr.0).into()
|
||||
}
|
||||
}
|
||||
|
||||
impl PartialEq<Fr> for CFr {
|
||||
fn eq(&self, other: &Fr) -> bool {
|
||||
self.0 == *other
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_cfr_zero() -> repr_c::Box<CFr> {
|
||||
CFr::from(Fr::from(0)).into()
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_cfr_one() -> repr_c::Box<CFr> {
|
||||
CFr::from(Fr::from(1)).into()
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_cfr_to_bytes_le(cfr: &CFr) -> repr_c::Vec<u8> {
|
||||
fr_to_bytes_le(&cfr.0).into()
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_cfr_to_bytes_be(cfr: &CFr) -> repr_c::Vec<u8> {
|
||||
fr_to_bytes_be(&cfr.0).into()
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_bytes_le_to_cfr(bytes: &repr_c::Vec<u8>) -> CResult<repr_c::Box<CFr>, repr_c::String> {
|
||||
match bytes_le_to_fr(bytes) {
|
||||
Ok((cfr, _)) => CResult {
|
||||
ok: Some(CFr(cfr).into()),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(format!("{:?}", err).into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_bytes_be_to_cfr(bytes: &repr_c::Vec<u8>) -> CResult<repr_c::Box<CFr>, repr_c::String> {
|
||||
match bytes_be_to_fr(bytes) {
|
||||
Ok((cfr, _)) => CResult {
|
||||
ok: Some(CFr(cfr).into()),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(format!("{:?}", err).into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_uint_to_cfr(value: u32) -> repr_c::Box<CFr> {
|
||||
CFr::from(Fr::from(value)).into()
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_cfr_debug(cfr: Option<&CFr>) -> repr_c::String {
|
||||
match cfr {
|
||||
Some(cfr) => format!("{:?}", cfr.0).into(),
|
||||
None => "None".into(),
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_cfr_free(cfr: repr_c::Box<CFr>) {
|
||||
drop(cfr);
|
||||
}
|
||||
|
||||
// Vec<CFr>
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_vec_cfr_new(capacity: usize) -> repr_c::Vec<CFr> {
|
||||
Vec::with_capacity(capacity).into()
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_vec_cfr_from_cfr(cfr: &CFr) -> repr_c::Vec<CFr> {
|
||||
vec![*cfr].into()
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_vec_cfr_push(v: &mut safer_ffi::Vec<CFr>, cfr: &CFr) {
|
||||
let mut new: Vec<CFr> = std::mem::replace(v, Vec::new().into()).into();
|
||||
if new.len() == new.capacity() {
|
||||
new.reserve_exact(1);
|
||||
}
|
||||
new.push(*cfr);
|
||||
*v = new.into();
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_vec_cfr_len(v: &repr_c::Vec<CFr>) -> usize {
|
||||
v.len()
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_vec_cfr_get(v: &repr_c::Vec<CFr>, i: usize) -> Option<&CFr> {
|
||||
v.get(i)
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_vec_cfr_to_bytes_le(vec: &repr_c::Vec<CFr>) -> repr_c::Vec<u8> {
|
||||
let vec_fr: Vec<Fr> = vec.iter().map(|cfr| cfr.0).collect();
|
||||
vec_fr_to_bytes_le(&vec_fr).into()
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_vec_cfr_to_bytes_be(vec: &repr_c::Vec<CFr>) -> repr_c::Vec<u8> {
|
||||
let vec_fr: Vec<Fr> = vec.iter().map(|cfr| cfr.0).collect();
|
||||
vec_fr_to_bytes_be(&vec_fr).into()
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_bytes_le_to_vec_cfr(
|
||||
bytes: &repr_c::Vec<u8>,
|
||||
) -> CResult<repr_c::Vec<CFr>, repr_c::String> {
|
||||
match bytes_le_to_vec_fr(bytes) {
|
||||
Ok((vec_fr, _)) => {
|
||||
let vec_cfr: Vec<CFr> = vec_fr.into_iter().map(CFr).collect();
|
||||
CResult {
|
||||
ok: Some(vec_cfr.into()),
|
||||
err: None,
|
||||
}
|
||||
}
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_bytes_be_to_vec_cfr(
|
||||
bytes: &repr_c::Vec<u8>,
|
||||
) -> CResult<repr_c::Vec<CFr>, repr_c::String> {
|
||||
match bytes_be_to_vec_fr(bytes) {
|
||||
Ok((vec_fr, _)) => {
|
||||
let vec_cfr: Vec<CFr> = vec_fr.into_iter().map(CFr).collect();
|
||||
CResult {
|
||||
ok: Some(vec_cfr.into()),
|
||||
err: None,
|
||||
}
|
||||
}
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_vec_cfr_debug(v: Option<&repr_c::Vec<CFr>>) -> repr_c::String {
|
||||
match v {
|
||||
Some(v) => {
|
||||
let vec_fr: Vec<Fr> = v.iter().map(|cfr| cfr.0).collect();
|
||||
format!("{:?}", vec_fr).into()
|
||||
}
|
||||
None => "None".into(),
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_vec_cfr_free(v: repr_c::Vec<CFr>) {
|
||||
drop(v);
|
||||
}
|
||||
|
||||
// Vec<u8>
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_vec_u8_to_bytes_le(vec: &repr_c::Vec<u8>) -> repr_c::Vec<u8> {
|
||||
vec_u8_to_bytes_le(vec).into()
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_vec_u8_to_bytes_be(vec: &repr_c::Vec<u8>) -> repr_c::Vec<u8> {
|
||||
vec_u8_to_bytes_be(vec).into()
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_bytes_le_to_vec_u8(bytes: &repr_c::Vec<u8>) -> CResult<repr_c::Vec<u8>, repr_c::String> {
|
||||
match bytes_le_to_vec_u8(bytes) {
|
||||
Ok((vec, _)) => CResult {
|
||||
ok: Some(vec.into()),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_bytes_be_to_vec_u8(bytes: &repr_c::Vec<u8>) -> CResult<repr_c::Vec<u8>, repr_c::String> {
|
||||
match bytes_be_to_vec_u8(bytes) {
|
||||
Ok((vec, _)) => CResult {
|
||||
ok: Some(vec.into()),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(err.to_string().into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_vec_u8_debug(v: Option<&repr_c::Vec<u8>>) -> repr_c::String {
|
||||
match v {
|
||||
Some(v) => format!("{:x?}", v.deref()).into(),
|
||||
None => "None".into(),
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_vec_u8_free(v: repr_c::Vec<u8>) {
|
||||
drop(v);
|
||||
}
|
||||
|
||||
// Utility APIs
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_hash_to_field_le(input: &repr_c::Vec<u8>) -> CResult<repr_c::Box<CFr>, repr_c::String> {
|
||||
match hash_to_field_le(input) {
|
||||
Ok(hash_result) => CResult {
|
||||
ok: Some(CFr::from(hash_result).into()),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(format!("{:?}", err).into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_hash_to_field_be(input: &repr_c::Vec<u8>) -> CResult<repr_c::Box<CFr>, repr_c::String> {
|
||||
match hash_to_field_be(input) {
|
||||
Ok(hash_result) => CResult {
|
||||
ok: Some(CFr::from(hash_result).into()),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(format!("{:?}", err).into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_poseidon_hash_pair(a: &CFr, b: &CFr) -> CResult<repr_c::Box<CFr>, repr_c::String> {
|
||||
match poseidon_hash(&[a.0, b.0]) {
|
||||
Ok(hash_result) => CResult {
|
||||
ok: Some(CFr::from(hash_result).into()),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(format!("{:?}", err).into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_key_gen() -> CResult<repr_c::Vec<CFr>, repr_c::String> {
|
||||
match keygen() {
|
||||
Ok((identity_secret, id_commitment)) => CResult {
|
||||
ok: Some(vec![CFr(*identity_secret), CFr(id_commitment)].into()),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(format!("{:?}", err).into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_seeded_key_gen(seed: &repr_c::Vec<u8>) -> CResult<repr_c::Vec<CFr>, repr_c::String> {
|
||||
match seeded_keygen(seed) {
|
||||
Ok((identity_secret, id_commitment)) => CResult {
|
||||
ok: Some(vec![CFr(identity_secret), CFr(id_commitment)].into()),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(format!("{:?}", err).into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_extended_key_gen() -> CResult<repr_c::Vec<CFr>, repr_c::String> {
|
||||
match extended_keygen() {
|
||||
Ok((identity_trapdoor, identity_nullifier, identity_secret, id_commitment)) => CResult {
|
||||
ok: Some(
|
||||
vec![
|
||||
CFr(identity_trapdoor),
|
||||
CFr(identity_nullifier),
|
||||
CFr(identity_secret),
|
||||
CFr(id_commitment),
|
||||
]
|
||||
.into(),
|
||||
),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(format!("{:?}", err).into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_seeded_extended_key_gen(
|
||||
seed: &repr_c::Vec<u8>,
|
||||
) -> CResult<repr_c::Vec<CFr>, repr_c::String> {
|
||||
match extended_seeded_keygen(seed) {
|
||||
Ok((identity_trapdoor, identity_nullifier, identity_secret, id_commitment)) => CResult {
|
||||
ok: Some(
|
||||
vec![
|
||||
CFr(identity_trapdoor),
|
||||
CFr(identity_nullifier),
|
||||
CFr(identity_secret),
|
||||
CFr(id_commitment),
|
||||
]
|
||||
.into(),
|
||||
),
|
||||
err: None,
|
||||
},
|
||||
Err(err) => CResult {
|
||||
ok: None,
|
||||
err: Some(format!("{:?}", err).into()),
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
#[ffi_export]
|
||||
pub fn ffi_c_string_free(s: repr_c::String) {
|
||||
drop(s);
|
||||
}
|
||||
10
rln/src/ffi/mod.rs
Normal file
10
rln/src/ffi/mod.rs
Normal file
@@ -0,0 +1,10 @@
|
||||
#![cfg(not(target_arch = "wasm32"))]
|
||||
|
||||
pub mod ffi_rln;
|
||||
pub mod ffi_tree;
|
||||
pub mod ffi_utils;
|
||||
|
||||
#[cfg(feature = "headers")]
|
||||
pub fn generate_headers() -> std::io::Result<()> {
|
||||
safer_ffi::headers::builder().to_file("rln.h")?.generate()
|
||||
}
|
||||
@@ -1,13 +1,19 @@
|
||||
/// This crate instantiates the Poseidon hash algorithm.
|
||||
use crate::{circuit::Fr, utils::bytes_le_to_fr};
|
||||
// This crate instantiates the Poseidon hash algorithm.
|
||||
|
||||
use once_cell::sync::Lazy;
|
||||
use tiny_keccak::{Hasher, Keccak};
|
||||
use utils::poseidon::Poseidon;
|
||||
use zerokit_utils::{error::HashError, poseidon::Poseidon};
|
||||
|
||||
use crate::{
|
||||
circuit::Fr,
|
||||
error::UtilsError,
|
||||
utils::{bytes_be_to_fr, bytes_le_to_fr},
|
||||
};
|
||||
|
||||
/// These indexed constants hardcode the supported round parameters tuples (t, RF, RN, SKIP_MATRICES) for the Bn254 scalar field.
|
||||
/// SKIP_MATRICES is the index of the randomly generated secure MDS matrix.
|
||||
/// TODO: generate these parameters
|
||||
pub const ROUND_PARAMS: [(usize, usize, usize, usize); 8] = [
|
||||
const ROUND_PARAMS: [(usize, usize, usize, usize); 8] = [
|
||||
(2, 8, 56, 0),
|
||||
(3, 8, 57, 0),
|
||||
(4, 8, 56, 0),
|
||||
@@ -21,10 +27,9 @@ pub const ROUND_PARAMS: [(usize, usize, usize, usize); 8] = [
|
||||
/// Poseidon Hash wrapper over above implementation.
|
||||
static POSEIDON: Lazy<Poseidon<Fr>> = Lazy::new(|| Poseidon::<Fr>::from(&ROUND_PARAMS));
|
||||
|
||||
pub fn poseidon_hash(input: &[Fr]) -> Fr {
|
||||
POSEIDON
|
||||
.hash(input.to_vec())
|
||||
.expect("hash with fixed input size can't fail")
|
||||
pub fn poseidon_hash(input: &[Fr]) -> Result<Fr, HashError> {
|
||||
let hash = POSEIDON.hash(input)?;
|
||||
Ok(hash)
|
||||
}
|
||||
|
||||
/// The zerokit RLN Merkle tree Hasher.
|
||||
@@ -32,20 +37,21 @@ pub fn poseidon_hash(input: &[Fr]) -> Fr {
|
||||
pub struct PoseidonHash;
|
||||
|
||||
/// The default Hasher trait used by Merkle tree implementation in utils.
|
||||
impl utils::merkle_tree::Hasher for PoseidonHash {
|
||||
impl zerokit_utils::merkle_tree::Hasher for PoseidonHash {
|
||||
type Fr = Fr;
|
||||
type Error = HashError;
|
||||
|
||||
fn default_leaf() -> Self::Fr {
|
||||
Self::Fr::from(0)
|
||||
}
|
||||
|
||||
fn hash(inputs: &[Self::Fr]) -> Self::Fr {
|
||||
fn hash(inputs: &[Self::Fr]) -> Result<Self::Fr, Self::Error> {
|
||||
poseidon_hash(inputs)
|
||||
}
|
||||
}
|
||||
|
||||
/// Hashes arbitrary signal to the underlying prime field.
|
||||
pub fn hash_to_field(signal: &[u8]) -> Fr {
|
||||
pub fn hash_to_field_le(signal: &[u8]) -> Result<Fr, UtilsError> {
|
||||
// We hash the input signal using Keccak256
|
||||
let mut hash = [0; 32];
|
||||
let mut hasher = Keccak::v256();
|
||||
@@ -53,6 +59,24 @@ pub fn hash_to_field(signal: &[u8]) -> Fr {
|
||||
hasher.finalize(&mut hash);
|
||||
|
||||
// We export the hash as a field element
|
||||
let (el, _) = bytes_le_to_fr(hash.as_ref());
|
||||
el
|
||||
let (el, _) = bytes_le_to_fr(hash.as_ref())?;
|
||||
|
||||
Ok(el)
|
||||
}
|
||||
|
||||
/// Hashes arbitrary signal to the underlying prime field.
|
||||
pub fn hash_to_field_be(signal: &[u8]) -> Result<Fr, UtilsError> {
|
||||
// We hash the input signal using Keccak256
|
||||
let mut hash = [0; 32];
|
||||
let mut hasher = Keccak::v256();
|
||||
hasher.update(signal);
|
||||
hasher.finalize(&mut hash);
|
||||
|
||||
// Reverse the bytes to get big endian representation
|
||||
hash.reverse();
|
||||
|
||||
// We export the hash as a field element
|
||||
let (el, _) = bytes_be_to_fr(hash.as_ref())?;
|
||||
|
||||
Ok(el)
|
||||
}
|
||||
|
||||
@@ -1,15 +1,31 @@
|
||||
#![allow(dead_code)]
|
||||
|
||||
pub mod circuit;
|
||||
pub mod error;
|
||||
pub mod ffi;
|
||||
pub mod hashers;
|
||||
#[cfg(feature = "pmtree-ft")]
|
||||
pub mod pm_tree_adapter;
|
||||
pub mod poseidon_tree;
|
||||
pub mod prelude;
|
||||
pub mod protocol;
|
||||
pub mod public;
|
||||
#[cfg(test)]
|
||||
pub mod public_api_tests;
|
||||
pub mod utils;
|
||||
|
||||
#[cfg(not(target_arch = "wasm32"))]
|
||||
pub mod ffi;
|
||||
// Ensure that only one Merkle tree feature is enabled at a time
|
||||
#[cfg(any(
|
||||
all(feature = "fullmerkletree", feature = "optimalmerkletree"),
|
||||
all(feature = "fullmerkletree", feature = "pmtree-ft"),
|
||||
all(feature = "optimalmerkletree", feature = "pmtree-ft"),
|
||||
))]
|
||||
compile_error!(
|
||||
"Only one of `fullmerkletree`, `optimalmerkletree`, or `pmtree-ft` can be enabled at a time."
|
||||
);
|
||||
|
||||
// Ensure that the `stateless` feature is not enabled with any Merkle tree features
|
||||
#[cfg(all(
|
||||
feature = "stateless",
|
||||
any(
|
||||
feature = "fullmerkletree",
|
||||
feature = "optimalmerkletree",
|
||||
feature = "pmtree-ft"
|
||||
)
|
||||
))]
|
||||
compile_error!("Cannot enable any Merkle tree features with stateless");
|
||||
|
||||
@@ -1,17 +1,24 @@
|
||||
use std::fmt::Debug;
|
||||
use std::path::PathBuf;
|
||||
use std::str::FromStr;
|
||||
#![cfg(feature = "pmtree-ft")]
|
||||
|
||||
use std::{fmt::Debug, path::PathBuf, str::FromStr};
|
||||
|
||||
use color_eyre::{Report, Result};
|
||||
use serde_json::Value;
|
||||
use tempfile::Builder;
|
||||
use zerokit_utils::{
|
||||
error::{FromConfigError, ZerokitMerkleTreeError},
|
||||
merkle_tree::{ZerokitMerkleProof, ZerokitMerkleTree},
|
||||
pm_tree::{
|
||||
pmtree,
|
||||
pmtree::{tree::Key, Database, Hasher, PmtreeErrorKind},
|
||||
Config, Mode, SledDB,
|
||||
},
|
||||
};
|
||||
|
||||
use utils::pmtree::tree::Key;
|
||||
use utils::pmtree::{Database, Hasher};
|
||||
use utils::*;
|
||||
|
||||
use crate::circuit::Fr;
|
||||
use crate::hashers::{poseidon_hash, PoseidonHash};
|
||||
use crate::utils::{bytes_le_to_fr, fr_to_bytes_le};
|
||||
use crate::{
|
||||
circuit::Fr,
|
||||
hashers::{poseidon_hash, PoseidonHash},
|
||||
utils::{bytes_le_to_fr, fr_to_bytes_le},
|
||||
};
|
||||
|
||||
const METADATA_KEY: [u8; 8] = *b"metadata";
|
||||
|
||||
@@ -39,7 +46,8 @@ impl Hasher for PoseidonHash {
|
||||
}
|
||||
|
||||
fn deserialize(value: pmtree::Value) -> Self::Fr {
|
||||
let (fr, _) = bytes_le_to_fr(&value);
|
||||
// TODO: allow to handle error properly in pmtree Hasher trait
|
||||
let (fr, _) = bytes_le_to_fr(&value).expect("Fr deserialization must be valid");
|
||||
fr
|
||||
}
|
||||
|
||||
@@ -48,24 +56,114 @@ impl Hasher for PoseidonHash {
|
||||
}
|
||||
|
||||
fn hash(inputs: &[Self::Fr]) -> Self::Fr {
|
||||
poseidon_hash(inputs)
|
||||
// TODO: allow to handle error properly in pmtree Hasher trait
|
||||
poseidon_hash(inputs).expect("Poseidon hash must be valid")
|
||||
}
|
||||
}
|
||||
|
||||
fn get_tmp_path() -> PathBuf {
|
||||
std::env::temp_dir().join(format!("pmtree-{}", rand::random::<u64>()))
|
||||
fn default_tmp_path() -> Result<PathBuf, std::io::Error> {
|
||||
Ok(Builder::new()
|
||||
.prefix("pmtree-")
|
||||
.tempfile()?
|
||||
.into_temp_path()
|
||||
.to_path_buf())
|
||||
}
|
||||
|
||||
fn get_tmp() -> bool {
|
||||
true
|
||||
const DEFAULT_TEMPORARY: bool = true;
|
||||
const DEFAULT_CACHE_CAPACITY: u64 = 1073741824; // 1 Gigabyte
|
||||
const DEFAULT_FLUSH_EVERY_MS: u64 = 500; // 500 Milliseconds
|
||||
const DEFAULT_MODE: Mode = Mode::HighThroughput;
|
||||
const DEFAULT_USE_COMPRESSION: bool = false;
|
||||
|
||||
pub struct PmtreeConfigBuilder {
|
||||
path: Option<PathBuf>,
|
||||
temporary: bool,
|
||||
cache_capacity: u64,
|
||||
flush_every_ms: u64,
|
||||
mode: Mode,
|
||||
use_compression: bool,
|
||||
}
|
||||
|
||||
impl Default for PmtreeConfigBuilder {
|
||||
fn default() -> Self {
|
||||
Self::new()
|
||||
}
|
||||
}
|
||||
|
||||
impl PmtreeConfigBuilder {
|
||||
pub fn new() -> Self {
|
||||
PmtreeConfigBuilder {
|
||||
path: None,
|
||||
temporary: DEFAULT_TEMPORARY,
|
||||
cache_capacity: DEFAULT_CACHE_CAPACITY,
|
||||
flush_every_ms: DEFAULT_FLUSH_EVERY_MS,
|
||||
mode: DEFAULT_MODE,
|
||||
use_compression: DEFAULT_USE_COMPRESSION,
|
||||
}
|
||||
}
|
||||
|
||||
pub fn path<P: Into<PathBuf>>(mut self, path: P) -> Self {
|
||||
self.path = Some(path.into());
|
||||
self
|
||||
}
|
||||
|
||||
pub fn temporary(mut self, temporary: bool) -> Self {
|
||||
self.temporary = temporary;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn cache_capacity(mut self, capacity: u64) -> Self {
|
||||
self.cache_capacity = capacity;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn flush_every_ms(mut self, ms: u64) -> Self {
|
||||
self.flush_every_ms = ms;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn mode(mut self, mode: Mode) -> Self {
|
||||
self.mode = mode;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn use_compression(mut self, compression: bool) -> Self {
|
||||
self.use_compression = compression;
|
||||
self
|
||||
}
|
||||
|
||||
pub fn build(self) -> Result<PmtreeConfig, FromConfigError> {
|
||||
let path = match (self.temporary, self.path) {
|
||||
(true, None) => default_tmp_path()?,
|
||||
(false, None) => return Err(FromConfigError::MissingPath),
|
||||
(true, Some(path)) if path.exists() => return Err(FromConfigError::PathExists),
|
||||
(_, Some(path)) => path,
|
||||
};
|
||||
|
||||
let config = Config::new()
|
||||
.temporary(self.temporary)
|
||||
.path(path)
|
||||
.cache_capacity(self.cache_capacity)
|
||||
.flush_every_ms(Some(self.flush_every_ms))
|
||||
.mode(self.mode)
|
||||
.use_compression(self.use_compression);
|
||||
|
||||
Ok(PmtreeConfig(config))
|
||||
}
|
||||
}
|
||||
|
||||
pub struct PmtreeConfig(Config);
|
||||
|
||||
impl FromStr for PmtreeConfig {
|
||||
type Err = Report;
|
||||
impl PmtreeConfig {
|
||||
pub fn builder() -> PmtreeConfigBuilder {
|
||||
PmtreeConfigBuilder::new()
|
||||
}
|
||||
}
|
||||
|
||||
fn from_str(s: &str) -> Result<Self> {
|
||||
impl FromStr for PmtreeConfig {
|
||||
type Err = FromConfigError;
|
||||
|
||||
fn from_str(s: &str) -> Result<Self, Self::Err> {
|
||||
let config: Value = serde_json::from_str(s)?;
|
||||
|
||||
let path = config["path"].as_str();
|
||||
@@ -80,21 +178,17 @@ impl FromStr for PmtreeConfig {
|
||||
};
|
||||
let use_compression = config["use_compression"].as_bool();
|
||||
|
||||
if temporary.is_some()
|
||||
&& path.is_some()
|
||||
&& temporary.unwrap()
|
||||
&& path.as_ref().unwrap().exists()
|
||||
{
|
||||
return Err(Report::msg(format!(
|
||||
"Path {:?} already exists, cannot use temporary",
|
||||
path.unwrap()
|
||||
)));
|
||||
if let (Some(true), Some(path)) = (temporary, path.as_ref()) {
|
||||
if path.exists() {
|
||||
return Err(FromConfigError::PathExists);
|
||||
}
|
||||
}
|
||||
|
||||
let default_tmp_path = default_tmp_path()?;
|
||||
let config = Config::new()
|
||||
.temporary(temporary.unwrap_or(get_tmp()))
|
||||
.path(path.unwrap_or(get_tmp_path()))
|
||||
.cache_capacity(cache_capacity.unwrap_or(1024 * 1024 * 1024))
|
||||
.temporary(temporary.unwrap_or(DEFAULT_TEMPORARY))
|
||||
.path(path.unwrap_or(default_tmp_path))
|
||||
.cache_capacity(cache_capacity.unwrap_or(DEFAULT_CACHE_CAPACITY))
|
||||
.flush_every_ms(flush_every_ms)
|
||||
.mode(mode)
|
||||
.use_compression(use_compression.unwrap_or(false));
|
||||
@@ -104,18 +198,12 @@ impl FromStr for PmtreeConfig {
|
||||
|
||||
impl Default for PmtreeConfig {
|
||||
fn default() -> Self {
|
||||
let tmp_path = get_tmp_path();
|
||||
PmtreeConfig(
|
||||
Config::new()
|
||||
.temporary(true)
|
||||
.path(tmp_path)
|
||||
.cache_capacity(150_000)
|
||||
.mode(Mode::HighThroughput)
|
||||
.use_compression(false)
|
||||
.flush_every_ms(Some(12_000)),
|
||||
)
|
||||
Self::builder()
|
||||
.build()
|
||||
.expect("Default PmtreeConfig must be valid")
|
||||
}
|
||||
}
|
||||
|
||||
impl Debug for PmtreeConfig {
|
||||
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
||||
self.0.fmt(f)
|
||||
@@ -133,12 +221,16 @@ impl ZerokitMerkleTree for PmTree {
|
||||
type Hasher = PoseidonHash;
|
||||
type Config = PmtreeConfig;
|
||||
|
||||
fn default(depth: usize) -> Result<Self> {
|
||||
fn default(depth: usize) -> Result<Self, ZerokitMerkleTreeError> {
|
||||
let default_config = PmtreeConfig::default();
|
||||
PmTree::new(depth, Self::Hasher::default_leaf(), default_config)
|
||||
}
|
||||
|
||||
fn new(depth: usize, _default_leaf: FrOf<Self::Hasher>, config: Self::Config) -> Result<Self> {
|
||||
fn new(
|
||||
depth: usize,
|
||||
_default_leaf: FrOf<Self::Hasher>,
|
||||
config: Self::Config,
|
||||
) -> Result<Self, ZerokitMerkleTreeError> {
|
||||
let tree_loaded = pmtree::MerkleTree::load(config.clone().0);
|
||||
let tree = match tree_loaded {
|
||||
Ok(tree) => tree,
|
||||
@@ -168,14 +260,12 @@ impl ZerokitMerkleTree for PmTree {
|
||||
self.tree.root()
|
||||
}
|
||||
|
||||
fn compute_root(&mut self) -> Result<FrOf<Self::Hasher>> {
|
||||
Ok(self.tree.root())
|
||||
}
|
||||
|
||||
fn set(&mut self, index: usize, leaf: FrOf<Self::Hasher>) -> Result<()> {
|
||||
self.tree
|
||||
.set(index, leaf)
|
||||
.map_err(|e| Report::msg(e.to_string()))?;
|
||||
fn set(
|
||||
&mut self,
|
||||
index: usize,
|
||||
leaf: FrOf<Self::Hasher>,
|
||||
) -> Result<(), ZerokitMerkleTreeError> {
|
||||
self.tree.set(index, leaf)?;
|
||||
self.cached_leaves_indices[index] = 1;
|
||||
Ok(())
|
||||
}
|
||||
@@ -184,38 +274,41 @@ impl ZerokitMerkleTree for PmTree {
|
||||
&mut self,
|
||||
start: usize,
|
||||
values: I,
|
||||
) -> Result<()> {
|
||||
) -> Result<(), ZerokitMerkleTreeError> {
|
||||
let v = values.into_iter().collect::<Vec<_>>();
|
||||
self.tree
|
||||
.set_range(start, v.clone().into_iter())
|
||||
.map_err(|e| Report::msg(e.to_string()))?;
|
||||
self.tree.set_range(start, v.clone().into_iter())?;
|
||||
for i in start..v.len() {
|
||||
self.cached_leaves_indices[i] = 1
|
||||
}
|
||||
Ok(())
|
||||
}
|
||||
|
||||
fn get(&self, index: usize) -> Result<FrOf<Self::Hasher>> {
|
||||
self.tree.get(index).map_err(|e| Report::msg(e.to_string()))
|
||||
fn get(&self, index: usize) -> Result<FrOf<Self::Hasher>, ZerokitMerkleTreeError> {
|
||||
self.tree
|
||||
.get(index)
|
||||
.map_err(ZerokitMerkleTreeError::PmtreeErrorKind)
|
||||
}
|
||||
|
||||
fn get_subtree_root(&self, n: usize, index: usize) -> Result<FrOf<Self::Hasher>> {
|
||||
fn get_subtree_root(
|
||||
&self,
|
||||
n: usize,
|
||||
index: usize,
|
||||
) -> Result<FrOf<Self::Hasher>, ZerokitMerkleTreeError> {
|
||||
if n > self.depth() {
|
||||
return Err(Report::msg("level exceeds depth size"));
|
||||
return Err(ZerokitMerkleTreeError::InvalidLevel);
|
||||
}
|
||||
if index >= self.capacity() {
|
||||
return Err(Report::msg("index exceeds set size"));
|
||||
return Err(ZerokitMerkleTreeError::InvalidLeaf);
|
||||
}
|
||||
if n == 0 {
|
||||
Ok(self.root())
|
||||
} else if n == self.depth() {
|
||||
self.get(index)
|
||||
} else {
|
||||
let node = self
|
||||
.tree
|
||||
.get_elem(Key::new(n, index >> (self.depth() - n)))
|
||||
.unwrap();
|
||||
Ok(node)
|
||||
match self.tree.get_elem(Key::new(n, index >> (self.depth() - n))) {
|
||||
Ok(value) => Ok(value),
|
||||
Err(_) => Err(ZerokitMerkleTreeError::InvalidSubTreeIndex),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -235,70 +328,86 @@ impl ZerokitMerkleTree for PmTree {
|
||||
start: usize,
|
||||
leaves: I,
|
||||
indices: J,
|
||||
) -> Result<()> {
|
||||
) -> Result<(), ZerokitMerkleTreeError> {
|
||||
let leaves = leaves.into_iter().collect::<Vec<_>>();
|
||||
let mut indices = indices.into_iter().collect::<Vec<_>>();
|
||||
indices.sort();
|
||||
|
||||
match (leaves.len(), indices.len()) {
|
||||
(0, 0) => Err(Report::msg("no leaves or indices to be removed")),
|
||||
(0, 0) => Err(ZerokitMerkleTreeError::InvalidLeaf),
|
||||
(1, 0) => self.set(start, leaves[0]),
|
||||
(0, 1) => self.delete(indices[0]),
|
||||
(_, 0) => self.set_range(start, leaves),
|
||||
(0, _) => self.remove_indices(&indices),
|
||||
(_, _) => self.remove_indices_and_set_leaves(start, leaves, &indices),
|
||||
(_, 0) => self.set_range(start, leaves.into_iter()),
|
||||
(0, _) => self
|
||||
.remove_indices(&indices)
|
||||
.map_err(ZerokitMerkleTreeError::PmtreeErrorKind),
|
||||
(_, _) => self
|
||||
.remove_indices_and_set_leaves(start, leaves, &indices)
|
||||
.map_err(ZerokitMerkleTreeError::PmtreeErrorKind),
|
||||
}
|
||||
}
|
||||
|
||||
fn update_next(&mut self, leaf: FrOf<Self::Hasher>) -> Result<()> {
|
||||
fn update_next(&mut self, leaf: FrOf<Self::Hasher>) -> Result<(), ZerokitMerkleTreeError> {
|
||||
self.tree
|
||||
.update_next(leaf)
|
||||
.map_err(|e| Report::msg(e.to_string()))
|
||||
.map_err(ZerokitMerkleTreeError::PmtreeErrorKind)
|
||||
}
|
||||
|
||||
fn delete(&mut self, index: usize) -> Result<()> {
|
||||
/// Delete a leaf in the merkle tree given its index
|
||||
///
|
||||
/// Deleting a leaf is done by resetting it to its default value. Note that the next_index field
|
||||
/// will not be changed (== previously used index cannot be reused - this to avoid replay
|
||||
/// attacks or unexpected and very hard to tackle issues)
|
||||
fn delete(&mut self, index: usize) -> Result<(), ZerokitMerkleTreeError> {
|
||||
self.tree
|
||||
.delete(index)
|
||||
.map_err(|e| Report::msg(e.to_string()))?;
|
||||
.map_err(ZerokitMerkleTreeError::PmtreeErrorKind)?;
|
||||
self.cached_leaves_indices[index] = 0;
|
||||
Ok(())
|
||||
}
|
||||
|
||||
fn proof(&self, index: usize) -> Result<Self::Proof> {
|
||||
fn proof(&self, index: usize) -> Result<Self::Proof, ZerokitMerkleTreeError> {
|
||||
let proof = self.tree.proof(index)?;
|
||||
Ok(PmTreeProof { proof })
|
||||
}
|
||||
|
||||
fn verify(&self, leaf: &FrOf<Self::Hasher>, witness: &Self::Proof) -> Result<bool> {
|
||||
if self.tree.verify(leaf, &witness.proof) {
|
||||
fn verify(
|
||||
&self,
|
||||
leaf: &FrOf<Self::Hasher>,
|
||||
merkle_proof: &Self::Proof,
|
||||
) -> Result<bool, ZerokitMerkleTreeError> {
|
||||
if self.tree.verify(leaf, &merkle_proof.proof) {
|
||||
Ok(true)
|
||||
} else {
|
||||
Err(Report::msg("verify failed"))
|
||||
Err(ZerokitMerkleTreeError::InvalidMerkleProof)
|
||||
}
|
||||
}
|
||||
|
||||
fn set_metadata(&mut self, metadata: &[u8]) -> Result<()> {
|
||||
self.tree.db.put(METADATA_KEY, metadata.to_vec())?;
|
||||
fn set_metadata(&mut self, metadata: &[u8]) -> Result<(), ZerokitMerkleTreeError> {
|
||||
self.tree
|
||||
.db
|
||||
.put(METADATA_KEY, metadata.to_vec())
|
||||
.map_err(ZerokitMerkleTreeError::PmtreeErrorKind)?;
|
||||
self.metadata = metadata.to_vec();
|
||||
Ok(())
|
||||
}
|
||||
|
||||
fn metadata(&self) -> Result<Vec<u8>> {
|
||||
fn metadata(&self) -> Result<Vec<u8>, ZerokitMerkleTreeError> {
|
||||
if !self.metadata.is_empty() {
|
||||
return Ok(self.metadata.clone());
|
||||
}
|
||||
// if empty, try searching the db
|
||||
let data = self.tree.db.get(METADATA_KEY)?;
|
||||
|
||||
if data.is_none() {
|
||||
// send empty Metadata
|
||||
return Ok(Vec::new());
|
||||
}
|
||||
Ok(data.unwrap())
|
||||
// Return empty metadata if not found, otherwise return the data
|
||||
Ok(data.unwrap_or_default())
|
||||
}
|
||||
|
||||
fn close_db_connection(&mut self) -> Result<()> {
|
||||
self.tree.db.close().map_err(|e| Report::msg(e.to_string()))
|
||||
fn close_db_connection(&mut self) -> Result<(), ZerokitMerkleTreeError> {
|
||||
self.tree
|
||||
.db
|
||||
.close()
|
||||
.map_err(ZerokitMerkleTreeError::PmtreeErrorKind)
|
||||
}
|
||||
}
|
||||
|
||||
@@ -306,15 +415,18 @@ type PmTreeHasher = <PmTree as ZerokitMerkleTree>::Hasher;
|
||||
type FrOfPmTreeHasher = FrOf<PmTreeHasher>;
|
||||
|
||||
impl PmTree {
|
||||
fn remove_indices(&mut self, indices: &[usize]) -> Result<()> {
|
||||
fn remove_indices(&mut self, indices: &[usize]) -> Result<(), PmtreeErrorKind> {
|
||||
if indices.is_empty() {
|
||||
return Err(PmtreeErrorKind::TreeError(
|
||||
pmtree::TreeErrorKind::InvalidKey,
|
||||
));
|
||||
}
|
||||
let start = indices[0];
|
||||
let end = indices.last().unwrap() + 1;
|
||||
let end = indices[indices.len() - 1] + 1;
|
||||
|
||||
let new_leaves = (start..end).map(|_| PmTreeHasher::default_leaf());
|
||||
|
||||
self.tree
|
||||
.set_range(start, new_leaves)
|
||||
.map_err(|e| Report::msg(e.to_string()))?;
|
||||
self.tree.set_range(start, new_leaves)?;
|
||||
|
||||
for i in start..end {
|
||||
self.cached_leaves_indices[i] = 0
|
||||
@@ -327,8 +439,13 @@ impl PmTree {
|
||||
start: usize,
|
||||
leaves: Vec<FrOfPmTreeHasher>,
|
||||
indices: &[usize],
|
||||
) -> Result<()> {
|
||||
let min_index = *indices.first().unwrap();
|
||||
) -> Result<(), PmtreeErrorKind> {
|
||||
if indices.is_empty() {
|
||||
return Err(PmtreeErrorKind::TreeError(
|
||||
pmtree::TreeErrorKind::InvalidKey,
|
||||
));
|
||||
}
|
||||
let min_index = indices[0];
|
||||
let max_index = start + leaves.len();
|
||||
|
||||
let mut set_values = vec![PmTreeHasher::default_leaf(); max_index - min_index];
|
||||
@@ -344,9 +461,7 @@ impl PmTree {
|
||||
set_values[start - min_index + i] = leaf;
|
||||
}
|
||||
|
||||
self.tree
|
||||
.set_range(start, set_values)
|
||||
.map_err(|e| Report::msg(e.to_string()))?;
|
||||
self.tree.set_range(start, set_values)?;
|
||||
|
||||
for i in indices {
|
||||
self.cached_leaves_indices[*i] = 0;
|
||||
@@ -378,7 +493,40 @@ impl ZerokitMerkleProof for PmTreeProof {
|
||||
fn get_path_index(&self) -> Vec<Self::Index> {
|
||||
self.proof.get_path_index()
|
||||
}
|
||||
fn compute_root_from(&self, leaf: &FrOf<Self::Hasher>) -> FrOf<Self::Hasher> {
|
||||
self.proof.compute_root_from(leaf)
|
||||
|
||||
fn compute_root_from(
|
||||
&self,
|
||||
leaf: &FrOf<Self::Hasher>,
|
||||
) -> Result<FrOf<Self::Hasher>, ZerokitMerkleTreeError> {
|
||||
Ok(self.proof.compute_root_from(leaf))
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
mod test {
|
||||
use super::*;
|
||||
|
||||
#[test]
|
||||
fn test_pmtree_json_config() {
|
||||
let json = r#"
|
||||
{
|
||||
"path": "pmtree-123456",
|
||||
"temporary": false,
|
||||
"cache_capacity": 1073741824,
|
||||
"flush_every_ms": 500,
|
||||
"mode": "HighThroughput",
|
||||
"use_compression": false
|
||||
}"#;
|
||||
|
||||
let _: PmtreeConfig = json.parse().unwrap();
|
||||
|
||||
let _ = PmtreeConfig::builder()
|
||||
.path(default_tmp_path().unwrap())
|
||||
.temporary(DEFAULT_TEMPORARY)
|
||||
.cache_capacity(DEFAULT_CACHE_CAPACITY)
|
||||
.mode(DEFAULT_MODE)
|
||||
.use_compression(DEFAULT_USE_COMPRESSION)
|
||||
.build()
|
||||
.unwrap();
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1,30 +1,32 @@
|
||||
// This crate defines the RLN module default Merkle tree implementation and its Hasher
|
||||
// Implementation inspired by https://github.com/worldcoin/semaphore-rs/blob/d462a4372f1fd9c27610f2acfe4841fab1d396aa/src/poseidon_tree.rs
|
||||
|
||||
// Implementation inspired by https://github.com/worldcoin/semaphore-rs/blob/d462a4372f1fd9c27610f2acfe4841fab1d396aa/src/poseidon_tree.rs (no differences)
|
||||
#![cfg(not(feature = "stateless"))]
|
||||
|
||||
use cfg_if::cfg_if;
|
||||
|
||||
cfg_if! {
|
||||
if #[cfg(feature = "pmtree-ft")] {
|
||||
use crate::pm_tree_adapter::*;
|
||||
} else {
|
||||
use crate::hashers::{PoseidonHash};
|
||||
use utils::merkle_tree::*;
|
||||
}
|
||||
}
|
||||
|
||||
// The zerokit RLN default Merkle tree implementation is the OptimalMerkleTree.
|
||||
// To switch to FullMerkleTree implementation, it is enough to enable the fullmerkletree feature
|
||||
// The zerokit RLN default Merkle tree implementation is the PMTree from the vacp2p_pmtree crate
|
||||
// To switch to FullMerkleTree or OptimalMerkleTree, enable the corresponding feature in the Cargo.toml file
|
||||
|
||||
cfg_if! {
|
||||
if #[cfg(feature = "fullmerkletree")] {
|
||||
use zerokit_utils::merkle_tree::{FullMerkleTree, FullMerkleProof};
|
||||
use crate::hashers::PoseidonHash;
|
||||
|
||||
pub type PoseidonTree = FullMerkleTree<PoseidonHash>;
|
||||
pub type MerkleProof = FullMerkleProof<PoseidonHash>;
|
||||
} else if #[cfg(feature = "pmtree-ft")] {
|
||||
pub type PoseidonTree = PmTree;
|
||||
pub type MerkleProof = PmTreeProof;
|
||||
} else {
|
||||
} else if #[cfg(feature = "optimalmerkletree")] {
|
||||
use zerokit_utils::merkle_tree::{OptimalMerkleTree, OptimalMerkleProof};
|
||||
use crate::hashers::PoseidonHash;
|
||||
|
||||
pub type PoseidonTree = OptimalMerkleTree<PoseidonHash>;
|
||||
pub type MerkleProof = OptimalMerkleProof<PoseidonHash>;
|
||||
} else if #[cfg(feature = "pmtree-ft")] {
|
||||
use crate::pm_tree_adapter::{PmTree, PmTreeProof};
|
||||
|
||||
pub type PoseidonTree = PmTree;
|
||||
pub type MerkleProof = PmTreeProof;
|
||||
} else {
|
||||
compile_error!("One of the features `fullmerkletree`, `optimalmerkletree`, or `pmtree-ft` must be enabled.");
|
||||
}
|
||||
}
|
||||
|
||||
37
rln/src/prelude.rs
Normal file
37
rln/src/prelude.rs
Normal file
@@ -0,0 +1,37 @@
|
||||
// This module re-exports the most commonly used types and functions from the RLN library
|
||||
|
||||
#[cfg(not(target_arch = "wasm32"))]
|
||||
pub use crate::circuit::{graph_from_folder, zkey_from_folder};
|
||||
#[cfg(feature = "pmtree-ft")]
|
||||
pub use crate::pm_tree_adapter::{FrOf, PmTree, PmTreeProof, PmtreeConfig, PmtreeConfigBuilder};
|
||||
#[cfg(not(feature = "stateless"))]
|
||||
pub use crate::poseidon_tree::{MerkleProof, PoseidonTree};
|
||||
#[cfg(not(feature = "stateless"))]
|
||||
pub use crate::protocol::compute_tree_root;
|
||||
#[cfg(not(target_arch = "wasm32"))]
|
||||
pub use crate::protocol::{generate_zk_proof, verify_zk_proof};
|
||||
pub use crate::{
|
||||
circuit::{
|
||||
zkey_from_raw, Curve, Fq, Fq2, Fr, G1Affine, G1Projective, G2Affine, G2Projective, Proof,
|
||||
VerifyingKey, Zkey, COMPRESS_PROOF_SIZE, DEFAULT_TREE_DEPTH,
|
||||
},
|
||||
error::{ProtocolError, RLNError, UtilsError, VerifyError},
|
||||
hashers::{hash_to_field_be, hash_to_field_le, poseidon_hash, PoseidonHash},
|
||||
protocol::{
|
||||
bytes_be_to_rln_proof, bytes_be_to_rln_proof_values, bytes_be_to_rln_witness,
|
||||
bytes_le_to_rln_proof, bytes_le_to_rln_proof_values, bytes_le_to_rln_witness,
|
||||
extended_keygen, extended_seeded_keygen, generate_zk_proof_with_witness, keygen,
|
||||
proof_values_from_witness, recover_id_secret, rln_proof_to_bytes_be, rln_proof_to_bytes_le,
|
||||
rln_proof_values_to_bytes_be, rln_proof_values_to_bytes_le, rln_witness_to_bigint_json,
|
||||
rln_witness_to_bytes_be, rln_witness_to_bytes_le, seeded_keygen, RLNProof, RLNProofValues,
|
||||
RLNWitnessInput,
|
||||
},
|
||||
public::RLN,
|
||||
utils::{
|
||||
bytes_be_to_fr, bytes_be_to_vec_fr, bytes_be_to_vec_u8, bytes_be_to_vec_usize,
|
||||
bytes_le_to_fr, bytes_le_to_vec_fr, bytes_le_to_vec_u8, bytes_le_to_vec_usize,
|
||||
fr_to_bytes_be, fr_to_bytes_le, normalize_usize_be, normalize_usize_le, str_to_fr,
|
||||
to_bigint, vec_fr_to_bytes_be, vec_fr_to_bytes_le, vec_u8_to_bytes_be, vec_u8_to_bytes_le,
|
||||
IdSecret, FR_BYTE_SIZE,
|
||||
},
|
||||
};
|
||||
@@ -1,807 +0,0 @@
|
||||
// This crate collects all the underlying primitives used to implement RLN
|
||||
|
||||
use ark_circom::{CircomReduction, WitnessCalculator};
|
||||
use ark_groth16::{prepare_verifying_key, Groth16, Proof as ArkProof, ProvingKey, VerifyingKey};
|
||||
use ark_relations::r1cs::ConstraintMatrices;
|
||||
use ark_relations::r1cs::SynthesisError;
|
||||
use ark_serialize::{CanonicalDeserialize, CanonicalSerialize};
|
||||
use ark_std::{rand::thread_rng, UniformRand};
|
||||
use color_eyre::{Report, Result};
|
||||
use num_bigint::BigInt;
|
||||
use rand::{Rng, SeedableRng};
|
||||
use rand_chacha::ChaCha20Rng;
|
||||
use serde::{Deserialize, Serialize};
|
||||
#[cfg(not(target_arch = "wasm32"))]
|
||||
use std::sync::Mutex;
|
||||
#[cfg(debug_assertions)]
|
||||
use std::time::Instant;
|
||||
use thiserror::Error;
|
||||
use tiny_keccak::{Hasher as _, Keccak};
|
||||
|
||||
use crate::circuit::{Curve, Fr};
|
||||
use crate::hashers::hash_to_field;
|
||||
use crate::hashers::poseidon_hash;
|
||||
use crate::poseidon_tree::*;
|
||||
use crate::public::RLN_IDENTIFIER;
|
||||
use crate::utils::*;
|
||||
use cfg_if::cfg_if;
|
||||
use utils::{ZerokitMerkleProof, ZerokitMerkleTree};
|
||||
|
||||
///////////////////////////////////////////////////////
|
||||
// RLN Witness data structure and utility functions
|
||||
///////////////////////////////////////////////////////
|
||||
|
||||
#[derive(Debug, PartialEq, Serialize, Deserialize)]
|
||||
pub struct RLNWitnessInput {
|
||||
#[serde(serialize_with = "ark_se", deserialize_with = "ark_de")]
|
||||
identity_secret: Fr,
|
||||
#[serde(serialize_with = "ark_se", deserialize_with = "ark_de")]
|
||||
user_message_limit: Fr,
|
||||
#[serde(serialize_with = "ark_se", deserialize_with = "ark_de")]
|
||||
message_id: Fr,
|
||||
#[serde(serialize_with = "ark_se", deserialize_with = "ark_de")]
|
||||
path_elements: Vec<Fr>,
|
||||
identity_path_index: Vec<u8>,
|
||||
#[serde(serialize_with = "ark_se", deserialize_with = "ark_de")]
|
||||
x: Fr,
|
||||
#[serde(serialize_with = "ark_se", deserialize_with = "ark_de")]
|
||||
external_nullifier: Fr,
|
||||
}
|
||||
|
||||
#[derive(Debug, PartialEq)]
|
||||
pub struct RLNProofValues {
|
||||
// Public outputs:
|
||||
pub y: Fr,
|
||||
pub nullifier: Fr,
|
||||
pub root: Fr,
|
||||
// Public Inputs:
|
||||
pub x: Fr,
|
||||
pub external_nullifier: Fr,
|
||||
}
|
||||
|
||||
pub fn serialize_field_element(element: Fr) -> Vec<u8> {
|
||||
fr_to_bytes_le(&element)
|
||||
}
|
||||
|
||||
pub fn deserialize_field_element(serialized: Vec<u8>) -> Fr {
|
||||
let (element, _) = bytes_le_to_fr(&serialized);
|
||||
|
||||
element
|
||||
}
|
||||
|
||||
pub fn deserialize_identity_pair(serialized: Vec<u8>) -> (Fr, Fr) {
|
||||
let (identity_secret_hash, read) = bytes_le_to_fr(&serialized);
|
||||
let (id_commitment, _) = bytes_le_to_fr(&serialized[read..]);
|
||||
|
||||
(identity_secret_hash, id_commitment)
|
||||
}
|
||||
|
||||
pub fn deserialize_identity_tuple(serialized: Vec<u8>) -> (Fr, Fr, Fr, Fr) {
|
||||
let mut all_read = 0;
|
||||
|
||||
let (identity_trapdoor, read) = bytes_le_to_fr(&serialized[all_read..]);
|
||||
all_read += read;
|
||||
|
||||
let (identity_nullifier, read) = bytes_le_to_fr(&serialized[all_read..]);
|
||||
all_read += read;
|
||||
|
||||
let (identity_secret_hash, read) = bytes_le_to_fr(&serialized[all_read..]);
|
||||
all_read += read;
|
||||
|
||||
let (identity_commitment, _) = bytes_le_to_fr(&serialized[all_read..]);
|
||||
|
||||
(
|
||||
identity_trapdoor,
|
||||
identity_nullifier,
|
||||
identity_secret_hash,
|
||||
identity_commitment,
|
||||
)
|
||||
}
|
||||
|
||||
/// Serializes witness
|
||||
///
|
||||
/// # Errors
|
||||
///
|
||||
/// Returns an error if `rln_witness.message_id` is not within `rln_witness.user_message_limit`.
|
||||
pub fn serialize_witness(rln_witness: &RLNWitnessInput) -> Result<Vec<u8>> {
|
||||
message_id_range_check(&rln_witness.message_id, &rln_witness.user_message_limit)?;
|
||||
|
||||
let mut serialized: Vec<u8> = Vec::new();
|
||||
|
||||
serialized.append(&mut fr_to_bytes_le(&rln_witness.identity_secret));
|
||||
serialized.append(&mut fr_to_bytes_le(&rln_witness.user_message_limit));
|
||||
serialized.append(&mut fr_to_bytes_le(&rln_witness.message_id));
|
||||
serialized.append(&mut vec_fr_to_bytes_le(&rln_witness.path_elements)?);
|
||||
serialized.append(&mut vec_u8_to_bytes_le(&rln_witness.identity_path_index)?);
|
||||
serialized.append(&mut fr_to_bytes_le(&rln_witness.x));
|
||||
serialized.append(&mut fr_to_bytes_le(&rln_witness.external_nullifier));
|
||||
|
||||
Ok(serialized)
|
||||
}
|
||||
|
||||
/// Deserializes witness
|
||||
///
|
||||
/// # Errors
|
||||
///
|
||||
/// Returns an error if `message_id` is not within `user_message_limit`.
|
||||
pub fn deserialize_witness(serialized: &[u8]) -> Result<(RLNWitnessInput, usize)> {
|
||||
let mut all_read: usize = 0;
|
||||
|
||||
let (identity_secret, read) = bytes_le_to_fr(&serialized[all_read..]);
|
||||
all_read += read;
|
||||
|
||||
let (user_message_limit, read) = bytes_le_to_fr(&serialized[all_read..]);
|
||||
all_read += read;
|
||||
|
||||
let (message_id, read) = bytes_le_to_fr(&serialized[all_read..]);
|
||||
all_read += read;
|
||||
|
||||
message_id_range_check(&message_id, &user_message_limit)?;
|
||||
|
||||
let (path_elements, read) = bytes_le_to_vec_fr(&serialized[all_read..])?;
|
||||
all_read += read;
|
||||
|
||||
let (identity_path_index, read) = bytes_le_to_vec_u8(&serialized[all_read..])?;
|
||||
all_read += read;
|
||||
|
||||
let (x, read) = bytes_le_to_fr(&serialized[all_read..]);
|
||||
all_read += read;
|
||||
|
||||
let (external_nullifier, read) = bytes_le_to_fr(&serialized[all_read..]);
|
||||
all_read += read;
|
||||
|
||||
if serialized.len() != all_read {
|
||||
return Err(Report::msg("serialized length is not equal to all_read"));
|
||||
}
|
||||
|
||||
Ok((
|
||||
RLNWitnessInput {
|
||||
identity_secret,
|
||||
path_elements,
|
||||
identity_path_index,
|
||||
x,
|
||||
external_nullifier,
|
||||
user_message_limit,
|
||||
message_id,
|
||||
},
|
||||
all_read,
|
||||
))
|
||||
}
|
||||
|
||||
// This function deserializes input for kilic's rln generate_proof public API
|
||||
// https://github.com/kilic/rln/blob/7ac74183f8b69b399e3bc96c1ae8ab61c026dc43/src/public.rs#L148
|
||||
// input_data is [ identity_secret<32> | id_index<8> | user_message_limit<32> | message_id<32> | external_nullifier<32> | signal_len<8> | signal<var> ]
|
||||
// return value is a rln witness populated according to this information
|
||||
pub fn proof_inputs_to_rln_witness(
|
||||
tree: &mut PoseidonTree,
|
||||
serialized: &[u8],
|
||||
) -> Result<(RLNWitnessInput, usize)> {
|
||||
let mut all_read: usize = 0;
|
||||
|
||||
let (identity_secret, read) = bytes_le_to_fr(&serialized[all_read..]);
|
||||
all_read += read;
|
||||
|
||||
let id_index = usize::try_from(u64::from_le_bytes(
|
||||
serialized[all_read..all_read + 8].try_into()?,
|
||||
))?;
|
||||
all_read += 8;
|
||||
|
||||
let (user_message_limit, read) = bytes_le_to_fr(&serialized[all_read..]);
|
||||
all_read += read;
|
||||
|
||||
let (message_id, read) = bytes_le_to_fr(&serialized[all_read..]);
|
||||
all_read += read;
|
||||
|
||||
let (external_nullifier, read) = bytes_le_to_fr(&serialized[all_read..]);
|
||||
all_read += read;
|
||||
|
||||
let signal_len = usize::try_from(u64::from_le_bytes(
|
||||
serialized[all_read..all_read + 8].try_into()?,
|
||||
))?;
|
||||
all_read += 8;
|
||||
|
||||
let signal: Vec<u8> = serialized[all_read..all_read + signal_len].to_vec();
|
||||
|
||||
let merkle_proof = tree.proof(id_index).expect("proof should exist");
|
||||
let path_elements = merkle_proof.get_path_elements();
|
||||
let identity_path_index = merkle_proof.get_path_index();
|
||||
|
||||
let x = hash_to_field(&signal);
|
||||
|
||||
Ok((
|
||||
RLNWitnessInput {
|
||||
identity_secret,
|
||||
path_elements,
|
||||
identity_path_index,
|
||||
user_message_limit,
|
||||
message_id,
|
||||
x,
|
||||
external_nullifier,
|
||||
},
|
||||
all_read,
|
||||
))
|
||||
}
|
||||
|
||||
/// Creates `RLNWitnessInput` from it's fields.
|
||||
///
|
||||
/// # Errors
|
||||
///
|
||||
/// Returns an error if `message_id` is not within `user_message_limit`.
|
||||
pub fn rln_witness_from_values(
|
||||
identity_secret: Fr,
|
||||
merkle_proof: &MerkleProof,
|
||||
x: Fr,
|
||||
external_nullifier: Fr,
|
||||
user_message_limit: Fr,
|
||||
message_id: Fr,
|
||||
) -> Result<RLNWitnessInput> {
|
||||
message_id_range_check(&message_id, &user_message_limit)?;
|
||||
|
||||
let path_elements = merkle_proof.get_path_elements();
|
||||
let identity_path_index = merkle_proof.get_path_index();
|
||||
|
||||
Ok(RLNWitnessInput {
|
||||
identity_secret,
|
||||
path_elements,
|
||||
identity_path_index,
|
||||
x,
|
||||
external_nullifier,
|
||||
user_message_limit,
|
||||
message_id,
|
||||
})
|
||||
}
|
||||
|
||||
pub fn random_rln_witness(tree_height: usize) -> RLNWitnessInput {
|
||||
let mut rng = thread_rng();
|
||||
|
||||
let identity_secret = hash_to_field(&rng.gen::<[u8; 32]>());
|
||||
let x = hash_to_field(&rng.gen::<[u8; 32]>());
|
||||
let epoch = hash_to_field(&rng.gen::<[u8; 32]>());
|
||||
let rln_identifier = hash_to_field(RLN_IDENTIFIER); //hash_to_field(&rng.gen::<[u8; 32]>());
|
||||
|
||||
let mut path_elements: Vec<Fr> = Vec::new();
|
||||
let mut identity_path_index: Vec<u8> = Vec::new();
|
||||
|
||||
for _ in 0..tree_height {
|
||||
path_elements.push(hash_to_field(&rng.gen::<[u8; 32]>()));
|
||||
identity_path_index.push(rng.gen_range(0..2) as u8);
|
||||
}
|
||||
|
||||
let user_message_limit = Fr::from(100);
|
||||
let message_id = Fr::from(1);
|
||||
|
||||
RLNWitnessInput {
|
||||
identity_secret,
|
||||
path_elements,
|
||||
identity_path_index,
|
||||
x,
|
||||
external_nullifier: poseidon_hash(&[epoch, rln_identifier]),
|
||||
user_message_limit,
|
||||
message_id,
|
||||
}
|
||||
}
|
||||
|
||||
pub fn proof_values_from_witness(rln_witness: &RLNWitnessInput) -> Result<RLNProofValues> {
|
||||
message_id_range_check(&rln_witness.message_id, &rln_witness.user_message_limit)?;
|
||||
|
||||
// y share
|
||||
let a_0 = rln_witness.identity_secret;
|
||||
let a_1 = poseidon_hash(&[a_0, rln_witness.external_nullifier, rln_witness.message_id]);
|
||||
let y = a_0 + rln_witness.x * a_1;
|
||||
|
||||
// Nullifier
|
||||
let nullifier = poseidon_hash(&[a_1]);
|
||||
|
||||
// Merkle tree root computations
|
||||
let root = compute_tree_root(
|
||||
&rln_witness.identity_secret,
|
||||
&rln_witness.user_message_limit,
|
||||
&rln_witness.path_elements,
|
||||
&rln_witness.identity_path_index,
|
||||
);
|
||||
|
||||
Ok(RLNProofValues {
|
||||
y,
|
||||
nullifier,
|
||||
root,
|
||||
x: rln_witness.x,
|
||||
external_nullifier: rln_witness.external_nullifier,
|
||||
})
|
||||
}
|
||||
|
||||
pub fn serialize_proof_values(rln_proof_values: &RLNProofValues) -> Vec<u8> {
|
||||
let mut serialized: Vec<u8> = Vec::new();
|
||||
|
||||
serialized.append(&mut fr_to_bytes_le(&rln_proof_values.root));
|
||||
serialized.append(&mut fr_to_bytes_le(&rln_proof_values.external_nullifier));
|
||||
serialized.append(&mut fr_to_bytes_le(&rln_proof_values.x));
|
||||
serialized.append(&mut fr_to_bytes_le(&rln_proof_values.y));
|
||||
serialized.append(&mut fr_to_bytes_le(&rln_proof_values.nullifier));
|
||||
|
||||
serialized
|
||||
}
|
||||
|
||||
// Note: don't forget to skip the 128 bytes ZK proof, if serialized contains it.
|
||||
// This proc deserialzies only proof _values_, i.e. circuit outputs, not the zk proof.
|
||||
pub fn deserialize_proof_values(serialized: &[u8]) -> (RLNProofValues, usize) {
|
||||
let mut all_read: usize = 0;
|
||||
|
||||
let (root, read) = bytes_le_to_fr(&serialized[all_read..]);
|
||||
all_read += read;
|
||||
|
||||
let (external_nullifier, read) = bytes_le_to_fr(&serialized[all_read..]);
|
||||
all_read += read;
|
||||
|
||||
let (x, read) = bytes_le_to_fr(&serialized[all_read..]);
|
||||
all_read += read;
|
||||
|
||||
let (y, read) = bytes_le_to_fr(&serialized[all_read..]);
|
||||
all_read += read;
|
||||
|
||||
let (nullifier, read) = bytes_le_to_fr(&serialized[all_read..]);
|
||||
all_read += read;
|
||||
|
||||
(
|
||||
RLNProofValues {
|
||||
y,
|
||||
nullifier,
|
||||
root,
|
||||
x,
|
||||
external_nullifier,
|
||||
},
|
||||
all_read,
|
||||
)
|
||||
}
|
||||
|
||||
pub fn prepare_prove_input(
|
||||
identity_secret: Fr,
|
||||
id_index: usize,
|
||||
external_nullifier: Fr,
|
||||
signal: &[u8],
|
||||
) -> Vec<u8> {
|
||||
let mut serialized: Vec<u8> = Vec::new();
|
||||
|
||||
serialized.append(&mut fr_to_bytes_le(&identity_secret));
|
||||
serialized.append(&mut normalize_usize(id_index));
|
||||
serialized.append(&mut fr_to_bytes_le(&external_nullifier));
|
||||
serialized.append(&mut normalize_usize(signal.len()));
|
||||
serialized.append(&mut signal.to_vec());
|
||||
|
||||
serialized
|
||||
}
|
||||
|
||||
#[allow(clippy::redundant_clone)]
|
||||
pub fn prepare_verify_input(proof_data: Vec<u8>, signal: &[u8]) -> Vec<u8> {
|
||||
let mut serialized: Vec<u8> = Vec::new();
|
||||
|
||||
serialized.append(&mut proof_data.clone());
|
||||
serialized.append(&mut normalize_usize(signal.len()));
|
||||
serialized.append(&mut signal.to_vec());
|
||||
|
||||
serialized
|
||||
}
|
||||
|
||||
///////////////////////////////////////////////////////
|
||||
// Merkle tree utility functions
|
||||
///////////////////////////////////////////////////////
|
||||
|
||||
pub fn compute_tree_root(
|
||||
identity_secret: &Fr,
|
||||
user_message_limit: &Fr,
|
||||
path_elements: &[Fr],
|
||||
identity_path_index: &[u8],
|
||||
) -> Fr {
|
||||
let id_commitment = poseidon_hash(&[*identity_secret]);
|
||||
let mut root = poseidon_hash(&[id_commitment, *user_message_limit]);
|
||||
|
||||
for i in 0..identity_path_index.len() {
|
||||
if identity_path_index[i] == 0 {
|
||||
root = poseidon_hash(&[root, path_elements[i]]);
|
||||
} else {
|
||||
root = poseidon_hash(&[path_elements[i], root]);
|
||||
}
|
||||
}
|
||||
|
||||
root
|
||||
}
|
||||
|
||||
///////////////////////////////////////////////////////
|
||||
// Protocol utility functions
|
||||
///////////////////////////////////////////////////////
|
||||
|
||||
// Generates a tuple (identity_secret_hash, id_commitment) where
|
||||
// identity_secret_hash is random and id_commitment = PoseidonHash(identity_secret_hash)
|
||||
// RNG is instantiated using thread_rng()
|
||||
pub fn keygen() -> (Fr, Fr) {
|
||||
let mut rng = thread_rng();
|
||||
let identity_secret_hash = Fr::rand(&mut rng);
|
||||
let id_commitment = poseidon_hash(&[identity_secret_hash]);
|
||||
(identity_secret_hash, id_commitment)
|
||||
}
|
||||
|
||||
// Generates a tuple (identity_trapdoor, identity_nullifier, identity_secret_hash, id_commitment) where
|
||||
// identity_trapdoor and identity_nullifier are random,
|
||||
// identity_secret_hash = PoseidonHash(identity_trapdoor, identity_nullifier),
|
||||
// id_commitment = PoseidonHash(identity_secret_hash),
|
||||
// RNG is instantiated using thread_rng()
|
||||
// Generated credentials are compatible with Semaphore credentials
|
||||
pub fn extended_keygen() -> (Fr, Fr, Fr, Fr) {
|
||||
let mut rng = thread_rng();
|
||||
let identity_trapdoor = Fr::rand(&mut rng);
|
||||
let identity_nullifier = Fr::rand(&mut rng);
|
||||
let identity_secret_hash = poseidon_hash(&[identity_trapdoor, identity_nullifier]);
|
||||
let id_commitment = poseidon_hash(&[identity_secret_hash]);
|
||||
(
|
||||
identity_trapdoor,
|
||||
identity_nullifier,
|
||||
identity_secret_hash,
|
||||
id_commitment,
|
||||
)
|
||||
}
|
||||
|
||||
// Generates a tuple (identity_secret_hash, id_commitment) where
|
||||
// identity_secret_hash is random and id_commitment = PoseidonHash(identity_secret_hash)
|
||||
// RNG is instantiated using 20 rounds of ChaCha seeded with the hash of the input
|
||||
pub fn seeded_keygen(signal: &[u8]) -> (Fr, Fr) {
|
||||
// ChaCha20 requires a seed of exactly 32 bytes.
|
||||
// We first hash the input seed signal to a 32 bytes array and pass this as seed to ChaCha20
|
||||
let mut seed = [0; 32];
|
||||
let mut hasher = Keccak::v256();
|
||||
hasher.update(signal);
|
||||
hasher.finalize(&mut seed);
|
||||
|
||||
let mut rng = ChaCha20Rng::from_seed(seed);
|
||||
let identity_secret_hash = Fr::rand(&mut rng);
|
||||
let id_commitment = poseidon_hash(&[identity_secret_hash]);
|
||||
(identity_secret_hash, id_commitment)
|
||||
}
|
||||
|
||||
// Generates a tuple (identity_trapdoor, identity_nullifier, identity_secret_hash, id_commitment) where
|
||||
// identity_trapdoor and identity_nullifier are random,
|
||||
// identity_secret_hash = PoseidonHash(identity_trapdoor, identity_nullifier),
|
||||
// id_commitment = PoseidonHash(identity_secret_hash),
|
||||
// RNG is instantiated using 20 rounds of ChaCha seeded with the hash of the input
|
||||
// Generated credentials are compatible with Semaphore credentials
|
||||
pub fn extended_seeded_keygen(signal: &[u8]) -> (Fr, Fr, Fr, Fr) {
|
||||
// ChaCha20 requires a seed of exactly 32 bytes.
|
||||
// We first hash the input seed signal to a 32 bytes array and pass this as seed to ChaCha20
|
||||
let mut seed = [0; 32];
|
||||
let mut hasher = Keccak::v256();
|
||||
hasher.update(signal);
|
||||
hasher.finalize(&mut seed);
|
||||
|
||||
let mut rng = ChaCha20Rng::from_seed(seed);
|
||||
let identity_trapdoor = Fr::rand(&mut rng);
|
||||
let identity_nullifier = Fr::rand(&mut rng);
|
||||
let identity_secret_hash = poseidon_hash(&[identity_trapdoor, identity_nullifier]);
|
||||
let id_commitment = poseidon_hash(&[identity_secret_hash]);
|
||||
(
|
||||
identity_trapdoor,
|
||||
identity_nullifier,
|
||||
identity_secret_hash,
|
||||
id_commitment,
|
||||
)
|
||||
}
|
||||
|
||||
pub fn compute_id_secret(share1: (Fr, Fr), share2: (Fr, Fr)) -> Result<Fr, String> {
|
||||
// Assuming a0 is the identity secret and a1 = poseidonHash([a0, external_nullifier]),
|
||||
// a (x,y) share satisfies the following relation
|
||||
// y = a_0 + x * a_1
|
||||
let (x1, y1) = share1;
|
||||
let (x2, y2) = share2;
|
||||
|
||||
// If the two input shares were computed for the same external_nullifier and identity secret, we can recover the latter
|
||||
// y1 = a_0 + x1 * a_1
|
||||
// y2 = a_0 + x2 * a_1
|
||||
let a_1 = (y1 - y2) / (x1 - x2);
|
||||
let a_0 = y1 - x1 * a_1;
|
||||
|
||||
// If shares come from the same polynomial, a0 is correctly recovered and a1 = poseidonHash([a0, external_nullifier])
|
||||
Ok(a_0)
|
||||
}
|
||||
|
||||
///////////////////////////////////////////////////////
|
||||
// zkSNARK utility functions
|
||||
///////////////////////////////////////////////////////
|
||||
|
||||
#[derive(Error, Debug)]
|
||||
pub enum ProofError {
|
||||
#[error("Error reading circuit key: {0}")]
|
||||
CircuitKeyError(#[from] Report),
|
||||
#[error("Error producing witness: {0}")]
|
||||
WitnessError(Report),
|
||||
#[error("Error producing proof: {0}")]
|
||||
SynthesisError(#[from] SynthesisError),
|
||||
}
|
||||
|
||||
fn calculate_witness_element<E: ark_ec::pairing::Pairing>(
|
||||
witness: Vec<BigInt>,
|
||||
) -> Result<Vec<E::ScalarField>> {
|
||||
use ark_ff::PrimeField;
|
||||
let modulus = <E::ScalarField as PrimeField>::MODULUS;
|
||||
|
||||
// convert it to field elements
|
||||
use num_traits::Signed;
|
||||
let mut witness_vec = vec![];
|
||||
for w in witness.into_iter() {
|
||||
let w = if w.sign() == num_bigint::Sign::Minus {
|
||||
// Need to negate the witness element if negative
|
||||
modulus.into()
|
||||
- w.abs()
|
||||
.to_biguint()
|
||||
.ok_or(Report::msg("not a biguint value"))?
|
||||
} else {
|
||||
w.to_biguint().ok_or(Report::msg("not a biguint value"))?
|
||||
};
|
||||
witness_vec.push(E::ScalarField::from(w))
|
||||
}
|
||||
|
||||
Ok(witness_vec)
|
||||
}
|
||||
|
||||
pub fn generate_proof_with_witness(
|
||||
witness: Vec<BigInt>,
|
||||
proving_key: &(ProvingKey<Curve>, ConstraintMatrices<Fr>),
|
||||
) -> Result<ArkProof<Curve>, ProofError> {
|
||||
// If in debug mode, we measure and later print time take to compute witness
|
||||
#[cfg(debug_assertions)]
|
||||
let now = Instant::now();
|
||||
|
||||
let full_assignment =
|
||||
calculate_witness_element::<Curve>(witness).map_err(ProofError::WitnessError)?;
|
||||
|
||||
#[cfg(debug_assertions)]
|
||||
println!("witness generation took: {:.2?}", now.elapsed());
|
||||
|
||||
// Random Values
|
||||
let mut rng = thread_rng();
|
||||
let r = Fr::rand(&mut rng);
|
||||
let s = Fr::rand(&mut rng);
|
||||
|
||||
// If in debug mode, we measure and later print time take to compute proof
|
||||
#[cfg(debug_assertions)]
|
||||
let now = Instant::now();
|
||||
|
||||
let proof = Groth16::<_, CircomReduction>::create_proof_with_reduction_and_matrices(
|
||||
&proving_key.0,
|
||||
r,
|
||||
s,
|
||||
&proving_key.1,
|
||||
proving_key.1.num_instance_variables,
|
||||
proving_key.1.num_constraints,
|
||||
full_assignment.as_slice(),
|
||||
)?;
|
||||
|
||||
#[cfg(debug_assertions)]
|
||||
println!("proof generation took: {:.2?}", now.elapsed());
|
||||
|
||||
Ok(proof)
|
||||
}
|
||||
|
||||
/// Formats inputs for witness calculation
|
||||
///
|
||||
/// # Errors
|
||||
///
|
||||
/// Returns an error if `rln_witness.message_id` is not within `rln_witness.user_message_limit`.
|
||||
pub fn inputs_for_witness_calculation(
|
||||
rln_witness: &RLNWitnessInput,
|
||||
) -> Result<[(&str, Vec<BigInt>); 7]> {
|
||||
message_id_range_check(&rln_witness.message_id, &rln_witness.user_message_limit)?;
|
||||
|
||||
// We convert the path indexes to field elements
|
||||
// TODO: check if necessary
|
||||
let mut path_elements = Vec::new();
|
||||
|
||||
for v in rln_witness.path_elements.iter() {
|
||||
path_elements.push(to_bigint(v)?);
|
||||
}
|
||||
|
||||
let mut identity_path_index = Vec::new();
|
||||
rln_witness
|
||||
.identity_path_index
|
||||
.iter()
|
||||
.for_each(|v| identity_path_index.push(BigInt::from(*v)));
|
||||
|
||||
Ok([
|
||||
(
|
||||
"identitySecret",
|
||||
vec![to_bigint(&rln_witness.identity_secret)?],
|
||||
),
|
||||
(
|
||||
"userMessageLimit",
|
||||
vec![to_bigint(&rln_witness.user_message_limit)?],
|
||||
),
|
||||
("messageId", vec![to_bigint(&rln_witness.message_id)?]),
|
||||
("pathElements", path_elements),
|
||||
("identityPathIndex", identity_path_index),
|
||||
("x", vec![to_bigint(&rln_witness.x)?]),
|
||||
(
|
||||
"externalNullifier",
|
||||
vec![to_bigint(&rln_witness.external_nullifier)?],
|
||||
),
|
||||
])
|
||||
}
|
||||
|
||||
/// Generates a RLN proof
|
||||
///
|
||||
/// # Errors
|
||||
///
|
||||
/// Returns a [`ProofError`] if proving fails.
|
||||
pub fn generate_proof(
|
||||
#[cfg(not(target_arch = "wasm32"))] witness_calculator: &Mutex<WitnessCalculator>,
|
||||
#[cfg(target_arch = "wasm32")] witness_calculator: &mut WitnessCalculator,
|
||||
proving_key: &(ProvingKey<Curve>, ConstraintMatrices<Fr>),
|
||||
rln_witness: &RLNWitnessInput,
|
||||
) -> Result<ArkProof<Curve>, ProofError> {
|
||||
let inputs = inputs_for_witness_calculation(rln_witness)?
|
||||
.into_iter()
|
||||
.map(|(name, values)| (name.to_string(), values));
|
||||
|
||||
// If in debug mode, we measure and later print time take to compute witness
|
||||
#[cfg(debug_assertions)]
|
||||
let now = Instant::now();
|
||||
|
||||
cfg_if! {
|
||||
if #[cfg(target_arch = "wasm32")] {
|
||||
let full_assignment = witness_calculator
|
||||
.calculate_witness_element::<Curve, _>(inputs, false)
|
||||
.map_err(ProofError::WitnessError)?;
|
||||
} else {
|
||||
let full_assignment = witness_calculator
|
||||
.lock()
|
||||
.expect("witness_calculator mutex should not get poisoned")
|
||||
.calculate_witness_element::<Curve, _>(inputs, false)
|
||||
.map_err(ProofError::WitnessError)?;
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(debug_assertions)]
|
||||
println!("witness generation took: {:.2?}", now.elapsed());
|
||||
|
||||
// Random Values
|
||||
let mut rng = thread_rng();
|
||||
let r = Fr::rand(&mut rng);
|
||||
let s = Fr::rand(&mut rng);
|
||||
|
||||
// If in debug mode, we measure and later print time take to compute proof
|
||||
#[cfg(debug_assertions)]
|
||||
let now = Instant::now();
|
||||
let proof = Groth16::<_, CircomReduction>::create_proof_with_reduction_and_matrices(
|
||||
&proving_key.0,
|
||||
r,
|
||||
s,
|
||||
&proving_key.1,
|
||||
proving_key.1.num_instance_variables,
|
||||
proving_key.1.num_constraints,
|
||||
full_assignment.as_slice(),
|
||||
)?;
|
||||
|
||||
#[cfg(debug_assertions)]
|
||||
println!("proof generation took: {:.2?}", now.elapsed());
|
||||
|
||||
Ok(proof)
|
||||
}
|
||||
|
||||
/// Verifies a given RLN proof
|
||||
///
|
||||
/// # Errors
|
||||
///
|
||||
/// Returns a [`ProofError`] if verifying fails. Verification failure does not
|
||||
/// necessarily mean the proof is incorrect.
|
||||
pub fn verify_proof(
|
||||
verifying_key: &VerifyingKey<Curve>,
|
||||
proof: &ArkProof<Curve>,
|
||||
proof_values: &RLNProofValues,
|
||||
) -> Result<bool, ProofError> {
|
||||
// We re-arrange proof-values according to the circuit specification
|
||||
let inputs = vec![
|
||||
proof_values.y,
|
||||
proof_values.root,
|
||||
proof_values.nullifier,
|
||||
proof_values.x,
|
||||
proof_values.external_nullifier,
|
||||
];
|
||||
|
||||
// Check that the proof is valid
|
||||
let pvk = prepare_verifying_key(verifying_key);
|
||||
//let pr: ArkProof<Curve> = (*proof).into();
|
||||
|
||||
// If in debug mode, we measure and later print time take to verify proof
|
||||
#[cfg(debug_assertions)]
|
||||
let now = Instant::now();
|
||||
|
||||
let verified = Groth16::<_, CircomReduction>::verify_proof(&pvk, proof, &inputs)?;
|
||||
|
||||
#[cfg(debug_assertions)]
|
||||
println!("verify took: {:.2?}", now.elapsed());
|
||||
|
||||
Ok(verified)
|
||||
}
|
||||
|
||||
// auxiliary function for serialisation Fr to json using ark serilize
|
||||
fn ark_se<S, A: CanonicalSerialize>(a: &A, s: S) -> Result<S::Ok, S::Error>
|
||||
where
|
||||
S: serde::Serializer,
|
||||
{
|
||||
let mut bytes = vec![];
|
||||
a.serialize_compressed(&mut bytes)
|
||||
.map_err(serde::ser::Error::custom)?;
|
||||
s.serialize_bytes(&bytes)
|
||||
}
|
||||
|
||||
// auxiliary function for deserialisation Fr to json using ark serilize
|
||||
fn ark_de<'de, D, A: CanonicalDeserialize>(data: D) -> Result<A, D::Error>
|
||||
where
|
||||
D: serde::de::Deserializer<'de>,
|
||||
{
|
||||
let s: Vec<u8> = serde::de::Deserialize::deserialize(data)?;
|
||||
let a = A::deserialize_compressed_unchecked(s.as_slice());
|
||||
a.map_err(serde::de::Error::custom)
|
||||
}
|
||||
|
||||
/// Converts a JSON value into [`RLNWitnessInput`](crate::protocol::RLNWitnessInput) object.
|
||||
///
|
||||
/// # Errors
|
||||
///
|
||||
/// Returns an error if `rln_witness.message_id` is not within `rln_witness.user_message_limit`.
|
||||
pub fn rln_witness_from_json(input_json: serde_json::Value) -> Result<RLNWitnessInput> {
|
||||
let rln_witness: RLNWitnessInput = serde_json::from_value(input_json).unwrap();
|
||||
message_id_range_check(&rln_witness.message_id, &rln_witness.user_message_limit)?;
|
||||
|
||||
Ok(rln_witness)
|
||||
}
|
||||
|
||||
/// Converts a [`RLNWitnessInput`](crate::protocol::RLNWitnessInput) object to the corresponding JSON serialization.
|
||||
///
|
||||
/// # Errors
|
||||
///
|
||||
/// Returns an error if `message_id` is not within `user_message_limit`.
|
||||
pub fn rln_witness_to_json(rln_witness: &RLNWitnessInput) -> Result<serde_json::Value> {
|
||||
message_id_range_check(&rln_witness.message_id, &rln_witness.user_message_limit)?;
|
||||
|
||||
let rln_witness_json = serde_json::to_value(rln_witness)?;
|
||||
Ok(rln_witness_json)
|
||||
}
|
||||
|
||||
/// Converts a [`RLNWitnessInput`](crate::protocol::RLNWitnessInput) object to the corresponding JSON serialization.
|
||||
/// Before serialisation the data should be translated into big int for further calculation in the witness calculator.
|
||||
///
|
||||
/// # Errors
|
||||
///
|
||||
/// Returns an error if `message_id` is not within `user_message_limit`.
|
||||
pub fn rln_witness_to_bigint_json(rln_witness: &RLNWitnessInput) -> Result<serde_json::Value> {
|
||||
message_id_range_check(&rln_witness.message_id, &rln_witness.user_message_limit)?;
|
||||
|
||||
let mut path_elements = Vec::new();
|
||||
|
||||
for v in rln_witness.path_elements.iter() {
|
||||
path_elements.push(to_bigint(v)?.to_str_radix(10));
|
||||
}
|
||||
|
||||
let mut identity_path_index = Vec::new();
|
||||
rln_witness
|
||||
.identity_path_index
|
||||
.iter()
|
||||
.for_each(|v| identity_path_index.push(BigInt::from(*v).to_str_radix(10)));
|
||||
|
||||
let inputs = serde_json::json!({
|
||||
"identitySecret": to_bigint(&rln_witness.identity_secret)?.to_str_radix(10),
|
||||
"userMessageLimit": to_bigint(&rln_witness.user_message_limit)?.to_str_radix(10),
|
||||
"messageId": to_bigint(&rln_witness.message_id)?.to_str_radix(10),
|
||||
"pathElements": path_elements,
|
||||
"identityPathIndex": identity_path_index,
|
||||
"x": to_bigint(&rln_witness.x)?.to_str_radix(10),
|
||||
"externalNullifier": to_bigint(&rln_witness.external_nullifier)?.to_str_radix(10),
|
||||
});
|
||||
|
||||
Ok(inputs)
|
||||
}
|
||||
|
||||
pub fn message_id_range_check(message_id: &Fr, user_message_limit: &Fr) -> Result<()> {
|
||||
if message_id > user_message_limit {
|
||||
return Err(color_eyre::Report::msg(
|
||||
"message_id is not within user_message_limit",
|
||||
));
|
||||
}
|
||||
Ok(())
|
||||
}
|
||||
80
rln/src/protocol/keygen.rs
Normal file
80
rln/src/protocol/keygen.rs
Normal file
@@ -0,0 +1,80 @@
|
||||
use ark_std::{rand::thread_rng, UniformRand};
|
||||
use rand::SeedableRng;
|
||||
use rand_chacha::ChaCha20Rng;
|
||||
use tiny_keccak::{Hasher as _, Keccak};
|
||||
use zerokit_utils::error::ZerokitMerkleTreeError;
|
||||
|
||||
use crate::{circuit::Fr, hashers::poseidon_hash, utils::IdSecret};
|
||||
|
||||
/// Generates a random RLN identity using a cryptographically secure RNG.
|
||||
///
|
||||
/// Returns `(identity_secret, id_commitment)` where the commitment is `PoseidonHash(identity_secret)`.
|
||||
pub fn keygen() -> Result<(IdSecret, Fr), ZerokitMerkleTreeError> {
|
||||
let mut rng = thread_rng();
|
||||
let identity_secret = IdSecret::rand(&mut rng);
|
||||
let id_commitment = poseidon_hash(&[*identity_secret.clone()])?;
|
||||
Ok((identity_secret, id_commitment))
|
||||
}
|
||||
|
||||
/// Generates an extended RLN identity compatible with Semaphore.
|
||||
///
|
||||
/// Returns `(identity_trapdoor, identity_nullifier, identity_secret, id_commitment)` where:
|
||||
/// - `identity_secret = PoseidonHash(identity_trapdoor, identity_nullifier)`
|
||||
/// - `id_commitment = PoseidonHash(identity_secret)`
|
||||
pub fn extended_keygen() -> Result<(Fr, Fr, Fr, Fr), ZerokitMerkleTreeError> {
|
||||
let mut rng = thread_rng();
|
||||
let identity_trapdoor = Fr::rand(&mut rng);
|
||||
let identity_nullifier = Fr::rand(&mut rng);
|
||||
let identity_secret = poseidon_hash(&[identity_trapdoor, identity_nullifier])?;
|
||||
let id_commitment = poseidon_hash(&[identity_secret])?;
|
||||
Ok((
|
||||
identity_trapdoor,
|
||||
identity_nullifier,
|
||||
identity_secret,
|
||||
id_commitment,
|
||||
))
|
||||
}
|
||||
|
||||
/// Generates a deterministic RLN identity from a seed.
|
||||
///
|
||||
/// Uses ChaCha20 RNG seeded with Keccak-256 hash of the input.
|
||||
/// Returns `(identity_secret, id_commitment)`. Same input always produces the same identity.
|
||||
pub fn seeded_keygen(signal: &[u8]) -> Result<(Fr, Fr), ZerokitMerkleTreeError> {
|
||||
// ChaCha20 requires a seed of exactly 32 bytes.
|
||||
// We first hash the input seed signal to a 32 bytes array and pass this as seed to ChaCha20
|
||||
let mut seed = [0; 32];
|
||||
let mut hasher = Keccak::v256();
|
||||
hasher.update(signal);
|
||||
hasher.finalize(&mut seed);
|
||||
|
||||
let mut rng = ChaCha20Rng::from_seed(seed);
|
||||
let identity_secret = Fr::rand(&mut rng);
|
||||
let id_commitment = poseidon_hash(&[identity_secret])?;
|
||||
Ok((identity_secret, id_commitment))
|
||||
}
|
||||
|
||||
/// Generates a deterministic extended RLN identity from a seed, compatible with Semaphore.
|
||||
///
|
||||
/// Uses ChaCha20 RNG seeded with Keccak-256 hash of the input.
|
||||
/// Returns `(identity_trapdoor, identity_nullifier, identity_secret, id_commitment)`.
|
||||
/// Same input always produces the same identity.
|
||||
pub fn extended_seeded_keygen(signal: &[u8]) -> Result<(Fr, Fr, Fr, Fr), ZerokitMerkleTreeError> {
|
||||
// ChaCha20 requires a seed of exactly 32 bytes.
|
||||
// We first hash the input seed signal to a 32 bytes array and pass this as seed to ChaCha20
|
||||
let mut seed = [0; 32];
|
||||
let mut hasher = Keccak::v256();
|
||||
hasher.update(signal);
|
||||
hasher.finalize(&mut seed);
|
||||
|
||||
let mut rng = ChaCha20Rng::from_seed(seed);
|
||||
let identity_trapdoor = Fr::rand(&mut rng);
|
||||
let identity_nullifier = Fr::rand(&mut rng);
|
||||
let identity_secret = poseidon_hash(&[identity_trapdoor, identity_nullifier])?;
|
||||
let id_commitment = poseidon_hash(&[identity_secret])?;
|
||||
Ok((
|
||||
identity_trapdoor,
|
||||
identity_nullifier,
|
||||
identity_secret,
|
||||
id_commitment,
|
||||
))
|
||||
}
|
||||
19
rln/src/protocol/mod.rs
Normal file
19
rln/src/protocol/mod.rs
Normal file
@@ -0,0 +1,19 @@
|
||||
// This crate collects all the underlying primitives used to implement RLN
|
||||
|
||||
mod keygen;
|
||||
mod proof;
|
||||
mod slashing;
|
||||
mod witness;
|
||||
|
||||
pub use keygen::{extended_keygen, extended_seeded_keygen, keygen, seeded_keygen};
|
||||
pub use proof::{
|
||||
bytes_be_to_rln_proof, bytes_be_to_rln_proof_values, bytes_le_to_rln_proof,
|
||||
bytes_le_to_rln_proof_values, generate_zk_proof, generate_zk_proof_with_witness,
|
||||
rln_proof_to_bytes_be, rln_proof_to_bytes_le, rln_proof_values_to_bytes_be,
|
||||
rln_proof_values_to_bytes_le, verify_zk_proof, RLNProof, RLNProofValues,
|
||||
};
|
||||
pub use slashing::recover_id_secret;
|
||||
pub use witness::{
|
||||
bytes_be_to_rln_witness, bytes_le_to_rln_witness, compute_tree_root, proof_values_from_witness,
|
||||
rln_witness_to_bigint_json, rln_witness_to_bytes_be, rln_witness_to_bytes_le, RLNWitnessInput,
|
||||
};
|
||||
345
rln/src/protocol/proof.rs
Normal file
345
rln/src/protocol/proof.rs
Normal file
@@ -0,0 +1,345 @@
|
||||
use ark_ff::PrimeField;
|
||||
use ark_groth16::{prepare_verifying_key, Groth16};
|
||||
use ark_serialize::{CanonicalDeserialize, CanonicalSerialize};
|
||||
use ark_std::{rand::thread_rng, UniformRand};
|
||||
use num_bigint::BigInt;
|
||||
use num_traits::Signed;
|
||||
|
||||
use super::witness::{inputs_for_witness_calculation, RLNWitnessInput};
|
||||
use crate::{
|
||||
circuit::{
|
||||
iden3calc::calc_witness, qap::CircomReduction, Curve, Fr, Proof, VerifyingKey, Zkey,
|
||||
COMPRESS_PROOF_SIZE,
|
||||
},
|
||||
error::ProtocolError,
|
||||
utils::{bytes_be_to_fr, bytes_le_to_fr, fr_to_bytes_be, fr_to_bytes_le, FR_BYTE_SIZE},
|
||||
};
|
||||
|
||||
/// Complete RLN proof.
|
||||
///
|
||||
/// Combines the Groth16 proof with its public values.
|
||||
#[derive(Debug, PartialEq, Clone)]
|
||||
pub struct RLNProof {
|
||||
pub proof: Proof,
|
||||
pub proof_values: RLNProofValues,
|
||||
}
|
||||
|
||||
/// Public values for RLN proof verification.
|
||||
///
|
||||
/// Contains the circuit's public inputs and outputs. Used in proof verification
|
||||
/// and identity secret recovery when rate limit violations are detected.
|
||||
#[derive(Debug, PartialEq, Clone, Copy)]
|
||||
pub struct RLNProofValues {
|
||||
// Public outputs:
|
||||
pub y: Fr,
|
||||
pub nullifier: Fr,
|
||||
pub root: Fr,
|
||||
// Public Inputs:
|
||||
pub x: Fr,
|
||||
pub external_nullifier: Fr,
|
||||
}
|
||||
|
||||
/// Serializes RLN proof values to little-endian bytes.
|
||||
pub fn rln_proof_values_to_bytes_le(rln_proof_values: &RLNProofValues) -> Vec<u8> {
|
||||
// Calculate capacity for Vec:
|
||||
// 5 field elements: root, external_nullifier, x, y, nullifier
|
||||
let mut bytes = Vec::with_capacity(FR_BYTE_SIZE * 5);
|
||||
|
||||
bytes.extend_from_slice(&fr_to_bytes_le(&rln_proof_values.root));
|
||||
bytes.extend_from_slice(&fr_to_bytes_le(&rln_proof_values.external_nullifier));
|
||||
bytes.extend_from_slice(&fr_to_bytes_le(&rln_proof_values.x));
|
||||
bytes.extend_from_slice(&fr_to_bytes_le(&rln_proof_values.y));
|
||||
bytes.extend_from_slice(&fr_to_bytes_le(&rln_proof_values.nullifier));
|
||||
|
||||
bytes
|
||||
}
|
||||
|
||||
/// Serializes RLN proof values to big-endian bytes.
|
||||
pub fn rln_proof_values_to_bytes_be(rln_proof_values: &RLNProofValues) -> Vec<u8> {
|
||||
// Calculate capacity for Vec:
|
||||
// 5 field elements: root, external_nullifier, x, y, nullifier
|
||||
let mut bytes = Vec::with_capacity(FR_BYTE_SIZE * 5);
|
||||
|
||||
bytes.extend_from_slice(&fr_to_bytes_be(&rln_proof_values.root));
|
||||
bytes.extend_from_slice(&fr_to_bytes_be(&rln_proof_values.external_nullifier));
|
||||
bytes.extend_from_slice(&fr_to_bytes_be(&rln_proof_values.x));
|
||||
bytes.extend_from_slice(&fr_to_bytes_be(&rln_proof_values.y));
|
||||
bytes.extend_from_slice(&fr_to_bytes_be(&rln_proof_values.nullifier));
|
||||
|
||||
bytes
|
||||
}
|
||||
|
||||
/// Deserializes RLN proof values from little-endian bytes.
|
||||
///
|
||||
/// Format: `[ root<32> | external_nullifier<32> | x<32> | y<32> | nullifier<32> ]`
|
||||
///
|
||||
/// Returns the deserialized proof values and the number of bytes read.
|
||||
pub fn bytes_le_to_rln_proof_values(
|
||||
bytes: &[u8],
|
||||
) -> Result<(RLNProofValues, usize), ProtocolError> {
|
||||
let mut read: usize = 0;
|
||||
|
||||
let (root, el_size) = bytes_le_to_fr(&bytes[read..])?;
|
||||
read += el_size;
|
||||
|
||||
let (external_nullifier, el_size) = bytes_le_to_fr(&bytes[read..])?;
|
||||
read += el_size;
|
||||
|
||||
let (x, el_size) = bytes_le_to_fr(&bytes[read..])?;
|
||||
read += el_size;
|
||||
|
||||
let (y, el_size) = bytes_le_to_fr(&bytes[read..])?;
|
||||
read += el_size;
|
||||
|
||||
let (nullifier, el_size) = bytes_le_to_fr(&bytes[read..])?;
|
||||
read += el_size;
|
||||
|
||||
Ok((
|
||||
RLNProofValues {
|
||||
y,
|
||||
nullifier,
|
||||
root,
|
||||
x,
|
||||
external_nullifier,
|
||||
},
|
||||
read,
|
||||
))
|
||||
}
|
||||
|
||||
/// Deserializes RLN proof values from big-endian bytes.
|
||||
///
|
||||
/// Format: `[ root<32> | external_nullifier<32> | x<32> | y<32> | nullifier<32> ]`
|
||||
///
|
||||
/// Returns the deserialized proof values and the number of bytes read.
|
||||
pub fn bytes_be_to_rln_proof_values(
|
||||
bytes: &[u8],
|
||||
) -> Result<(RLNProofValues, usize), ProtocolError> {
|
||||
let mut read: usize = 0;
|
||||
|
||||
let (root, el_size) = bytes_be_to_fr(&bytes[read..])?;
|
||||
read += el_size;
|
||||
|
||||
let (external_nullifier, el_size) = bytes_be_to_fr(&bytes[read..])?;
|
||||
read += el_size;
|
||||
|
||||
let (x, el_size) = bytes_be_to_fr(&bytes[read..])?;
|
||||
read += el_size;
|
||||
|
||||
let (y, el_size) = bytes_be_to_fr(&bytes[read..])?;
|
||||
read += el_size;
|
||||
|
||||
let (nullifier, el_size) = bytes_be_to_fr(&bytes[read..])?;
|
||||
read += el_size;
|
||||
|
||||
Ok((
|
||||
RLNProofValues {
|
||||
y,
|
||||
nullifier,
|
||||
root,
|
||||
x,
|
||||
external_nullifier,
|
||||
},
|
||||
read,
|
||||
))
|
||||
}
|
||||
|
||||
/// Serializes RLN proof to little-endian bytes.
|
||||
///
|
||||
/// Note: The Groth16 proof is always serialized in LE format (arkworks behavior),
|
||||
/// while proof_values are serialized in LE format.
|
||||
pub fn rln_proof_to_bytes_le(rln_proof: &RLNProof) -> Result<Vec<u8>, ProtocolError> {
|
||||
// Calculate capacity for Vec:
|
||||
// - 128 bytes for compressed Groth16 proof
|
||||
// - 5 field elements for proof values (root, external_nullifier, x, y, nullifier)
|
||||
let mut bytes = Vec::with_capacity(COMPRESS_PROOF_SIZE + FR_BYTE_SIZE * 5);
|
||||
|
||||
// Serialize proof (always LE format from arkworks)
|
||||
rln_proof.proof.serialize_compressed(&mut bytes)?;
|
||||
|
||||
// Serialize proof values in LE
|
||||
let proof_values_bytes = rln_proof_values_to_bytes_le(&rln_proof.proof_values);
|
||||
bytes.extend_from_slice(&proof_values_bytes);
|
||||
|
||||
Ok(bytes)
|
||||
}
|
||||
|
||||
/// Serializes RLN proof to big-endian bytes.
|
||||
///
|
||||
/// Note: The Groth16 proof is always serialized in LE format (arkworks behavior),
|
||||
/// while proof_values are serialized in BE format. This creates a mixed-endian format.
|
||||
pub fn rln_proof_to_bytes_be(rln_proof: &RLNProof) -> Result<Vec<u8>, ProtocolError> {
|
||||
// Calculate capacity for Vec:
|
||||
// - 128 bytes for compressed Groth16 proof
|
||||
// - 5 field elements for proof values (root, external_nullifier, x, y, nullifier)
|
||||
let mut bytes = Vec::with_capacity(COMPRESS_PROOF_SIZE + FR_BYTE_SIZE * 5);
|
||||
|
||||
// Serialize proof (always LE format from arkworks)
|
||||
rln_proof.proof.serialize_compressed(&mut bytes)?;
|
||||
|
||||
// Serialize proof values in BE
|
||||
let proof_values_bytes = rln_proof_values_to_bytes_be(&rln_proof.proof_values);
|
||||
bytes.extend_from_slice(&proof_values_bytes);
|
||||
|
||||
Ok(bytes)
|
||||
}
|
||||
|
||||
/// Deserializes RLN proof from little-endian bytes.
|
||||
///
|
||||
/// Format: `[ proof<128,LE> | root<32,LE> | external_nullifier<32,LE> | x<32,LE> | y<32,LE> | nullifier<32,LE> ]`
|
||||
///
|
||||
/// Returns the deserialized proof and the number of bytes read.
|
||||
pub fn bytes_le_to_rln_proof(bytes: &[u8]) -> Result<(RLNProof, usize), ProtocolError> {
|
||||
let mut read: usize = 0;
|
||||
|
||||
// Deserialize proof (always LE from arkworks)
|
||||
let proof = Proof::deserialize_compressed(&bytes[read..read + COMPRESS_PROOF_SIZE])?;
|
||||
read += COMPRESS_PROOF_SIZE;
|
||||
|
||||
// Deserialize proof values
|
||||
let (values, el_size) = bytes_le_to_rln_proof_values(&bytes[read..])?;
|
||||
read += el_size;
|
||||
|
||||
Ok((
|
||||
RLNProof {
|
||||
proof,
|
||||
proof_values: values,
|
||||
},
|
||||
read,
|
||||
))
|
||||
}
|
||||
|
||||
/// Deserializes RLN proof from big-endian bytes.
|
||||
///
|
||||
/// Format: `[ proof<128,LE> | root<32,BE> | external_nullifier<32,BE> | x<32,BE> | y<32,BE> | nullifier<32,BE> ]`
|
||||
///
|
||||
/// Note: Mixed-endian format - proof is LE (arkworks), proof_values are BE.
|
||||
///
|
||||
/// Returns the deserialized proof and the number of bytes read.
|
||||
pub fn bytes_be_to_rln_proof(bytes: &[u8]) -> Result<(RLNProof, usize), ProtocolError> {
|
||||
let mut read: usize = 0;
|
||||
|
||||
// Deserialize proof (always LE from arkworks)
|
||||
let proof = Proof::deserialize_compressed(&bytes[read..read + COMPRESS_PROOF_SIZE])?;
|
||||
read += COMPRESS_PROOF_SIZE;
|
||||
|
||||
// Deserialize proof values
|
||||
let (values, el_size) = bytes_be_to_rln_proof_values(&bytes[read..])?;
|
||||
read += el_size;
|
||||
|
||||
Ok((
|
||||
RLNProof {
|
||||
proof,
|
||||
proof_values: values,
|
||||
},
|
||||
read,
|
||||
))
|
||||
}
|
||||
|
||||
// zkSNARK proof generation and verification
|
||||
|
||||
/// Converts calculated witness (BigInt) to field elements.
|
||||
fn calculated_witness_to_field_elements<E: ark_ec::pairing::Pairing>(
|
||||
calculated_witness: Vec<BigInt>,
|
||||
) -> Result<Vec<E::ScalarField>, ProtocolError> {
|
||||
let modulus = <E::ScalarField as PrimeField>::MODULUS;
|
||||
|
||||
// convert it to field elements
|
||||
let mut field_elements = vec![];
|
||||
for w in calculated_witness.into_iter() {
|
||||
let w = if w.sign() == num_bigint::Sign::Minus {
|
||||
// Need to negate the witness element if negative
|
||||
modulus.into()
|
||||
- w.abs()
|
||||
.to_biguint()
|
||||
.ok_or(ProtocolError::BigUintConversion(w))?
|
||||
} else {
|
||||
w.to_biguint().ok_or(ProtocolError::BigUintConversion(w))?
|
||||
};
|
||||
field_elements.push(E::ScalarField::from(w))
|
||||
}
|
||||
|
||||
Ok(field_elements)
|
||||
}
|
||||
|
||||
/// Generates a zkSNARK proof from pre-calculated witness values.
|
||||
///
|
||||
/// Use this when witness calculation is performed externally.
|
||||
pub fn generate_zk_proof_with_witness(
|
||||
calculated_witness: Vec<BigInt>,
|
||||
zkey: &Zkey,
|
||||
) -> Result<Proof, ProtocolError> {
|
||||
let full_assignment = calculated_witness_to_field_elements::<Curve>(calculated_witness)?;
|
||||
|
||||
// Random Values
|
||||
let mut rng = thread_rng();
|
||||
let r = Fr::rand(&mut rng);
|
||||
let s = Fr::rand(&mut rng);
|
||||
|
||||
let proof = Groth16::<_, CircomReduction>::create_proof_with_reduction_and_matrices(
|
||||
&zkey.0,
|
||||
r,
|
||||
s,
|
||||
&zkey.1,
|
||||
zkey.1.num_instance_variables,
|
||||
zkey.1.num_constraints,
|
||||
full_assignment.as_slice(),
|
||||
)?;
|
||||
|
||||
Ok(proof)
|
||||
}
|
||||
|
||||
/// Generates a zkSNARK proof from witness input using the provided circuit data.
|
||||
pub fn generate_zk_proof(
|
||||
zkey: &Zkey,
|
||||
witness: &RLNWitnessInput,
|
||||
graph_data: &[u8],
|
||||
) -> Result<Proof, ProtocolError> {
|
||||
let inputs = inputs_for_witness_calculation(witness)?
|
||||
.into_iter()
|
||||
.map(|(name, values)| (name.to_string(), values));
|
||||
|
||||
let full_assignment = calc_witness(inputs, graph_data)?;
|
||||
|
||||
// Random Values
|
||||
let mut rng = thread_rng();
|
||||
let r = Fr::rand(&mut rng);
|
||||
let s = Fr::rand(&mut rng);
|
||||
|
||||
let proof = Groth16::<_, CircomReduction>::create_proof_with_reduction_and_matrices(
|
||||
&zkey.0,
|
||||
r,
|
||||
s,
|
||||
&zkey.1,
|
||||
zkey.1.num_instance_variables,
|
||||
zkey.1.num_constraints,
|
||||
full_assignment.as_slice(),
|
||||
)?;
|
||||
|
||||
Ok(proof)
|
||||
}
|
||||
|
||||
/// Verifies a zkSNARK proof against the verifying key and public values.
|
||||
///
|
||||
/// Returns `true` if the proof is cryptographically valid, `false` if verification fails.
|
||||
/// Note: Verification failure may occur due to proof computation errors, not necessarily malicious proofs.
|
||||
pub fn verify_zk_proof(
|
||||
verifying_key: &VerifyingKey,
|
||||
proof: &Proof,
|
||||
proof_values: &RLNProofValues,
|
||||
) -> Result<bool, ProtocolError> {
|
||||
// We re-arrange proof-values according to the circuit specification
|
||||
let inputs = vec![
|
||||
proof_values.y,
|
||||
proof_values.root,
|
||||
proof_values.nullifier,
|
||||
proof_values.x,
|
||||
proof_values.external_nullifier,
|
||||
];
|
||||
|
||||
// Check that the proof is valid
|
||||
let pvk = prepare_verifying_key(verifying_key);
|
||||
|
||||
let verified = Groth16::<_, CircomReduction>::verify_proof(&pvk, proof, &inputs)?;
|
||||
|
||||
Ok(verified)
|
||||
}
|
||||
55
rln/src/protocol/slashing.rs
Normal file
55
rln/src/protocol/slashing.rs
Normal file
@@ -0,0 +1,55 @@
|
||||
use ark_ff::AdditiveGroup;
|
||||
|
||||
use super::proof::RLNProofValues;
|
||||
use crate::{circuit::Fr, error::ProtocolError, utils::IdSecret};
|
||||
|
||||
/// Computes identity secret from two (x, y) shares.
|
||||
fn compute_id_secret(share1: (Fr, Fr), share2: (Fr, Fr)) -> Result<IdSecret, ProtocolError> {
|
||||
// Assuming a0 is the identity secret and a1 = poseidonHash([a0, external_nullifier]),
|
||||
// a (x,y) share satisfies the following relation
|
||||
// y = a_0 + x * a_1
|
||||
let (x1, y1) = share1;
|
||||
let (x2, y2) = share2;
|
||||
|
||||
// If the two input shares were computed for the same external_nullifier and identity secret, we can recover the latter
|
||||
// y1 = a_0 + x1 * a_1
|
||||
// y2 = a_0 + x2 * a_1
|
||||
|
||||
if (x1 - x2) != Fr::ZERO {
|
||||
let a_1 = (y1 - y2) / (x1 - x2);
|
||||
let mut a_0 = y1 - x1 * a_1;
|
||||
|
||||
// If shares come from the same polynomial, a0 is correctly recovered and a1 = poseidonHash([a0, external_nullifier])
|
||||
let id_secret = IdSecret::from(&mut a_0);
|
||||
Ok(id_secret)
|
||||
} else {
|
||||
Err(ProtocolError::DivisionByZero)
|
||||
}
|
||||
}
|
||||
|
||||
/// Recovers identity secret from two proof shares with the same external nullifier.
|
||||
///
|
||||
/// When a user violates rate limits by generating multiple proofs in the same epoch,
|
||||
/// their shares can be used to recover their identity secret through polynomial interpolation.
|
||||
pub fn recover_id_secret(
|
||||
rln_proof_values_1: &RLNProofValues,
|
||||
rln_proof_values_2: &RLNProofValues,
|
||||
) -> Result<IdSecret, ProtocolError> {
|
||||
let external_nullifier_1 = rln_proof_values_1.external_nullifier;
|
||||
let external_nullifier_2 = rln_proof_values_2.external_nullifier;
|
||||
|
||||
// We continue only if the proof values are for the same external nullifier
|
||||
if external_nullifier_1 != external_nullifier_2 {
|
||||
return Err(ProtocolError::ExternalNullifierMismatch(
|
||||
external_nullifier_1,
|
||||
external_nullifier_2,
|
||||
));
|
||||
}
|
||||
|
||||
// We extract the two shares
|
||||
let share1 = (rln_proof_values_1.x, rln_proof_values_1.y);
|
||||
let share2 = (rln_proof_values_2.x, rln_proof_values_2.y);
|
||||
|
||||
// We recover the secret
|
||||
compute_id_secret(share1, share2)
|
||||
}
|
||||
354
rln/src/protocol/witness.rs
Normal file
354
rln/src/protocol/witness.rs
Normal file
@@ -0,0 +1,354 @@
|
||||
use zeroize::Zeroize;
|
||||
|
||||
use super::proof::RLNProofValues;
|
||||
use crate::{
|
||||
circuit::Fr,
|
||||
error::ProtocolError,
|
||||
hashers::poseidon_hash,
|
||||
utils::{
|
||||
bytes_be_to_fr, bytes_be_to_vec_fr, bytes_be_to_vec_u8, bytes_le_to_fr, bytes_le_to_vec_fr,
|
||||
bytes_le_to_vec_u8, fr_to_bytes_be, fr_to_bytes_le, to_bigint, vec_fr_to_bytes_be,
|
||||
vec_fr_to_bytes_le, vec_u8_to_bytes_be, vec_u8_to_bytes_le, FrOrSecret, IdSecret,
|
||||
FR_BYTE_SIZE,
|
||||
},
|
||||
};
|
||||
|
||||
/// Witness input for RLN proof generation.
|
||||
///
|
||||
/// Contains the identity credentials, merkle proof, rate-limiting parameters,
|
||||
/// and signal binding data required to generate a Groth16 proof for the RLN protocol.
|
||||
#[derive(Debug, PartialEq, Clone)]
|
||||
pub struct RLNWitnessInput {
|
||||
identity_secret: IdSecret,
|
||||
user_message_limit: Fr,
|
||||
message_id: Fr,
|
||||
path_elements: Vec<Fr>,
|
||||
identity_path_index: Vec<u8>,
|
||||
x: Fr,
|
||||
external_nullifier: Fr,
|
||||
}
|
||||
|
||||
impl RLNWitnessInput {
|
||||
pub fn new(
|
||||
identity_secret: IdSecret,
|
||||
user_message_limit: Fr,
|
||||
message_id: Fr,
|
||||
path_elements: Vec<Fr>,
|
||||
identity_path_index: Vec<u8>,
|
||||
x: Fr,
|
||||
external_nullifier: Fr,
|
||||
) -> Result<Self, ProtocolError> {
|
||||
// Message ID range check
|
||||
if message_id > user_message_limit {
|
||||
return Err(ProtocolError::InvalidMessageId(
|
||||
message_id,
|
||||
user_message_limit,
|
||||
));
|
||||
}
|
||||
|
||||
// Merkle proof length check
|
||||
let path_elements_len = path_elements.len();
|
||||
let identity_path_index_len = identity_path_index.len();
|
||||
if path_elements_len != identity_path_index_len {
|
||||
return Err(ProtocolError::InvalidMerkleProofLength(
|
||||
path_elements_len,
|
||||
identity_path_index_len,
|
||||
));
|
||||
}
|
||||
|
||||
Ok(Self {
|
||||
identity_secret,
|
||||
user_message_limit,
|
||||
message_id,
|
||||
path_elements,
|
||||
identity_path_index,
|
||||
x,
|
||||
external_nullifier,
|
||||
})
|
||||
}
|
||||
|
||||
pub fn identity_secret(&self) -> &IdSecret {
|
||||
&self.identity_secret
|
||||
}
|
||||
|
||||
pub fn user_message_limit(&self) -> &Fr {
|
||||
&self.user_message_limit
|
||||
}
|
||||
|
||||
pub fn message_id(&self) -> &Fr {
|
||||
&self.message_id
|
||||
}
|
||||
|
||||
pub fn path_elements(&self) -> &[Fr] {
|
||||
&self.path_elements
|
||||
}
|
||||
|
||||
pub fn identity_path_index(&self) -> &[u8] {
|
||||
&self.identity_path_index
|
||||
}
|
||||
|
||||
pub fn x(&self) -> &Fr {
|
||||
&self.x
|
||||
}
|
||||
|
||||
pub fn external_nullifier(&self) -> &Fr {
|
||||
&self.external_nullifier
|
||||
}
|
||||
}
|
||||
|
||||
/// Serializes an RLN witness to little-endian bytes.
|
||||
pub fn rln_witness_to_bytes_le(witness: &RLNWitnessInput) -> Result<Vec<u8>, ProtocolError> {
|
||||
// Calculate capacity for Vec:
|
||||
// - 5 fixed field elements: identity_secret, user_message_limit, message_id, x, external_nullifier
|
||||
// - variable number of path elements
|
||||
// - identity_path_index (variable size)
|
||||
let mut bytes: Vec<u8> = Vec::with_capacity(
|
||||
FR_BYTE_SIZE * (5 + witness.path_elements.len()) + witness.identity_path_index.len(),
|
||||
);
|
||||
bytes.extend_from_slice(&witness.identity_secret.to_bytes_le());
|
||||
bytes.extend_from_slice(&fr_to_bytes_le(&witness.user_message_limit));
|
||||
bytes.extend_from_slice(&fr_to_bytes_le(&witness.message_id));
|
||||
bytes.extend_from_slice(&vec_fr_to_bytes_le(&witness.path_elements));
|
||||
bytes.extend_from_slice(&vec_u8_to_bytes_le(&witness.identity_path_index));
|
||||
bytes.extend_from_slice(&fr_to_bytes_le(&witness.x));
|
||||
bytes.extend_from_slice(&fr_to_bytes_le(&witness.external_nullifier));
|
||||
|
||||
Ok(bytes)
|
||||
}
|
||||
|
||||
/// Serializes an RLN witness to big-endian bytes.
|
||||
pub fn rln_witness_to_bytes_be(witness: &RLNWitnessInput) -> Result<Vec<u8>, ProtocolError> {
|
||||
// Calculate capacity for Vec:
|
||||
// - 5 fixed field elements: identity_secret, user_message_limit, message_id, x, external_nullifier
|
||||
// - variable number of path elements
|
||||
// - identity_path_index (variable size)
|
||||
let mut bytes: Vec<u8> = Vec::with_capacity(
|
||||
FR_BYTE_SIZE * (5 + witness.path_elements.len()) + witness.identity_path_index.len(),
|
||||
);
|
||||
bytes.extend_from_slice(&witness.identity_secret.to_bytes_be());
|
||||
bytes.extend_from_slice(&fr_to_bytes_be(&witness.user_message_limit));
|
||||
bytes.extend_from_slice(&fr_to_bytes_be(&witness.message_id));
|
||||
bytes.extend_from_slice(&vec_fr_to_bytes_be(&witness.path_elements));
|
||||
bytes.extend_from_slice(&vec_u8_to_bytes_be(&witness.identity_path_index));
|
||||
bytes.extend_from_slice(&fr_to_bytes_be(&witness.x));
|
||||
bytes.extend_from_slice(&fr_to_bytes_be(&witness.external_nullifier));
|
||||
|
||||
Ok(bytes)
|
||||
}
|
||||
|
||||
/// Deserializes an RLN witness from little-endian bytes.
|
||||
///
|
||||
/// Format: `[ identity_secret<32> | user_message_limit<32> | message_id<32> | path_elements<var> | identity_path_index<var> | x<32> | external_nullifier<32> ]`
|
||||
///
|
||||
/// Returns the deserialized witness and the number of bytes read.
|
||||
pub fn bytes_le_to_rln_witness(bytes: &[u8]) -> Result<(RLNWitnessInput, usize), ProtocolError> {
|
||||
let mut read: usize = 0;
|
||||
|
||||
let (identity_secret, el_size) = IdSecret::from_bytes_le(&bytes[read..])?;
|
||||
read += el_size;
|
||||
|
||||
let (user_message_limit, el_size) = bytes_le_to_fr(&bytes[read..])?;
|
||||
read += el_size;
|
||||
|
||||
let (message_id, el_size) = bytes_le_to_fr(&bytes[read..])?;
|
||||
read += el_size;
|
||||
|
||||
let (path_elements, el_size) = bytes_le_to_vec_fr(&bytes[read..])?;
|
||||
read += el_size;
|
||||
|
||||
let (identity_path_index, el_size) = bytes_le_to_vec_u8(&bytes[read..])?;
|
||||
read += el_size;
|
||||
|
||||
let (x, el_size) = bytes_le_to_fr(&bytes[read..])?;
|
||||
read += el_size;
|
||||
|
||||
let (external_nullifier, el_size) = bytes_le_to_fr(&bytes[read..])?;
|
||||
read += el_size;
|
||||
|
||||
if bytes.len() != read {
|
||||
return Err(ProtocolError::InvalidReadLen(bytes.len(), read));
|
||||
}
|
||||
|
||||
Ok((
|
||||
RLNWitnessInput::new(
|
||||
identity_secret,
|
||||
user_message_limit,
|
||||
message_id,
|
||||
path_elements,
|
||||
identity_path_index,
|
||||
x,
|
||||
external_nullifier,
|
||||
)?,
|
||||
read,
|
||||
))
|
||||
}
|
||||
|
||||
/// Deserializes an RLN witness from big-endian bytes.
|
||||
///
|
||||
/// Format: `[ identity_secret<32> | user_message_limit<32> | message_id<32> | path_elements<var> | identity_path_index<var> | x<32> | external_nullifier<32> ]`
|
||||
///
|
||||
/// Returns the deserialized witness and the number of bytes read.
|
||||
pub fn bytes_be_to_rln_witness(bytes: &[u8]) -> Result<(RLNWitnessInput, usize), ProtocolError> {
|
||||
let mut read: usize = 0;
|
||||
|
||||
let (identity_secret, el_size) = IdSecret::from_bytes_be(&bytes[read..])?;
|
||||
read += el_size;
|
||||
|
||||
let (user_message_limit, el_size) = bytes_be_to_fr(&bytes[read..])?;
|
||||
read += el_size;
|
||||
|
||||
let (message_id, el_size) = bytes_be_to_fr(&bytes[read..])?;
|
||||
read += el_size;
|
||||
|
||||
let (path_elements, el_size) = bytes_be_to_vec_fr(&bytes[read..])?;
|
||||
read += el_size;
|
||||
|
||||
let (identity_path_index, el_size) = bytes_be_to_vec_u8(&bytes[read..])?;
|
||||
read += el_size;
|
||||
|
||||
let (x, el_size) = bytes_be_to_fr(&bytes[read..])?;
|
||||
read += el_size;
|
||||
|
||||
let (external_nullifier, el_size) = bytes_be_to_fr(&bytes[read..])?;
|
||||
read += el_size;
|
||||
|
||||
if bytes.len() != read {
|
||||
return Err(ProtocolError::InvalidReadLen(bytes.len(), read));
|
||||
}
|
||||
|
||||
Ok((
|
||||
RLNWitnessInput::new(
|
||||
identity_secret,
|
||||
user_message_limit,
|
||||
message_id,
|
||||
path_elements,
|
||||
identity_path_index,
|
||||
x,
|
||||
external_nullifier,
|
||||
)?,
|
||||
read,
|
||||
))
|
||||
}
|
||||
|
||||
/// Converts RLN witness to JSON with BigInt string representation for witness calculator.
|
||||
pub fn rln_witness_to_bigint_json(
|
||||
witness: &RLNWitnessInput,
|
||||
) -> Result<serde_json::Value, ProtocolError> {
|
||||
use num_bigint::BigInt;
|
||||
|
||||
let mut path_elements = Vec::new();
|
||||
|
||||
for v in witness.path_elements.iter() {
|
||||
path_elements.push(to_bigint(v).to_str_radix(10));
|
||||
}
|
||||
|
||||
let mut identity_path_index = Vec::new();
|
||||
witness
|
||||
.identity_path_index
|
||||
.iter()
|
||||
.for_each(|v| identity_path_index.push(BigInt::from(*v).to_str_radix(10)));
|
||||
|
||||
let inputs = serde_json::json!({
|
||||
"identitySecret": to_bigint(&witness.identity_secret).to_str_radix(10),
|
||||
"userMessageLimit": to_bigint(&witness.user_message_limit).to_str_radix(10),
|
||||
"messageId": to_bigint(&witness.message_id).to_str_radix(10),
|
||||
"pathElements": path_elements,
|
||||
"identityPathIndex": identity_path_index,
|
||||
"x": to_bigint(&witness.x).to_str_radix(10),
|
||||
"externalNullifier": to_bigint(&witness.external_nullifier).to_str_radix(10),
|
||||
});
|
||||
|
||||
Ok(inputs)
|
||||
}
|
||||
|
||||
/// Computes RLN proof values from witness input.
|
||||
///
|
||||
/// Calculates the public outputs (y, nullifier, root) that will be part of the proof.
|
||||
pub fn proof_values_from_witness(
|
||||
witness: &RLNWitnessInput,
|
||||
) -> Result<RLNProofValues, ProtocolError> {
|
||||
// y share
|
||||
let a_0 = &witness.identity_secret;
|
||||
let mut to_hash = [**a_0, witness.external_nullifier, witness.message_id];
|
||||
let a_1 = poseidon_hash(&to_hash)?;
|
||||
let y = *(a_0.clone()) + witness.x * a_1;
|
||||
|
||||
// Nullifier
|
||||
let nullifier = poseidon_hash(&[a_1])?;
|
||||
to_hash[0].zeroize();
|
||||
|
||||
// Merkle tree root computations
|
||||
let root = compute_tree_root(
|
||||
&witness.identity_secret,
|
||||
&witness.user_message_limit,
|
||||
&witness.path_elements,
|
||||
&witness.identity_path_index,
|
||||
)?;
|
||||
|
||||
Ok(RLNProofValues {
|
||||
y,
|
||||
nullifier,
|
||||
root,
|
||||
x: witness.x,
|
||||
external_nullifier: witness.external_nullifier,
|
||||
})
|
||||
}
|
||||
|
||||
/// Computes the Merkle tree root from identity credentials and Merkle membership proof.
|
||||
pub fn compute_tree_root(
|
||||
identity_secret: &IdSecret,
|
||||
user_message_limit: &Fr,
|
||||
path_elements: &[Fr],
|
||||
identity_path_index: &[u8],
|
||||
) -> Result<Fr, ProtocolError> {
|
||||
let mut to_hash = [*identity_secret.clone()];
|
||||
let id_commitment = poseidon_hash(&to_hash)?;
|
||||
to_hash[0].zeroize();
|
||||
|
||||
let mut root = poseidon_hash(&[id_commitment, *user_message_limit])?;
|
||||
|
||||
for i in 0..identity_path_index.len() {
|
||||
if identity_path_index[i] == 0 {
|
||||
root = poseidon_hash(&[root, path_elements[i]])?;
|
||||
} else {
|
||||
root = poseidon_hash(&[path_elements[i], root])?;
|
||||
}
|
||||
}
|
||||
|
||||
Ok(root)
|
||||
}
|
||||
|
||||
/// Prepares inputs for witness calculation from RLN witness input.
|
||||
pub(super) fn inputs_for_witness_calculation(
|
||||
witness: &RLNWitnessInput,
|
||||
) -> Result<[(&str, Vec<FrOrSecret>); 7], ProtocolError> {
|
||||
let mut identity_path_index = Vec::with_capacity(witness.identity_path_index.len());
|
||||
witness
|
||||
.identity_path_index
|
||||
.iter()
|
||||
.for_each(|v| identity_path_index.push(Fr::from(*v)));
|
||||
|
||||
Ok([
|
||||
(
|
||||
"identitySecret",
|
||||
vec![witness.identity_secret.clone().into()],
|
||||
),
|
||||
("userMessageLimit", vec![witness.user_message_limit.into()]),
|
||||
("messageId", vec![witness.message_id.into()]),
|
||||
(
|
||||
"pathElements",
|
||||
witness
|
||||
.path_elements
|
||||
.iter()
|
||||
.cloned()
|
||||
.map(Into::into)
|
||||
.collect(),
|
||||
),
|
||||
(
|
||||
"identityPathIndex",
|
||||
identity_path_index.into_iter().map(Into::into).collect(),
|
||||
),
|
||||
("x", vec![witness.x.into()]),
|
||||
("externalNullifier", vec![witness.external_nullifier.into()]),
|
||||
])
|
||||
}
|
||||
1499
rln/src/public.rs
1499
rln/src/public.rs
File diff suppressed because it is too large
Load Diff
File diff suppressed because it is too large
Load Diff
537
rln/src/utils.rs
537
rln/src/utils.rs
@@ -1,27 +1,37 @@
|
||||
// This crate provides cross-module useful utilities (mainly type conversions) not necessarily specific to RLN
|
||||
|
||||
use crate::circuit::Fr;
|
||||
use std::ops::Deref;
|
||||
|
||||
use ark_ff::PrimeField;
|
||||
use color_eyre::{Report, Result};
|
||||
use ark_serialize::{CanonicalDeserialize, CanonicalSerialize};
|
||||
use ark_std::UniformRand;
|
||||
use num_bigint::{BigInt, BigUint};
|
||||
use num_traits::Num;
|
||||
use serde_json::json;
|
||||
use std::io::Cursor;
|
||||
use std::iter::Extend;
|
||||
use rand::Rng;
|
||||
use ruint::aliases::U256;
|
||||
use zeroize::{Zeroize, ZeroizeOnDrop, Zeroizing};
|
||||
|
||||
pub fn to_bigint(el: &Fr) -> Result<BigInt> {
|
||||
let res: BigUint = (*el).into();
|
||||
Ok(res.into())
|
||||
use crate::{circuit::Fr, error::UtilsError};
|
||||
|
||||
/// Byte size of a field element aligned to 64-bit boundary, computed once at compile time.
|
||||
pub const FR_BYTE_SIZE: usize = {
|
||||
// Get the modulus bit size of the field
|
||||
let modulus_bits: u32 = Fr::MODULUS_BIT_SIZE;
|
||||
// Alignment boundary in bits for field element serialization
|
||||
let alignment_bits: u32 = 64;
|
||||
// Align to the next multiple of alignment_bits and convert to bytes
|
||||
((modulus_bits + alignment_bits - (modulus_bits % alignment_bits)) / 8) as usize
|
||||
};
|
||||
|
||||
#[inline(always)]
|
||||
pub fn to_bigint(el: &Fr) -> BigInt {
|
||||
BigUint::from(*el).into()
|
||||
}
|
||||
|
||||
pub fn fr_byte_size() -> usize {
|
||||
let mbs = <Fr as PrimeField>::MODULUS_BIT_SIZE;
|
||||
((mbs + 64 - (mbs % 64)) / 8) as usize
|
||||
}
|
||||
|
||||
pub fn str_to_fr(input: &str, radix: u32) -> Result<Fr> {
|
||||
#[inline(always)]
|
||||
pub fn str_to_fr(input: &str, radix: u32) -> Result<Fr, UtilsError> {
|
||||
if !(radix == 10 || radix == 16) {
|
||||
return Err(Report::msg("wrong radix"));
|
||||
return Err(UtilsError::WrongRadix);
|
||||
}
|
||||
|
||||
// We remove any quote present and we trim
|
||||
@@ -37,298 +47,413 @@ pub fn str_to_fr(input: &str, radix: u32) -> Result<Fr> {
|
||||
}
|
||||
}
|
||||
|
||||
pub fn bytes_le_to_fr(input: &[u8]) -> (Fr, usize) {
|
||||
let el_size = fr_byte_size();
|
||||
(
|
||||
#[inline(always)]
|
||||
pub fn bytes_le_to_fr(input: &[u8]) -> Result<(Fr, usize), UtilsError> {
|
||||
let el_size = FR_BYTE_SIZE;
|
||||
if input.len() < el_size {
|
||||
return Err(UtilsError::InsufficientData {
|
||||
expected: el_size,
|
||||
actual: input.len(),
|
||||
});
|
||||
}
|
||||
Ok((
|
||||
Fr::from(BigUint::from_bytes_le(&input[0..el_size])),
|
||||
el_size,
|
||||
)
|
||||
))
|
||||
}
|
||||
|
||||
pub fn bytes_be_to_fr(input: &[u8]) -> (Fr, usize) {
|
||||
let el_size = fr_byte_size();
|
||||
(
|
||||
#[inline(always)]
|
||||
pub fn bytes_be_to_fr(input: &[u8]) -> Result<(Fr, usize), UtilsError> {
|
||||
let el_size = FR_BYTE_SIZE;
|
||||
if input.len() < el_size {
|
||||
return Err(UtilsError::InsufficientData {
|
||||
expected: el_size,
|
||||
actual: input.len(),
|
||||
});
|
||||
}
|
||||
Ok((
|
||||
Fr::from(BigUint::from_bytes_be(&input[0..el_size])),
|
||||
el_size,
|
||||
)
|
||||
))
|
||||
}
|
||||
|
||||
#[inline(always)]
|
||||
pub fn fr_to_bytes_le(input: &Fr) -> Vec<u8> {
|
||||
let input_biguint: BigUint = (*input).into();
|
||||
let mut res = input_biguint.to_bytes_le();
|
||||
//BigUint conversion ignores most significant zero bytes. We restore them otherwise serialization will fail (length % 8 != 0)
|
||||
while res.len() != fr_byte_size() {
|
||||
res.push(0);
|
||||
}
|
||||
res.resize(FR_BYTE_SIZE, 0);
|
||||
res
|
||||
}
|
||||
|
||||
#[inline(always)]
|
||||
pub fn fr_to_bytes_be(input: &Fr) -> Vec<u8> {
|
||||
let input_biguint: BigUint = (*input).into();
|
||||
let mut res = input_biguint.to_bytes_be();
|
||||
// BigUint conversion ignores most significant zero bytes. We restore them otherwise serialization might fail
|
||||
// Fr elements are stored using 64 bits nimbs
|
||||
while res.len() != fr_byte_size() {
|
||||
res.insert(0, 0);
|
||||
// For BE, insert 0 at the start of the Vec (see also fr_to_bytes_le comments)
|
||||
let to_insert_count = FR_BYTE_SIZE.saturating_sub(res.len());
|
||||
if to_insert_count > 0 {
|
||||
// Insert multi 0 at index 0
|
||||
res.splice(0..0, std::iter::repeat_n(0, to_insert_count));
|
||||
}
|
||||
res
|
||||
}
|
||||
|
||||
pub fn vec_fr_to_bytes_le(input: &[Fr]) -> Result<Vec<u8>> {
|
||||
let mut bytes: Vec<u8> = Vec::new();
|
||||
//We store the vector length
|
||||
bytes.extend(u64::try_from(input.len())?.to_le_bytes().to_vec());
|
||||
#[inline(always)]
|
||||
pub fn vec_fr_to_bytes_le(input: &[Fr]) -> Vec<u8> {
|
||||
// Calculate capacity for Vec:
|
||||
// - 8 bytes for normalized vector length (usize)
|
||||
// - each Fr element requires FR_BYTE_SIZE bytes (typically 32 bytes)
|
||||
let mut bytes = Vec::with_capacity(8 + input.len() * FR_BYTE_SIZE);
|
||||
|
||||
// We store the vector length
|
||||
bytes.extend_from_slice(&normalize_usize_le(input.len()));
|
||||
|
||||
// We store each element
|
||||
input.iter().for_each(|el| bytes.extend(fr_to_bytes_le(el)));
|
||||
for el in input {
|
||||
bytes.extend_from_slice(&fr_to_bytes_le(el));
|
||||
}
|
||||
|
||||
Ok(bytes)
|
||||
bytes
|
||||
}
|
||||
|
||||
pub fn vec_fr_to_bytes_be(input: &[Fr]) -> Result<Vec<u8>> {
|
||||
let mut bytes: Vec<u8> = Vec::new();
|
||||
//We store the vector length
|
||||
bytes.extend(u64::try_from(input.len())?.to_be_bytes().to_vec());
|
||||
#[inline(always)]
|
||||
pub fn vec_fr_to_bytes_be(input: &[Fr]) -> Vec<u8> {
|
||||
// Calculate capacity for Vec:
|
||||
// - 8 bytes for normalized vector length (usize)
|
||||
// - each Fr element requires FR_BYTE_SIZE bytes (typically 32 bytes)
|
||||
let mut bytes = Vec::with_capacity(8 + input.len() * FR_BYTE_SIZE);
|
||||
|
||||
// We store the vector length
|
||||
bytes.extend_from_slice(&normalize_usize_be(input.len()));
|
||||
|
||||
// We store each element
|
||||
input.iter().for_each(|el| bytes.extend(fr_to_bytes_be(el)));
|
||||
for el in input {
|
||||
bytes.extend_from_slice(&fr_to_bytes_be(el));
|
||||
}
|
||||
|
||||
Ok(bytes)
|
||||
bytes
|
||||
}
|
||||
|
||||
pub fn vec_u8_to_bytes_le(input: &[u8]) -> Result<Vec<u8>> {
|
||||
let mut bytes: Vec<u8> = Vec::new();
|
||||
//We store the vector length
|
||||
bytes.extend(u64::try_from(input.len())?.to_le_bytes().to_vec());
|
||||
#[inline(always)]
|
||||
pub fn vec_u8_to_bytes_le(input: &[u8]) -> Vec<u8> {
|
||||
// Calculate capacity for Vec:
|
||||
// - 8 bytes for normalized vector length (usize)
|
||||
// - variable length input data
|
||||
let mut bytes = Vec::with_capacity(8 + input.len());
|
||||
|
||||
bytes.extend(input);
|
||||
// We store the vector length
|
||||
bytes.extend_from_slice(&normalize_usize_le(input.len()));
|
||||
|
||||
Ok(bytes)
|
||||
// We store the input
|
||||
bytes.extend_from_slice(input);
|
||||
|
||||
bytes
|
||||
}
|
||||
|
||||
pub fn vec_u8_to_bytes_be(input: Vec<u8>) -> Result<Vec<u8>> {
|
||||
let mut bytes: Vec<u8> = Vec::new();
|
||||
//We store the vector length
|
||||
bytes.extend(u64::try_from(input.len())?.to_be_bytes().to_vec());
|
||||
#[inline(always)]
|
||||
pub fn vec_u8_to_bytes_be(input: &[u8]) -> Vec<u8> {
|
||||
// Calculate capacity for Vec:
|
||||
// - 8 bytes for normalized vector length (usize)
|
||||
// - variable length input data
|
||||
let mut bytes = Vec::with_capacity(8 + input.len());
|
||||
|
||||
bytes.extend(input);
|
||||
// We store the vector length
|
||||
bytes.extend_from_slice(&normalize_usize_be(input.len()));
|
||||
|
||||
Ok(bytes)
|
||||
// We store the input
|
||||
bytes.extend_from_slice(input);
|
||||
|
||||
bytes
|
||||
}
|
||||
|
||||
pub fn bytes_le_to_vec_u8(input: &[u8]) -> Result<(Vec<u8>, usize)> {
|
||||
#[inline(always)]
|
||||
pub fn bytes_le_to_vec_u8(input: &[u8]) -> Result<(Vec<u8>, usize), UtilsError> {
|
||||
let mut read: usize = 0;
|
||||
|
||||
if input.len() < 8 {
|
||||
return Err(UtilsError::InsufficientData {
|
||||
expected: 8,
|
||||
actual: input.len(),
|
||||
});
|
||||
}
|
||||
let len = usize::try_from(u64::from_le_bytes(input[0..8].try_into()?))?;
|
||||
read += 8;
|
||||
|
||||
if input.len() < 8 + len {
|
||||
return Err(UtilsError::InsufficientData {
|
||||
expected: 8 + len,
|
||||
actual: input.len(),
|
||||
});
|
||||
}
|
||||
let res = input[8..8 + len].to_vec();
|
||||
read += res.len();
|
||||
|
||||
Ok((res, read))
|
||||
}
|
||||
|
||||
pub fn bytes_be_to_vec_u8(input: &[u8]) -> Result<(Vec<u8>, usize)> {
|
||||
#[inline(always)]
|
||||
pub fn bytes_be_to_vec_u8(input: &[u8]) -> Result<(Vec<u8>, usize), UtilsError> {
|
||||
let mut read: usize = 0;
|
||||
|
||||
if input.len() < 8 {
|
||||
return Err(UtilsError::InsufficientData {
|
||||
expected: 8,
|
||||
actual: input.len(),
|
||||
});
|
||||
}
|
||||
let len = usize::try_from(u64::from_be_bytes(input[0..8].try_into()?))?;
|
||||
read += 8;
|
||||
|
||||
if input.len() < 8 + len {
|
||||
return Err(UtilsError::InsufficientData {
|
||||
expected: 8 + len,
|
||||
actual: input.len(),
|
||||
});
|
||||
}
|
||||
let res = input[8..8 + len].to_vec();
|
||||
|
||||
read += res.len();
|
||||
|
||||
Ok((res, read))
|
||||
}
|
||||
|
||||
pub fn bytes_le_to_vec_fr(input: &[u8]) -> Result<(Vec<Fr>, usize)> {
|
||||
#[inline(always)]
|
||||
pub fn bytes_le_to_vec_fr(input: &[u8]) -> Result<(Vec<Fr>, usize), UtilsError> {
|
||||
let mut read: usize = 0;
|
||||
let mut res: Vec<Fr> = Vec::new();
|
||||
|
||||
if input.len() < 8 {
|
||||
return Err(UtilsError::InsufficientData {
|
||||
expected: 8,
|
||||
actual: input.len(),
|
||||
});
|
||||
}
|
||||
let len = usize::try_from(u64::from_le_bytes(input[0..8].try_into()?))?;
|
||||
read += 8;
|
||||
|
||||
let el_size = fr_byte_size();
|
||||
let el_size = FR_BYTE_SIZE;
|
||||
if input.len() < 8 + len * el_size {
|
||||
return Err(UtilsError::InsufficientData {
|
||||
expected: 8 + len * el_size,
|
||||
actual: input.len(),
|
||||
});
|
||||
}
|
||||
let mut res: Vec<Fr> = Vec::with_capacity(len);
|
||||
for i in 0..len {
|
||||
let (curr_el, _) = bytes_le_to_fr(&input[8 + el_size * i..8 + el_size * (i + 1)]);
|
||||
let (curr_el, _) = bytes_le_to_fr(&input[8 + el_size * i..8 + el_size * (i + 1)])?;
|
||||
res.push(curr_el);
|
||||
read += el_size;
|
||||
}
|
||||
|
||||
Ok((res, read))
|
||||
}
|
||||
|
||||
pub fn bytes_be_to_vec_fr(input: &[u8]) -> Result<(Vec<Fr>, usize)> {
|
||||
#[inline(always)]
|
||||
pub fn bytes_be_to_vec_fr(input: &[u8]) -> Result<(Vec<Fr>, usize), UtilsError> {
|
||||
let mut read: usize = 0;
|
||||
let mut res: Vec<Fr> = Vec::new();
|
||||
|
||||
if input.len() < 8 {
|
||||
return Err(UtilsError::InsufficientData {
|
||||
expected: 8,
|
||||
actual: input.len(),
|
||||
});
|
||||
}
|
||||
let len = usize::try_from(u64::from_be_bytes(input[0..8].try_into()?))?;
|
||||
read += 8;
|
||||
|
||||
let el_size = fr_byte_size();
|
||||
let el_size = FR_BYTE_SIZE;
|
||||
if input.len() < 8 + len * el_size {
|
||||
return Err(UtilsError::InsufficientData {
|
||||
expected: 8 + len * el_size,
|
||||
actual: input.len(),
|
||||
});
|
||||
}
|
||||
let mut res: Vec<Fr> = Vec::with_capacity(len);
|
||||
for i in 0..len {
|
||||
let (curr_el, _) = bytes_be_to_fr(&input[8 + el_size * i..8 + el_size * (i + 1)]);
|
||||
let (curr_el, _) = bytes_be_to_fr(&input[8 + el_size * i..8 + el_size * (i + 1)])?;
|
||||
res.push(curr_el);
|
||||
read += el_size;
|
||||
}
|
||||
|
||||
Ok((res, read))
|
||||
}
|
||||
|
||||
pub fn normalize_usize(input: usize) -> Vec<u8> {
|
||||
let mut normalized_usize = input.to_le_bytes().to_vec();
|
||||
normalized_usize.resize(8, 0);
|
||||
normalized_usize
|
||||
}
|
||||
|
||||
pub fn bytes_le_to_vec_usize(input: &[u8]) -> Result<Vec<usize>> {
|
||||
#[inline(always)]
|
||||
pub fn bytes_le_to_vec_usize(input: &[u8]) -> Result<Vec<usize>, UtilsError> {
|
||||
if input.len() < 8 {
|
||||
return Err(UtilsError::InsufficientData {
|
||||
expected: 8,
|
||||
actual: input.len(),
|
||||
});
|
||||
}
|
||||
let nof_elem = usize::try_from(u64::from_le_bytes(input[0..8].try_into()?))?;
|
||||
if nof_elem == 0 {
|
||||
Ok(vec![])
|
||||
} else {
|
||||
let elements: Vec<usize> = input[8..]
|
||||
.chunks(8)
|
||||
.map(|ch| usize::from_le_bytes(ch[0..8].try_into().unwrap()))
|
||||
.collect();
|
||||
Ok(elements)
|
||||
if input.len() < 8 + nof_elem * 8 {
|
||||
return Err(UtilsError::InsufficientData {
|
||||
expected: 8 + nof_elem * 8,
|
||||
actual: input.len(),
|
||||
});
|
||||
}
|
||||
input[8..]
|
||||
.chunks_exact(8)
|
||||
.take(nof_elem)
|
||||
.map(|ch| {
|
||||
ch.try_into()
|
||||
.map(usize::from_le_bytes)
|
||||
.map_err(UtilsError::FromSlice)
|
||||
})
|
||||
.collect()
|
||||
}
|
||||
}
|
||||
|
||||
// using for test
|
||||
pub fn generate_input_buffer() -> Cursor<String> {
|
||||
Cursor::new(json!({}).to_string())
|
||||
}
|
||||
|
||||
/* Old conversion utilities between different libraries data types
|
||||
|
||||
// Conversion Utilities between poseidon-rs Field and arkworks Fr (in order to call directly poseidon-rs' poseidon_hash)
|
||||
|
||||
use ff::{PrimeField as _, PrimeFieldRepr as _};
|
||||
use poseidon_rs::Fr as PosFr;
|
||||
|
||||
pub fn fr_to_posfr(value: Fr) -> PosFr {
|
||||
let mut bytes = [0_u8; 32];
|
||||
let byte_vec = value.into_repr().to_bytes_be();
|
||||
bytes.copy_from_slice(&byte_vec[..]);
|
||||
let mut repr = <PosFr as ff::PrimeField>::Repr::default();
|
||||
repr.read_be(&bytes[..])
|
||||
.expect("read from correctly sized slice always succeeds");
|
||||
PosFr::from_repr(repr).expect("value is always in range")
|
||||
}
|
||||
|
||||
pub fn posfr_to_fr(value: PosFr) -> Fr {
|
||||
let mut bytes = [0u8; 32];
|
||||
value
|
||||
.into_repr()
|
||||
.write_be(&mut bytes[..])
|
||||
.expect("write to correctly sized slice always succeeds");
|
||||
Fr::from_be_bytes_mod_order(&bytes)
|
||||
}
|
||||
|
||||
|
||||
// Conversion Utilities between semaphore-rs Field and arkworks Fr
|
||||
|
||||
use semaphore::Field;
|
||||
|
||||
pub fn to_fr(el: &Field) -> Fr {
|
||||
Fr::try_from(*el).unwrap()
|
||||
}
|
||||
|
||||
pub fn to_field(el: &Fr) -> Field {
|
||||
(*el).try_into().unwrap()
|
||||
}
|
||||
|
||||
pub fn vec_to_fr(v: &[Field]) -> Vec<Fr> {
|
||||
v.iter().map(|el| to_fr(el)).collect()
|
||||
}
|
||||
|
||||
pub fn vec_to_field(v: &[Fr]) -> Vec<Field> {
|
||||
v.iter().map(|el| to_field(el)).collect()
|
||||
}
|
||||
|
||||
pub fn vec_fr_to_field(input: &[Fr]) -> Vec<Field> {
|
||||
input.iter().map(|el| to_field(el)).collect()
|
||||
}
|
||||
|
||||
pub fn vec_field_to_fr(input: &[Field]) -> Vec<Fr> {
|
||||
input.iter().map(|el| to_fr(el)).collect()
|
||||
}
|
||||
|
||||
pub fn str_to_field(input: String, radix: i32) -> Field {
|
||||
assert!((radix == 10) || (radix == 16));
|
||||
|
||||
// We remove any quote present and we trim
|
||||
let single_quote: char = '\"';
|
||||
let input_clean = input.replace(single_quote, "");
|
||||
let input_clean = input_clean.trim();
|
||||
|
||||
if radix == 10 {
|
||||
Field::from_str(&format!(
|
||||
"{:01$x}",
|
||||
BigUint::from_str(input_clean).unwrap(),
|
||||
64
|
||||
))
|
||||
.unwrap()
|
||||
#[inline(always)]
|
||||
pub fn bytes_be_to_vec_usize(input: &[u8]) -> Result<Vec<usize>, UtilsError> {
|
||||
if input.len() < 8 {
|
||||
return Err(UtilsError::InsufficientData {
|
||||
expected: 8,
|
||||
actual: input.len(),
|
||||
});
|
||||
}
|
||||
let nof_elem = usize::try_from(u64::from_be_bytes(input[0..8].try_into()?))?;
|
||||
if nof_elem == 0 {
|
||||
Ok(vec![])
|
||||
} else {
|
||||
let input_clean = input_clean.replace("0x", "");
|
||||
Field::from_str(&format!("{:0>64}", &input_clean)).unwrap()
|
||||
if input.len() < 8 + nof_elem * 8 {
|
||||
return Err(UtilsError::InsufficientData {
|
||||
expected: 8 + nof_elem * 8,
|
||||
actual: input.len(),
|
||||
});
|
||||
}
|
||||
input[8..]
|
||||
.chunks_exact(8)
|
||||
.take(nof_elem)
|
||||
.map(|ch| {
|
||||
ch.try_into()
|
||||
.map(usize::from_be_bytes)
|
||||
.map_err(UtilsError::FromSlice)
|
||||
})
|
||||
.collect()
|
||||
}
|
||||
}
|
||||
|
||||
pub fn bytes_le_to_field(input: &[u8]) -> (Field, usize) {
|
||||
let (fr_el, read) = bytes_le_to_fr(input);
|
||||
(to_field(&fr_el), read)
|
||||
/// Normalizes a `usize` into an 8-byte array, ensuring consistency across architectures.
|
||||
/// On 32-bit systems, the result is zero-padded to 8 bytes.
|
||||
/// On 64-bit systems, it directly represents the `usize` value.
|
||||
#[inline(always)]
|
||||
pub fn normalize_usize_le(input: usize) -> [u8; 8] {
|
||||
let mut bytes = [0u8; 8];
|
||||
let input_bytes = input.to_le_bytes();
|
||||
bytes[..input_bytes.len()].copy_from_slice(&input_bytes);
|
||||
bytes
|
||||
}
|
||||
|
||||
pub fn bytes_be_to_field(input: &[u8]) -> (Field, usize) {
|
||||
let (fr_el, read) = bytes_be_to_fr(input);
|
||||
(to_field(&fr_el), read)
|
||||
/// Normalizes a `usize` into an 8-byte array, ensuring consistency across architectures.
|
||||
/// On 32-bit systems, the result is zero-padded to 8 bytes.
|
||||
/// On 64-bit systems, it directly represents the `usize` value.
|
||||
#[inline(always)]
|
||||
pub fn normalize_usize_be(input: usize) -> [u8; 8] {
|
||||
let mut bytes = [0u8; 8];
|
||||
let input_bytes = input.to_be_bytes();
|
||||
let offset = 8 - input_bytes.len();
|
||||
bytes[offset..].copy_from_slice(&input_bytes);
|
||||
bytes
|
||||
}
|
||||
|
||||
#[derive(
|
||||
Debug, Zeroize, ZeroizeOnDrop, Clone, PartialEq, CanonicalSerialize, CanonicalDeserialize,
|
||||
)]
|
||||
pub struct IdSecret(Fr);
|
||||
|
||||
pub fn field_to_bytes_le(input: &Field) -> Vec<u8> {
|
||||
fr_to_bytes_le(&to_fr(input))
|
||||
impl IdSecret {
|
||||
pub fn rand<R: Rng + ?Sized>(rng: &mut R) -> Self {
|
||||
let mut fr = Fr::rand(rng);
|
||||
let res = Self::from(&mut fr);
|
||||
// No need to zeroize fr (already zeroiz'ed in from implementation)
|
||||
#[allow(clippy::let_and_return)]
|
||||
res
|
||||
}
|
||||
|
||||
pub fn from_bytes_le(input: &[u8]) -> Result<(Self, usize), UtilsError> {
|
||||
let el_size = FR_BYTE_SIZE;
|
||||
if input.len() < el_size {
|
||||
return Err(UtilsError::InsufficientData {
|
||||
expected: el_size,
|
||||
actual: input.len(),
|
||||
});
|
||||
}
|
||||
let b_uint = BigUint::from_bytes_le(&input[0..el_size]);
|
||||
let mut fr = Fr::from(b_uint);
|
||||
let res = IdSecret::from(&mut fr);
|
||||
// Note: no zeroize on b_uint as it has been moved
|
||||
Ok((res, el_size))
|
||||
}
|
||||
|
||||
pub fn from_bytes_be(input: &[u8]) -> Result<(Self, usize), UtilsError> {
|
||||
let el_size = FR_BYTE_SIZE;
|
||||
if input.len() < el_size {
|
||||
return Err(UtilsError::InsufficientData {
|
||||
expected: el_size,
|
||||
actual: input.len(),
|
||||
});
|
||||
}
|
||||
let b_uint = BigUint::from_bytes_be(&input[0..el_size]);
|
||||
let mut fr = Fr::from(b_uint);
|
||||
let res = IdSecret::from(&mut fr);
|
||||
// Note: no zeroize on b_uint as it has been moved
|
||||
Ok((res, el_size))
|
||||
}
|
||||
|
||||
pub(crate) fn to_bytes_le(&self) -> Zeroizing<Vec<u8>> {
|
||||
let input_biguint: BigUint = self.0.into();
|
||||
let mut res = input_biguint.to_bytes_le();
|
||||
res.resize(FR_BYTE_SIZE, 0);
|
||||
Zeroizing::new(res)
|
||||
}
|
||||
|
||||
pub(crate) fn to_bytes_be(&self) -> Zeroizing<Vec<u8>> {
|
||||
let input_biguint: BigUint = self.0.into();
|
||||
let mut res = input_biguint.to_bytes_be();
|
||||
let to_insert_count = FR_BYTE_SIZE.saturating_sub(res.len());
|
||||
if to_insert_count > 0 {
|
||||
// Insert multi 0 at index 0
|
||||
res.splice(0..0, std::iter::repeat_n(0, to_insert_count));
|
||||
}
|
||||
Zeroizing::new(res)
|
||||
}
|
||||
|
||||
/// Warning: this can leak the secret value
|
||||
/// Warning: Leaked value is of type 'U256' which implement Copy (every copy will not be zeroized)
|
||||
pub(crate) fn to_u256(&self) -> U256 {
|
||||
let mut big_int = self.0.into_bigint();
|
||||
let res = U256::from_limbs(big_int.0);
|
||||
big_int.zeroize();
|
||||
res
|
||||
}
|
||||
}
|
||||
|
||||
pub fn field_to_bytes_be(input: &Field) -> Vec<u8> {
|
||||
fr_to_bytes_be(&to_fr(input))
|
||||
impl From<&mut Fr> for IdSecret {
|
||||
fn from(value: &mut Fr) -> Self {
|
||||
let id_secret = Self(*value);
|
||||
value.zeroize();
|
||||
id_secret
|
||||
}
|
||||
}
|
||||
|
||||
impl Deref for IdSecret {
|
||||
type Target = Fr;
|
||||
|
||||
pub fn vec_field_to_bytes_le(input: &[Field]) -> Vec<u8> {
|
||||
vec_fr_to_bytes_le(&vec_field_to_fr(input))
|
||||
/// Deref to &Fr
|
||||
///
|
||||
/// Warning: this can leak the secret value
|
||||
/// Warning: Leaked value is of type 'Fr' which implement Copy (every copy will not be zeroized)
|
||||
fn deref(&self) -> &Self::Target {
|
||||
&self.0
|
||||
}
|
||||
}
|
||||
|
||||
pub fn vec_field_to_bytes_be(input: &[Field]) -> Vec<u8> {
|
||||
vec_fr_to_bytes_be(&vec_field_to_fr(input))
|
||||
#[derive(Debug, Zeroize, ZeroizeOnDrop)]
|
||||
pub(crate) enum FrOrSecret {
|
||||
IdSecret(IdSecret),
|
||||
Fr(Fr),
|
||||
}
|
||||
|
||||
|
||||
pub fn bytes_le_to_vec_field(input: &[u8]) -> (Vec<Field>, usize) {
|
||||
let (vec_fr, read) = bytes_le_to_vec_fr(input);
|
||||
(vec_fr_to_field(&vec_fr), read)
|
||||
impl From<Fr> for FrOrSecret {
|
||||
fn from(value: Fr) -> Self {
|
||||
FrOrSecret::Fr(value)
|
||||
}
|
||||
}
|
||||
|
||||
pub fn bytes_be_to_vec_field(input: &[u8]) -> (Vec<Field>, usize) {
|
||||
let (vec_fr, read) = bytes_be_to_vec_fr(input);
|
||||
(vec_fr_to_field(&vec_fr), read)
|
||||
impl From<IdSecret> for FrOrSecret {
|
||||
fn from(value: IdSecret) -> Self {
|
||||
FrOrSecret::IdSecret(value)
|
||||
}
|
||||
}
|
||||
|
||||
// Arithmetic over Field elements (wrapped over arkworks algebra crate)
|
||||
|
||||
pub fn add(a: &Field, b: &Field) -> Field {
|
||||
to_field(&(to_fr(a) + to_fr(b)))
|
||||
}
|
||||
|
||||
pub fn mul(a: &Field, b: &Field) -> Field {
|
||||
to_field(&(to_fr(a) * to_fr(b)))
|
||||
}
|
||||
|
||||
pub fn div(a: &Field, b: &Field) -> Field {
|
||||
to_field(&(to_fr(a) / to_fr(b)))
|
||||
}
|
||||
|
||||
pub fn inv(a: &Field) -> Field {
|
||||
to_field(&(Fr::from(1) / to_fr(a)))
|
||||
}
|
||||
*/
|
||||
|
||||
1577
rln/tests/ffi.rs
1577
rln/tests/ffi.rs
File diff suppressed because it is too large
Load Diff
288
rln/tests/ffi_utils.rs
Normal file
288
rln/tests/ffi_utils.rs
Normal file
@@ -0,0 +1,288 @@
|
||||
#[cfg(test)]
|
||||
mod test {
|
||||
use rand::Rng;
|
||||
use rln::{ffi::ffi_utils::*, prelude::*};
|
||||
|
||||
#[test]
|
||||
// Tests hash to field using FFI APIs
|
||||
fn test_seeded_keygen_ffi() {
|
||||
// We generate a new identity pair from an input seed
|
||||
let seed_bytes: Vec<u8> = vec![0, 1, 2, 3, 4, 5, 6, 7, 8, 9];
|
||||
let res = match ffi_seeded_key_gen(&seed_bytes.into()) {
|
||||
CResult {
|
||||
ok: Some(vec_cfr),
|
||||
err: None,
|
||||
} => vec_cfr,
|
||||
CResult {
|
||||
ok: None,
|
||||
err: Some(err),
|
||||
} => panic!("ffi_seeded_key_gen call failed: {}", err),
|
||||
_ => unreachable!(),
|
||||
};
|
||||
let identity_secret = res.first().unwrap();
|
||||
let id_commitment = res.get(1).unwrap();
|
||||
|
||||
// We check against expected values
|
||||
let expected_identity_secret_seed_bytes = str_to_fr(
|
||||
"0x766ce6c7e7a01bdf5b3f257616f603918c30946fa23480f2859c597817e6716",
|
||||
16,
|
||||
)
|
||||
.unwrap();
|
||||
let expected_id_commitment_seed_bytes = str_to_fr(
|
||||
"0xbf16d2b5c0d6f9d9d561e05bfca16a81b4b873bb063508fae360d8c74cef51f",
|
||||
16,
|
||||
)
|
||||
.unwrap();
|
||||
|
||||
assert_eq!(*identity_secret, expected_identity_secret_seed_bytes);
|
||||
assert_eq!(*id_commitment, expected_id_commitment_seed_bytes);
|
||||
}
|
||||
|
||||
#[test]
|
||||
// Tests hash to field using FFI APIs
|
||||
fn test_seeded_extended_keygen_ffi() {
|
||||
// We generate a new identity tuple from an input seed
|
||||
let seed_bytes: Vec<u8> = vec![0, 1, 2, 3, 4, 5, 6, 7, 8, 9];
|
||||
let key_gen = match ffi_seeded_extended_key_gen(&seed_bytes.into()) {
|
||||
CResult {
|
||||
ok: Some(vec_cfr),
|
||||
err: None,
|
||||
} => vec_cfr,
|
||||
CResult {
|
||||
ok: None,
|
||||
err: Some(err),
|
||||
} => panic!("ffi_seeded_extended_key_gen call failed: {}", err),
|
||||
_ => unreachable!(),
|
||||
};
|
||||
let identity_trapdoor = *key_gen[0];
|
||||
let identity_nullifier = *key_gen[1];
|
||||
let identity_secret = *key_gen[2];
|
||||
let id_commitment = *key_gen[3];
|
||||
|
||||
// We check against expected values
|
||||
let expected_identity_trapdoor_seed_bytes = str_to_fr(
|
||||
"0x766ce6c7e7a01bdf5b3f257616f603918c30946fa23480f2859c597817e6716",
|
||||
16,
|
||||
)
|
||||
.unwrap();
|
||||
let expected_identity_nullifier_seed_bytes = str_to_fr(
|
||||
"0x1f18714c7bc83b5bca9e89d404cf6f2f585bc4c0f7ed8b53742b7e2b298f50b4",
|
||||
16,
|
||||
)
|
||||
.unwrap();
|
||||
let expected_identity_secret_seed_bytes = str_to_fr(
|
||||
"0x2aca62aaa7abaf3686fff2caf00f55ab9462dc12db5b5d4bcf3994e671f8e521",
|
||||
16,
|
||||
)
|
||||
.unwrap();
|
||||
let expected_id_commitment_seed_bytes = str_to_fr(
|
||||
"0x68b66aa0a8320d2e56842581553285393188714c48f9b17acd198b4f1734c5c",
|
||||
16,
|
||||
)
|
||||
.unwrap();
|
||||
|
||||
assert_eq!(identity_trapdoor, expected_identity_trapdoor_seed_bytes);
|
||||
assert_eq!(identity_nullifier, expected_identity_nullifier_seed_bytes);
|
||||
assert_eq!(identity_secret, expected_identity_secret_seed_bytes);
|
||||
assert_eq!(id_commitment, expected_id_commitment_seed_bytes);
|
||||
}
|
||||
|
||||
#[test]
|
||||
// Test CFr FFI functions
|
||||
fn test_cfr_ffi() {
|
||||
let cfr_zero = ffi_cfr_zero();
|
||||
let fr_zero = rln::circuit::Fr::from(0u8);
|
||||
assert_eq!(*cfr_zero, fr_zero);
|
||||
|
||||
let cfr_one = ffi_cfr_one();
|
||||
let fr_one = rln::circuit::Fr::from(1u8);
|
||||
assert_eq!(*cfr_one, fr_one);
|
||||
|
||||
let cfr_int = ffi_uint_to_cfr(42);
|
||||
let fr_int = rln::circuit::Fr::from(42u8);
|
||||
assert_eq!(*cfr_int, fr_int);
|
||||
|
||||
let cfr_debug_str = ffi_cfr_debug(Some(&cfr_int));
|
||||
assert_eq!(cfr_debug_str.to_string(), "42");
|
||||
|
||||
let key_gen = match ffi_key_gen() {
|
||||
CResult {
|
||||
ok: Some(vec_cfr),
|
||||
err: None,
|
||||
} => vec_cfr,
|
||||
CResult {
|
||||
ok: None,
|
||||
err: Some(err),
|
||||
} => panic!("ffi_key_gen call failed: {}", err),
|
||||
_ => unreachable!(),
|
||||
};
|
||||
let mut id_secret_fr = *key_gen[0];
|
||||
let id_secret_hash = IdSecret::from(&mut id_secret_fr);
|
||||
let id_commitment = *key_gen[1];
|
||||
let cfr_id_secret_hash = ffi_vec_cfr_get(&key_gen, 0).unwrap();
|
||||
assert_eq!(*cfr_id_secret_hash, *id_secret_hash);
|
||||
let cfr_id_commitment = ffi_vec_cfr_get(&key_gen, 1).unwrap();
|
||||
assert_eq!(*cfr_id_commitment, id_commitment);
|
||||
}
|
||||
|
||||
#[test]
|
||||
// Test Vec<u8> FFI functions
|
||||
fn test_vec_u8_ffi() {
|
||||
let mut rng = rand::thread_rng();
|
||||
let signal_gen: [u8; 32] = rng.gen();
|
||||
let signal: Vec<u8> = signal_gen.to_vec();
|
||||
|
||||
let bytes_le = ffi_vec_u8_to_bytes_le(&signal.clone().into());
|
||||
let expected_le = vec_u8_to_bytes_le(&signal);
|
||||
assert_eq!(bytes_le.iter().copied().collect::<Vec<_>>(), expected_le);
|
||||
|
||||
let bytes_be = ffi_vec_u8_to_bytes_be(&signal.clone().into());
|
||||
let expected_be = vec_u8_to_bytes_be(&signal);
|
||||
assert_eq!(bytes_be.iter().copied().collect::<Vec<_>>(), expected_be);
|
||||
|
||||
let signal_from_le = match ffi_bytes_le_to_vec_u8(&bytes_le) {
|
||||
CResult {
|
||||
ok: Some(vec_u8),
|
||||
err: None,
|
||||
} => vec_u8,
|
||||
CResult {
|
||||
ok: None,
|
||||
err: Some(err),
|
||||
} => panic!("ffi_bytes_le_to_vec_u8 call failed: {}", err),
|
||||
_ => unreachable!(),
|
||||
};
|
||||
assert_eq!(signal_from_le.iter().copied().collect::<Vec<_>>(), signal);
|
||||
|
||||
let signal_from_be = match ffi_bytes_be_to_vec_u8(&bytes_be) {
|
||||
CResult {
|
||||
ok: Some(vec_u8),
|
||||
err: None,
|
||||
} => vec_u8,
|
||||
CResult {
|
||||
ok: None,
|
||||
err: Some(err),
|
||||
} => panic!("ffi_bytes_be_to_vec_u8 call failed: {}", err),
|
||||
_ => unreachable!(),
|
||||
};
|
||||
assert_eq!(signal_from_be.iter().copied().collect::<Vec<_>>(), signal);
|
||||
}
|
||||
|
||||
#[test]
|
||||
// Test Vec<CFr> FFI functions
|
||||
fn test_vec_cfr_ffi() {
|
||||
let vec_fr = [Fr::from(1u8), Fr::from(2u8), Fr::from(3u8), Fr::from(4u8)];
|
||||
let vec_cfr: Vec<CFr> = vec_fr.iter().map(|fr| CFr::from(*fr)).collect();
|
||||
|
||||
let bytes_le = ffi_vec_cfr_to_bytes_le(&vec_cfr.clone().into());
|
||||
let expected_le = vec_fr_to_bytes_le(&vec_fr);
|
||||
assert_eq!(bytes_le.iter().copied().collect::<Vec<_>>(), expected_le);
|
||||
|
||||
let bytes_be = ffi_vec_cfr_to_bytes_be(&vec_cfr.clone().into());
|
||||
let expected_be = vec_fr_to_bytes_be(&vec_fr);
|
||||
assert_eq!(bytes_be.iter().copied().collect::<Vec<_>>(), expected_be);
|
||||
|
||||
let vec_cfr_from_le = match ffi_bytes_le_to_vec_cfr(&bytes_le) {
|
||||
CResult {
|
||||
ok: Some(vec_cfr),
|
||||
err: None,
|
||||
} => vec_cfr,
|
||||
CResult {
|
||||
ok: None,
|
||||
err: Some(err),
|
||||
} => panic!("ffi_bytes_le_to_vec_cfr call failed: {}", err),
|
||||
_ => unreachable!(),
|
||||
};
|
||||
assert_eq!(vec_cfr_from_le.iter().copied().collect::<Vec<_>>(), vec_cfr);
|
||||
|
||||
let vec_cfr_from_be = match ffi_bytes_be_to_vec_cfr(&bytes_be) {
|
||||
CResult {
|
||||
ok: Some(vec_cfr),
|
||||
err: None,
|
||||
} => vec_cfr,
|
||||
CResult {
|
||||
ok: None,
|
||||
err: Some(err),
|
||||
} => panic!("ffi_bytes_be_to_vec_cfr call failed: {}", err),
|
||||
_ => unreachable!(),
|
||||
};
|
||||
assert_eq!(vec_cfr_from_be.iter().copied().collect::<Vec<_>>(), vec_cfr);
|
||||
}
|
||||
|
||||
#[test]
|
||||
// Tests hash to field using FFI APIs
|
||||
fn test_hash_to_field_ffi() {
|
||||
let mut rng = rand::thread_rng();
|
||||
let signal_gen: [u8; 32] = rng.gen();
|
||||
let signal: Vec<u8> = signal_gen.to_vec();
|
||||
|
||||
let cfr_le_1 = match ffi_hash_to_field_le(&signal.clone().into()) {
|
||||
CResult {
|
||||
ok: Some(cfr),
|
||||
err: None,
|
||||
} => cfr,
|
||||
CResult {
|
||||
ok: None,
|
||||
err: Some(err),
|
||||
} => panic!("ffi_hash_to_field_le call failed: {}", err),
|
||||
_ => unreachable!(),
|
||||
};
|
||||
let fr_le_2 = hash_to_field_le(&signal).unwrap();
|
||||
assert_eq!(*cfr_le_1, fr_le_2);
|
||||
|
||||
let cfr_be_1 = match ffi_hash_to_field_be(&signal.clone().into()) {
|
||||
CResult {
|
||||
ok: Some(cfr),
|
||||
err: None,
|
||||
} => cfr,
|
||||
CResult {
|
||||
ok: None,
|
||||
err: Some(err),
|
||||
} => panic!("ffi_hash_to_field_be call failed: {}", err),
|
||||
_ => unreachable!(),
|
||||
};
|
||||
let fr_be_2 = hash_to_field_be(&signal).unwrap();
|
||||
assert_eq!(*cfr_le_1, fr_be_2);
|
||||
|
||||
assert_eq!(*cfr_le_1, *cfr_be_1);
|
||||
assert_eq!(fr_le_2, fr_be_2);
|
||||
|
||||
let hash_cfr_le_1 = ffi_cfr_to_bytes_le(&cfr_le_1)
|
||||
.iter()
|
||||
.copied()
|
||||
.collect::<Vec<_>>();
|
||||
let hash_fr_le_2 = fr_to_bytes_le(&fr_le_2);
|
||||
assert_eq!(hash_cfr_le_1, hash_fr_le_2);
|
||||
|
||||
let hash_cfr_be_1 = ffi_cfr_to_bytes_be(&cfr_be_1)
|
||||
.iter()
|
||||
.copied()
|
||||
.collect::<Vec<_>>();
|
||||
let hash_fr_be_2 = fr_to_bytes_be(&fr_be_2);
|
||||
assert_eq!(hash_cfr_be_1, hash_fr_be_2);
|
||||
|
||||
assert_ne!(hash_cfr_le_1, hash_cfr_be_1);
|
||||
assert_ne!(hash_fr_le_2, hash_fr_be_2);
|
||||
}
|
||||
|
||||
#[test]
|
||||
// Test Poseidon hash FFI
|
||||
fn test_poseidon_hash_pair_ffi() {
|
||||
let input_1 = Fr::from(42u8);
|
||||
let input_2 = Fr::from(99u8);
|
||||
|
||||
let expected_hash = poseidon_hash(&[input_1, input_2]).unwrap();
|
||||
let received_hash_cfr =
|
||||
match ffi_poseidon_hash_pair(&CFr::from(input_1), &CFr::from(input_2)) {
|
||||
CResult {
|
||||
ok: Some(cfr),
|
||||
err: None,
|
||||
} => cfr,
|
||||
CResult {
|
||||
ok: None,
|
||||
err: Some(err),
|
||||
} => panic!("ffi_poseidon_hash_pair call failed: {}", err),
|
||||
_ => unreachable!(),
|
||||
};
|
||||
assert_eq!(*received_hash_cfr, expected_hash);
|
||||
}
|
||||
}
|
||||
@@ -1,29 +1,35 @@
|
||||
////////////////////////////////////////////////////////////
|
||||
/// Tests
|
||||
////////////////////////////////////////////////////////////
|
||||
// Tests
|
||||
|
||||
#![cfg(not(feature = "stateless"))]
|
||||
|
||||
#[cfg(test)]
|
||||
mod test {
|
||||
use rln::hashers::{poseidon_hash, PoseidonHash};
|
||||
use rln::{circuit::*, poseidon_tree::PoseidonTree};
|
||||
use utils::{FullMerkleTree, OptimalMerkleTree, ZerokitMerkleProof, ZerokitMerkleTree};
|
||||
use rln::prelude::*;
|
||||
use zerokit_utils::merkle_tree::{
|
||||
FullMerkleTree, OptimalMerkleTree, ZerokitMerkleProof, ZerokitMerkleTree,
|
||||
};
|
||||
|
||||
#[test]
|
||||
// The test is checked correctness for `FullMerkleTree` and `OptimalMerkleTree` with Poseidon hash
|
||||
// The test checked correctness for `FullMerkleTree` and `OptimalMerkleTree` with Poseidon hash
|
||||
fn test_zerokit_merkle_implementations() {
|
||||
let sample_size = 100;
|
||||
let leaves: Vec<Fr> = (0..sample_size).map(|s| Fr::from(s)).collect();
|
||||
let leaves: Vec<Fr> = (0..sample_size).map(Fr::from).collect();
|
||||
|
||||
let mut tree_full = FullMerkleTree::<PoseidonHash>::default(TEST_TREE_HEIGHT).unwrap();
|
||||
let mut tree_opt = OptimalMerkleTree::<PoseidonHash>::default(TEST_TREE_HEIGHT).unwrap();
|
||||
let mut tree_full = FullMerkleTree::<PoseidonHash>::default(DEFAULT_TREE_DEPTH).unwrap();
|
||||
let mut tree_opt = OptimalMerkleTree::<PoseidonHash>::default(DEFAULT_TREE_DEPTH).unwrap();
|
||||
|
||||
for i in 0..sample_size.try_into().unwrap() {
|
||||
tree_full.set(i, leaves[i]).unwrap();
|
||||
let proof = tree_full.proof(i).expect("index should be set");
|
||||
for (i, leave) in leaves
|
||||
.into_iter()
|
||||
.enumerate()
|
||||
.take(sample_size.try_into().unwrap())
|
||||
{
|
||||
tree_full.set(i, leave).unwrap();
|
||||
let proof = tree_full.proof(i).unwrap();
|
||||
assert_eq!(proof.leaf_index(), i);
|
||||
|
||||
tree_opt.set(i, leaves[i]).unwrap();
|
||||
let proof = tree_opt.proof(i).expect("index should be set");
|
||||
tree_opt.set(i, leave).unwrap();
|
||||
assert_eq!(tree_opt.root(), tree_full.root());
|
||||
let proof = tree_opt.proof(i).unwrap();
|
||||
assert_eq!(proof.leaf_index(), i);
|
||||
}
|
||||
|
||||
@@ -37,11 +43,11 @@ mod test {
|
||||
#[test]
|
||||
fn test_subtree_root() {
|
||||
const DEPTH: usize = 3;
|
||||
const LEAVES_LEN: usize = 6;
|
||||
const LEAVES_LEN: usize = 8;
|
||||
|
||||
let mut tree = PoseidonTree::default(DEPTH).unwrap();
|
||||
let leaves: Vec<Fr> = (0..LEAVES_LEN).map(|s| Fr::from(s as i32)).collect();
|
||||
let _ = tree.set_range(0, leaves);
|
||||
let _ = tree.set_range(0, leaves.into_iter());
|
||||
|
||||
for i in 0..LEAVES_LEN {
|
||||
// check leaves
|
||||
@@ -64,7 +70,7 @@ mod test {
|
||||
let prev_r = tree.get_subtree_root(n, idx_r).unwrap();
|
||||
let subroot = tree.get_subtree_root(n - 1, idx_sr).unwrap();
|
||||
|
||||
assert_eq!(poseidon_hash(&[prev_l, prev_r]), subroot);
|
||||
assert_eq!(poseidon_hash(&[prev_l, prev_r]).unwrap(), subroot);
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -78,7 +84,7 @@ mod test {
|
||||
let leaves: Vec<Fr> = (0..nof_leaves).map(|s| Fr::from(s as i32)).collect();
|
||||
|
||||
// check set_range
|
||||
let _ = tree.set_range(0, leaves.clone());
|
||||
let _ = tree.set_range(0, leaves.clone().into_iter());
|
||||
assert!(tree.get_empty_leaves_indices().is_empty());
|
||||
|
||||
let mut vec_idxs = Vec::new();
|
||||
@@ -97,27 +103,29 @@ mod test {
|
||||
|
||||
// check remove_indices_and_set_leaves inside override_range function
|
||||
assert!(tree.get_empty_leaves_indices().is_empty());
|
||||
let leaves_2: Vec<Fr> = (0..2).map(|s| Fr::from(s as i32)).collect();
|
||||
tree.override_range(0, leaves_2.clone(), [0, 1, 2, 3])
|
||||
let leaves_2: Vec<Fr> = (0..2).map(Fr::from).collect();
|
||||
tree.override_range(0, leaves_2.clone().into_iter(), [0, 1, 2, 3].into_iter())
|
||||
.unwrap();
|
||||
assert_eq!(tree.get_empty_leaves_indices(), vec![2, 3]);
|
||||
|
||||
// check remove_indices inside override_range function
|
||||
tree.override_range(0, [], [0, 1]).unwrap();
|
||||
tree.override_range(0, [].into_iter(), [0, 1].into_iter())
|
||||
.unwrap();
|
||||
assert_eq!(tree.get_empty_leaves_indices(), vec![0, 1, 2, 3]);
|
||||
|
||||
// check set_range inside override_range function
|
||||
tree.override_range(0, leaves_2.clone(), []).unwrap();
|
||||
tree.override_range(0, leaves_2.clone().into_iter(), [].into_iter())
|
||||
.unwrap();
|
||||
assert_eq!(tree.get_empty_leaves_indices(), vec![2, 3]);
|
||||
|
||||
let leaves_4: Vec<Fr> = (0..4).map(|s| Fr::from(s as i32)).collect();
|
||||
let leaves_4: Vec<Fr> = (0..4).map(Fr::from).collect();
|
||||
// check if the indexes for write and delete are the same
|
||||
tree.override_range(0, leaves_4.clone(), [0, 1, 2, 3])
|
||||
tree.override_range(0, leaves_4.clone().into_iter(), [0, 1, 2, 3].into_iter())
|
||||
.unwrap();
|
||||
assert!(tree.get_empty_leaves_indices().is_empty());
|
||||
|
||||
// check if indexes for deletion are before indexes for overwriting
|
||||
tree.override_range(4, leaves_4.clone(), [0, 1, 2, 3])
|
||||
tree.override_range(4, leaves_4.clone().into_iter(), [0, 1, 2, 3].into_iter())
|
||||
.unwrap();
|
||||
// The result will be like this, because in the set_range function in pmtree
|
||||
// the next_index value is increased not by the number of elements to insert,
|
||||
@@ -128,7 +136,7 @@ mod test {
|
||||
);
|
||||
|
||||
// check if the indices for write and delete do not overlap completely
|
||||
tree.override_range(2, leaves_4.clone(), [0, 1, 2, 3])
|
||||
tree.override_range(2, leaves_4.clone().into_iter(), [0, 1, 2, 3].into_iter())
|
||||
.unwrap();
|
||||
// The result will be like this, because in the set_range function in pmtree
|
||||
// the next_index value is increased not by the number of elements to insert,
|
||||
|
||||
@@ -1,13 +1,10 @@
|
||||
#![cfg(not(feature = "stateless"))]
|
||||
|
||||
#[cfg(test)]
|
||||
mod test {
|
||||
use ark_ff::BigInt;
|
||||
use rln::circuit::zkey_from_folder;
|
||||
use rln::circuit::{circom_from_folder, vk_from_folder, Fr, TEST_TREE_HEIGHT};
|
||||
use rln::hashers::{hash_to_field, poseidon_hash};
|
||||
use rln::poseidon_tree::PoseidonTree;
|
||||
use rln::protocol::*;
|
||||
use rln::utils::str_to_fr;
|
||||
use utils::{ZerokitMerkleProof, ZerokitMerkleTree};
|
||||
use rln::prelude::*;
|
||||
use zerokit_utils::merkle_tree::{ZerokitMerkleProof, ZerokitMerkleTree};
|
||||
|
||||
type ConfigOf<T> = <T as ZerokitMerkleTree>::Config;
|
||||
|
||||
@@ -17,19 +14,19 @@ mod test {
|
||||
let leaf_index = 3;
|
||||
|
||||
// generate identity
|
||||
let identity_secret_hash = hash_to_field(b"test-merkle-proof");
|
||||
let id_commitment = poseidon_hash(&[identity_secret_hash]);
|
||||
let rate_commitment = poseidon_hash(&[id_commitment, 100.into()]);
|
||||
let identity_secret = hash_to_field_le(b"test-merkle-proof").unwrap();
|
||||
let id_commitment = poseidon_hash(&[identity_secret]).unwrap();
|
||||
let rate_commitment = poseidon_hash(&[id_commitment, 100.into()]).unwrap();
|
||||
|
||||
// generate merkle tree
|
||||
let default_leaf = Fr::from(0);
|
||||
let mut tree = PoseidonTree::new(
|
||||
TEST_TREE_HEIGHT,
|
||||
DEFAULT_TREE_DEPTH,
|
||||
default_leaf,
|
||||
ConfigOf::<PoseidonTree>::default(),
|
||||
)
|
||||
.unwrap();
|
||||
tree.set(leaf_index, rate_commitment.into()).unwrap();
|
||||
tree.set(leaf_index, rate_commitment).unwrap();
|
||||
|
||||
// We check correct computation of the root
|
||||
let root = tree.root();
|
||||
@@ -45,7 +42,7 @@ mod test {
|
||||
.into()
|
||||
);
|
||||
|
||||
let merkle_proof = tree.proof(leaf_index).expect("proof should exist");
|
||||
let merkle_proof = tree.proof(leaf_index).unwrap();
|
||||
let path_elements = merkle_proof.get_path_elements();
|
||||
let identity_path_index = merkle_proof.get_path_index();
|
||||
|
||||
@@ -72,7 +69,7 @@ mod test {
|
||||
"0x0f57c5571e9a4eab49e2c8cf050dae948aef6ead647392273546249d1c1ff10f",
|
||||
"0x1830ee67b5fb554ad5f63d4388800e1cfe78e310697d46e43c9ce36134f72cca",
|
||||
]
|
||||
.map(|e| str_to_fr(e, 16).unwrap())
|
||||
.map(|str| str_to_fr(str, 16).unwrap())
|
||||
.to_vec();
|
||||
|
||||
let expected_identity_path_index: Vec<u8> =
|
||||
@@ -88,106 +85,77 @@ mod test {
|
||||
fn get_test_witness() -> RLNWitnessInput {
|
||||
let leaf_index = 3;
|
||||
// Generate identity pair
|
||||
let (identity_secret_hash, id_commitment) = keygen();
|
||||
let (identity_secret, id_commitment) = keygen().unwrap();
|
||||
let user_message_limit = Fr::from(100);
|
||||
let rate_commitment = poseidon_hash(&[id_commitment, user_message_limit]);
|
||||
let rate_commitment = poseidon_hash(&[id_commitment, user_message_limit]).unwrap();
|
||||
|
||||
//// generate merkle tree
|
||||
let default_leaf = Fr::from(0);
|
||||
let mut tree = PoseidonTree::new(
|
||||
TEST_TREE_HEIGHT,
|
||||
DEFAULT_TREE_DEPTH,
|
||||
default_leaf,
|
||||
ConfigOf::<PoseidonTree>::default(),
|
||||
)
|
||||
.unwrap();
|
||||
tree.set(leaf_index, rate_commitment.into()).unwrap();
|
||||
tree.set(leaf_index, rate_commitment).unwrap();
|
||||
|
||||
let merkle_proof = tree.proof(leaf_index).expect("proof should exist");
|
||||
let merkle_proof = tree.proof(leaf_index).unwrap();
|
||||
|
||||
let signal = b"hey hey";
|
||||
let x = hash_to_field(signal);
|
||||
let x = hash_to_field_le(signal).unwrap();
|
||||
|
||||
// We set the remaining values to random ones
|
||||
let epoch = hash_to_field(b"test-epoch");
|
||||
let rln_identifier = hash_to_field(b"test-rln-identifier");
|
||||
let external_nullifier = poseidon_hash(&[epoch, rln_identifier]);
|
||||
let epoch = hash_to_field_le(b"test-epoch").unwrap();
|
||||
let rln_identifier = hash_to_field_le(b"test-rln-identifier").unwrap();
|
||||
let external_nullifier = poseidon_hash(&[epoch, rln_identifier]).unwrap();
|
||||
|
||||
rln_witness_from_values(
|
||||
identity_secret_hash,
|
||||
&merkle_proof,
|
||||
let message_id = Fr::from(1);
|
||||
|
||||
RLNWitnessInput::new(
|
||||
identity_secret,
|
||||
user_message_limit,
|
||||
message_id,
|
||||
merkle_proof.get_path_elements(),
|
||||
merkle_proof.get_path_index(),
|
||||
x,
|
||||
external_nullifier,
|
||||
user_message_limit,
|
||||
Fr::from(1),
|
||||
)
|
||||
.unwrap()
|
||||
}
|
||||
|
||||
#[test]
|
||||
// We test a RLN proof generation and verification
|
||||
fn test_witness_from_json() {
|
||||
// We generate all relevant keys
|
||||
let proving_key = zkey_from_folder();
|
||||
let verification_key = vk_from_folder();
|
||||
let builder = circom_from_folder();
|
||||
|
||||
// We compute witness from the json input
|
||||
let rln_witness = get_test_witness();
|
||||
let rln_witness_json = rln_witness_to_json(&rln_witness).unwrap();
|
||||
let rln_witness_deser = rln_witness_from_json(rln_witness_json).unwrap();
|
||||
assert_eq!(rln_witness_deser, rln_witness);
|
||||
|
||||
// Let's generate a zkSNARK proof
|
||||
let proof = generate_proof(builder, &proving_key, &rln_witness_deser).unwrap();
|
||||
let proof_values = proof_values_from_witness(&rln_witness_deser).unwrap();
|
||||
|
||||
// Let's verify the proof
|
||||
let verified = verify_proof(&verification_key, &proof, &proof_values);
|
||||
|
||||
assert!(verified.unwrap());
|
||||
}
|
||||
|
||||
#[test]
|
||||
// We test a RLN proof generation and verification
|
||||
fn test_end_to_end() {
|
||||
let rln_witness = get_test_witness();
|
||||
let rln_witness_json = rln_witness_to_json(&rln_witness).unwrap();
|
||||
let rln_witness_deser = rln_witness_from_json(rln_witness_json).unwrap();
|
||||
assert_eq!(rln_witness_deser, rln_witness);
|
||||
let witness = get_test_witness();
|
||||
|
||||
// We generate all relevant keys
|
||||
let proving_key = zkey_from_folder();
|
||||
let verification_key = vk_from_folder();
|
||||
let builder = circom_from_folder();
|
||||
let graph_data = graph_from_folder();
|
||||
|
||||
// Let's generate a zkSNARK proof
|
||||
let proof = generate_proof(builder, &proving_key, &rln_witness_deser).unwrap();
|
||||
let proof = generate_zk_proof(proving_key, &witness, graph_data).unwrap();
|
||||
|
||||
let proof_values = proof_values_from_witness(&rln_witness_deser).unwrap();
|
||||
let proof_values = proof_values_from_witness(&witness).unwrap();
|
||||
|
||||
// Let's verify the proof
|
||||
let success = verify_proof(&verification_key, &proof, &proof_values).unwrap();
|
||||
let success = verify_zk_proof(&proving_key.0.vk, &proof, &proof_values).unwrap();
|
||||
|
||||
assert!(success);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_witness_serialization() {
|
||||
// We test witness JSON serialization
|
||||
let rln_witness = get_test_witness();
|
||||
let rln_witness_json = rln_witness_to_json(&rln_witness).unwrap();
|
||||
let rln_witness_deser = rln_witness_from_json(rln_witness_json).unwrap();
|
||||
assert_eq!(rln_witness_deser, rln_witness);
|
||||
let witness = get_test_witness();
|
||||
|
||||
// We test witness serialization
|
||||
let ser = serialize_witness(&rln_witness).unwrap();
|
||||
let (deser, _) = deserialize_witness(&ser).unwrap();
|
||||
assert_eq!(rln_witness, deser);
|
||||
let ser = rln_witness_to_bytes_le(&witness).unwrap();
|
||||
let (deser, _) = bytes_le_to_rln_witness(&ser).unwrap();
|
||||
assert_eq!(witness, deser);
|
||||
|
||||
// We test Proof values serialization
|
||||
let proof_values = proof_values_from_witness(&rln_witness).unwrap();
|
||||
let ser = serialize_proof_values(&proof_values);
|
||||
let (deser, _) = deserialize_proof_values(&ser);
|
||||
let proof_values = proof_values_from_witness(&witness).unwrap();
|
||||
let ser = rln_proof_values_to_bytes_le(&proof_values);
|
||||
let (deser, _) = bytes_le_to_rln_proof_values(&ser).unwrap();
|
||||
assert_eq!(proof_values, deser);
|
||||
}
|
||||
|
||||
@@ -197,10 +165,10 @@ mod test {
|
||||
fn test_seeded_keygen() {
|
||||
// Generate identity pair using a seed phrase
|
||||
let seed_phrase: &str = "A seed phrase example";
|
||||
let (identity_secret_hash, id_commitment) = seeded_keygen(seed_phrase.as_bytes());
|
||||
let (identity_secret, id_commitment) = seeded_keygen(seed_phrase.as_bytes()).unwrap();
|
||||
|
||||
// We check against expected values
|
||||
let expected_identity_secret_hash_seed_phrase = str_to_fr(
|
||||
let expected_identity_secret_seed_phrase = str_to_fr(
|
||||
"0x20df38f3f00496f19fe7c6535492543b21798ed7cb91aebe4af8012db884eda3",
|
||||
16,
|
||||
)
|
||||
@@ -211,18 +179,15 @@ mod test {
|
||||
)
|
||||
.unwrap();
|
||||
|
||||
assert_eq!(
|
||||
identity_secret_hash,
|
||||
expected_identity_secret_hash_seed_phrase
|
||||
);
|
||||
assert_eq!(identity_secret, expected_identity_secret_seed_phrase);
|
||||
assert_eq!(id_commitment, expected_id_commitment_seed_phrase);
|
||||
|
||||
// Generate identity pair using an byte array
|
||||
let seed_bytes: &[u8] = &[0, 1, 2, 3, 4, 5, 6, 7, 8, 9];
|
||||
let (identity_secret_hash, id_commitment) = seeded_keygen(seed_bytes);
|
||||
let (identity_secret, id_commitment) = seeded_keygen(seed_bytes).unwrap();
|
||||
|
||||
// We check against expected values
|
||||
let expected_identity_secret_hash_seed_bytes = str_to_fr(
|
||||
let expected_identity_secret_seed_bytes = str_to_fr(
|
||||
"0x766ce6c7e7a01bdf5b3f257616f603918c30946fa23480f2859c597817e6716",
|
||||
16,
|
||||
)
|
||||
@@ -233,19 +198,13 @@ mod test {
|
||||
)
|
||||
.unwrap();
|
||||
|
||||
assert_eq!(
|
||||
identity_secret_hash,
|
||||
expected_identity_secret_hash_seed_bytes
|
||||
);
|
||||
assert_eq!(identity_secret, expected_identity_secret_seed_bytes);
|
||||
assert_eq!(id_commitment, expected_id_commitment_seed_bytes);
|
||||
|
||||
// We check again if the identity pair generated with the same seed phrase corresponds to the previously generated one
|
||||
let (identity_secret_hash, id_commitment) = seeded_keygen(seed_phrase.as_bytes());
|
||||
let (identity_secret, id_commitment) = seeded_keygen(seed_phrase.as_bytes()).unwrap();
|
||||
|
||||
assert_eq!(
|
||||
identity_secret_hash,
|
||||
expected_identity_secret_hash_seed_phrase
|
||||
);
|
||||
assert_eq!(identity_secret, expected_identity_secret_seed_phrase);
|
||||
assert_eq!(id_commitment, expected_id_commitment_seed_phrase);
|
||||
}
|
||||
}
|
||||
|
||||
1283
rln/tests/public.rs
1283
rln/tests/public.rs
File diff suppressed because it is too large
Load Diff
424
rln/tests/utils.rs
Normal file
424
rln/tests/utils.rs
Normal file
@@ -0,0 +1,424 @@
|
||||
#[cfg(test)]
|
||||
mod test {
|
||||
use ark_std::{rand::thread_rng, UniformRand};
|
||||
use rln::prelude::*;
|
||||
|
||||
#[test]
|
||||
fn test_normalize_usize_le() {
|
||||
// Test basic cases
|
||||
assert_eq!(normalize_usize_le(0), [0, 0, 0, 0, 0, 0, 0, 0]);
|
||||
assert_eq!(normalize_usize_le(1), [1, 0, 0, 0, 0, 0, 0, 0]);
|
||||
assert_eq!(normalize_usize_le(255), [255, 0, 0, 0, 0, 0, 0, 0]);
|
||||
assert_eq!(normalize_usize_le(256), [0, 1, 0, 0, 0, 0, 0, 0]);
|
||||
assert_eq!(normalize_usize_le(65535), [255, 255, 0, 0, 0, 0, 0, 0]);
|
||||
assert_eq!(normalize_usize_le(65536), [0, 0, 1, 0, 0, 0, 0, 0]);
|
||||
|
||||
// Test 32-bit boundary
|
||||
assert_eq!(
|
||||
normalize_usize_le(4294967295),
|
||||
[255, 255, 255, 255, 0, 0, 0, 0]
|
||||
);
|
||||
assert_eq!(normalize_usize_le(4294967296), [0, 0, 0, 0, 1, 0, 0, 0]);
|
||||
|
||||
// Test maximum value
|
||||
assert_eq!(
|
||||
normalize_usize_le(usize::MAX),
|
||||
[255, 255, 255, 255, 255, 255, 255, 255]
|
||||
);
|
||||
|
||||
// Test that result is always 8 bytes
|
||||
assert_eq!(normalize_usize_le(0).len(), 8);
|
||||
assert_eq!(normalize_usize_le(usize::MAX).len(), 8);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_normalize_usize_be() {
|
||||
// Test basic cases
|
||||
assert_eq!(normalize_usize_be(0), [0, 0, 0, 0, 0, 0, 0, 0]);
|
||||
assert_eq!(normalize_usize_be(1), [0, 0, 0, 0, 0, 0, 0, 1]);
|
||||
assert_eq!(normalize_usize_be(255), [0, 0, 0, 0, 0, 0, 0, 255]);
|
||||
assert_eq!(normalize_usize_be(256), [0, 0, 0, 0, 0, 0, 1, 0]);
|
||||
assert_eq!(normalize_usize_be(65535), [0, 0, 0, 0, 0, 0, 255, 255]);
|
||||
assert_eq!(normalize_usize_be(65536), [0, 0, 0, 0, 0, 1, 0, 0]);
|
||||
|
||||
// Test 32-bit boundary
|
||||
assert_eq!(
|
||||
normalize_usize_be(4294967295),
|
||||
[0, 0, 0, 0, 255, 255, 255, 255]
|
||||
);
|
||||
assert_eq!(normalize_usize_be(4294967296), [0, 0, 0, 1, 0, 0, 0, 0]);
|
||||
|
||||
// Test maximum value
|
||||
assert_eq!(
|
||||
normalize_usize_be(usize::MAX),
|
||||
[255, 255, 255, 255, 255, 255, 255, 255]
|
||||
);
|
||||
|
||||
// Test that result is always 8 bytes
|
||||
assert_eq!(normalize_usize_be(0).len(), 8);
|
||||
assert_eq!(normalize_usize_be(usize::MAX).len(), 8);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_normalize_usize_endianness() {
|
||||
// Test that little-endian and big-endian produce different results for non-zero values
|
||||
let test_values = vec![1, 255, 256, 65535, 65536, 4294967295, 4294967296];
|
||||
|
||||
for &value in &test_values {
|
||||
let le_result = normalize_usize_le(value);
|
||||
let be_result = normalize_usize_be(value);
|
||||
|
||||
// For non-zero values, LE and BE should be different
|
||||
assert_ne!(
|
||||
le_result, be_result,
|
||||
"LE and BE should differ for value {value}"
|
||||
);
|
||||
|
||||
// Both should be 8 bytes
|
||||
assert_eq!(le_result.len(), 8);
|
||||
assert_eq!(be_result.len(), 8);
|
||||
}
|
||||
|
||||
// Zero should be the same in both endianness
|
||||
assert_eq!(normalize_usize_le(0), normalize_usize_be(0));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_normalize_usize_roundtrip() {
|
||||
// Test that we can reconstruct the original value from the normalized bytes
|
||||
let test_values = vec![
|
||||
0,
|
||||
1,
|
||||
255,
|
||||
256,
|
||||
65535,
|
||||
65536,
|
||||
4294967295,
|
||||
4294967296,
|
||||
usize::MAX,
|
||||
];
|
||||
|
||||
for &value in &test_values {
|
||||
let le_bytes = normalize_usize_le(value);
|
||||
let be_bytes = normalize_usize_be(value);
|
||||
|
||||
// Reconstruct from little-endian bytes
|
||||
let reconstructed_le = usize::from_le_bytes(le_bytes);
|
||||
assert_eq!(
|
||||
reconstructed_le, value,
|
||||
"LE roundtrip failed for value {value}"
|
||||
);
|
||||
|
||||
// Reconstruct from big-endian bytes
|
||||
let reconstructed_be = usize::from_be_bytes(be_bytes);
|
||||
assert_eq!(
|
||||
reconstructed_be, value,
|
||||
"BE roundtrip failed for value {value}"
|
||||
);
|
||||
}
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_normalize_usize_edge_cases() {
|
||||
// Test edge cases and boundary values
|
||||
let edge_cases = vec![
|
||||
0,
|
||||
1,
|
||||
255,
|
||||
256,
|
||||
65535,
|
||||
65536,
|
||||
16777215, // 2^24 - 1
|
||||
16777216, // 2^24
|
||||
4294967295, // 2^32 - 1
|
||||
4294967296, // 2^32
|
||||
1099511627775, // 2^40 - 1
|
||||
1099511627776, // 2^40
|
||||
281474976710655, // 2^48 - 1
|
||||
281474976710656, // 2^48
|
||||
72057594037927935, // 2^56 - 1
|
||||
72057594037927936, // 2^56
|
||||
usize::MAX,
|
||||
];
|
||||
|
||||
for &value in &edge_cases {
|
||||
let le_result = normalize_usize_le(value);
|
||||
let be_result = normalize_usize_be(value);
|
||||
|
||||
// Both should be 8 bytes
|
||||
assert_eq!(le_result.len(), 8);
|
||||
assert_eq!(be_result.len(), 8);
|
||||
|
||||
// Roundtrip should work
|
||||
assert_eq!(usize::from_le_bytes(le_result), value);
|
||||
assert_eq!(usize::from_be_bytes(be_result), value);
|
||||
}
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_normalize_usize_architecture_independence() {
|
||||
// Test that the functions work consistently regardless of the underlying architecture
|
||||
// This test ensures that the functions provide consistent 8-byte output
|
||||
// even on 32-bit systems where usize might be 4 bytes
|
||||
|
||||
let test_values = vec![0, 1, 255, 256, 65535, 65536, 4294967295, 4294967296];
|
||||
|
||||
for &value in &test_values {
|
||||
let le_result = normalize_usize_le(value);
|
||||
let be_result = normalize_usize_be(value);
|
||||
|
||||
// Always 8 bytes regardless of architecture
|
||||
assert_eq!(le_result.len(), 8);
|
||||
assert_eq!(be_result.len(), 8);
|
||||
|
||||
// The result should be consistent with the original value
|
||||
assert_eq!(usize::from_le_bytes(le_result), value);
|
||||
assert_eq!(usize::from_be_bytes(be_result), value);
|
||||
}
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_fr_serialization_roundtrip() {
|
||||
let mut rng = thread_rng();
|
||||
|
||||
// Test multiple random Fr values
|
||||
for _ in 0..10 {
|
||||
let fr = Fr::rand(&mut rng);
|
||||
|
||||
// Test little-endian roundtrip
|
||||
let le_bytes = fr_to_bytes_le(&fr);
|
||||
let (reconstructed_le, _) = bytes_le_to_fr(&le_bytes).unwrap();
|
||||
assert_eq!(fr, reconstructed_le);
|
||||
|
||||
// Test big-endian roundtrip
|
||||
let be_bytes = fr_to_bytes_be(&fr);
|
||||
let (reconstructed_be, _) = bytes_be_to_fr(&be_bytes).unwrap();
|
||||
assert_eq!(fr, reconstructed_be);
|
||||
}
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_vec_fr_serialization_roundtrip() {
|
||||
let mut rng = thread_rng();
|
||||
|
||||
// Test with different vector sizes
|
||||
for size in [0, 1, 5, 10] {
|
||||
let fr_vec: Vec<Fr> = (0..size).map(|_| Fr::rand(&mut rng)).collect();
|
||||
|
||||
// Test little-endian roundtrip
|
||||
let le_bytes = vec_fr_to_bytes_le(&fr_vec);
|
||||
let (reconstructed_le, _) = bytes_le_to_vec_fr(&le_bytes).unwrap();
|
||||
assert_eq!(fr_vec, reconstructed_le);
|
||||
|
||||
// Test big-endian roundtrip
|
||||
let be_bytes = vec_fr_to_bytes_be(&fr_vec);
|
||||
let (reconstructed_be, _) = bytes_be_to_vec_fr(&be_bytes).unwrap();
|
||||
assert_eq!(fr_vec, reconstructed_be);
|
||||
}
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_vec_u8_serialization_roundtrip() {
|
||||
// Test with different vector sizes and content
|
||||
let test_cases = vec![
|
||||
vec![],
|
||||
vec![0],
|
||||
vec![255],
|
||||
vec![1, 2, 3, 4, 5],
|
||||
vec![0, 255, 128, 64, 32, 16, 8, 4, 2, 1],
|
||||
(0..100).collect::<Vec<u8>>(),
|
||||
];
|
||||
|
||||
for test_case in test_cases {
|
||||
// Test little-endian roundtrip
|
||||
let le_bytes = vec_u8_to_bytes_le(&test_case);
|
||||
let (reconstructed_le, _) = bytes_le_to_vec_u8(&le_bytes).unwrap();
|
||||
assert_eq!(test_case, reconstructed_le);
|
||||
|
||||
// Test big-endian roundtrip
|
||||
let be_bytes = vec_u8_to_bytes_be(&test_case);
|
||||
let (reconstructed_be, _) = bytes_be_to_vec_u8(&be_bytes).unwrap();
|
||||
assert_eq!(test_case, reconstructed_be);
|
||||
}
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_vec_usize_serialization_roundtrip() {
|
||||
// Test with different vector sizes and content
|
||||
let test_cases = vec![
|
||||
vec![],
|
||||
vec![0],
|
||||
vec![usize::MAX],
|
||||
vec![1, 2, 3, 4, 5],
|
||||
vec![0, 255, 65535, 4294967295, usize::MAX],
|
||||
(0..10).collect::<Vec<usize>>(),
|
||||
];
|
||||
|
||||
for test_case in test_cases {
|
||||
// Test little-endian roundtrip
|
||||
let le_bytes = {
|
||||
let mut bytes = Vec::new();
|
||||
bytes.extend_from_slice(&normalize_usize_le(test_case.len()));
|
||||
for &value in &test_case {
|
||||
bytes.extend_from_slice(&normalize_usize_le(value));
|
||||
}
|
||||
bytes
|
||||
};
|
||||
let reconstructed_le = bytes_le_to_vec_usize(&le_bytes).unwrap();
|
||||
assert_eq!(test_case, reconstructed_le);
|
||||
|
||||
// Test big-endian roundtrip
|
||||
let be_bytes = {
|
||||
let mut bytes = Vec::new();
|
||||
bytes.extend_from_slice(&normalize_usize_be(test_case.len()));
|
||||
for &value in &test_case {
|
||||
bytes.extend_from_slice(&normalize_usize_be(value));
|
||||
}
|
||||
bytes
|
||||
};
|
||||
let reconstructed_be = bytes_be_to_vec_usize(&be_bytes).unwrap();
|
||||
assert_eq!(test_case, reconstructed_be);
|
||||
}
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_str_to_fr() {
|
||||
// Test valid hex strings
|
||||
let test_cases = vec![
|
||||
("0x0", 16, Fr::from(0u64)),
|
||||
("0x1", 16, Fr::from(1u64)),
|
||||
("0xff", 16, Fr::from(255u64)),
|
||||
("0x100", 16, Fr::from(256u64)),
|
||||
];
|
||||
|
||||
for (input, radix, expected) in test_cases {
|
||||
let result = str_to_fr(input, radix).unwrap();
|
||||
assert_eq!(result, expected);
|
||||
}
|
||||
|
||||
// Test invalid inputs
|
||||
assert!(str_to_fr("invalid", 16).is_err());
|
||||
assert!(str_to_fr("0x", 16).is_err());
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_endianness_differences() {
|
||||
let mut rng = thread_rng();
|
||||
let fr = Fr::rand(&mut rng);
|
||||
|
||||
// Test that LE and BE produce different byte representations
|
||||
let le_bytes = fr_to_bytes_le(&fr);
|
||||
let be_bytes = fr_to_bytes_be(&fr);
|
||||
|
||||
// They should be different (unless the value is symmetric)
|
||||
if le_bytes != be_bytes {
|
||||
// Verify they can both be reconstructed correctly
|
||||
let (reconstructed_le, _) = bytes_le_to_fr(&le_bytes).unwrap();
|
||||
let (reconstructed_be, _) = bytes_be_to_fr(&be_bytes).unwrap();
|
||||
assert_eq!(fr, reconstructed_le);
|
||||
assert_eq!(fr, reconstructed_be);
|
||||
}
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_error_handling() {
|
||||
// Test bytes_le_to_fr and bytes_be_to_fr with insufficient data
|
||||
let short_bytes = vec![0u8; 10]; // Less than FR_BYTE_SIZE (32 bytes)
|
||||
assert!(bytes_le_to_fr(&short_bytes).is_err());
|
||||
assert!(bytes_be_to_fr(&short_bytes).is_err());
|
||||
|
||||
// Test with empty bytes
|
||||
let empty_bytes = vec![];
|
||||
assert!(bytes_le_to_fr(&empty_bytes).is_err());
|
||||
assert!(bytes_be_to_fr(&empty_bytes).is_err());
|
||||
|
||||
// Test with exact size - should succeed
|
||||
let exact_bytes = vec![0u8; FR_BYTE_SIZE];
|
||||
assert!(bytes_le_to_fr(&exact_bytes).is_ok());
|
||||
assert!(bytes_be_to_fr(&exact_bytes).is_ok());
|
||||
|
||||
// Test with more than enough data - should succeed
|
||||
let extra_bytes = vec![0u8; FR_BYTE_SIZE + 10];
|
||||
assert!(bytes_le_to_fr(&extra_bytes).is_ok());
|
||||
assert!(bytes_be_to_fr(&extra_bytes).is_ok());
|
||||
|
||||
// Test with valid length but insufficient data for vector deserialization
|
||||
let valid_length_invalid_data = vec![0u8; 8]; // Length 0, but no data
|
||||
assert!(bytes_le_to_vec_u8(&valid_length_invalid_data).is_ok());
|
||||
assert!(bytes_be_to_vec_u8(&valid_length_invalid_data).is_ok());
|
||||
assert!(bytes_le_to_vec_fr(&valid_length_invalid_data).is_ok());
|
||||
assert!(bytes_be_to_vec_fr(&valid_length_invalid_data).is_ok());
|
||||
assert!(bytes_le_to_vec_usize(&valid_length_invalid_data).is_ok());
|
||||
assert!(bytes_be_to_vec_usize(&valid_length_invalid_data).is_ok());
|
||||
|
||||
// Test with reasonable length but insufficient data for vector deserialization
|
||||
let reasonable_length = {
|
||||
let mut bytes = vec![0u8; 8];
|
||||
bytes[0] = 1; // Length 1
|
||||
bytes
|
||||
};
|
||||
// This should fail because we don't have enough data for the vector elements
|
||||
assert!(bytes_le_to_vec_u8(&reasonable_length).is_err());
|
||||
assert!(bytes_be_to_vec_u8(&reasonable_length).is_err());
|
||||
assert!(bytes_le_to_vec_fr(&reasonable_length).is_err());
|
||||
assert!(bytes_be_to_vec_fr(&reasonable_length).is_err());
|
||||
assert!(bytes_le_to_vec_usize(&reasonable_length).is_err());
|
||||
assert!(bytes_be_to_vec_usize(&reasonable_length).is_err());
|
||||
|
||||
// Test with valid data for u8 vector
|
||||
let valid_u8_data_le = {
|
||||
let mut bytes = vec![0u8; 9];
|
||||
bytes[..8].copy_from_slice(&(1u64.to_le_bytes())); // Length 1, little-endian
|
||||
bytes[8] = 42; // One byte of data
|
||||
bytes
|
||||
};
|
||||
let valid_u8_data_be = {
|
||||
let mut bytes = vec![0u8; 9];
|
||||
bytes[..8].copy_from_slice(&(1u64.to_be_bytes())); // Length 1, big-endian
|
||||
bytes[8] = 42; // One byte of data
|
||||
bytes
|
||||
};
|
||||
assert!(bytes_le_to_vec_u8(&valid_u8_data_le).is_ok());
|
||||
assert!(bytes_be_to_vec_u8(&valid_u8_data_be).is_ok());
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_empty_vectors() {
|
||||
// Test empty vector serialization/deserialization
|
||||
let empty_fr: Vec<Fr> = vec![];
|
||||
let empty_u8: Vec<u8> = vec![];
|
||||
let empty_usize: Vec<usize> = vec![];
|
||||
|
||||
// Test Fr vectors
|
||||
let le_fr_bytes = vec_fr_to_bytes_le(&empty_fr);
|
||||
let be_fr_bytes = vec_fr_to_bytes_be(&empty_fr);
|
||||
let (reconstructed_le_fr, _) = bytes_le_to_vec_fr(&le_fr_bytes).unwrap();
|
||||
let (reconstructed_be_fr, _) = bytes_be_to_vec_fr(&be_fr_bytes).unwrap();
|
||||
assert_eq!(empty_fr, reconstructed_le_fr);
|
||||
assert_eq!(empty_fr, reconstructed_be_fr);
|
||||
|
||||
// Test u8 vectors
|
||||
let le_u8_bytes = vec_u8_to_bytes_le(&empty_u8);
|
||||
let be_u8_bytes = vec_u8_to_bytes_be(&empty_u8);
|
||||
let (reconstructed_le_u8, _) = bytes_le_to_vec_u8(&le_u8_bytes).unwrap();
|
||||
let (reconstructed_be_u8, _) = bytes_be_to_vec_u8(&be_u8_bytes).unwrap();
|
||||
assert_eq!(empty_u8, reconstructed_le_u8);
|
||||
assert_eq!(empty_u8, reconstructed_be_u8);
|
||||
|
||||
// Test usize vectors
|
||||
let le_usize_bytes = {
|
||||
let mut bytes = Vec::new();
|
||||
bytes.extend_from_slice(&normalize_usize_le(0));
|
||||
bytes
|
||||
};
|
||||
let be_usize_bytes = {
|
||||
let mut bytes = Vec::new();
|
||||
bytes.extend_from_slice(&normalize_usize_be(0));
|
||||
bytes
|
||||
};
|
||||
let reconstructed_le_usize = bytes_le_to_vec_usize(&le_usize_bytes).unwrap();
|
||||
let reconstructed_be_usize = bytes_be_to_vec_usize(&be_usize_bytes).unwrap();
|
||||
assert_eq!(empty_usize, reconstructed_le_usize);
|
||||
assert_eq!(empty_usize, reconstructed_be_usize);
|
||||
}
|
||||
}
|
||||
6
rustfmt.toml
Normal file
6
rustfmt.toml
Normal file
@@ -0,0 +1,6 @@
|
||||
# Run cargo +nightly fmt to format with this configuration
|
||||
edition = "2021" # use Rust 2021 edition
|
||||
unstable_features = true # needed for group_imports
|
||||
reorder_imports = true # sort imports alphabetically
|
||||
imports_granularity = "Crate" # keep items from the same crate grouped together
|
||||
group_imports = "StdExternalCrate" # group std, external, and local imports separately
|
||||
@@ -1,6 +1,6 @@
|
||||
[package]
|
||||
name = "zerokit_utils"
|
||||
version = "0.5.1"
|
||||
version = "1.0.0"
|
||||
edition = "2021"
|
||||
license = "MIT OR Apache-2.0"
|
||||
description = "Various utilities for Zerokit"
|
||||
@@ -12,29 +12,34 @@ repository = "https://github.com/vacp2p/zerokit"
|
||||
bench = false
|
||||
|
||||
[dependencies]
|
||||
ark-ff = { version = "=0.4.1", default-features = false, features = ["asm"] }
|
||||
num-bigint = { version = "=0.4.3", default-features = false, features = [
|
||||
"rand",
|
||||
] }
|
||||
color-eyre = "=0.6.2"
|
||||
pmtree = { package = "vacp2p_pmtree", version = "=2.0.2", optional = true }
|
||||
sled = "=0.34.7"
|
||||
serde = "=1.0.163"
|
||||
lazy_static = "1.4.0"
|
||||
hex = "0.4"
|
||||
ark-ff = { version = "0.5.0", default-features = false }
|
||||
num-bigint = { version = "0.4.6", default-features = false }
|
||||
pmtree = { package = "vacp2p_pmtree", version = "2.0.3", optional = true }
|
||||
sled = "0.34.7"
|
||||
serde_json = "1.0.145"
|
||||
rayon = "1.11.0"
|
||||
thiserror = "2.0"
|
||||
|
||||
[dev-dependencies]
|
||||
ark-bn254 = "=0.4.0"
|
||||
num-traits = "=0.2.15"
|
||||
hex-literal = "=0.3.4"
|
||||
tiny-keccak = { version = "=2.0.2", features = ["keccak"] }
|
||||
criterion = { version = "=0.4.0", features = ["html_reports"] }
|
||||
hex = "0.4.3"
|
||||
hex-literal = "1.1.0"
|
||||
ark-bn254 = { version = "0.5.0", features = ["std"] }
|
||||
num-traits = "0.2.19"
|
||||
tiny-keccak = { version = "2.0.2", features = ["keccak"] }
|
||||
criterion = { version = "0.8.0", features = ["html_reports"] }
|
||||
|
||||
[features]
|
||||
default = ["parallel"]
|
||||
default = []
|
||||
parallel = ["ark-ff/parallel"]
|
||||
pmtree-ft = ["pmtree"]
|
||||
|
||||
[[bench]]
|
||||
name = "merkle_tree_benchmark"
|
||||
harness = false
|
||||
|
||||
[[bench]]
|
||||
name = "poseidon_benchmark"
|
||||
harness = false
|
||||
|
||||
[package.metadata.docs.rs]
|
||||
all-features = true
|
||||
|
||||
@@ -4,8 +4,8 @@ args = ["build", "--release"]
|
||||
|
||||
[tasks.test]
|
||||
command = "cargo"
|
||||
args = ["test", "--release"]
|
||||
args = ["test", "--release", "--", "--nocapture"]
|
||||
|
||||
[tasks.bench]
|
||||
command = "cargo"
|
||||
args = ["bench"]
|
||||
args = ["bench"]
|
||||
|
||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user