Skip to content

hyperledger-identus/neoprism

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

NeoPRISM

Rust Edition Unit tests PRISM tests GitHub release GitHub tag License Docker Pulls Discord


📚 Documentation


Overview

NeoPRISM is an open-source implementation of a PRISM node, providing a robust and extensible platform for managing PRISM Decentralized Identifiers (DIDs) anchored on the Cardano blockchain.

A PRISM node continuously monitors the Cardano blockchain for DID operations, validates and indexes them, and enables efficient lookup of DID Documents. It also allows users to create, update, or deactivate DIDs by submitting operations, ensuring all protocol rules and security checks are enforced. By running a PRISM node, organizations and individuals can independently manage and verify DIDs, supporting a decentralized identity system without reliance on a central authority.

Features

  • 🛠️ Multiple Deployment Modes

    • Supports three operational modes:
      • Indexer: Resolves and indexes DIDs for verification services.
      • Submitter: Publishes DID operations to the Cardano blockchain.
      • Standalone: Combines indexing and submission capabilities into a single process.
  • 🔗 Cardano Data Source Integration

    • Ingests DID operations from various Cardano data sources, including Oura and DBSync.
  • 🆔 W3C-Compliant DID Resolution

    • Provides a universal-resolver compatible HTTP endpoint.
    • Resolves PRISM DIDs to DID Documents according to the W3C DID specification.
  • 📤 DID Operation Publishing

    • Publishes PRISM DID operations to the Cardano blockchain.
    • Integrates with Cardano-wallet; future support for additional publishing methods.
  • 📚 OpenAPI Documentation

    • Documents all HTTP endpoints using the OpenAPI specification.
  • 🗄️ Verifiable Data Registry (VDR) Support

    • Enables reading and writing arbitrary data to the Cardano blockchain using transaction metadata.
    • Supports indexing and retrieval of data from transaction metadata for verifiable data use cases.
  • 🦀 Rust Implementation

    • Developed in Rust for efficient resource usage and reliable performance.

Introduction to PRISM DID

The PRISM DID method (did:prism) is a protocol for creating and managing Decentralized Identifiers (DIDs) built on the Cardano blockchain. This specification defines the operations, serialization formats, and rules for managing the lifecycle of DIDs and their associated DID documents.

At its core, PRISM works by encoding identity operations (Create, Update, Deactivate) as Protocol Buffer messages that are signed, batched into blocks, and published as metadata in Cardano transactions. PRISM nodes continuously monitor the blockchain, processing these operations to maintain an up-to-date map of DIDs and their states. The method requires a "secure depth" of 112 block confirmations before considering operations final. During this confirmation period, users can leverage long form DIDs immediately without waiting for on-chain anchoring, providing flexibility while maintaining the ability to later publish these DIDs to the blockchain.

The protocol defines clear rules for constructing and validating operations, translating internal state to W3C-compliant DID Documents, and resolving DIDs. Security is enforced through cryptographic signatures, with each DID having at least one master key for operation signing. PRISM is designed to be scalable and secure, with support for various verification methods, services, and key types including secp256k1, Ed25519, and X25519.

Examples:

Short-form DID:

did:prism:9b5118411248d9663b6ab15128fba8106511230ff654e7514cdcc4ce919bde9b

Long-form DID:

did:prism:9b5118411248d9663b6ab15128fba8106511230ff654e7514cdcc4ce919bde9b:Cj8KPRI7CgdtYXN0ZXIwEAFKLgoJc2VjcDI1NmsxEiEDHpf-yhIns-LP3tLvA8icC5FJ1ZlBwbllPtIdNZ3q0jU

Quickstart

Try It Now

Quickly experiment with NeoPRISM using an in-memory database and in-memory blockchain:

docker run --rm -p 8080:8080 -e NPRISM_DB_URL=sqlite::memory: hyperledgeridentus/identus-neoprism:<VERSION> dev

Note: Replace <VERSION> with a specific release tag (e.g., 0.1.0).

This command starts a NeoPRISM node in development mode with:

  • In-memory SQLite database (data is lost when container stops)
  • Mock Cardano data source (no real blockchain connection)
  • Web UI and API server on port 8080

Access the Web UI at http://localhost:8080 to explore the interface.

Submit a sample DID operation (optional)

Once the container is running, you can submit a sample DID operation from another terminal:

Note: The hex string below is a pre-generated test signed operation for demonstration purposes.

curl -X 'POST' \
  'http://localhost:8080/api/signed-operation-submissions' \
  -H 'accept: application/json' \
  -H 'Content-Type: application/json' \
  -d '{
  "signed_operations": [
    "0a086d61737465722d3012473045022100d54e1b7ce250c6d1333348c21aa162703793809d8751f07f82f5f9584b91850902201b2b3b43013a8bac123d9e32531aa6d8a065c23ff3e4160d7f4f0ce8c84c44581a7e0a7c0a7a123c0a086d61737465722d3010014a2e0a09736563703235366b31122103b20404f350d87eec98982131c176acfea520f26f8901fe08b619a56a0dd9e417123a0a06617574682d3010044a2e0a09736563703235366b31122103868f518859c09aa1a045c215d3a141b042299f09157192f67bc5c8a97e175b4d"
  ]
}'

Then navigate to http://localhost:8080/explorer to see the DIDs on the in-memory blockchain.

Public Demo Instance

Try NeoPRISM without any setup using these hosted instances connected to live Cardano networks:

Docker Compose

This example setup runs a NeoPRISM node that connects to the Cardano mainnet public relay using Oura. It fetches DID operations from the blockchain, synchronizes and indexes them into a local PostgreSQL database. Once operations are indexed, you can browse them using the Explorer page in the Web UI.

Prerequisites:

Supported Platforms:

  • The official NeoPRISM Docker image supports both x86_64 and arm64 architectures.
  • Compatible with Linux, macOS, and Windows hosts that can run Docker.

Steps:

  1. Clone the repository and navigate to the example directory:
    cd docker/mainnet-relay
  2. Start the node and sync block metadata:
    docker-compose up
  3. Access the Web UI at http://localhost:8080. The Explorer page allows you to browse indexed DID operations.
  4. Use the resolver endpoint to resolve DIDs:
    curl http://localhost:8080/api/dids/<did>

Development guide

This project uses Nix to provide a consistent local development environment and to package build artifacts. To get started, install Nix by following the instructions here.

Entering the development shell

If you already have cargo and other required dependencies (such as protoc) installed, you can use your own environment. You can review the nix shell file to see the required dependencies and adapt them to your setup.

We recommend using the nix develop command to enter the development shell. This ensures a consistent environment and uses the same library versions for building and testing.

nix develop

# You can now run commands like "cargo version"

Note: You may need to enable experimental flake commands. Please follow the instructions here.

You can also use --unset <ENV> to disable host environment variables and make the development shell more pure. For example:

nix develop --unset PATH

This disables all binaries available on the host PATH.

Development quickstart

To set up and run NeoPRISM for development, follow these steps:

  1. Enter the development shell:
    nix develop --unset PATH
  2. Install web UI dependencies (only required the first time):
    just init
  3. Start the local PostgreSQL database (required for NeoPRISM to store data):
    just postgres-up
  4. Run the NeoPRISM node (this will automatically generate all required assets):
    just run
  5. Access the Web UI and API at http://localhost:8080.

Cleaning up services

  • To stop and remove the local database:
    just postgres-down

Notes

  • Default port used is 8080.
  • No need to run just build-assets manually; just run handles asset generation automatically.
  • You can run just build-assets separately if you only want to generate web UI assets without starting the node.
  • All docker/*/compose*.yml files are auto-generated from Python sources in tools/compose_gen/. Do not edit these YAML files directly. Instead, modify the Python sources and run just build-config to regenerate them.

Running E2E Tests

NeoPRISM includes end-to-end conformance tests that verify the PRISM DID protocol implementation. These tests validate DID operations (create, update, deactivate) against a local Cardano testnet.

Prerequisites:

  • Nix with flakes enabled

Running the tests:

# Enter the development shell
nix develop

# Run the full e2e test suite
just e2e::run

This command will:

  1. Build the NeoPRISM and Cardano testnet Docker images
  2. Start a local test environment with a Cardano node
  3. Run the Scala-based conformance test suite
  4. Tear down the test environment

Important notes:

  • The first run may take several minutes to build all Docker images. Subsequent runs will be faster as images are cached.
  • The Docker images are built using Nix, which only includes files tracked by git. If you add new files to the codebase, make sure to stage them with git add before running the tests, otherwise they won't be included in the Docker image build.

Building tests only:

To compile the test suite without running it:

just e2e::build

Running specific E2E stacks

The e2e harness now ships four compose variants that differ by backend (PostgreSQL or SQLite) and target audience (developer-friendly vs CI-hardening). You can exercise any stack individually by combining the up, down, and run recipes:

# Start only the SQLite developer stack
just e2e::up sqlite-dev
# Run the Scala tests that target the running stack
(cd tests/prism-test && sbt test)
# Clean up once finished
just e2e::down sqlite-dev

just e2e::run iterates over every stack in sequence—sqlite-dev, ci, and ci-sqlite—so that both backends are covered automatically.

Configuration matrix

Mode Backend Compose file just target Recommended usage Pros Trade-offs
sqlite-dev SQLite docker/prism-test/compose-sqlite-dev.yml just e2e::up sqlite-dev Fast local iterations or laptop CI checks Zero extra services, file-backed DB keeps resource usage low Single-writer semantics; best for smoke tests
ci PostgreSQL docker/prism-test/compose-ci.yml just e2e::up ci Full CI pipeline and release rehearsal Mirrors production topology with stricter health checks Pulls more images and runs longer
ci SQLite docker/prism-test/compose-ci-sqlite.yml just e2e::up ci-sqlite CI shard that validates the embedded backend Ensures SQLite stays first-class even under CI load Cardano/I/O limits are similar to ci, so still slower than dev

Pick the stack that matches your goal; for example, run sqlite-dev while iterating on backend logic, and keep ci or ci-sqlite in nightly pipelines for extra safety.

Full QA helper script

The justfile provides a full-check recipe that strings the common QA steps together:

just full-check

This recipe performs cargo clean, cargo build --all-features, regenerates compose files, runs just test, builds the Docker artifacts, and executes just e2e::run. Use it right before sending a PR to replicate the checks we run manually.

SQLx schema checks

The development shell now bundles sqlx-cli, sqlite, and all required headers. Whenever you change migrations or entity definitions, run both backends to ensure they stay valid:

# Postgres schema (point to your local instance)
nix develop -c 'DATABASE_URL=postgres://postgres:postgres@localhost:5432/postgres sqlx migrate run'

# Embedded SQLite schema (creates a throwaway file)
nix develop -c 'DATABASE_URL=sqlite://$(pwd)/.tmp/neoprism-dev.sqlite sqlx migrate run'

Remove the temporary SQLite file afterwards if you do not need it anymore.

Offline sqlx metadata (optional)

If you rely on cargo sqlx prepare for offline builds, regenerate metadata for each backend:

# Postgres metadata
DATABASE_URL=postgres://postgres:postgres@localhost:5432/postgres \
  cargo sqlx prepare --bin neoprism-node

# SQLite metadata
DATABASE_URL=sqlite://$(pwd)/.tmp/neoprism-dev.sqlite \
  cargo sqlx prepare --bin neoprism-node

The generated sqlx-data.json reflects the currently enabled features, so keep both variants in sync if you commit the file.

Database backends

NeoPRISM supports both PostgreSQL and embedded SQLite storage. The backend is inferred from --db-url / NPRISM_DB_URL: use a postgres:// URL for Postgres or a sqlite:// URL for SQLite. If you omit the flag entirely, the node falls back to an embedded SQLite file in your platform's application-data directory (for example ~/Library/Application Support/NeoPRISM/<network>/neoprism.db on macOS or $XDG_DATA_HOME/NeoPRISM/<network>/neoprism.db on Linux).

SQLite is great for local development, demos, and CI smoke tests, but remember:

  • It is effectively single-writer; keep the node as the only process touching the file.
  • WAL mode is enforced automatically to keep reads non-blocking.
  • Long running readers can delay checkpointing and grow the file; run VACUUM occasionally if you prune data manually.

Frequently used commands

These are commands you can run outside the development shell:

command description
nix build .#neoprism-docker Use nix to build the docker image (output available at ./result)
nix build .#neoprism-docker && docker load < ./result Use nix to build the docker image and load it using docker

Assuming you are in the development shell, here are some frequently used commands:

command description
just init Install the npm dependencies (first time only)
cargo build Build the cargo workspace
cargo clean Clean the cargo workspace
cargo r -p neoprism-node -- -h See neoprism-node service CLI options
cargo test --all-features Run tests that enable all crate features

The following justfile commands are available to automate the local development workflow:

command description
just format Run the formatter on everything (Rust, Nix, TOML, Python, SQL, Hurl)
just build Build the whole project (assets + cargo)
just build-assets Build the Web UI assets (CSS, JavaScript, static assets)
just build-config Generate Docker Compose configs from the Python sources in tools/
just test Run all tests with all features enabled
just check Run the full Nix checks suite (format, lint, test, clippy)
just clean Clean all build artifacts
just postgres-up Spin up the local PostgreSQL database (Docker)
just postgres-down Tear down the local PostgreSQL database
just postgres-dump Dump the local PostgreSQL database to the postgres.dump file
just postgres-restore Restore the local PostgreSQL database from the postgres.dump file
just sqlite-init Create or migrate the embedded SQLite database (default path)
just sqlite-clean Delete the embedded SQLite database file
just run indexer Run the indexer node, connecting to the local database
just run indexer --cardano-addr <ADDR> Run the indexer node, connecting to the Cardano relay at <ADDR>
just run indexer --dbsync-url <URL> Run the indexer node, connecting to the DB Sync instance at <URL>
just tools format Format the Python code in tools/
just tools check Type-check and validate the Python tooling code
just e2e::build Build the PRISM conformance (end-to-end) test suite
just e2e::run Run the PRISM conformance (end-to-end) test suite