All contributions are highly appreciated! Start by forking the repository on GitHub and setting up Deepnote Toolkit for local development.
mise automatically manages Python, Java, and other tool versions:
-
Install mise: Getting started
-
Run setup:
mise install # Installs Python 3.12 and Java 17 mise run setup # Installs dependencies and pre-commit hooks
-
Install poetry: Installation
-
Install Java 17 (required for PySpark tests):
- macOS:
brew install openjdk@17 - Ubuntu/Debian:
sudo apt-get install openjdk-17-jdk - RHEL/Fedora:
sudo dnf install java-17-openjdk-devel
- macOS:
-
Set up venv for development package:
# If Python 3.10 is available, point Poetry to it $ poetry env use 3.10 -
Verify the virtual environment location:
$ poetry env info
-
Install dependencies:
$ poetry install
-
Install Poe Poetry add-on:
$ poetry self add 'poethepoet[poetry_plugin]' -
Install pre-commit hooks:
$ poetry poe setup-hooks
-
Verify installation:
$ poetry poe lint $ poetry poe format
-
If
poetry installfails with errorlibrary 'ssl' not found:env LDFLAGS="-I/opt/homebrew/opt/openssl/include -L/opt/homebrew/opt/openssl/lib" poetry install -
If
poetry installfails installingpymssql, installfreetdsvia Homebrew.
Tests run against all supported Python versions using nox in Docker for reproducible environments.
# Run unit tests (no coverage by default)
mise run test
# Run unit tests with coverage
mise run test:coverage
# Run tests quickly without nox/coverage overhead
mise run test:quick tests/unit/test_file.py
mise run test:quick tests/unit/test_file.py::TestClass::test_method -v
# Pass custom arguments (including --coverage)
mise run test -- --coverage tests/unit/test_file.py# Run unit tests without coverage
poetry run nox -s unit
# Run unit tests with coverage
poetry run nox -s unit -- --coverage
# Run specific test file
poetry run nox -s unit -- tests/unit/test_file.py# Run unit tests
TEST_TYPE="unit" TOOLKIT_VERSION="local-build" ./bin/test
# Run integration tests
TEST_TYPE="integration" TOOLKIT_VERSION="local-build" TOOLKIT_INDEX_URL="http://localhost:8000" ./bin/test
# Or use the test-local script for both unit tests and integration tests
./bin/test-local
# Run a specific file with test-local
./bin/test-local tests/unit/test_file.py
# ... or specific test
./bin/test-local tests/unit/test_file.py::TestClass::test_method- Kernel dependencies: Add to
[tool.poetry.dependencies]in pyproject.toml
# Add a package to kernel bundle (available in notebooks)
$ poetry add pandas
# Add a package with specific version
$ poetry add "pandas>=2.0.0"# Add a development dependency
$ poetry add --group dev pytestAfter adding dependencies, run tests to verify compatibility:
$ ./bin/test-localWhen you push a commit, a new version of deepnote/jupyter-for-local is built with your commit hash (shortened!). Use it in projects by updating common.yml:
jupyter:
image: "deepnote/jupyter-for-local:SHORTENED_COMMIT_SHA"Alternatively, to develop against a local copy of Deepnote Toolkit, first run this command to build the image:
docker build \
--build-arg "FROM_PYTHON_TAG=3.11" \
-t deepnote/deepnote-toolkit-local-hotreload \
-f ./dockerfiles/jupyter-for-local-hotreload/Dockerfile .Then start the container:
# To include server logs in the output add this argument
# -e WITH_SERVER_LOGS=1 \
# Some toolkit features (e.g. feature flags support) require
# DEEPNOTE_PROJECT_ID to be set to work correctly. Add this
# argument with your project id
# -e DEEPNOTE_PROJECT_ID=981af2c1-fe8b-41b7-94bf-006b74cf0641 \
docker run \
-v "$(pwd)":/deepnote-toolkit \
-v /tmp/deepnote-mounts:/deepnote-mounts:shared \
-p 8888:8888 \
-p 2087:2087 \
-p 8051:8051 \
-w /deepnote-toolkit \
--add-host=localstack.dev.deepnote.org:host-gateway \
--rm \
--name deepnote-toolkit-local-hotreload-container \
deepnote/deepnote-toolkit-local-hotreloadThis will start a container with Deepnote Toolkit mounted inside and expose all required ports. If you change code that runs in the kernel (e.g. you updated the DataFrame formatter), you only need to restart the kernel from Deepnote's UI. If you update code that starts Jupyter itself, you need to restart the container. And if you add or modify dependencies you need to rebuild the image.
Now, you need to modify common.yml in the Deepnote app. First, replace jupyter service with noop image:
jupyter:
image: 'screwdrivercd/noop-container'And change JUPYTER_HOST variable of executor to point to host machine:
executor:
environment:
JUPYTER_HOST: host.docker.internalEach PR creates a review application for testing. Access it via GitHub checks. Monitor logs in Grafana:
{pod="p-PROJECT_ID", container="notebook"}
We use Docker to ensure reproducible environments due to Jupyter libraries' binary dependencies:
-
builder.Dockerfile: Creates Python package bundles for different versions (3.9-3.12), generates kernel and server bundles, and packages the toolkit for distribution using Poetry. -
test.Dockerfile: Provides consistent test environment for running unit and integration tests across Python versions using nox. Used both locally and in CI/CD pipeline. -
jupyter-for-local.Dockerfile: Creates development environment with Jupyter integration, used for local development from docker-compose used in main monorepo.
To release a new version to production:
- Merge your changes to main. This will automatically trigger a GitHub Actions workflow that runs the test suite and a staging deployment.
- Trigger a new GitHub Release in the GitHub UI.
- Monitor the GitHub Actions workflows and ensure a successful production deployment.
Note: The production release pipeline automatically creates two PRs in the ops and app-config repositories:
- A staging PR that updates staging values and is auto-merged
- A production PR that updates production values and requires manual approval and merge
Important: Always test the changes in the staging environment before approving and merging the production PR to ensure everything works as expected.