Back to data-science
data-science v1.0.0 3.5 min read 171 lines

jupyter-live-kernel

라이브 Jupyter 커널로 상태 유지형 반복 Python 분석

Hermes Agent
MIT

Jupyter Live Kernel (hamelnb)

Gives you a stateful Python REPL via a live Jupyter kernel. Variables persist
across executions. Use this instead of execute_code when you need to build up
state incrementally, explore APIs, inspect DataFrames, or iterate on complex code.

When to Use This vs Other Tools

| Tool | Use When |
|------|----------|
| This skill | Iterative exploration, state across steps, data science, ML, "let me try this and check" |
| execute_code | One-shot scripts needing hermes tool access (web_search, file ops). Stateless. |
| terminal | Shell commands, builds, installs, git, process management |

Rule of thumb: If you'd want a Jupyter notebook for the task, use this skill.

Prerequisites

  • uv must be installed (check: which uv)
  • JupyterLab must be installed: uv tool install jupyterlab
  • A Jupyter server must be running (see Setup below)

Setup

The hamelnb script location:

SCRIPT="$HOME/.agent-skills/hamelnb/skills/jupyter-live-kernel/scripts/jupyter_live_kernel.py"

If not cloned yet:

git clone https://github.com/hamelsmu/hamelnb.git ~/.agent-skills/hamelnb

Starting JupyterLab

Check if a server is already running:

uv run "$SCRIPT" servers

If no servers found, start one:

jupyter-lab --no-browser --port=8888 --notebook-dir=$HOME/notebooks \
--IdentityProvider.token='' --ServerApp.password='' > /tmp/jupyter.log 2>&1 &
sleep 3

Note: Token/password disabled for local agent access. The server runs headless.

Creating a Notebook for REPL Use

If you just need a REPL (no existing notebook), create a minimal notebook file:

mkdir -p ~/notebooks

Write a minimal .ipynb JSON file with one empty code cell, then start a kernel
session via the Jupyter REST API:
curl -s -X POST http://127.0.0.1:8888/api/sessions \
-H "Content-Type: application/json" \
-d '{"path":"scratch.ipynb","type":"notebook","name":"scratch.ipynb","kernel":{"name":"python3"}}'

Core Workflow

All commands return structured JSON. Always use --compact to save tokens.

1. Discover servers and notebooks

uv run "$SCRIPT" servers --compact
uv run "$SCRIPT" notebooks --compact

2. Execute code (primary operation)

uv run "$SCRIPT" execute --path  --code '' --compact

State persists across execute calls. Variables, imports, objects all survive.

Multi-line code works with $'...' quoting:

uv run "$SCRIPT" execute --path scratch.ipynb --code $'import os\nfiles = os.listdir(".")\nprint(f"Found {len(files)} files")' --compact

3. Inspect live variables

uv run "$SCRIPT" variables --path  list --compact
uv run "$SCRIPT" variables --path preview --name --compact

4. Edit notebook cells

# View current cells
uv run "$SCRIPT" contents --path --compact

Insert a new cell


uv run "$SCRIPT" edit --path insert \
--at-index --cell-type code --source '' --compact

Replace cell source (use cell-id from contents output)


uv run "$SCRIPT" edit --path replace-source \
--cell-id --source '' --compact

Delete a cell


uv run "$SCRIPT" edit --path delete --cell-id --compact

5. Verification (restart + run all)

Only use when the user asks for a clean verification or you need to confirm
the notebook runs top-to-bottom:

uv run "$SCRIPT" restart-run-all --path  --save-outputs --compact

Practical Tips from Experience

  • First execution after server start may timeout — the kernel needs a moment
to initialize. If you get a timeout, just retry.

  • The kernel Python is JupyterLab's Python — packages must be installed in
that environment. If you need additional packages, install them into the
JupyterLab tool environment first.

  • --compact flag saves significant tokens — always use it. JSON output can
be very verbose without it.

  • For pure REPL use, create a scratch.ipynb and don't bother with cell editing.
Just use execute repeatedly.

  • Argument order matters — subcommand flags like --path go BEFORE the
sub-subcommand. E.g.: variables --path nb.ipynb list not variables list --path nb.ipynb.

  • If a session doesn't exist yet, you need to start one via the REST API
(see Setup section). The tool can't execute without a live kernel session.

  • Errors are returned as JSON with traceback — read the ename and evalue
fields to understand what went wrong.

  • Occasional websocket timeouts — some operations may timeout on first try,
especially after a kernel restart. Retry once before escalating.

Timeout Defaults

The script has a 30-second default timeout per execution. For long-running
operations, pass --timeout 120. Use generous timeouts (60+) for initial
setup or heavy computation.

Related Skills / 관련 스킬

data-analysis

Use this skill when the user uploads Excel (.xlsx/.xls) or CSV files and wants to perform data analysis, generate statistics, create summaries, pivot tables, SQL queries, or any form of structured data exploration. Supports multi-sheet Excel workbooks, aggregation, filtering, joins, and exporting results to CSV/JSON/Markdown.

smart-data-analyzer

CSV, Excel, JSON, Parquet 데이터 파일 자동 분석 — 통계 요약, 결측치 탐지, 상관관계, 시각화, 인사이트 도출까지 한번에