working-in-notebooks
Working in Notebooks
Use this skill to create, maintain, and choose between notebook environments (Jupyter, marimo, Colab) for data work. Covers tool selection, reproducibility patterns, and workflow best practices.
When to use this skill
- Setting up a notebook environment — choosing between Jupyter, marimo, VS Code, or Colab
- Converting between notebook formats — Jupyter to marimo, .ipynb to .py, or vice versa
- Making notebooks reproducible — pinning dependencies, managing random seeds, avoiding hardcoded paths
- Improving notebook structure — organizing cells, refactoring code, adding tests
- Publishing or sharing notebooks — nbconvert, Quarto, Voilà, or Git workflows
- Jupyter-specific features — magic commands, widgets, extensions, kernel management
- Marimo-specific workflows — reactive execution, UI components, version control patterns
When NOT to use this skill
Use a different skill for these related but distinct tasks:
| Instead of... | Use this skill | Because... |
More from legout/data-platform-agent-skills
data-science-eda
Exploratory Data Analysis (EDA): profiling, visualization, correlation analysis, and data quality checks. Use when understanding dataset structure, distributions, relationships, or preparing for feature engineering and modeling.
13data-science-visualization
Data visualization for Python: Matplotlib, Seaborn, Plotly, Altair, hvPlot/HoloViz, and Bokeh. Use when creating exploratory charts, interactive dashboards, publication-quality figures, or choosing the right library for your data and audience.
12data-engineering-core
Core Python data engineering: Polars, DuckDB, PyArrow, PostgreSQL, ETL patterns, performance tuning, and resilient pipeline construction. Use when building or reviewing batch ETL/dataframe/SQL pipelines in Python.
10data-science-feature-engineering
Feature engineering for machine learning: encoding, scaling, transformations, datetime features, text features, and feature selection. Use when preparing data for modeling or improving model performance through better representations.
10data-science-notebooks
Interactive notebooks for data science: Jupyter, JupyterLab, and marimo. Use for exploratory analysis, reproducible research, documentation, and sharing insights with stakeholders.
9data-engineering-best-practices
Data engineering best practices: medallion architecture, dataset lifecycle, partitioning, file sizing, schema evolution, and append/overwrite/merge patterns across Polars, PyArrow, DuckDB, Delta Lake, and Iceberg. Use when designing production data pipelines or reviewing data platform decisions.
8