← Back to registry
Skill
senior-data-engineer
"Data engineering workflows for designing reliable pipelines and datasets: ingestion, transforms, orchestration, schema evolution, data contracts, quality checks, and observability. Use when building ETL/ELT, reviewing pipelines, defining warehouse/lake schemas, or diagnosing data quality incidents."
Install
One-line setup
Copy and run this in your terminal to install the skill. Re-run to reinstall and update an existing install.
npx codex-skills-registry@latest --skill=data/senior-data-engineer --yesSenior Data Engineer
Make data pipelines boring: predictable, observable, and recoverable.
Quick Start
- Define the data contract (schema + semantics + freshness + ownership).
- Design the pipeline:
- Inputs, transformations, outputs, backfills, and failure handling
- Data quality: checks for nulls, ranges, uniqueness, and referential integrity.
- Operational story: retries, checkpoints, alerting, and lineage.
Optional tool: lightweight profiling for CSV/JSONL
python ~/.codex/skills/senior-data-engineer/scripts/data_quality_scan.py path/to/data.csv --out /tmp/data_profile.json
References
- Data contract template:
references/data-contract.md - Pipeline checklist:
references/pipeline-checklist.md