Skip to main content

Extraction

5 skills for extracting schemas, data, and scripts from existing databases — essential for migration and reverse engineering.

Skills

extract

Combined schema and data extraction. The all-in-one extraction skill that handles both structure and data.

You: /extract Extract the complete analytics schema from Snowflake including data samples

extract-schema

JDBC schema extraction with support for custom remarks, column selection, and filtering. Generates Starlake YAML configurations from existing database schemas.

You: /extract-schema Reverse-engineer my PostgreSQL analytics schema into Starlake YAML

Key features:

  • Extracts table structures, column types, and constraints
  • Custom remark handling for documentation
  • Column filtering and selection
  • Generates ready-to-use domain and table YAML files

extract-data

Extract data to files. Export data from databases to CSV, JSON, or Parquet files.

You: /extract-data Export the customers table from BigQuery to Parquet files

extract-bq-schema

BigQuery-specific schema extraction. Optimized for BigQuery's nested and repeated field structures.

You: /extract-bq-schema Extract all table schemas from the analytics dataset in BigQuery

extract-script

Generate extraction scripts. Creates reusable extraction configurations and shell scripts.

You: /extract-script Generate an extraction script for nightly exports from Snowflake

Example: Reverse-Engineering a Database

# metadata/extract/analytics.sl.yml
extract:
connectionRef: my-snowflake
jdbcSchemas:
- schema: ANALYTICS
tables:
- name: CUSTOMERS
columns:
- name: "*"
- name: ORDERS
columns:
- name: "*"
tableTypes:
- TABLE
- VIEW

The extraction generates Starlake-compatible YAML that you can immediately use for ingestion or transformation pipelines.