Build end-to-end data products on Databricks with AI-assisted development
This template is a complete data product builder powered by vibe coding. Start from a raw customer schema or existing data, and build your way through the full Databricks stack β Databricks Apps with Lakebase, medallion architecture (Bronze, Silver, Gold), semantic layer, Genie Spaces, ML pipelines, and GenAI agents β all guided by 77 accelerator skills, a standalone GenAI agent course, and your AI coding assistant.
Vibe Coding is an AI-assisted development approach where you collaborate with AI tools (like Cursor, GitHub Copilot, Claude Code, Windsurf, or similar) to rapidly build, iterate, and deploy production-quality data products. Instead of writing every line from scratch, you describe what you want and let the AI handle the implementation β guided by structured agent skills that encode best practices.
Workshop participants: See PRE-REQUISITES.md for the full setup checklist (workspace access, CLI, IDE, and authentication).
| Tool | Required | Installation |
|---|---|---|
| Databricks Workspace | Yes | Access to a Databricks workspace with Unity Catalog |
| Databricks CLI >= 0.295.0 | Yes | curl -fsSL https://raw.githubusercontent.com/databricks/setup-cli/main/install.sh | sh |
| Node.js v22+ | Yes (Path A) | nodejs.org β required by AppKit |
| AI-Powered IDE | Yes | Cursor, Windsurf, VS Code + Copilot, or similar |
| Python 3.10+ | Yes | python.org |
| Git | Yes | git-scm.com |
git clone https://github.com/databricks-solutions/vibe-coding-workshop-template.git my-project
cd my-projectBuild a full-stack TypeScript app on Databricks AppKit, guided by 10 agent skills:
- Read the workshop guide: apps_lakebase/Instructions.md
- Open your AI coding assistant and prompt:
I want to build a Databricks App. Read @apps_lakebase/skills/01-appkit-scaffold/SKILL.md and scaffold a new AppKit project.
- Follow the branch-aware workflow in Instructions.md:
- Phase 1: Scaffold + build UI from a PRD, test locally (uses
01-appkit-scaffoldand02-appkit-buildskills) - Phase 2: Deploy to Databricks Apps with mock data (uses
03-appkit-deployskill) - Phase 3: Setup Lakebase bundle resources (uses
04-appkit-plugin-addandapps_lakebase/prompts/03-setup-lakebase.md) - Phase 4: Wire Lakebase backend locally (uses
04-appkit-plugin-addand05-appkit-lakebase-wiringskills) - Phase 4b (optional): Wire Model Serving / Agent endpoint (uses
04-appkit-plugin-addand06-appkit-serving-wiringskills) - Phase 4c (optional): Wire AppKit to a separate Agent App with OBO proxying (uses
06d-appkit-agent-app-proxy) - Phase 4d (optional): Add persistent chat history and feedback (uses
07-appkit-chat-historyand08-appkit-feedback) - Phase 5: Deploy + E2E test with Lakebase (uses
03-appkit-deployskill)
- Phase 1: Scaffold + build UI from a PRD, test locally (uses
Take a raw schema CSV through the full medallion architecture to production AI agents β one prompt per stage:
- Drop your schema CSV into
data_product_accelerator/context/ - Open your AI coding assistant and prompt:
I have a customer schema at @data_product_accelerator/context/Wanderbricks_Schema.csv.
Please design the Gold layer using @data_product_accelerator/skills/gold/00-gold-layer-design/SKILL.md
- Follow the 9-stage pipeline guide β one prompt per stage, one new conversation per stage.
Both paths work together. Build your data pipeline first, then deploy a Databricks App on top of it β or start with the app and add data products incrementally.
vibe-coding-workshop-template/
β
βββ README.md # This file
βββ QUICKSTART.md # Quick-start guide with two pathways
βββ AGENTS.md # AI assistant routing (universal entry point)
βββ PRE-REQUISITES.md # Workshop prerequisites checklist
βββ CONTRIBUTING.md # Contribution guidelines
βββ LICENSE.md # License
βββ SECURITY.md # Security policy
βββ env.example # Environment variable template
β
βββ apps_lakebase/ # Databricks AppKit Workshop (10 agent skills)
β βββ Instructions.md # Comprehensive workshop guide
β βββ prompts/ # Numbered prompt files for each workshop step
β βββ skills/ # Agent skills for the full app lifecycle
β βββ 00-appkit-navigator/ # Entry-point navigator (read first)
β βββ 01-appkit-scaffold/ # Scaffold new AppKit projects (+ agent skills install)
β βββ 02-appkit-build/ # Build UI + backend from a PRD
β βββ 03-appkit-deploy/ # Deploy to Databricks Apps
β βββ 04-appkit-plugin-add/ # Add plugins (Lakebase, Analytics, Genie, Files, Serving)
β βββ 05-appkit-lakebase-wiring/ # Wire Lakebase DDL, API routes, frontend hooks
β βββ 06-appkit-serving-wiring/ # Wire Model Serving / Agent endpoint to frontend
β βββ 06d-appkit-agent-app-proxy/ # Wire AppKit frontend to a separate Agent App
β βββ 07-appkit-chat-history/ # Persist chat conversations in Lakebase
β βββ 08-appkit-feedback/ # Add thumbs feedback linked to MLflow assessments
β
βββ presentations/ # Workshop slide deck (Marp, HTML, PDF, PPTX)
β
βββ genai-agents/ # GenAI Agent Development Course
β βββ 00-course-orchestrator/ # Current GenAI course navigator
β βββ PROMPT-GUIDE.md # Canonical Track A + AppKit 2-Apps prompts
β βββ foundation/ # UC resources, MLflow, tracing, tools, AI Gateway
β βββ tracks/ # Track A custom Agent Apps
β βββ sdlc/ # Prompt registry, evals, deployment, monitoring
β
βββ data_product_accelerator/ # 77 Agent Skills for End-to-End Data Products
β βββ AGENTS.md # Detailed skill routing table
β βββ QUICKSTART.md # One-prompt-per-stage guide (9 stages)
β βββ README.md # Accelerator overview
β βββ context/ # Schema CSV inputs (starting point)
β βββ skills/ # 77 skills across 12 domains
β β βββ admin/ # Skill creation, auditing (4)
β β βββ bronze/ # Bronze layer + Faker data (2)
β β βββ common/ # Cross-cutting shared skills (8)
β β βββ exploration/ # Ad-hoc notebooks (1)
β β βββ genai-agents/ # GenAI course mirror, tracks, SDLC, capstone
β β βββ gold/ # Gold design, implementation, workers (14)
β β βββ ml/ # MLflow pipelines (1)
β β βββ monitoring/ # Monitors, dashboards, alerts (5)
β β βββ planning/ # Project planning (1)
β β βββ semantic-layer/ # Metric Views, TVFs, Genie, optimization (10)
β β βββ silver/ # DLT pipelines, DQ rules (3)
β β βββ skill-navigator/ # Master routing system (1)
β βββ docs/ # Framework design documentation
β
βββ agentic-framework/ # Multi-Agent Build Framework
βββ agents/ # Agent prompts for building multi-agent systems
β βββ prd-analyzer.md # Parse PRDs, map to agent capabilities
β βββ skill-scaffolder.md # Create new Agent Skills (SKILL.md)
β βββ tool-builder.md # Build runtime Python tools
β βββ agent-tester.md # Configure agent behavior tests
β βββ agent-ui-wiring-prompt.md # Guide agent-to-UI wiring
β βββ multi-agent-build-prompt.md # Orchestrator build with Foundation Models
β βββ databricks-deployer.md # Deployment guidance
β βββ prd-template.md # PRD template
βββ skills/
βββ foundation-model-agent-loop/
βββ SKILL.md # Tool-calling loop with Foundation Models
This template supports a unified workflow from raw data to production data products:
Raw Schema CSV or Existing Data
β
βββΊ Gold Design β dimensional model, ERDs, YAML schemas
βββΊ Bronze Layer β source tables + test data (Faker)
βββΊ Silver Layer β DLT pipelines + data quality expectations
βββΊ Gold Layer β tables, MERGE scripts, FK constraints
βββΊ Semantic Layer β Metric Views, TVFs, Genie Spaces
βββΊ Observability β Lakehouse Monitors, AI/BI Dashboards, SQL Alerts
βββΊ ML Pipelines β MLflow experiments, training, inference
βββΊ GenAI Agents β ResponsesAgent, evaluation, deployment
β
βββΊ Databricks App β AppKit (full-stack TypeScript), deployed on Databricks Apps
Each stage is driven by a single prompt to your AI coding assistant. The 77 agent skills in data_product_accelerator/ encode production-tested patterns so you get governed, high-quality output at every step.
The data_product_accelerator/ directory contains 77 agent skills organized by domain that guide your AI assistant through the entire pipeline:
| Domain | Skills | Focus |
|---|---|---|
| Gold | 14 | Dimensional modeling, design workers, pipeline workers, ERDs, MERGE scripts |
| Semantic Layer | 5 | Metric Views, TVFs, Genie Spaces, export/import API |
| GenAI Agents | 31 | Course orchestrator, foundation, Track A/B/C, SDLC, capstone, simple agent scaffold |
| Common | 8 | Asset Bundles, naming, constraints, imports |
| Monitoring | 5 | Lakehouse Monitors, dashboards, SQL alerts |
| Admin | 4 | Skill creation, auditing, docs |
| Silver | 3 | DLT pipelines, expectations, DQX |
| Bronze | 2 | Bronze tables, Faker data generation |
| ML | 1 | MLflow pipelines |
| Planning | 1 | Project planning |
| Exploration | 1 | Ad-hoc notebooks |
| Skill Navigator | 1 | Master routing system |
See data_product_accelerator/QUICKSTART.md for the step-by-step 9-stage guide.
The apps_lakebase/ directory contains 10 agent skills and a comprehensive workshop guide for building full-stack TypeScript apps on Databricks AppKit. The app is not pre-built β it gets scaffolded at runtime via databricks apps init and built iteratively with your AI coding assistant.
What gets built:
- Full-stack TypeScript app (React + Tailwind CSS frontend, AppKit backend)
- Optional SQL Warehouse integration for analytics queries
- Lakebase (managed PostgreSQL) persistence (wired after the mock-data app is deployed)
- Optional agent chat with either Model Serving or a separate Agent App proxy
- Optional persistent chat history and MLflow-linked feedback
- Deployed to Databricks Apps with hot reload for local dev
| Skill | Purpose |
|---|---|
00-appkit-navigator |
Entry-point navigator β routes tasks to the correct skill |
01-appkit-scaffold |
Scaffold new AppKit projects with plugins (analytics, lakebase, genie, files) |
02-appkit-build |
Build UI and backend from a PRD β components, queries, type generation |
03-appkit-deploy |
Deploy to Databricks Apps, validate configuration |
04-appkit-plugin-add |
Add plugins to an existing AppKit project |
05-appkit-lakebase-wiring |
Wire Lakebase DDL, Express API routes, frontend hooks, mock fallback |
06-appkit-serving-wiring |
Wire Model Serving / Agent endpoint to frontend (streaming, invoke) |
06d-appkit-agent-app-proxy |
Wire an AppKit frontend to a separate Agent App with OBO forwarding |
07-appkit-chat-history |
Persist agent chat conversations in Lakebase with session auth |
08-appkit-feedback |
Add thumbs up/down feedback linked to MLflow assessments |
| Service | URL |
|---|---|
| App + API | http://localhost:8000 |
| Health Check | http://localhost:8000/health |
Start the dev server from your scaffolded app directory with npm run dev.
The root genai-agents/ directory is the current course entry point for production GenAI agents on Databricks. Start with genai-agents/00-course-orchestrator/SKILL.md, then use genai-agents/PROMPT-GUIDE.md for the canonical Track A + AppKit 2-Apps walkthrough.
Canonical flow:
Foundation (UC, MLflow, tracing, tools)
β Track A custom Agent App on Databricks Apps
β AppKit frontend proxy with OBO (`06d`)
β Chat history + feedback (`07`, `08`)
β SDLC pipeline (prompt registry, evals, deployment, monitoring)
The canonical GenAI agent skills live in the root genai-agents/ tree; stage 9 of the data product pipeline routes into the same agent-building patterns via genai-agents/00-course-orchestrator/SKILL.md. For Databricks-platform reference patterns (Knowledge Assistants, Genie Spaces, Model Serving, etc.) see the upstream registry databricks/databricks-agent-skills; local skills here extend or specialize those for our workshop and accelerator use cases.
The agentic-framework/ directory provides prompts and patterns for building multi-agent systems with Databricks Foundation Models.
Agent prompts (in agentic-framework/agents/):
| Agent | Purpose |
|---|---|
| prd-analyzer | Parse PRDs, map requirements to agent capabilities |
| skill-scaffolder | Create new Agent Skills (SKILL.md) for any domain |
| tool-builder | Build runtime Python tools for agents |
| agent-tester | Configure tests for agent behavior |
| agent-ui-wiring-prompt | Guide agent-to-UI integration |
| multi-agent-build-prompt | Build multi-agent orchestrators with Foundation Models |
| databricks-deployer | Deployment guidance for agents |
| prd-template | PRD template for agent projects |
Foundation Model Agent Loop (in agentic-framework/skills/foundation-model-agent-loop/):
- Pattern for tool-calling loops with Databricks Foundation Models (e.g.,
databricks-meta-llama-3-3-70b-instruct) - Supports function calling without custom model deployment
After scaffolding your AppKit app, the workflow progresses from mock data to a wired backend, with optional branches for agent chat, a separate Agent App proxy, chat history, and feedback:
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
β AppKit Workshop Lifecycle β
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ€
β β
β Phase 1: SCAFFOLD + BUILD Phase 2: DEPLOY (mock data) β
β ββββββββββββββββββββββββ ββββββββββββββββββββββββββ β
β databricks apps init npm run build β
β Build UI from PRD databricks apps deploy --profile <P> β
β npm run dev (localhost:8000) Verify at Databricks Apps URL β
β β
β Phase 3: SETUP LAKEBASE Phase 4: WIRE LAKEBASE β
β ββββββββββββββββββββββ βββββββββββββββββββββ β
β Add Lakebase bundle config Add Lakebase plugin (skill 04) β
β Configure endpoint + compute DDL, API routes, frontend (skill 05)β
β Record host in state file Test locally with mock fallback β
β β
β Optional Agent Branches β
β βββββββββββββββββββββββ β
β Skill 06: Model Serving / Agent endpoint β
β Skill 06d: Separate Agent App proxy with OBO forwarding β
β Skill 07: Persistent chat history β
β Skill 08: User feedback linked to MLflow assessments β
β β
β Phase 5: DEPLOY + E2E TEST β
β ββββββββββββββββββββββββ β
β databricks apps deploy (with Lakebase config) β
β Verify live data end-to-end β
β β
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
| File | Purpose |
|---|---|
app.yaml |
Defines how Databricks starts your app |
databricks.yml |
Databricks bundle configuration |
server/server.ts |
AppKit backend entry point |
client/src/ |
React + Tailwind CSS frontend |
package.json |
Node.js dependencies |
See the 03-appkit-deploy skill for the full deployment workflow.
Configure a Databricks CLI profile to authenticate:
databricks auth login --host https://your-workspace.cloud.databricks.comVerify it works:
databricks current-user meTo use a named profile (useful when working with multiple workspaces):
databricks auth login --host https://your-workspace.cloud.databricks.com --profile myprofile
databricks current-user me --profile myprofileAll skills and CLI commands accept a --profile flag to target a specific workspace.
After scaffolding, your generated AppKit app is a full-stack TypeScript project. Customize it using standard AppKit patterns:
Edit server/server.ts in your generated app directory to add tRPC routes or custom endpoints. See the 02-appkit-build skill for patterns.
Use the 04-appkit-plugin-add skill to add capabilities:
Read @apps_lakebase/skills/04-appkit-plugin-add/SKILL.md and add the Lakebase plugin to my app.
Available plugins: analytics, lakebase, genie, files, serving
npm install your-packagenpx @databricks/appkit docs # documentation index
npx @databricks/appkit docs "<query>" # search for a specific topicdatabricks --version # Should be >= 0.295.0
databricks current-user me # Verify authentication
databricks auth profiles # List configured profilesdatabricks auth login --host https://your-workspace.cloud.databricks.comlsof -ti:8000 | xargs kill -9 2>/dev/null || true
npm run devdatabricks apps get <APP_NAME> --profile <PROFILE>cd <your-app-directory>
npm run dev
# Open http://localhost:8000- PRE-REQUISITES.md β Workshop prerequisites checklist
- Workshop Presentation β Marp slide deck (HTML, PDF, PPTX)
- AppKit Workshop Guide β Databricks App workshop guide
- GenAI Agent Prompt Guide β canonical Track A + AppKit 2-Apps course prompts
- Data Product Accelerator QUICKSTART β 9-stage pipeline guide
- Databricks AppKit Documentation β AppKit SDK reference
- Databricks Apps Documentation
- Databricks CLI Installation
- Agent Skills (SKILL.md) Format
- Unity Catalog
- Delta Live Tables
- Metric Views
For canonical Databricks-platform skills (Apps, Lakebase, Model Serving, DABs, Pipelines, Core), see databricks/databricks-agent-skills. The local skills in this repo extend or specialize those for our workshop and accelerator use cases. Where a local skill genuinely derives from an upstream skill, its frontmatter records a structured upstream_sources entry that the freshness scanner audits for drift; lighter "see also" references appear as a ## See Also footer in the skill body. The databricks-solutions/ai-dev-kit registry remains a parallel authoritative source for accelerator-specific medallion / semantic-layer / ML patterns.
This is a Git template repository. To use it:
- Click "Use this template" on GitHub, or clone directly
- Choose your starting point:
- Build a Databricks App: Follow the AppKit workshop guide β scaffold, build, deploy, and wire Lakebase
- Build a data product: Drop a schema CSV in
data_product_accelerator/context/and follow the 9-stage guide - Build a GenAI agent: Start with genai-agents/00-course-orchestrator/SKILL.md and use genai-agents/PROMPT-GUIDE.md for the AppKit 2-Apps walkthrough
- Build agents: Use the prompts in
agentic-framework/agents/to scaffold multi-agent systems
- Iterate with your AI coding assistant β the agent skills handle the patterns
Ready to build? Let's go!
git clone https://github.com/databricks-solutions/vibe-coding-workshop-template.git my-project
cd my-project