SQL & Analytics
SQL Execution & Warehouses
Section titled “SQL Execution & Warehouses”Skills: databricks-dbsql
MCP Tools: execute_sql, execute_sql_multi, list_warehouses, get_best_warehouse
List all SQL warehouses in my workspace and show which ones are running.Run this SQL query: SELECT * FROM main.default.my_table LIMIT 10Execute these SQL statements in order:1. CREATE SCHEMA IF NOT EXISTS main.analytics2. CREATE TABLE main.analytics.daily_summary AS SELECT date, COUNT(*) as cnt FROM main.raw.events GROUP BY dateFind the best available SQL warehouse and run a query to show the top 10 largesttables in the main catalog.Run a SQL query to get the row count and column count for every table in themain.sales schema.Execute a DESCRIBE EXTENDED on main.production.customers and explain what eachfield means.AI Functions (SQL)
Section titled “AI Functions (SQL)”Skills: databricks-ai-functions
MCP Tools: execute_sql
Classification & Extraction
Section titled “Classification & Extraction”Run a SQL query that uses ai_classify to categorize customer feedback inmain.support.tickets into ['bug', 'feature_request', 'question', 'complaint'].Use ai_extract to pull out product names, prices, and quantities fromunstructured order description text in main.sales.raw_orders.Classify the sentiment of customer reviews in main.reviews.comments usingai_analyze_sentiment and store results in a new column.Summarization & Generation
Section titled “Summarization & Generation”Summarize the top 50 longest support tickets in main.support.cases usingai_summarize, keeping summaries under 100 words each.Use ai_gen to generate product descriptions for items in main.catalog.productsthat are missing descriptions, based on the product name and category.Use ai_fix_grammar to clean up user-submitted text in main.content.posts.Advanced AI Functions
Section titled “Advanced AI Functions”Use ai_query to call my fine-tuned model endpoint with a custom prompt templatethat includes data from multiple columns.Run ai_forecast on main.sales.daily_revenue to predict the next 30 days ofrevenue with 90% prediction intervals.Build a document parsing pipeline:1. Use ai_parse_document to extract text from PDFs stored in a volume2. Chunk the extracted text into 500-token segments3. Store the chunks in a table ready for Vector Search indexingUse ai_mask to redact PII (names, emails, phone numbers) from customercommunications before sharing with the analytics team.Use ai_translate to translate product descriptions in main.catalog.products fromEnglish to Spanish, French, and German.Use ai_similarity to find duplicate customer records by comparing name + addressfields between two tables.AI/BI Dashboards
Section titled “AI/BI Dashboards”Skills: databricks-aibi-dashboards
MCP Tools: create_or_update_dashboard, get_dashboard, delete_dashboard, publish_dashboard
Create an AI/BI dashboard with a dataset querying main.sales.transactions and add:- A counter widget showing total revenue- A bar chart showing revenue by product category- A line chart showing daily revenue trends over the last 90 daysBuild a dashboard with two pages:- Page 1 "Overview": KPI counters for total users, active users, and churn rate- Page 2 "Details": a table widget with user-level data and a filter for date rangeCreate a dashboard with a global date filter that applies to all datasets,showing customer acquisition metrics.Get the JSON definition of my existing dashboard and help me add a new chartwidget to it.Publish my dashboard so it's accessible to all workspace users.List all dashboards in my workspace and show which ones are published vs draft.Create a dashboard that uses pre-aggregated datasets for performance — thesource table has 500M rows so I need the SQL to aggregate before visualization.Genie Spaces
Section titled “Genie Spaces”Skills: databricks-genie
MCP Tools: create_or_update_genie, get_genie, delete_genie, ask_genie, migrate_genie
Creating Genie Spaces
Section titled “Creating Genie Spaces”Create a Genie Space called "Sales Explorer" that connects tomain.sales.transactions and main.sales.customers with sample questions:"What was total revenue last month?", "Who are our top 10 customers?","Show monthly revenue trends".Create a Genie Space for our finance team that includes tablesmain.finance.gl_entries, main.finance.budgets, and main.finance.forecastswith instructions to always filter by the current fiscal year.Querying Genie Spaces
Section titled “Querying Genie Spaces”Ask my Genie Space "Sales Explorer": What were the top 5 product categories byrevenue in Q4?Ask the Genie Space: Show me customer churn rate by region for the last 6 monthsas a trend.Have a multi-turn conversation with my Genie Space — first ask about totalrevenue, then drill down into the top-performing region.Managing Genie Spaces
Section titled “Managing Genie Spaces”List all Genie Spaces in my workspace.Export the Genie Space "Sales Explorer" so I can import it into another workspace.Import a Genie Space configuration from a JSON file into my current workspace.Migrate my Genie Space from the dev workspace to production.Metric Views
Section titled “Metric Views”Skills: databricks-metric-views
MCP Tools: manage_metric_views, execute_sql
Create a metric view for "total_revenue" defined as SUM(amount) frommain.sales.transactions, with dimensions for product_category, region,and order_date.Define a metric view for "monthly_active_users" that counts distinct user_idswith activity in a calendar month, with dimensions for platform and country.Create a set of related metric views for e-commerce KPIs:- total_orders (COUNT)- average_order_value (AVG)- total_revenue (SUM)- customer_count (COUNT DISTINCT)All sourced from main.sales.orders with shared dimensions.Query my metric view "total_revenue" grouped by region and month for thelast year.List all metric views in the main.metrics schema.Databricks SQL Advanced Features
Section titled “Databricks SQL Advanced Features”Skills: databricks-dbsql
MCP Tools: execute_sql, execute_sql_multi
SQL Scripting & Stored Procedures
Section titled “SQL Scripting & Stored Procedures”Write a SQL stored procedure that accepts a date parameter, runs a data qualitycheck on main.production.orders for that date, and returns a summary of anyviolations found.Create a SQL script with variables, loops, and conditional logic that backfillsdaily aggregation tables for a specified date range.Pipe Syntax
Section titled “Pipe Syntax”Rewrite this SQL query using Databricks pipe syntax (|>):SELECT department, AVG(salary) as avg_salaryFROM main.hr.employeesWHERE status = 'active'GROUP BY departmentHAVING AVG(salary) > 100000ORDER BY avg_salary DESCGeospatial
Section titled “Geospatial”Write a geospatial query using H3 functions to aggregate delivery data byhexagonal regions at resolution 7 and find the top 10 busiest zones.Use ST_ spatial functions to find all stores within a 10km radius of a givenlatitude/longitude coordinate.Materialized Views & Recursive CTEs
Section titled “Materialized Views & Recursive CTEs”Create a materialized view main.gold.daily_revenue_summary that pre-aggregatesrevenue by day, product category, and region from main.silver.transactions.Write a recursive CTE to traverse an employee-manager hierarchy inmain.hr.employees and show the full reporting chain for each employee.