Validate data against schemas, business rules, and data quality standards.
Skills(SKILL.md)は、AIエージェント(Claude Code、Cursor、Codexなど)に特定の能力を追加するための設定ファイルです。
詳しく見る →Validate data against schemas, business rules, and data quality standards.
Design dimensional models and fact tables.
Snowflake, BigQuery, Redshift, dimensional modeling, and modern data warehouse architecture
'Use SQL (PostgreSQL) when:'
You are a database expert with deep knowledge of relational databases, NoSQL, query optimization, schema design, and ORM frameworks.
>-
Database design, optimization, and operations expert
Allows read-only access to the SQL database to allow querying and analysis using natural language
>-
Understanding and implementing database locking mechanisms for concurrency control.
Implement SQLite database patterns using the database.py interface with complete SQL isolation. MANDATORY for all database projects. Use when working with databases, data persistence, or SQLite.
Create Prisma schema models with relations and indexes. Use when designing database schemas, adding models, or defining entity relationships.
Instrument database queries, connection pools, and detect N+1 queries
database-patterns
Query database safely with parameterized statements
| Atributo | Valor |
Эксперт DB replication. Используй для настройки репликации MySQL, PostgreSQL, MongoDB, failover и high availability.
Document database schemas, ERD diagrams, table relationships, indexes, and constraints. Use when documenting database schema, creating ERD diagrams, or writing table documentation.
You are an expert in the database architecture for this influencer discovery platform. This skill provides comprehensive knowledge about the Postgres schema, Drizzle ORM patterns, normalized user tabl
VaultCPA database schema reference with 50+ Prisma models, relationships, and query patterns. Use when working with database models, designing features, or understanding data structure.
Enforces project database schema conventions when creating or modifying Drizzle ORM table definitions, including constraints, indexes, relations, and column patterns. This skill should be used proactively whenever working with schema files to ensure consistent schema design for PostgreSQL with Neon serverless.
Create robust, error-proof Prisma seed scripts with comprehensive error handling and validation.
Implement database sharding for horizontal scalability. Use when scaling large databases, distributing data across multiple servers, or designing sharded architectures.
Configure and operate TinyDB for JSON-based document storage. Use when setting up databases, creating queries, or managing data persistence.
RDBMS access patterns for DuckDB, MySQL (keycloak), PostgreSQL (dw, x3rocs), SQL Server (sage1000, x3), and DBISAM (Exportmaster) using ODBC and native drivers
SQL database querying, optimization, and data management for analytics
Guide for authenticating with Databricks Apps using cookie-based auth when OAuth/PAT tokens don't work. Use when connecting to Databricks Apps with User Authorization enabled.
Databricks documentation reference. Use as a lookup resource alongside other skills and MCP tools for comprehensive guidance.
Unity Catalog system tables for lineage, audit logs, billing, compute, jobs, and query history. Use when querying system.access.audit, system.access.table_lineage, system.billing.usage, system.compute.clusters, system.lakeflow.jobs, system.lakeflow.job_run_timeline, or system.query.history.
Troubleshoot Datadog API authentication issues (401/403 errors), understand API keys vs app keys, and configure correct regions. Use when hitting auth errors or setting up Datadog API access.
Work with Datadog Incidents API - fetch incidents with enrichment, update incident fields, understand incident data structures. Use when working with incidents or the dd get-incident/update-incident commands.
Search Datadog logs via API - query syntax, storage tiers (indexes, flex, online-archives), pagination. Use when searching logs or using the dd search-logs command.
Convert documents (PDF, EPUB, PPTX, DOCX, XLSX, HTML, images) to Markdown using Datalab cloud API. Use when user wants to use Datalab API for document conversion, or prefers cloud-based processing over local marker CLI.
description: Create or update Dataverse schema components (entities, attributes, relationships, option sets) via Web API
Deploy solutions, PCF controls, and web resources to Dataverse using PAC CLI
This skill provides guidance on designing Dataverse table schemas and data models. Use when users ask about "Dataverse table design", "Dataverse schema", "Dataverse relationships", "Dataverse columns"
Handle datetime and timezone conversions correctly across frontend, API, and database. Use this skill when working with datetime-local inputs, scheduling features, due dates, reminders, or any time-sensitive functionality. Covers the critical pitfall of browser datetime-local inputs lacking timezone info.
Query Deutsche Bahn train connections and prices.
提供通用的数据库查询工作流程、SQL 模板和结果分析方法。使用 exec_sql 工具执行查询。
db-designer
提供v3_e-report数据库的SQL查询模板,包括数据分析告警配置等表的查询。查询分析告警配置时使用。使用 exec_sql 工具执行查询。
Gestión del esquema PostgreSQL, Auto-Healing y arquitectura RAG híbrida.
MongoDB database exploration and querying. Use when you need to understand database structure, view existing data, check collection schemas, count documents, or run queries to investigate the database state. (project)
Initialize database infrastructure for new projects
提供v3_metadata数据库的SQL查询模板,包括对象编码(object_code)、对象名称、事件、按钮配置、自定义字段、元数据字段、插件中心等表的查询。查询对象编码、对象名称、自定义对象时使用。使用 exec_sql 工具执行查询。
Comprehensive guide to optimizing database queries, understanding EXPLAIN plans, indexing strategies, and eliminating N+1 queries
データベースクエリ・分析支援。SQLクエリの作成、実行、結果の分析を行う。BigQuery、PostgreSQL、MySQL対応。トリガー: /db-query, SQL, クエリ, データ分析, BigQuery
查询租户和组织信息(v3_user数据库)。使用 exec_sql 工具执行查询。
Laravel 데이터베이스 설정을 검증하고 실제 연결을 테스트하는 Skill입니다.
Guide for recovering data from SQLite Write-Ahead Log (WAL) files that may be corrupted, encrypted, or inaccessible through standard methods. This skill should be used when tasks involve SQLite database recovery, WAL file analysis, encrypted database files, or discrepancies between tool outputs and filesystem access.