data-profilerProfile construction data to understand characteristics, distributions, quality metrics, and patterns. Essential for data quality assessment and ETL planning.
Install via ClawdBot CLI:
clawdbot install datadrivenconstruction/data-profilerGrade Fair — based on market validation, documentation quality, package completeness, maintenance status, and authenticity signals.
Calls external URL not in known-safe list
https://datadrivenconstruction.ioAudited Apr 16, 2026 · audit v1.0
Generated Mar 20, 2026
A construction firm receives raw data from various project management systems and needs to assess its quality before analysis. This skill profiles the data to identify missing values, outliers, and inconsistencies, ensuring reliable insights for decision-making.
Before integrating BIM data into a centralized database, the skill analyzes data characteristics and distributions to design efficient ETL processes. It helps optimize storage and processing by identifying data types and patterns specific to construction.
Government agencies or contractors use this skill to profile financial and operational data from large infrastructure projects. It detects anomalies and referential integrity issues, ensuring data accuracy for regulatory audits and reporting.
A construction company profiles supplier and material data to understand cost distributions, quantity patterns, and quality metrics. This supports inventory management and cost control by identifying data issues early.
In facility management, historical equipment and maintenance data is profiled to uncover patterns and quality issues. This enables better predictive models by ensuring clean, consistent data for analysis.
Offer the data profiler as a cloud-based service where construction firms pay a monthly or annual fee per user or data volume. This model provides recurring revenue and scalability for handling diverse datasets.
Provide tailored consulting services to integrate the profiler into clients' existing systems, with customizations for specific construction data needs. Revenue comes from project-based fees and ongoing support contracts.
Sell perpetual or long-term licenses to large construction enterprises, including training and premium support. This model targets organizations with extensive data profiling requirements across multiple projects.
💬 Integration Tip
Ensure Python3 is installed and accessible in the system PATH; use pandas for data loading to leverage the profiler's DataFrame compatibility.
Scored Apr 19, 2026
Use the @steipete/oracle CLI to bundle a prompt plus the right files and get a second-model review (API or browser) for debugging, refactors, design checks, or cross-validation.
Local search/indexing CLI (BM25 + vectors + rerank) with MCP mode.
Use when designing database schemas, writing migrations, optimizing SQL queries, fixing N+1 problems, creating indexes, setting up PostgreSQL, configuring EF Core, implementing caching, partitioning tables, or any database performance question.
Connect to Supabase for database operations, vector search, and storage. Use for storing data, running SQL queries, similarity search with pgvector, and managing tables. Triggers on requests involving databases, vector stores, embeddings, or Supabase specifically.
Use SQLite correctly with proper concurrency, pragmas, and type handling.
Write correct MySQL queries avoiding common pitfalls with character sets, indexes, and locking.