Why use the Agent?
The Agent accelerates pipeline development by reducing the manual steps required to build and modify data workflows. Instead of navigating multiple interfaces, you describe intent. The Agent translates that intent into concrete, inspectable project changes within your active environment. The Agent assists development; it does not autonomously deploy or promote changes.Scope and control
The Agent operates within the currently active project and under the permissions of the user who invoked it.- All SQL execution and pipeline runs occur under your existing credentials.
- All changes remain visible, versioned, and editable within Prophecy.
- Transform Agent (default) — Builds and modifies pipelines, analyses, and documentation using natural language.
- Harmonization Agent — Automates mapping source data to a defined Common Data Model (CDM).
- Documentation Agent — Generates complete project and pipeline documentation.
Workflow
Agent workflows vary depending on the active agent.Transform Agent workflow (default)
The Transform Agent focuses on building and modifying pipelines, analyses, and related project artifacts within the active project. When you submit a request, the Agent:- Interprets your intent.
- Inspects the current project graph and schema metadata.
- Generates or modifies pipeline logic.
- Validates transformations and project structure.
- Surfaces changes for review.
- Exploration for discovering and understanding data sources.
- Transformation for building and modifying pipeline logic.
Harmonization workflow
The Harmonization Agent focuses on mapping source schemas to a defined Common Data Model (CDM). The workflow typically includes:- Defining or selecting a CDM.
- Generating source-to-target mappings.
- Reviewing confidence indicators and data quality tests.
Agent architecture
The Prophecy Agent operates within Prophecy’s structured execution environment and is equipped with tools to query warehouses, compile code, run pipelines, and modify project artifacts directly. Rather than generating text alone, the Agent uses controlled internal tools to inspect, modify, and validate project assets directly within the active project. This enables the Agent to reason over:- Your project’s structure (pipelines, analyses, datasets, documentation).
- Schema metadata and dataset relationships.
- Version-controlled project artifacts.
Tool-based execution
The Agent uses a set of internal tools to take concrete actions inside your project, including:- Querying your connected warehouse.
- Retrieving metadata and sample data.
- Compiling and validating code.
- Running pipelines.
- Creating and editing project artifacts.
Schema and lineage awareness
The Agent understands your project’s schema metadata and dataset relationships. When generating or modifying transformations, it:- Inspects upstream schema definitions.
- Validates generated changes against your project structure before applying them.
- Surfaces potential conflicts or mismatches when detected.
Project boundary
All Agent activity is strictly contained within the active project and cannot affect external projects or platform-level configuration. It can:- Create and refactor pipelines.
- Generate and update analyses.
- Modify documentation.
- Edit datasets and other project artifacts.
Single-agent editing
Only one Agent session can edit a project at a time. Collaborative agent editing across multiple users is not supported.Human in the loop
The Agent generates and can apply changes inside your project, but you remain in control. All modifications are visible in the editor and can be inspected, edited, or reverted. You can:- Inspect every transformation.
- Validate data samples.
- Modify generated logic.
- Re-run pipelines as needed.
Available features
Search through your data environment
Search through your data environment
Use table metadata—table names, schemas, owners, and tags—to locate datasets from your fabric.
When you don’t know the exact table name or work with many tables, searching through metadata
eliminates guesswork and reduces time spent browsing schema lists. You can find data from your
connected data warehouse, cloud storage, reporting platforms, and other sources.
Retrieve and visualize sample data from your data environment
Retrieve and visualize sample data from your data environment
Preview data samples and visualizations before selecting datasets for your pipeline. Understanding
column structures, data patterns, and potential quality issues helps you make informed decisions
about which tables to use.
Build pipelines step-by-step or all in one prompt
Build pipelines step-by-step or all in one prompt
Generate complete pipelines from a single description when requirements are clear, or build
incrementally gem by gem, adding one transformation at a time in a linear sequence. Instead of
manually dragging gems onto the canvas and configuring each step, describe your goal and let the
Agent handle the setup.
Transform data automatically to match a defined schema
Transform data automatically to match a defined schema
The Harmonization Agent lets you define a Common Data Model (CDM) and generate mappings that conform input data to the CDM. Learn more in the Harmonization documentation.
Remove redundant steps and improve readability
Remove redundant steps and improve readability
Clean up existing pipelines by removing unnecessary transformations and consolidating logic.
Easier-to-maintain pipelines execute faster, and clearer logic helps teammates understand your
work without deciphering complex transformation chains.
Summarize pipeline logic and datasets
Summarize pipeline logic and datasets
Get a high-level explanation of what a pipeline does and which datasets it uses without reading
through every transformation. Understanding existing pipelines quickly helps with onboarding,
while documenting your own work makes it easier for others to use and modify later.
Features in progress
Features in progress are subject to the same project-scope and permission constraints as existing capabilities.Find tables based on similar datasets
Find tables based on similar datasets
Provide sample data to the Agent and retrieve similar datasets from your fabric.
Make targeted updates to an existing pipeline
Make targeted updates to an existing pipeline
Iterate on granular aspects of the pipeline while ensuring previous work remains intact.
Find pipeline optimizations and simplifications
Find pipeline optimizations and simplifications
Identify opportunities to improve pipeline performance using methods such as query optimizations,
caching, or simplified joins.
Prompt a reindex of the knowledge graph
Prompt a reindex of the knowledge graph
Ask the Agent to start crawling sources to ensure the most up-to-date metadata in the knowledge
graph.
Build and run data tests and unit tests
Build and run data tests and unit tests
Generate tests to validate data quality and catch errors before pipelines run in production.
Recommend packages for your project
Recommend packages for your project
Find relevant packages that help build out your pipeline for your specific use case.
Schedule and deploy projects
Schedule and deploy projects
Automate moving pipelines from development to production.
Perform version control actions
Perform version control actions
Track changes, create branches, and manage pipeline versions through the Agent interface.

