Prerequisites
To complete this quickstart, you need a Prophecy fabric that uses Prophecy In Memory or Databricks as the compute engine.Prophecy automatically creates a compatible fabric for Free and Professional Edition users. If you
do not have any existing fabrics, you’ll need to create
one.
Set up a new project
First, you need to create the project where you will build your pipeline. You’ll also need to add data to the project for this quickstart.1
Create a project
- Click on the Create Entity button in the left navigation bar.
- Hover over the Project tile and click Create.
-
Give your project a Name, such as
Prophecy_Quickstart. - Under Team, select your personal team. (It will match your user email.)
- Under Select Template, choose Prophecy for Analysts.
- Click Complete.
2
Connect to the execution environment
- Select the default fabric or your own fabric.
- Click Save.
3
Add a new pipeline
- On the project landing page, click Create Pipeline.
- For the Pipeline Name, enter
patient_analytics. - Leave the default Directory Path of
pipelines. Prophecy saves your compiled pipeline code in this folder of the project repository. - Click Create.
Add data to your project
Next, you’ll add some data to the project so the Agent can find and transform it.1
Create the seed data file
Load some data into the project as a :
- Open the Source/Target gem category.
- Click Table. This adds a new Table gem to the canvas.
- Hover over the gem and click Open.
- Select + New Table.
- For the Type and Format, choose Seed.
-
Name the seed
patients_raw_data. - For the Seed path, choose seeds. Prophecy saves your seed file in this folder of the project repository.
- Click Next.
-
In the Properties tab, paste the following data.
patients_raw_data.csv
- Click Next.
- Click Load Data to preview the data in tabular format.
- Click Save.
2
Run the seed
To materialize the seed data into the SQL warehouse, run the pipeline once. To do so, click the play button in the bottom right corner of the canvas.
3
Reindex your connection
Prophecy should automatically index the table when you save the seed. This allows the Agent to discover and use the seed data. If you have trouble finding the table in later steps, you can also manually reindex the knowledge graph.
- Open the Environment tab in the left sidebar.
- Below your connections, you’ll see a Missing Tables? callout.
- Click Refresh to trigger the knowledge graph indexer.
Verify the table was indexed by checking the Environment tab in the left sidebar.
Explore the data
Now that you have data in your project, you can explore it using the Agent.1
Access the Agent
Locate and open the Chat tab in the left sidebar.This is where you’ll interact with the Agent.
2
Explore your data
Ask the Agent to search for the seed data table.Enter the following prompt in the chat:The Agent returns:
- A short list of relevant datasets with descriptions.
- A full list of matching datasets.
- The option to add datasets directly to your pipeline on hover.
-
Click on
patients_raw_data(the data you uploaded) in the chat to open a detailed preview dialog where you can:- View the table location.
- Examine the schema and column structure.
- Preview sample data.
- Review data profiles.
- Open an Explore session for dataset-specific queries.
- Close the preview dialog to return to the chat.
If the Agent doesn’t find your table, verify that the table was indexed by checking the
Environment tab in the left sidebar. Additionally, if you have other patient data in your
warehouse, the Agent may return other datasets that match your query instead.
3
Get data samples
Request specific data samples to validate your understanding.Enter the following prompt, replacing The Agent returns:
@patients_raw_data with your actual table path:- A table with the requested data.
- An option to preview the table for detailed examination.
- An option to add the table to your pipeline.
- SQL execution logs showing the query used.
Verify the results show 4 patients: John Smith, Sarah Johnson, Emily Brown, and Jessica Garcia,
all from Boston.
4
Create visualizations
Generate charts and insights from your data. Enter the following prompt:The Agent returns:
- An embedded chart in the chat showing patient counts by city.
- An option to preview the table for detailed examination.
- An option to add the table to your pipeline.
- SQL execution logs showing the query used.
- Visualization tab: View larger charts and download charts as images.
- Data tab: Examine underlying data and download data as JSON/Excel/CSV.
The chart should show Boston with 5 patients, Cambridge with 2, Worcester with 2, and Springfield
with 1.
Build your pipeline
Now, you’ll build your pipeline transformations using the Agent. Keep the Visual view open (rather than the Code view) to see updates in real-time as the Agent adds or modifies gems on the canvas.1
Build your first transformation
Describe the transformation you want to perform. Enter the following prompt:Replace
@patient_records with the actual gem label from your canvas if it differs.The Agent will:- Add the appropriate gem(s) to your pipeline canvas. This prompt should produce an Aggregate gem.
- Execute the pipeline, generating data samples that you can review.
- Provide a description of the changes made.
- Show options to inspect, preview, or restore changes.
- Display SQL execution logs.
The output should show 3 counties: Suffolk with 5 patients, Middlesex with 2, Worcester with 2,
and Hampden with 1.
Each change that the Agent makes can be viewed in the project version
history. You can
revert the changes at any time.
2
Inspect pipeline changes
To understand what the Agent built:
- Click Inspect on the transformation response.
- Review the configuration panel starting with the first modified gem (highlighted in yellow).
- Use the Previous and Next buttons to navigate through modified gems.
- Examine input and output data to verify the transformation produces expected results.
- Understand the Agent’s approach.
- Verify that the transformation logic matches your expectations.
3
Add another transformation
Build a more complex transformation. Enter the following prompt:The Agent adds another transformation that calculates the average cost. This demonstrates how you can chain transformations together.
Verify the results show average costs for each county. Suffolk should have an average around
1,544, Middlesex around 2,410, Worcester around 1,800, and Hampden 3,800.50.4
Save your results
After building your pipeline, ask the Agent to save the final output. Enter the following prompt:The Agent adds a Table gem to the end of your pipeline. When you run the pipeline, the Table gem writes the data to your default database and schema defined in your fabric, allowing you to persist results.
Explore further
Try these additional tasks to extend your pipeline.1
Filter data
Enter the following prompt in the chat:
2
Join with additional data
Create a second Seed file named Then, prompt the Agent:
county_info with the following content:county_info.csv
The Agent will infer join keys, but you can also specify them explicitly.
3
Calculate derived metrics
Enter the following prompt in the chat:
Connect your own data
This quickstart uses a Seed file as the source data. When you start building your own pipelines, you’ll likely want to use your own data from files or external systems. You can do this by:- Uploading files directly from your local filesystem using the upload file feature.
- Ingesting data from external systems using connections.
Sample prompts reference
Use these prompts as templates for your own pipelines.| Task | Prompt Example |
|---|---|
| Find data | Find datasets containing customer information |
| Sample data | Show me 5 random records from @sales_data |
| Filter data | Filter to only include orders from 2024 |
| Transform data | Calculate total revenue as quantity * price |
| Join data | Join the orders and customers tables |
| Parse data | Extract the fields from json_data as columns — This works best if you provide a sample JSON object. |
| Clean data | Remove rows where email is null |
| Aggregate data | Group by region and calculate average sales |
| Visualize data | Create a bar chart of monthly sales |
| Save results | Save the final output as a table |
Tips
Be specific with your requests
Be specific with your requests
How to do it: Instead of
Clean the data → Try Remove duplicate customer records and fill null values in the email columnWhy it helps: Reduces ambiguity so the Agent applies the correct operations.Use step-by-step instructions
Use step-by-step instructions
How to do it: Break complex transformations into smaller requests. For example:
Aggregate orders by customer → Join with customers → Filter for orders made in 2024Why it helps: Improves reliability and makes it easier to debug or adjust each step.Review changes before proceeding
Review changes before proceeding
How to do it: After a response, click Inspect and use Previous/Next to review highlighted gem configuration and outputWhy it helps: Ensures the transformation matches expectations before you continue.
Leverage both the Agent and visual interface
Leverage both the Agent and visual interface
How to do it: Use chat to scaffold transformations, then switch to the Visual canvas to fine-tune or add gemsWhy it helps: Combines speed (AI) with precision and control (visual editor).
Troubleshooting
Agent can't find tables
Agent can't find tables
Probable cause: Knowledge graph is out of dateHow to fix: Reindex your Databricks connection so the knowledge graph includes the table
Conversation feels stuck or off track
Conversation feels stuck or off track
Probable cause: Chat context isn’t relevant anymoreHow to fix: Click
… > Reset in the chat interface to start a fresh sessionNeed to report a bad response
Need to report a bad response
Probable cause: You want to provide feedback to improve resultsHow to fix: Click the thumbs-up or thumbs-down on the Agent’s message. This helps us improve the Agent’s behavior.
Results don't match expected values
Results don't match expected values
Probable cause: Seed data wasn’t created correctlyHow to fix: Verify the table exists and contains the expected 10 rows of data

