How Tabular Editor supports agentic development for semantic models

Key Takeaways

  • We anticipate that agents will play a major role in semantic model development in the future. Agents can help you make better semantic models faster already today, if they have the right tools and context. We want to enable these workflows to augment existing ways-of-working and to save you time, cost, and effort. We are introducing the AI assistant, the Tabular Editor 3 semantic model CLI, and collections of AI skills to ensure you and your agents can make better semantic models faster with Tabular Editor.
  • The AI assistant in Tabular Editor is an entry point for simple, AI-assisted development. It has some specific knowledge about Tabular Editor and semantic models to help you with C# scripting, DAX queries, and Q&A about various topics. It can also use the VertiPaq Analyzer and Best Practice Analyzer. The AI assistant is designed to be non-intrusive and can be used with both local and organizational LLMs for scenarios where you need this for compliance. However, it is scoped to Tabular Editor, only; it’s not an MCP server or coding agent. It’s an integrated UI feature in Tabular Editor, like Copilot in Power BI Desktop.
  • The AI assistant is a bring-your-own model experience; you need to provide an API key. This allows you to choose your own provider and model, including local or organizational LLMs. Crucially, however, API costs are more expensive than personal, subsidized subscriptions from these providers.
  • The Tabular Editor CLI is a semantic model command-line tool built for humans but optimized for agents; it’s not yet available and coming soon in April: Command-line tools are usable by both humans (manually or in automation pipelines) and agents. They’re a flexible, simple way enable agents with structured tools without flooding the context window. We’ve developed a new Tabular Editor CLI that works on Mac and Windows to facilitate agentic semantic model development with frontier coding agents, but also automation or interactive use.

This summary is produced by the author, and not by AI.


Semantic model development and AI agents

If you work with Power BI, Fabric, or Analysis Services, then you likely are aware of (or already use) AI with semantic models in some capacity. Be it helping you write DAX or M code, querying your model, or even end-to-end agentic and asynchronous development. If you want an introduction to agentic development with semantic models, you can read about it in our previous series.

Agentic development can save you time and help you produce better semantic models. Therefore, we anticipate that AI and agents will play a major role in semantic model development (as well as BI and data analytics as a whole). It’s clear that AI and agents will augment BI professionals and self-service users to build better models, visualizations, and end-to-end data solutions in a fraction of the time that it used to take. In our opinion, this value is demonstrated already for early adopters of other tools: the Fabric command-line interface (CLI), semantic model MCP servers released by Maxim Anatsko and Microsoft, or just AI skills that promote agentic development.

Our mission is to provide professionals the best tools available for building, optimizing, and managing semantic models. We will make sure that both you and the agents that you use and orchestrate can use these tools to get the best possible semantic models quickly and efficiently.

In this article, we share our view of tooling for agentic development of semantic models. We also announce the release of a built-in AI assistant integrated in the Tabular Editor 3 UI, and a new, enhanced Tabular Editor CLI for sophisticated end-to-end development.

K037 Figure 2 - Comparison of two approaches for AI agents to use Tabular Editor: the AI assistant built into the UI with bring-your-own-model API support and custom instructions, and the Tabular Editor 3 Semantic Model CLI that works cross-platform and is usable by any coding agent with a Bash tool

NOTE

If you do not want to or can’t use AI and agents in your own development work, then that is fine. However, we would encourage you to remain open-minded and explore these new tools; we are already past the time of “slop and vibes”, provided that you invest in the right tools and context.

How we see tooling for agentic development of semantic models… and in general

For agents to perform well on any task, they need the right tools and context. This is also true with semantic models, where the agent needs ways to interact with a semantic model and good context about the relevant concepts, techniques, and workflows.

K037 Figure 3 - Diagram showing approaches for AI to work with and modify semantic models, including an AI agent connecting to model metadata via TMDL, MCP servers, and code with CLIs, alongside built-in assistants and integrations connecting directly to the semantic model

Like traditional semantic model development, there is no “best” tool; each tool is useful for different scenarios and addressing different problems. We see this as follows:

  • Built-in assistants: Assistants like Copilot and the AI assistant in Tabular Editor provide an entry-level experience that is more accessible for non-technical users. They’re simpler than a coding agent and more limited in scope. Depending on the level of AI integration, the assistant can still accomplish many tasks (the assistant in Tabular Editor can make any change with C# scripts, for instance), however these are not coding agents nor MCP servers. They also have limited interaction with their external environment, such as using web search or interacting with local files and directories.
  • MCP servers: The model context protocol (MCP) is new, and the spec is still evolving. Initially, MCP servers made a strong impression to extend agentic chatbot tools like Claude Desktop and GitHub Copilot in VS Code. This is especially true for tool discovery and transport. Recently, however, companies have demonstrated reluctance to develop and support MCP servers over traditional approaches with APIs and CLIs, or even custom agents. We still believe that MCP servers are a good choice when an agent can’t call commands or execute code, when a good CLI does not already exist, or when you just prefer to use an MCP server.
  • CLIs: Command-line interfaces (CLIs) have predated agent use. They are simple and, unlike MCP servers, usable by both agents and humans. You can use a command-line tool manually or incorporate it in a CI/CD pipeline. This means that a CLI has some extra benefits out the gate:
    • With a good CLI, you enable both an agent and human, at the same time.
    • CLIs can also be more flexible since they don’t adhere to a specification. CLIs can also be more creative in their design and execution to solve novel problems.
    • CLIs can also be broader in scope than MCP servers, since they don’t risk overflowing the context window like MCP server tool descriptions and resources do.
    • Research has shown that agents are better at using code than calling tools. In our observations and testing, we’ve seen that agents perform best by using CLIs in combination with relevant context (such as memory and skills) and access to metadata files.
  • Skills and other context: To perform well with any approach, agents require good context and instructions. The best approach for this now is agent skills, which are text files with references, scripts, and examples that agents can refer to or use. Good skills must be well-written by human experts and not generated by AI and iteratively refined over time. This is something already provided by Databricks with their AI dev kit, for instance.
  • Other approaches: This area is evolving quickly, and there are more approaches than what we list here. For instance, we’ve observed Codex and Claude Code with the most recent models can connect to, query, and modify models open in Power BI Desktop without any special tools or engineering (no MCP server; no CLI). You can just tell Claude that you have a model open in Power BI Desktop and to go find and modify it… it will just do it. Agents can discover and connect to the local Analysis Services instance via ADOMD.NET, and query or modify it using DAX and TOM, directly; it works quite well, actually. What’s crazy is that you can even do this on a Mac, connecting to Power BI Desktop open on a VM. You can see an example of this, below, in an approach using this skill:

K037 Figure 4 - Animated demonstration of Claude Code in the terminal using the Tabular Editor CLI to discover, explore, query, and modify a semantic model connected via the Power BI REST API

In the previous GIF, you can see Claude Code in the terminal, discovering, exploring, querying, and modifying a model open in Power BI Desktop without an MCP server or any other tools. It’s not a prescriptive workflow; Claude just figures it out.

This is an emerging area, so it’s unclear how this space will look in a few years… or even a few months. However, we want to ensure that any approaches we take to AI integration will benefit both human users and agents both equally and effectively.

Available March 2026: The AI assistant in Tabular Editor

In the March release of Tabular Editor, we’re releasing the AI assistant. This is a new feature that lets you connect an LLM and let it perform certain tasks in Tabular Editor:

  • Asking questions about Tabular Editor or semantic models.
  • Writing and executing DAX to query your model.
  • Analyzing or optimizing the model using TE features like VertiPaq Analyzer, BPA, etc.
  • Modifying the model via C# scripts, which you can save and re-use as macros.
  • Uses a bring-your-own-model approach; you must provide an API key, and can't use your coding agent subscriptions.
NOTE

Here’s an interactive walkthrough of the AI assistant, teaching you how to set up and use it.

The following is a brief demonstration:

The AI assistant is intended to be an entry-level helper for areas where we’ve seen people use AI with Tabular Editor. The purpose of the AI assistant is to make these more advanced or time-consuming tasks more convenient for those who choose to use AI. The assistant isn’t designed to be as broad as an MCP server or coding agent; rather, to make certain parts of the tool more accessible.

The most interesting application of the AI assistant is to help you write C# scripts. These scripts let the assistant make changes to the semantic model in Tabular Editor, which means that you can see the changes in real-time, and Tabular Editor validates them. You also can undo any changes with Ctrl+Z and preview those changes in a new model “diff” view. The best part, however, is that working scripts can be generalized to re-use as saved scripts or macros without AI at all… so you can quite easily build up a library of re-usable macros without knowing or writing any C# code.

diff

The diff view is particularly interesting, as it’s a new view in Tabular Editor that will eventually allow you to compare two semantic models. In the future, we intend this to work just like other popular diff tools, such as ALM toolkit. For the March release, however, the model diff view only works with C# scripts, irrespective of how you author them (i.e. with or without AI).

View what a C# script will do in your model before you run it with the new model change diff view.

However, we also want to ensure that you can ignore the assistant altogether, if you choose to. You can choose to install the AI assistant or not when you update Tabular Editor. Furthermore, this feature will never bother or intrude on your work, and when you close it, it stays closed. You’re fully in control.

With the AI assistant, we’ve opted for a bring your own model approach. This means that you must supply connection information in the preferences, like an API key. This affords you the flexibility to choose providers, models, and even use local or organizational LLMs if available. This is particularly helpful to organizations that have stricter compliance or residency requirements.

NOTE

You cannot use your subscriptions for other coding agents and AI tools (like Claude, Open AI, GitHub Copilot) with the AI assistant, as that’s against their terms of service.

These personal subscriptions are also cheaper than using the APIs, because they're subsidized. Therefore, using the same model in a coding agent with your subscription is cheaper than using it in the AI assistant via an API key. There are users who may prefer to use GitHub Copilot or Claude Code; for these users, we recommend you read about the Tabular Editor semantic model CLI in the next section, as this tool is specifically made to be the best possible semantic model tool for these agents.

To change the provider or model, you just must click the “gear” icon. You can choose from multiple providers, including Anthropic, OpenAI, and OpenRouter. For local and organizational LLMs, see our documentation for a full walkthrough of how to set this up.

Click the gear icon to go to preferences and set the model

The AI assistant also supports custom instructions, which function like skills. These are simple markdown files which you can add to an AI folder:

K037 Figure 8 - The custom instructions folder browser in Tabular Editor, showing markdown files for AI assistant skills that can be viewed, edited, or extended by clicking the folder icon

These instructions are conditionally loaded to the assistant based on keywords that you configure, or you can invoke them like you would in other tools by pressing “/”. The AI assistant comes with several custom instructions pre-loaded. You can see some examples, below:

K037 Figure 9 - The AI assistant chat input in Tabular Editor showing built-in slash command skills including model-modification, model-design, organize-model, and optimize-model

The purpose of these custom instructions is to provide good context to improve the outputs that you get from the AI assistant. Some of these instructions are workflows, like organize-model or optimize-model. Others are general information to teach the LLM about important concepts related to semantic models, like model-design.

Notably, the AI assistant isn’t designed to let you do everything in Tabular Editor. Here’s some examples of what you can’t do:

  • Connect to external files or services or search the web.
  • Add or use MCP servers or other tools.
  • Connect to another model from an AI assistant (you do this in the user interface).
  • Manage or change preferences for Tabular Editor or the AI assistant, itself.

To reiterate, the AI assistant is meant to be a helper with advanced tasks that we’ve observed cost users time or require specific skills, like writing DAX queries or C# scripts. The AI assistant isn’t designed or intended for more complex or end-to-end agentic development scenarios. For these scenarios, we’ll provide a new, improved semantic model command-line interface (CLI).

Available April 2026: The Tabular Editor semantic model CLI

In April, we’ll introduce a new semantic model command-line interface (CLI) in preview that’s built for human developers but optimized for agents:

  • Works on any operating system, including MacOS and Linux, and with any coding agent, including GitHub Copilot and Claude Code.
  • Commands that give you full control over a semantic model.
  • Get headless use of the Tabular Editor 3 features you know like C# scripts, the VertiPaq Analyzer, the BPA, DAX code assistance and code actions, and more.
  • Backward compatible with the Tabular Editor 2 CLI
  • While in preview, the CLI will be available to try for anyone with a Tabular Editor account. Later this year, we will conclude this preview, whereupon the CLI will be available only for specific Tabular Editor licenses. We will be fully transparent about this timeline and the details in our CLI, docs, communications, and website!

A CLI can be used by humans manually or headlessly in CI/CD pipelines. An example of this is the Tabular Editor 2 CLI, which is used by many people to automate the development, testing, and deployment of semantic models. However, CLI tools can also be used by agents. We’ve written about several use-cases of this with the Fabric CLI, which is (in the author’s opinion!) the best Power BI or Fabric feature Microsoft released in 2025.

We’d like to announce a new semantic model CLI that has all the features that you expect and need to develop, manage, and optimize semantic models… and some extras. It’s been designed from the ground up to provide you everything that’s in Tabular Editor, and even new, more advanced features specifically designed for agentic development scenarios.

The biggest feature of this CLI is that it works on Windows, Mac, and Linux. This means that you can now develop, manage, and optimize your semantic models from any operating system:

K037 Figure 10 - The Tabular Editor CLI running on macOS, connected to a remote semantic model and displaying VertiPaq Analyzer results with table sizes, column counts, and row statistics

In the screenshot, you can see the te command line connecting to a semantic model, and then running the VertiPaq Analyzer to identify memory bottlenecks. The CLI is running on a MacOS in the cmux terminal, manually.

The CLI has full backward compatibility with Tabular Editor 2’s CLI. However, it’s also been enhanced with a variety of new commands and features.

WARNING

While the CLI is technically fully backward compatible, it’s using the enhanced scripting and BPA engine of Tabular Editor 3. Therefore, you will still want to test it rigorously before integrating it into your pipelines. Note that schema compare will not be available at the time of first release.

First, to introduce it, you can concisely browse and summarize a model:

K037 Figure 11 - The Tabular Editor CLI listing a semantic model summary with object counts for tables, measures, columns, relationships, calculation groups, and other model metadata

As well as query a model with DAX, and even get query traces or benchmarks like you would with DAX Studio:

K037 Figure 12 - The Tabular Editor CLI executing a DAX SUMMARIZECOLUMNS query against a semantic model, showing query results in a formatted table followed by a performance trace with Formula Engine and Storage Engine timing breakdowns

This means that the CLI can be a valuable tool not just for writing DAX but also optimizing it in the hands of both humans and agents. It provides a rich set of tools for exploring and querying your model in an elegant and effective way but also one that will still be token-lean and straightforward for AI coding agents, too.

Of course, you can also develop a semantic model with this CLI. Search, find, and modify any object or property, individually or in bulk. You also have all ofe the typical features of Tabular Editor, like formatting DAX and even Power Query:

The tool has full access to the functionality in Tabular Editor that you know and love, like the VertiPaq Analyzer for memory optimization, the Best Practice Analyzer for finding and resolving issues and Code Actions, which provide smart warnings and can even enable auto-fix on model save and deploy:

K037 Figure 13 - The Tabular Editor CLI running Best Practice Analyzer rules against a semantic model, displaying BPA violations with severity, rule name, affected object, and object type in a formatted table

What’s also unique to the CLI is that you can set up skills to help agents better use the CLI and develop semantic models. This agent setup works with any provider, from Claude Code to GitHub Copilot, and can be scoped to the user or project level. We’ve tested the CLI with various coding agents and experienced positive results, and we look forward to hearing your experiences, as well:

In this demonstration, you can see Claude Code using the Tabular Editor CLI without any skills or context to identify optimization opportunities. It can do this quite well out-of-the-box because it has access to the tools in the Tabular Editor UI designed and used for optimization, and familiar to the LLM from training data, alone. With skills this becomes even more powerful.

NOTE

To reiterate, the CLI is previewed in this article, but is not yet available. It will be available in our next release in April.

These are just some of the features of the AI assistant and this new semantic model CLI; less than 1/5th of what’s possible. The best is yet to come…

Further recommended reading

In conclusion

We anticipate that agents will become a key part of semantic model development in the future. Already today, we see many examples of agents enhancing or expediting semantic model development workflows. We’re releasing two new features:

  • The AI assistant is an entry to point for helping you work with AI in Tabular Editor. You use it in the user interface, where it can write and execute C# scripts, DAX queries, and use tools like the VertiPaq Analyzer and BPA.
  • The Tabular Editor semantic model CLI, built for human developers first, and optimized to give agents the best tools available for model development and optimization. It can do almost everything that you can do in the Tabular Editor GUI, and works on any operating system.

We hope these new features help you develop better semantic models, faster!

Related articles