Skip to main content

AI-powered “talk-to-your-data” applications promise a compelling future: business users making natural-language data queries and instantly receiving accurate, contextual insights from complex enterprise data. In practice, delivering this experience—especially inside an existing enterprise product—is more challenging than the marketing suggests.

Cleartelligence’s work building an AI-driven conversational analytics feature for one of our clients—a financial software company—delivered a number of high-impact benefits. For our client’s customers, it opened the door to fast and easy access to accurate insights from very complex data sets. All they have to do is ask a natural language question and the information they need is delivered in seconds, so it’s clearly a beneficial and differentiating product feature. For our client, this talk-to-your-data solution not only optimizes their investment in Databricks by operationalizing the AI/BI Genie functionality, it is accelerating their customers’ migration to a high-performing cloud-based version of the company’s product offering.

But the process for activating this level of performance and impact surfaced a number of important lessons about architecture, data readiness, security, and the realities of emerging AI platforms. It also revealed some genuinely pleasant surprises about what modern data and AI tooling can do when thoughtfully applied.

This article highlights six key lessons we learned deploying this solution, including the unique challenges of talk-to-your-data solutions and the workarounds that made the solution viable, production-ready, and scalable.

Six lessons learned from deploying a real-world talk-to-your-data solution.

1. “Talk-to-Your-Data” Solutions Are Really “Talk to Your Data Architecture”

One of the earliest lessons we learned was that conversational analytics solutions live or die on data architecture, not just AI models. The client’s existing reporting layer relied on Microsoft SQL Server views that had evolved over many years. At first glance, the AI solution seemed simple: Replicate the core reporting views into the client’s Databricks instance and point the AI at them. In reality, those core views were the product of dozens of interdependent views, with logic distributed across layers of technical debt.

Recreating this structure in Databricks turned into a significant effort, requiring:

  • Replication of raw application data from SQL Server into Databricks
  • Rebuilding and validating dozens of reporting views
  • Ensuring functional parity with existing customer-facing reports

Lesson learned: AI doesn’t simplify data complexity—it exposes it. If your reporting logic is brittle, undocumented, or fragmented, a “talk-to-your-data” layer will force you to confront that reality head-on.

2. Data Performance Problems Don’t Magically Disappear with AI

When you’re working in a complex, big data environment where test environments can contain millions and millions of rows in the largest tables, and existing workflow performance is challenged, especially for large data exports, introducing AI can add new considerations, including:

  • Data replication latency
  • Query execution speed
  • AI response times when generating SQL, executing queries, and summarizing results

To counteract the environmental challenges for this project, we put significant effort into optimizing Databricks queries, tuning Genie performance, and ensuring acceptable end-to-end response times even during conversational interactions

Lesson learned: AI sits on top of your data platform—it doesn’t replace the need for strong performance engineering. If your data queries are slow without AI, they’ll be slow with AI too.

3. Early AI-Native Capabilities Are Rapidly Enhancing the Value of Data Platforms

Databricks Genie proved to be extremely strong at a critical capability: accurate SQL generation across complex schemas. In fact, the team found it to be among the best tools available for this purpose, especially when paired with strong configuration. The Databricks Genie team was incredibly collaborative in helping us layer in Claude (Anthropic hosted natively in Databricks) as guardrails, and to expand upon the answers that Genie provided. This layered AI approach delivered the best end-user experience, while maintaining acceptable performance.

Lesson learned: Production AI solutions often require working with multiple AI services, each doing what it does best. Flexibility and creativity matter as much as model choice.

4. Security Is One of the Hardest Problems in Conversational Analytics

And it proved true for this project as well. The client wanted to be able to establish row-based access controls so users could only see data for companies they were authorized to access.

In a traditional BI tool, security is enforced through user context and database permissions. In a “talk-to-your-data” solution, the AI effectively “sees” the entire database making strict security requirements a challenge to implement. When you use an API that’s run by a service principal account, it cannot take on the identity of the end user. Its identity is static. Since that was the situation for this project, we had to figure out how to apply dynamic permissions across all of the data.

The Workaround: Prompt-Driven Row-Level Security

In this case, our team implemented a novel approach:

  • The user ID was passed from the UI into the AI prompt (hidden from the user).
  • Genie was instructed to join against a permissions table and filter results accordingly.
  • A guardrail ensured that the user id came through in the query result.

Lesson learned: While this approach has proved effective, hard guardrails in code should always be the long-term goal, rather than relying solely on AI-generated logic. Security cannot be an afterthought. In AI-driven systems, you must design for the reality that the model can “see everything” unless explicitly constrained. We’re implementing deterministic guardrails in the next phase of the project.

5. Instructing Your Model With Domain-Specific Knowledge Dramatically Improves AI Accuracy

One of the biggest upsides to Databricks Genie is how configurable it is, which is particularly important for a client whose business lexicon is complicated and unique. Genie allows you to go beyond using the LLM as a generic language model, establishing a metadata layer on top of your data model. This enabled our team to:

  • Provide valuable business domain context that helps the AI determine how to answer questions
  • Define business concepts directly in SQL (e.g., specific definitions of client and industry terms)
  • Explicitly define joins and relationships
  • Provide curated example questions and answers

This domain grounding dramatically improved the accuracy and relevance of AI responses, even across wide tables with dozens of columns and highly specialized terminology.

Lesson learned: Talk-to-your-data solutions work best when the AI is taught to speak the business’s language, not just SQL.

6. The Right Architecture Enables Future Growth

Despite a few technical hurdles, the final architecture establishes a strong foundation for broader AI adoption across the client’s product portfolio. The orchestration API, logging capabilities, and AI layering features we developed will support additional use cases moving forward. Importantly, the solution was designed so the client can independently own, operate, and extend it over time.

Lesson learned: A successful conversational analytics solution is not a one-off experiment—it’s an architectural investment.

The Ultimate Lesson from Talking to Your Data

Building a production-grade talk-to-your-data experience is equal parts AI, data engineering, security design, and pragmatism. This project reinforced that while AI capabilities are advancing rapidly, success still depends on fundamentals: clean data flows, performance tuning, strong governance, and thoughtful system design. When these pieces come together, conversational analytics solutions can move beyond novelty into real business value and deliver a genuinely transformative user experience.

Learn more about how Cleartelligence can help you supercharge business insights with AI.

 

This is article was written with the assistance of LLM editing capabilities.

Conversational Analytics FAQs

Answers To Your Data & AI Challenges

Find quick answers to some of the most common questions about AI-powered conversational analytics.

What is a “talk-to-your-data” solution?

It is a conversational analytics system that lets users ask questions in natural language and receive answers generated from enterprise data using AI.

How do AI-powered conversational analytics work?

Conversational analytics works by translating natural language questions into SQL queries, running those queries against a data platform, and returning summarized results in plain language.

What are the biggest challenges in building a talk-to-your-data solution?

The biggest challenges include data architecture, performance and scalability, security and data access control, domain understanding and context, and accuracy and reliability of AI responses across large, complex datasets.

Why is data architecture important for a successful talk-to-your-data solution?

Data architecture is critical because AI systems—like conversational AI solution—rely on clean, well-structured data models. Poorly designed or fragmented data layers can lead to inaccurate results and slow performance.

How do you ensure security in a conversational AI analytics system?

In AI-driven systems, you must design for the reality that the model can “see everything” unless explicitly constrained, which means hard guardrails in code is always preferrable to AI-generated logic.

Keith Meyer, AI Engineering Practice Leader

As head of Cleartelligence’s AI Engineering practice, Keith is both a business and technical strategist with the rare ability to engage in solutioning from the 50,000-foot level down to the 5-foot level.  Keith has been the leader for many first-in-industry initiatives in tech, eCommerce, Life Sciences, and SaaS.  At Cleartelligence, Keith works to clients to identify impactful data, application, and AI strategies and solutions. He follows through on recommendations by leading integrated teams in analytics, AI and software development projects.