Now in Beta · insightsassist.ai

AI that only knows
what you tell it.

InsightsAssist embeds a citation-first AI into any website. Users explore your data through natural language — every answer grounded in your dataset, never hallucinated, never invented.

→ “Which departments led external funding from 2019–2024, and who were the top principal investigators?”

Across 4,847 funded research initiatives in this dataset, three departments account for 62% of total awarded funding. DEPT-ENG-042 leads with $284.3M across 1,203 active grants — PI-0891 is the highest-funded investigator at $47.3M in awarded federal funding. DEPT-MED-017 follows at $198.1M, up 43% from the prior five-year period. DEPT-ENV-008 recorded a 340% increase since 2019, driven by 14 new federal climate initiatives tracked under GRANT-FED-2023-C. Full rankings available across all 4,847 funded projects.

DEPT-ENG-042 · $284.3M · 1,203 grantsPI-0891 · $47.3M · Federal4,847 records · 2019–2024
HOW IT WORKS

From dataset to live AI in under ten minutes.

01

Connect your dataset

Upload a CSV, JSON, or connect a live database. InsightsAssist indexes your records, runs semantic embeddings, and builds SQL aggregation queries automatically. Supports Google Datasets, Supabase, PostgreSQL, Airtable, and custom uploads.

02

Drop one script tag

Paste a single line of code onto any page. Works on WordPress, Webflow, Shopify, React, Framer, or plain HTML — no rebuild required.

<script
  src="insightsassist.ai/embed.js"
  data-key="your-api-key"
  data-dataset="ds_abc123">
</script>
03

Users explore. Facts stay grounded.

Every answer the AI gives cites the exact record IDs it drew from. Users can verify any claim in one click. If your dataset doesn't support an answer, InsightsAssist says so — it never fills the gap with a guess.

FEATURES

Built for accuracy. Designed for users.

🔐

Grounded by architecture, not instruction

The AI is structurally prohibited from using training data. Your records are the only source. There is no fallback, no invented answer — because there's no pathway for hallucination to occur.

📎

Citations on every single claim

Record IDs appear inline with every statement. Tap any citation to see the source record in full. Verification is one click — not one Google search.

🧮

Semantic + SQL dual-mode search

"How many X?" runs an exact SQL count. "Tell me about Y" runs semantic search. The engine picks the right mode automatically — both types of queries are accurate.

One embed tag. Any platform.

Shadow DOM widget deploys on WordPress, Webflow, Shopify, React, Framer, or raw HTML — without touching your CSS or breaking your existing layout.

🎛️

Full admin dashboard

Upload datasets, configure appearance, set scope restrictions, manage API keys, monitor usage. Built for non-technical site owners.

📊

Query analytics and coverage gaps

See exactly what your users are asking. Questions your dataset can't answer appear in a gap report — a direct signal for what content to add next.

TECHNOLOGY

Instant RAG.
No infrastructure required.

Retrieval-Augmented Generation — RAG — is the technique that makes AI answers trustworthy: instead of asking the model to answer from memory, you first retrieve the relevant records from your dataset, then instruct the model to reason only from those records. Every serious AI data product is built on RAG. Building it yourself takes a senior engineer four to eight weeks. InsightsAssist is pre-built RAG infrastructure — connect your dataset and it's live in minutes.

QUERY PIPELINE — FROM USER QUESTION TO CITED ANSWER
INGESTData indexing
CSV · JSON · SQL · GBQChunking strategyRecord metadata tagging
EMBEDVector embedding
text-embedding-3-large1,536 dimensions / recordpgvector storage
RETRIEVEDual-mode search
Semantic: cosine similarityStructured: SQL aggregationTop-K + threshold tuning
GROUNDHallucination lock
Retrieved records onlyTool-use enforcementNo training data fallback
GENERATECited answer
Claude SonnetRecord IDs extractedSource cards rendered
🔧

Building RAG yourself

estimated: 4–8 weeks, senior engineer

  • Choose and configure an embedding model
  • Stand up a vector database (Pinecone, Weaviate, pgvector)
  • Build an ingestion and chunking pipeline
  • Write retrieval logic with similarity thresholds
  • Engineer a hallucination-prevention prompt layer
  • Handle semantic vs. structured query routing
  • Build citation extraction and rendering
  • Design and ship an end-user chat UI
  • Wire up usage analytics and monitoring
  • Maintain all of the above as models and APIs evolve

...then repeat this for every new dataset.

InsightsAssist

estimated: under 10 minutes

  • Upload your dataset (CSV, JSON, or DB connection)
  • Done.

Embedding model selection, vector storage, retrieval tuning, hallucination prevention, citation rendering, analytics, and UI — all pre-built, pre-tuned, and maintained by InsightsAssist. Every new dataset you connect gets the same production-grade RAG pipeline in the same ten minutes.

# That's the entire integration.
<script
  src="insightsassist.ai/embed.js"
  data-key="ia_live_••••••••"
  data-dataset="ds_••••••••">
</script>
1,536Vector dimensions per record
< 200msMedian query latency
Dual-modeSemantic + SQL routing
ZeroTraining data fallback

What makes RAG work

When a user submits a question, InsightsAssist converts it into a vector embedding using the same model that indexed your dataset. It then runs a cosine similarity search across your stored record embeddings to find the most semantically relevant matches. Those top records — and only those records — are passed to Claude as context. The model is given a strict system prompt: answer only from the provided records, cite every claim with a record ID, and respond ‘the dataset does not contain data to answer that’ if the records don't support the question. Claude's training knowledge is not a fallback. It has no access pathway.

Why dual-mode retrieval matters

Pure semantic search — the approach most RAG implementations use — performs well for open-ended, exploratory questions. But it breaks down on analytical queries. Ask ‘how many members traded technology stocks in Q1?’ and a vector search returns similar records, not an exact count. InsightsAssist detects query intent and routes accordingly: semantic questions go through vector similarity, counting and comparison questions go through parameterized SQL against your indexed dataset. Both paths enforce the same citation requirement. The result is a system that handles the full range of questions real users actually ask — not just the ones RAG demos are designed to look good on.

LIVE IN BETA

Real data. Real questions. Right now.

Two very different datasets. The same grounded AI.

VOTERHQ.COM

Congressional Stock Trade Tracker

VoterHQ tracks STOCK Act filings, congressional trading activity, and member financial disclosures in real time. InsightsAssist lets any voter ask natural-language questions across the full trading database — 'Which members traded pharmaceutical stocks before drug pricing votes?' — with every answer linked directly to the filing.

50K+Trade records indexed
535Members tracked
ITALIANAMERICANS.COM

Notable Italian Americans Database

A verified database of 250+ notable Italian Americans across entertainment, sports, science, business, law, and public service — spanning 1776 to today. InsightsAssist enables rich exploration: 'Who were the first Italian Americans in each field?' or 'Compare contributions by decade.' All citations trace to verified records.

250+Notable figures indexed
1776Earliest record
PRICING

Simple, transparent pricing. Lock in beta rates today.

All beta participants receive permanent rate lock — your price never increases as long as your account is active.

FREE
$0
always free
  • 1 dataset
  • Up to 1,000 records
  • 500 queries/month
  • Citation engine
  • Standard embed widget
  • Community support
Apply →
STARTER
$49
per month
  • 3 datasets
  • Up to 10,000 records
  • 5,000 queries/month
  • Citation engine
  • Custom widget appearance
  • Query analytics
  • Email support
Apply →
Most PopularPRO
$149
per month
  • Unlimited datasets
  • Up to 100,000 records
  • 25,000 queries/month
  • Citation engine
  • Full widget customization
  • Advanced analytics + coverage gaps
  • Google Datasets connector
  • Priority support
Apply →
ENTERPRISE
Custom
private deployment available
  • Unlimited everything
  • Private cloud deployment
  • Guaranteed SLA
  • Oracle/Salesforce/SAP connectors
  • AWS/Azure Marketplace billing
  • White-label option
  • Dedicated account manager
Contact Us →

Nonprofits and 501(c)(3) organizations receive the Free tier permanently — no application required, no time limit.

DISTRIBUTION

Coming to every major marketplace.

Install InsightsAssist from the platforms you already use. One-click integrations. No custom development needed.

🤖Anthropic Partner Marketplace✓ Available Now
🌊Webflow App MarketplaceQ2 2026
🖼️Framer PluginsQ2 2026
🛍️Shopify App StoreQ3 2026
📝WordPress Plugin DirectoryQ3 2026
🔶HubSpot App MarketplaceQ3 2026
☁️Oracle Cloud MarketplaceQ4 2026
Salesforce AppExchangeQ4 2026
🔵Microsoft Azure MarketplaceQ4 2026
🟠AWS MarketplaceQ4 2026
APPLY FOR BETA

Join the waitlist. Shape the product.

Beta is limited to 50 sites. Participants get permanent rate lock, direct access to the founding team, and their use case prioritized in the development roadmap.

No credit card required. Beta access is limited to 50 sites. We respond within 48 hours. Questions? [email protected]

FAQ

Common questions.

Most AI tools use your data as context but can still fall back on training knowledge when that data doesn't answer — that's where hallucination happens. InsightsAssist is architecturally constrained: if the answer isn't in your dataset, it responds "The dataset doesn't contain data to answer that" explicitly, every time. There is no fallback, no invented answer, because the architecture doesn't allow one.

Your data never trains any AI model. It is indexed in an isolated environment scoped to your account, retrieved only at query time, never shared with other customers or Anthropic. Enterprise customers can deploy entirely within their own AWS or Azure infrastructure with zero data leaving their environment.

Currently: CSV, JSON, Supabase/PostgreSQL direct connection, and Google BigQuery public datasets (Pro and Enterprise). Q3 2026 roadmap includes Airtable, Notion databases, Salesforce objects, and Snowflake.

One natural language question from a user equals one query, regardless of how many records were searched to formulate the answer. Follow-up questions in the same conversation count individually. Page loads and widget opens do not count.

Datasets under 10,000 records are fully indexed within 2–5 minutes of upload. The widget is available for testing immediately while indexing completes in the background.

Yes — white-label is available on Enterprise plans. Fully customize widget branding, remove all InsightsAssist attribution, deploy under your own domain, and manage multiple client datasets from one admin account.