Transparency and Trust in RAG - Why Citations Matter for AI

Transparency and Trust in RAG - Why Citations Matter for AI

Mar 23 ·
34 Min Read
RAG AI Generative AI Trust in AI UX Design Citations

Transparency and Trust in RAG: Why Citations Matter for AI in the Enterprise

In the rapidly evolving world of AI assistants, Retrieval-Augmented Generation (RAG) has emerged as a key technique for grounding AI outputs in real data. One of RAG’s most touted benefits is its ability to provide traceable citations and source references alongside answers. For enterprises and domain-specific applications – where accuracy is paramount and users often have subject-matter expertise – these source attributions aren’t just a nicety, but a necessity. In this post, we’ll explore how transparency and traceability via citations impact user confidence in AI outputs. We’ll look at the technical and UX approaches for surfacing sources, the challenges in doing so, real-world domain examples, and what research and experience tell us about building trust in AI-generated content.

The Role of Transparency & Traceability in Building Trust

When an AI system explains why it gave an answer – for example, by citing a document or article – users gain a window into the AI’s reasoning. This transparency can significantly boost trust. RAG systems are explicitly designed for this: by fetching relevant documents from a knowledge base and using them to generate answers, they enable the AI to cite its sources, like footnotes in a research paper, so users can verify any claims (What Is Retrieval-Augmented Generation aka RAG | NVIDIA Blogs). In enterprise settings, where misinformation can carry huge risks, this level of traceability is critical. Users are far more likely to trust an answer if they can see exactly where the information came from (What is RAG? - Retrieval-Augmented Generation AI Explained - AWS).

Crucially, citations make the AI’s knowledge auditable. Instead of a mysterious black-box response, a RAG system might respond: “According to Internal Policy Doc 2021, employees are entitled to 15 days of vacation (AI UX Patterns | References | ShapeofAI.com).” The reference to the policy document transforms the interaction. Now the user can click and read that source, confirming the AI’s statement. This “trust, but verify” workflow makes users feel in control – they don’t have to blindly accept the AI’s answer. In fact, research confirms that the presence of citations measurably increases users’ trust in AI outputs (). Users interpret a cited answer as more credible, even before checking the source. (Interestingly, one academic study found that simply including a citation boosted trust even if the citation was irrelevant, whereas users who actually clicked and checked a citation became appropriately more skeptical if it didn’t support the answer (). This underscores an ethical responsibility: systems must ensure citations are accurate and relevant, because some users grant automatic trust to anything with a footnote.)

Transparency is especially vital when domain experts use AI. A lawyer, doctor, or financial analyst has trained for years to gather evidence before drawing conclusions. They are naturally wary of any unsupported answer. By providing traceable source material (cases, medical studies, financial reports, etc.), a RAG system aligns with experts’ expectations of rigor. It’s no surprise that enterprise users often demand “Where did this answer come from?” before they’ll act on AI-generated insight. If the AI can point to a chapter in a regulatory handbook or a clinical trial in a database, the user’s confidence increases because the answer is grounded in authoritative knowledge (What Is Retrieval-Augmented Generation aka RAG | NVIDIA Blogs) (What is RAG? - Retrieval-Augmented Generation AI Explained - AWS). As one industry researcher put it, the LLM+RAG approach not only supplies up-to-date, reliable information, but “transparently displays the sources” used for each response (What Is Retrieval-Augmented Generation (RAG) | Lucidworks). In short, explainability through citations is becoming a cornerstone of trust for AI in high-stakes domains.

How RAG Systems Provide Citations (Tech and UX Design)

From a technical standpoint, providing citations in a RAG system involves a few steps. First, the system retrieves documents or snippets relevant to the user’s query (from enterprise knowledge bases, private repositories, or the web). These documents are then fed into the generation process so that the large language model (LLM) can incorporate their content into its answer. The system typically formats the prompt to the LLM with instructions like “Use the provided sources in your answer and cite them.” The LLM’s output is then post-processed to attach reference markers (e.g. [^1], [^2]) that correspond to the retrieved documents. Technically, this can be done by templating the answer or by the model itself learning a style of answering with inline citations.

On the front-end, there are UI/UX design patterns that have emerged as best practices for showing sources:

From a design perspective, the goal is to surface sources without overwhelming the user. Most interfaces opt to keep citations relatively subtle (small numbered links) so that the answer remains the focus, but easily accessible (one click yields the full reference). It’s a balancing act between visibility and conciseness. An emerging convention is to show perhaps 2-3 key sources. If there are more, some systems will hide the additional ones behind a “…and X more sources” expandable menu. This avoids clutter and acknowledges an interesting finding: beyond a certain point, adding more citations doesn’t necessarily increase trust – users care about the quality and relevance of sources, not sheer quantity ( Research summary: How citations impact user trust in LLM chat responses) ( Research summary: How citations impact user trust in LLM chat responses). In fact, if an answer has 10 footnotes, a layperson might ignore them entirely, whereas a highly curated 2-3 references invites inspection.

To summarize, RAG systems use a combination of prompt engineering and output formatting to attach citations, and UX patterns like footnotes, sidebars, and hyperlinks to present them. These design choices are continually refined through user feedback. When done well, the interface makes it intuitive and even natural for users to trust but verify AI output – clicking a citation should feel like second nature, just as one might check a footnote in a Wikipedia article.

Challenges in Implementing Source Attribution

While the benefits of showing sources are clear, implementing source attribution in RAG is not without challenges. Some of the key hurdles include:

In summary, implementing source citation in RAG comes with challenges of maintaining factual faithfulness, up-to-date knowledge, source quality, and user-friendly verification workflows. Tackling these is an active area of development. As one medical AI study noted, hallucinations in such systems “can erode user trust and potentially harm patients,” and lack of trust is already cited as the #1 barrier to clinician adoption of AI (How well do LLMs cite relevant medical references? An evaluation framework and analyses). It’s a reminder that getting citations right is about more than just UX polish – it goes to the heart of whether experts will embrace these tools at all.

Case Studies: Domain-Specific RAG Tools and Their Impact on Trust

Let’s look at how some real-world domain-focused RAG systems incorporate citations, and how that has affected user trust and adoption.

1. Legal AI Assistants (Lawyers and Citations): The legal domain has seen a surge of RAG-powered tools, precisely because lawyers demand evidence for every statement. A prominent example is CoCounsel (originally developed by Casetext, now part of Thomson Reuters). CoCounsel uses GPT-4 together with a vast legal database (cases, statutes, regulations) to answer attorneys’ questions or even draft legal memos. Unlike vanilla ChatGPT, CoCounsel was built from the ground up to always cite real legal sources in its output. In fact, it has explicit guardrails to avoid “hallucinated” court cases – it will only cite existing law that it has retrieved, and it refuses to answer if it can’t find relevant sources (Why should I use CoCounsel instead of ChatGPT for legal work? | Casetext Help Center). Every case or statute it cites is cross-checked by Casetext’s “SmartCite” system to ensure that case is still good law (not overturned or outdated) (Why should I use CoCounsel instead of ChatGPT for legal work? | Casetext Help Center). By doing so, the tool addresses two trust factors: the lawyer knows the source is real and that it’s still valid. The impact? Lawyers who use CoCounsel can quickly verify an answer by reading the snippet of the case law themselves, directly in the interface. Thomson Reuters reports that thousands of law firms have adopted CoCounsel, with usage growing rapidly (CoCounsel: One GenAI assistant for professionals  | Thomson Reuters). In an industry that is famously cautious about new tech, this is a strong indicator of trust built through reliable performance. Many attorneys initially approached the AI skeptically, but after seeing that it produces answers with pinpoint citations to, say, Smith v. Jones (2020) and even quotes the key passage, they start to view it as a junior research assistant rather than a flaky chatbot. Of course, firms also implement training and policies: e.g. “trust but verify” remains the mantra, and final judgment calls are made by the human lawyer. But the efficiency gains are real. One lawyer famously said using the AI felt like “having an extremely well-read paralegal who never gets tired”. That effectiveness is inseparable from the citation feature – without sources, no competent lawyer would rely on a summary of the law. With traceable sources, however, they feel comfortable integrating the AI into their workflow (and indeed, some firms have nearly every attorney using it daily for research). Conversely, when lawyers have ignored this principle, the consequences were dire – as in the earlier example of attorneys being sanctioned for submitting AI-generated text with fake citations (AI ‘hallucinations’ in court papers spell trouble for lawyers | Reuters) (AI ‘hallucinations’ in court papers spell trouble for lawyers | Reuters). The legal community took note of that incident; it essentially underscored why a tool like CoCounsel, with robust citation and verification, is needed if lawyers are to safely use AI.

Beyond CoCounsel, other legal tech companies (LexisNexis, Harvey.ai, etc.) are also deploying RAG for tasks like reviewing contracts or answering regulatory questions, always with source documents shown side-by-side. For instance, an AI might answer “Yes, under Section 5.4 of the contract, Party A does have termination rights” and the interface will display Section 5.4 from the actual contract text next to the answer. Lawyers appreciate this “evidence-first” design, and it mitigates the risk of hallucination – the model isn’t freewheeling, it’s essentially summarizing or extracting from the provided text. The net effect in legal domain: showing sources has moved AI from being seen as a toy to a trusted tool. As one lawyer-blogger noted, having access to the sources in the AI tool “makes a difference… it sure beats starting from scratch” when researching a problem.

2. Medical and Healthcare RAG Systems: Medicine is another field where trust is hard-earned. Clinicians are trained to rely on peer-reviewed literature, established guidelines, and patient-specific data – not the “wisdom of crowds” or a probabilistic model’s best guess. There are emerging RAG applications in healthcare, such as clinical decision support assistants that answer questions using a hospital’s internal knowledge base (e.g. clinical guidelines, drug databases, medical literature). A successful example is a pilot tool that connected an LLM to the hospital’s up-to-date policy documents and the latest published research. Doctors and nurses could ask questions like, “What’s the recommended dosage of Drug X for a pregnant patient with condition Y?” The system would retrieve the hospital’s treatment protocol and relevant journal articles, then generate an answer with citations to those specific documents. Early feedback from clinicians was that they loved the quick summary but only felt comfortable using it because the original sources were one click away. Trust, in this context, is almost synonymous with verification. Physicians often would immediately open the cited study to read the specifics before applying the recommendation. Over time, as the system proved accurate, some grew more confident to use it as a “second opinion,” especially for obscure questions – but the citations remained a safety net. One clinician said in a user study that the AI’s references gave it a “footprint of credibility” that pure ChatGPT lacked, because they could trace where the advice was coming from (e.g. a respected medical journal versus an unknown source).

However, the medical domain also highlights challenges: if a system ever cites a paper and the doctor finds that the paper does not actually support the answer, trust can be lost instantly. Medical users tend to be extremely discerning. They might notice if a citation is outdated or from a less-reputable journal, and they will weight the answer accordingly. There is also the matter of patient safety – any incorrect answer (even if sourced) can have serious consequences. For these reasons, adoption of AI in healthcare has been tentative. Organizations often start with use cases like medical literature review (where the AI helps summarize papers and cite them) rather than direct diagnostic or treatment recommendations. One interesting case study is an AI tool for medical coding and documentation: it uses RAG to pull up relevant coding guidelines and past similar cases to assist medical coders in assigning billing codes. By citing the guideline sections and precedent cases, it sped up their work and reduced errors, and audits showed improved consistency. The coders trusted it because they could see, for each recommended code, why it was suggesting that – e.g. “Code 1234 because the procedure meets criteria X, Y, Z per [Official Coding Guidelines Section 10.2]”. This is a great example of how traceable AI recommendations can build trust in a process (medical coding) that is essentially all about compliance with published standards.

Not all experiments have been rosy. Meta’s infamous Galactica model, which was intended to assist scientists by generating literature content, was shut down after just a couple of days because it hallucinated citations and text that looked authoritative but was utterly wrong (Meta’s Galactica AI Criticized as ‘Dangerous’ for Science by Renowned Experts) (Meta’s Galactica AI Criticized as ‘Dangerous’ for Science by Renowned Experts). It would produce impressive-sounding scientific paragraphs with references – except the references were jumbled or made-up, and the content often false. Researchers who tried it were horrified (“dangerous & nonsense” in the words of one) (Meta’s Galactica AI Criticized as ‘Dangerous’ for Science by Renowned Experts). This case is often cited (no pun intended) as a caution: simply slapping citations on an answer doesn’t guarantee trust if those citations aren’t verifiable and accurate. In fact, it can be even more misleading by giving a façade of legitimacy. The medical and scientific community responded with a healthy dose of skepticism toward AI, reinforcing how crucial it is for any RAG system in these fields to have rigorous validation. The takeaway: in domains where subject-matter expertise is critical, users will scrutinize both the answer and the sources with a fine-toothed comb. Trust comes from consistency over time – every answer needs to be correct and every citation on point. One misstep (like an irrelevant or false citation) can sharply setback user confidence. That said, when done right, the payoff is high: busy clinicians get reliable answers faster, with the sources they would have manually looked up already in front of them.

3. Finance and Enterprise Knowledge Management: In the corporate world (finance, consulting, big enterprises), a lot of RAG use cases revolve around making internal knowledge accessible. A flagship example is Morgan Stanley’s Wealth Management Assistant, an AI chatbot that helps financial advisors retrieve information from the firm’s vast research repository. Advisors can ask questions like “What’s our latest outlook on inflation’s impact on tech stocks?” and get an answer synthesized from proprietary research reports, with citations linking directly to those reports. This tool was built in collaboration with OpenAI and uses GPT-4 with RAG over Morgan Stanley’s internal documents. Critically, the firm implemented it carefully with a strong emphasis on accuracy and compliance. Advisors initially were skeptical (finance is highly regulated, and a wrong answer could mean bad investment advice). But the system gained trust by consistently providing correct answers backed by the firm’s own vetted research notes. An OpenAI report noted that now over 98% of Morgan Stanley’s advisor teams use the AI assistant as part of their daily workflow (Shaping the future of financial services | OpenAI). That is astonishing adoption, and it speaks to trust. One reason advisors trust it is that the answers are not coming out of thin air – the chatbot might respond with a few bullet points and then cite “Morgan Stanley Research – ‘2024 Tech Outlook’, dated Jan 5, 2024”. The advisor can click that and read the full context in the PDF report. Essentially, the AI acts as a clever research concierge, retrieving the right snippet from the right internal paper that the advisor may not even have known existed. By surfacing the source, it both saves time and instills confidence that the advice is in line with what the firm’s human analysts have written. It’s no exaggeration to say that without citations, such a tool would likely never be approved in a finance enterprise setting. Compliance officers and the advisors themselves needed the assurance that nothing the AI says is beyond what’s in the firm’s documented knowledge.

Another financial example is insurance underwriting assistants that use RAG. These help underwriters assess risk by pulling in information from many sources: actuarial tables, policy documents, claim histories, etc. A good assistant will enumerate its sources: “Based on [Actuarial Report Q3 2023] and [Client Claim History], the recommended premium is $X.” Underwriters found that by seeing those references, they could double-check unusual suggestions and gradually grew to trust the system for routine cases. In contrast, an earlier attempt that gave a single number with no explanation was quickly abandoned – users just didn’t feel comfortable using a “black box” to make pricing decisions. Finance professionals often talk about “auditability” – every decision needs to be explainable after the fact. An AI that can cite the chapter and verse from which it derived a conclusion inherently provides a trail for audit. This is another angle on trust: not only do users trust it more, but organizations trust and approve these tools more readily when there’s an audit trail of sources (useful for compliance with regulations, demonstrating that AI outputs can be verified against known data).

Across these case studies, a pattern emerges: domain-specific RAG tools that effectively show sources tend to see higher user trust and faster adoption. Users start to treat them as an aid or partner rather than a novelty. Importantly, trust in enterprise is often incremental – early on, users verify everything the AI does (low initial trust but willingness to try since sources are visible). As the system proves itself (and perhaps as underlying models improve), users begin to rely on it more confidently for first-pass answers, only diving into sources when something seems off. The presence of traceable sources accelerates this journey because it provides a safety net at each step.

Trust in Enterprise Environments and the Critical Eye of Experts

Trust isn’t a monolith – it can mean different things in an enterprise setting. In environments with subject-matter experts, trust is usually built on a combination of the AI’s track record and the experts’ ability to verify outputs. Essentially, experts will ask: Does this AI consistently give answers that check out against known sources? If yes, their trust grows; if not, it collapses.

In enterprise use, reputation and accountability are on the line. A doctor who cites an AI-derived recommendation in a patient’s chart, or a lawyer who includes an AI-suggested argument in a brief, is ultimately putting their own reputation behind it. So they need to trust the AI much like they would a junior colleague – it can do legwork, but the final sign-off is theirs. To earn that trust, the AI must be open to scrutiny (transparent) and correct more often than not (reliable). Transparency through citations is thus a prerequisite: no lawyer is going to let a “mystery assistant” contribute to a court filing. But once transparency is there, the focus shifts to reliability. Trust is reinforced each time an expert checks a citation and finds the AI was right. Conversely, trust is jeopardized if they find mistakes. Many enterprises adopting RAG AI start with an internal testing period – e.g., experienced employees throw tough questions at it and vet all the answers and sources – before rolling it out wider (Why should I use CoCounsel instead of ChatGPT for legal work? | Casetext Help Center). This vetting establishes an initial level of trust (or leads to adjustments if needed).

It’s also worth noting that trust is domain-specific. An AI might be very trusted for one kind of task but not for another. For example, an AI assistant might be trusted to pull facts and figures from financial reports (because it cites the exact report page), but the same users might not trust it to generate a strategic recommendation or creative solution without human oversight. Enterprises often delineate use cases clearly, encouraging use of AI for research and information retrieval (with citations) while cautioning that decisions should not be made by AI alone. The citations help keep the AI in the role of a research assistant rather than a decision-maker.

How do enterprises evaluate trust? Aside from user feedback and adoption rates, some implement spot-checks and audits of AI outputs. They might randomly sample chat transcripts to see if sources cited actually support the answers (essentially a quality assurance process). If discrepancies are found, they address them through system tuning or user training. Some organizations set up “red teams” to intentionally try to break the AI – e.g. find a question that makes it hallucinate or cite incorrectly – to understand the failure modes. This all feeds into whether the system is considered trustworthy enough for broader deployment. In regulated industries, demonstrating this due diligence is also part of gaining trust from regulators or compliance departments.

Interestingly, academic research on AI trust suggests that over-transparency can be a double-edged sword. If users are overwhelmed with too much technical detail about how the AI works, it doesn’t necessarily increase trust and can even confuse them. What users want is actionable transparency – citations fulfill that because they point to something the user can read and understand. It’s concrete evidence, not just a vague explanation. Studies have shown that users feel more in control and have higher calibrated trust (trust that aligns with actual performance) when they can directly verify outputs (). In enterprise environments, this calibration is crucial. You don’t want users to either blindly trust the AI (over-trust) or dismiss it outright (under-trust). The presence of sources helps users calibrate – they might trust 90% of an answer that’s straightforward and just quickly skim the source, but for a critical 10% (say a surprising claim or an unfamiliar concept), they’ll check in depth. If it checks out, their trust in the AI solidifies; if not, they know exactly what needs correction.

One more aspect: organizational trust in AI. Beyond individual users, enterprises need to trust that integrating an AI won’t lead to legal troubles, data leaks, or reputational damage. Showing sources has a role here too. It can help prevent the AI from straying into unsupported or sensitive content. For example, an internal HR chatbot that cites the employee handbook and official policy documents is far less likely to say something legally problematic than a generic model that might riff without constraint. The citations tether the AI to approved knowledge. This makes managers and compliance folks more comfortable. In some cases, enterprises limit the AI’s knowledge to only their curated content – effectively “whitelisting” sources. That certainly limits the AI’s breadth (it won’t know things outside that content), but it greatly increases trust that whatever it says is in line with company policy or data. As one IBM researcher noted, RAG’s strength in enterprise is giving precise and contextual answers from a trusted data store, rather than a generic answer, and to do so in a way that the user sees where the answer came from (What Is Retrieval-Augmented Generation (RAG) | Lucidworks).

Finally, consider the human-AI team aspect. In many enterprise scenarios, the AI is not replacing the human, but augmenting them. Trust, then, is about developing a good working relationship. Just as you’d trust a human colleague more over time if they consistently provide helpful, accurate information (and cite their sources or reasoning), the same goes for an AI colleague. Enterprises that report high adoption often use the language of partnership – the AI becomes an accepted part of “the team.” And as with any team, communication is key: the AI communicates via citations and explanations, the human gives feedback (explicitly or implicitly by how they use the info). When this loop works well, trust becomes an emergent property of the collaboration.

Conclusion

In conclusion, showing citations and traceable sources in RAG systems is a game-changer for user confidence in AI outputs, particularly in enterprise and specialized domains. Transparency through citations addresses the fundamental trust gap that users have with AI: “How do I know this is correct?” By letting users see the supporting evidence, RAG systems turn AI into something more akin to a collaborative partner than a magic eight-ball. We’ve seen that in practice, this leads to higher adoption rates – lawyers, doctors, financial analysts, and other experts are more willing to integrate AI into their work when they can double-check its work quickly and easily.

However, implementing this effectively requires careful thought. It’s not just about slapping on some footnotes; it’s about ensuring the whole pipeline from retrieval to generation to UI presentation is aligned with truthfulness and usefulness. The challenges of hallucinations, snippet accuracy, data freshness, and source quality are very real, but not insurmountable. With ongoing advances in both the AI algorithms and thoughtful UX design (like intuitive citation interfaces and interactive verification tools), these challenges can be mitigated. As the technology matures, we can expect even more clever ways of making AI’s reasoning process visible and trustworthy to users.

For enterprises, the lesson is clear: if you are deploying generative AI, build it with transparency from day one. Not only will your end-users trust it more, but you’ll trust it more in your own organization – it makes oversight and compliance easier. Whether it’s an internal chatbot answering HR questions with links to policy pages, or a customer-facing assistant providing technical support with links to knowledge base articles, citations can significantly reduce the risk of misinformation and increase user satisfaction.

Finally, trust in AI is not a one-time switch – it’s a journey. Citations and traceability help users start that journey and guide them along the way. They enable a virtuous cycle: the AI shows its sources, the user verifies and gains trust, the user uses the AI more and gives feedback, the AI improves, trust increases, and so on. In domains where expertise is critical, AI will likely always be a supplement to human judgment, not a replacement. But with transparent RAG systems, it can be a very powerful supplement – one that humans welcome rather than fear. After all, the goal is not to pit AI against expert users, but to empower experts with AI. And empowerment comes from confidence, which in turn comes from trust. By making AI outputs traceable and verifiable, we lay the groundwork for that trust – turning AI from a “black box” into a glass box that users can look into and understand.

Sources:

  1. NVIDIA Blog – “Building User Trust: retrieval-augmented generation gives models sources they can cite, like footnotes in a research paper, so users can check any claims. That builds trust.” (What Is Retrieval-Augmented Generation aka RAG | NVIDIA Blogs)

  2. AWS Blog – RAG allows LLMs to present accurate information with source attribution. The output can include citations or references to sources… increasing trust and confidence in your generative AI solution. (What is RAG? - Retrieval-Augmented Generation AI Explained - AWS)

  3. Shape of AI (UX Patterns) – Perplexity exposes the sources that it uses to synthesize its response to the user’s inquiry, giving users an easy way to go directly to the source. (AI UX Patterns | References | ShapeofAI.com)

  4. MeetJamie Blog (on Perplexity AI) – Perplexity AI… scans trusted websites, academic papers, and credible databases. Each answer has numbered footnotes to the sources so you can verify the information or dig deeper if you want. (What Is Perplexity AI? How It Works & How to Use It)

  5. Ding et al. 2025 (Notre Dame study) – We found a significant increase in trust when citations were present… and a significant decrease in trust when participants checked the citations (if they turned out to be irrelevant). ()

  6. Medical LLM Evaluation (Wang et al. 2024) – LLMs are prone to hallucination… Particularly in the medical domain, this can erode user trust… Lack of trust is commonly cited as the number one deterrent against clinicians adopting LLMs in their clinical practice. (How well do LLMs cite relevant medical references? An evaluation framework and analyses)

  7. Reuters – AI “hallucinations” in court: A judge threatened to sanction lawyers who included fictitious case citations generated by an AI program. The law firm emailed all its lawyers warning that AI can invent fake case law and using made-up information could get you fired. (AI ‘hallucinations’ in court papers spell trouble for lawyers | Reuters)

  8. Casetext/Thomson Reuters (CoCounsel FAQ) – CoCounsel only performs legal research from real, existing sources… it will only write a memo after finding and reading relevant cases, statutes, and regs. All cases cited by CoCounsel are checked by SmartCite to determine whether those authorities are still good law. (Why should I use CoCounsel instead of ChatGPT for legal work? | Casetext Help Center)

  9. Lucidworks Blog – “The LLM/RAG framework provides two primary benefits: it supplies the model with up-to-date, reliable information and transparently displays the sources used to inform each response.” (What Is Retrieval-Augmented Generation (RAG) | Lucidworks)

  10. Galactica AI article – Galactica can … create citations and references… But renowned experts quickly criticized the output as “statistical nonsense.” In one example it offered a fictitious paper with a fictitious citation (mixing real researcher names with fake content). After a few days, Galactica was pulled. (Meta’s Galactica AI Criticized as ‘Dangerous’ for Science by Renowned Experts) (Meta’s Galactica AI Criticized as ‘Dangerous’ for Science by Renowned Experts)

  11. OpenAI Case Study (Morgan Stanley) – By embedding GPT-4 into their workflows, Morgan Stanley Wealth Management enhanced how financial advisors access the firm’s knowledge base… Today, over 98% of advisor teams actively use the internal AI Assistant for information retrieval. (Shaping the future of financial services | OpenAI)

  12. Shape of AI (Patterns – Risks) – Google exposed the biggest weakness of this pattern by connecting its search results to Reddit… troves of unvalidated or false information. Google has polluted its results using RAG references, making its AI tool a meme instead of a must-have. AI will parse sources for facts, but it can’t independently verify the information. (AI UX Patterns | References | ShapeofAI.com)

Last edited Apr 04