Back to home

Compare

Comparing: LLMs Keep Hallucinating: RAG Becomes the Enterprise Standard Config & 大模型总爱瞎编:RAG 让 AI 先查资料再回答,这正成为企业落地标配

AEN
RAGVector DatabaseLLM Hallucination·

LLMs Keep Hallucinating: RAG Becomes the Enterprise Standard Config

Over 70% of enterprise AI projects are stuck on LLMs "hallucinating with a straight face." RAG (Retrieval-Augmented Generation: making LLMs check external data before answering) is becoming the standard operation to stop the bleeding. We note that having AI memorize all knowledge is no longer realistic; equipping it with a reference database it can consult at any time is a more reliable path to deployment.

What This Is

Ask an LLM directly, and it is limited by its training data cutoff date, ignorant of internal company rules, and prone to fabricating plausible but incorrect information just to answer—this is hallucination. The core logic of RAG is simple: an open-book exam. Before the AI generates an answer, it first retrieves relevant information from an external knowledge base, feeds the retrieved results and the question to the AI together, and makes it speak based on facts.

The complete RAG pipeline is divided into seven steps: Document processing (extracting text from PDFs, etc.) → Chunking (splitting long documents into small pieces) → Embedding (vectorization: turning text into coordinate points in mathematical space) → Retrieval → Reranking → Generation. In the retrieval phase, relying purely on semantic vector retrieval easily misses exact words like status codes or professional abbreviations; currently, the industry favors hybrid retrieval (vector retrieval plus keyword retrieval) as a safety net.

Industry View

The industry generally regards RAG as the most cost-effective solution for current LLM deployment. Compared to Fine-tuning (retraining the model with new data), RAG requires no expensive computing power; simply updating the external database allows the AI to master new knowledge, and the data always stays inside the enterprise.

But it is worth our concern that RAG is no panacea. Critics point out that once the retrieval phase recalls wrong data, the LLM will still output based on wrong facts, and because it is endorsed by "reference materials," the hallucinations are often more deceptive. Furthermore, how to chunk documents, choose vector models, and tune reranking algorithms—each step is an engineering pitfall. Although there are many vector databases on the market like Milvus, pulling the accuracy of the entire pipeline up to production level incurs an operational cost that is by no means low.

Impact on Regular People

For enterprise IT: The focus of infrastructure shifts from simply choosing an LLM to data cleaning and knowledge base construction. Garbage in, garbage out; the quality of internal documents directly determines the AI's intelligence.

For the individual workplace: The ability to organize and consolidate business materials is becoming valuable. Whoever can structure tacit knowledge will be able to train an AI assistant that better understands the business.

For the consumer market: The novelty of general-purpose chatbots is fading. Professional assistants tied to personal notes or enterprise knowledge bases will become the standard form of the next wave of AI products.

Source: juejin.cn
BZH
RAG向量数据库大模型幻觉·

大模型总爱瞎编:RAG 让 AI 先查资料再回答,这正成为企业落地标配

超过 70% 的企业 AI 项目卡在大模型“一本正经地胡说”,RAG(检索增强生成:让大模型先查外部资料再回答)正成为止血的标准操作。我们注意到,当下让 AI 直接背下所有知识已不现实,给它配一个能随时翻阅的资料库,才是更靠谱的落地路径。

这是什么

直接向大模型提问,它会受限于训练数据的截止日期,对公司内部规矩一无所知,且为了答题容易编造看似合理的错误信息,这就是幻觉。RAG 的核心思路很简单:开卷考试。在 AI 生成回答前,先从外部知识库检索相关信息,把检索结果和问题一起喂给 AI,让它基于事实说话。

完整的 RAG 链路分为七步:文档处理(提取 PDF 等文本)→ 切片(把长文档切小块)→ Embedding(向量化:把文本变成数学空间里的坐标点)→ 检索 → 重排 → 生成。其中检索环节,纯靠语义的向量检索容易漏掉状态码或专业缩写等精确词,目前行业更推崇混合检索(向量检索加关键词检索)来兜底。

行业怎么看

行业普遍将 RAG 视为目前大模型落地性价比最高的方案。相比微调(Fine-tuning:用新数据重新训练模型),RAG 无需高昂算力,只需更新外部数据库就能让 AI 掌握新知识,数据也始终留在企业内部。

但值得我们关心的是,RAG 并非万能药。反对声音指出,检索环节一旦召回错误资料,大模型依然会基于错的事实输出,且因为有了“参考资料”背书,输出的幻觉往往更具欺骗性。此外,文档怎么切、向量模型怎么选、重排算法怎么调,每一步都是工程坑,目前市面上 Milvus 等向量数据库虽多,但要把整个链路的准确率拉到生产级别,运维成本并不低。

对普通人的影响

对企业 IT:基建重心从单纯选大模型,转向数据清洗和知识库搭建,垃圾进只会垃圾出,内部文档的质量直接决定 AI 的智商。

对个人职场:整理和沉淀业务资料的能力正变得值钱,谁能把隐性知识结构化,谁就能调教出更懂业务的 AI 助手。

对消费市场:通用聊天机器人的新鲜感正在褪去,绑定个人笔记或企业知识库的专业助理,将成为下一波 AI 产品的标配形态。

Source: juejin.cn