Skip to main content
Open In ColabOpen on GitHub

Vectara

概览

Vectara 是值得信赖的 AI 助手和代理平台,专注于为企业关键任务应用程序提供支持。 Vectara 无服务器 RAG 即服务通过一个易于使用的 API 提供了 RAG 的所有组件,包括:

  1. 一种从文件(PDF、PPT、DOCX 等)中提取文本的方法
  2. 基于机器学习的分块技术,提供最先进的性能。
  3. 回旋镖 嵌入模型。
  4. 它自己的内部向量数据库,用于存储文本块和嵌入向量。
  5. 一种查询服务,可自动将查询编码为嵌入,并检索最相关的文本片段,支持混合搜索以及多种重排序选项,例如多语言相关性重排序器MMRUDF重排序器
  6. 一个基于检索到的文档(上下文)创建生成式摘要的大型语言模型,包含引用。

更多信息:

本笔记本展示了如何使用Vectara的聊天功能,该功能可自动存储对话历史记录,并确保后续问题会考虑该历史记录。

设置

要使用 VectaraVectorStore,首先需要安装配套包。

!uv pip install -U pip && uv pip install -qU langchain-vectara

入门指南

开始使用以下步骤:

  1. 如果您还没有账号,注册免费的Vectara试用。
  2. 在您的账户中,您可以创建一个或多个语料库。每个语料库代表一个存储从输入文档中提取的文本数据的区域。要创建语料库,请使用“创建语料库”按钮。然后,您需要为语料库提供名称和描述。您还可以选择定义过滤属性并应用一些高级选项。如果单击您创建的语料库,您可以在顶部直接看到其名称和语料库ID。
  3. 接下来,您需要创建API密钥以访问语料库。在语料库视图中点击“访问控制”选项卡,然后点击“创建API密钥”按钮。为您的密钥命名,并选择您希望密钥是仅用于查询还是查询+索引。点击“创建”,您现在已拥有一个有效的API密钥。请妥善保密此密钥。

要将LangChain与Vectara一起使用,您需要具备以下两个值:corpus_keyapi_key。 您可以通过两种方式向LangChain提供VECTARA_API_KEY

实例化

  1. 在您的环境中包含这两个变量: VECTARA_API_KEY

    例如,您可以使用 os.environ 和 getpass 设置这些变量,如下所示:

import os
import getpass

os.environ["VECTARA_API_KEY"] = getpass.getpass("Vectara API Key:")
  1. 将它们添加到 Vectara 向量存储构造函数中:
vectara = Vectara(
vectara_api_key=vectara_api_key
)

在本笔记本中,我们假设它们是在环境中提供的。

import os

os.environ["VECTARA_API_KEY"] = "<VECTARA_API_KEY>"
os.environ["VECTARA_CORPUS_KEY"] = "<VECTARA_CORPUS_KEY>"

from langchain_vectara import Vectara
from langchain_vectara.vectorstores import (
CorpusConfig,
GenerationConfig,
MmrReranker,
SearchConfig,
VectaraQueryConfig,
)

Vectara Chat 解释

在大多数使用LangChain创建聊天机器人的场景中,必须集成一个特殊的memory组件,该组件维护聊天会话的历史记录,然后利用该历史记录确保聊天机器人能够感知对话历史。

使用Vectara聊天功能——所有这些都在后台由Vectara自动完成。您可以查看聊天文档以了解更多细节,学习这一功能的内部实现方式。但通过LangChain,您只需在Vectara向量存储中启用该功能即可。

让我们来看一个例子。首先,我们加载国情咨文文档(记住,在Vectara平台上,文本提取和分块都会自动完成):

from langchain_community.document_loaders import TextLoader

loader = TextLoader("../document_loaders/example_data/state_of_the_union.txt")
documents = loader.load()

corpus_key = os.getenv("VECTARA_CORPUS_KEY")
vectara = Vectara.from_documents(documents, embedding=None, corpus_key=corpus_key)
API 参考:TextLoader

现在,我们使用 as_chat 方法创建一个聊天 Runnable:

generation_config = GenerationConfig(
max_used_search_results=7,
response_language="eng",
generation_preset_name="vectara-summary-ext-24-05-med-omni",
enable_factual_consistency_score=True,
)
search_config = SearchConfig(
corpora=[CorpusConfig(corpus_key=corpus_key, limit=25)],
reranker=MmrReranker(diversity_bias=0.2),
)

config = VectaraQueryConfig(
search=search_config,
generation=generation_config,
)


bot = vectara.as_chat(config)

调用

以下是一个没有聊天记录的情况下提问的示例

bot.invoke("What did the president say about Ketanji Brown Jackson?")["answer"]
'The president stated that nominating someone to serve on the United States Supreme Court is one of the most serious constitutional responsibilities. He nominated Circuit Court of Appeals Judge Ketanji Brown Jackson, describing her as one of the nation’s top legal minds who will continue Justice Breyer’s legacy of excellence and noting her experience as a former top litigator in private practice [1].'

以下是一个结合聊天历史提问的示例

bot.invoke("Did he mention who she suceeded?")["answer"]
'Yes, the president mentioned that Ketanji Brown Jackson succeeded Justice Breyer [1].'

使用流式传输进行聊天

当然,聊天机器人界面也支持流式传输。 您只需使用stream方法代替invoke方法:

output = {}
curr_key = None
for chunk in bot.stream("what did he said about the covid?"):
for key in chunk:
if key not in output:
output[key] = chunk[key]
else:
output[key] += chunk[key]
if key == "answer":
print(chunk[key], end="", flush=True)
curr_key = key
The president acknowledged the significant impact of COVID-19 on the nation, expressing understanding of the public's fatigue and frustration. He emphasized the need to view COVID-19 not as a partisan issue but as a serious disease, urging unity among Americans. The president highlighted the progress made, noting that severe cases have decreased significantly, and mentioned new CDC guidelines allowing most Americans to be mask-free. He also pointed out the efforts to vaccinate the nation and provide economic relief, and the ongoing commitment to vaccinate the world [2], [3], [5].

链式调用

对于额外的功能,您可以使用链式操作。

from langchain_core.output_parsers import StrOutputParser
from langchain_core.prompts import ChatPromptTemplate
from langchain_openai.chat_models import ChatOpenAI

llm = ChatOpenAI(temperature=0)

prompt = ChatPromptTemplate.from_messages(
[
(
"system",
"You are a helpful assistant that explains the stuff to a five year old. Vectara is providing the answer.",
),
("human", "{vectara_response}"),
]
)


def get_vectara_response(question: dict) -> str:
"""
Calls Vectara as_chat and returns the answer string. This encapsulates
the Vectara call.
"""
try:
response = bot.invoke(question["question"])
return response["answer"]
except Exception as e:
return "I'm sorry, I couldn't get an answer from Vectara."


# Create the chain
chain = get_vectara_response | prompt | llm | StrOutputParser()


# Invoke the chain
result = chain.invoke({"question": "what did he say about the covid?"})
print(result)
So, the president talked about how the COVID-19 sickness has affected a lot of people in the country. He said that it's important for everyone to work together to fight the sickness, no matter what political party they are in. The president also mentioned that they are working hard to give vaccines to people to help protect them from getting sick. They are also giving money and help to people who need it, like food, housing, and cheaper health insurance. The president also said that they are sending vaccines to many other countries to help people all around the world stay healthy.

API 参考

你可以查看聊天文档以获取详细信息。