概览
LlamaIndex 是一个用于 LLM 应用的数据框架,在构建 RAG(检索增强生成)系统方面功能尤为强大。LemonData 可以与 LlamaIndex 的 OpenAI 集成无缝协作。安装
复制
pip install llama-index llama-index-llms-openai llama-index-embeddings-openai
基础配置
复制
from llama_index.llms.openai import OpenAI
from llama_index.core import Settings
# Configure LLM
llm = OpenAI(
model="gpt-4o",
api_key="sk-your-lemondata-key",
api_base="https://api.lemondata.cc/v1"
)
# Set as default
Settings.llm = llm
# Simple query
response = llm.complete("What is LemonData?")
print(response.text)
使用不同模型
复制
# OpenAI GPT-4o
gpt4 = OpenAI(
model="gpt-4o",
api_key="sk-your-key",
api_base="https://api.lemondata.cc/v1"
)
# Anthropic Claude (via OpenAI-compatible endpoint)
claude = OpenAI(
model="claude-sonnet-4-5",
api_key="sk-your-key",
api_base="https://api.lemondata.cc/v1"
)
# Google Gemini
gemini = OpenAI(
model="gemini-2.5-flash",
api_key="sk-your-key",
api_base="https://api.lemondata.cc/v1"
)
对话接口
复制
from llama_index.core.llms import ChatMessage
messages = [
ChatMessage(role="system", content="You are a helpful assistant."),
ChatMessage(role="user", content="What is the capital of France?")
]
response = llm.chat(messages)
print(response.message.content)
流式传输
复制
# Streaming completion
for chunk in llm.stream_complete("Write a poem about AI"):
print(chunk.delta, end="", flush=True)
# Streaming chat
for chunk in llm.stream_chat(messages):
print(chunk.delta, end="", flush=True)
Embeddings
复制
from llama_index.embeddings.openai import OpenAIEmbedding
embed_model = OpenAIEmbedding(
model="text-embedding-3-small",
api_key="sk-your-lemondata-key",
api_base="https://api.lemondata.cc/v1"
)
# Set as default
Settings.embed_model = embed_model
# Get embeddings
embeddings = embed_model.get_text_embedding("Hello, world!")
print(f"Embedding dimension: {len(embeddings)}")
使用文档构建 RAG
复制
from llama_index.core import VectorStoreIndex, SimpleDirectoryReader
# Configure settings
Settings.llm = llm
Settings.embed_model = embed_model
# Load documents
documents = SimpleDirectoryReader("./data").load_data()
# Create index
index = VectorStoreIndex.from_documents(documents)
# Query
query_engine = index.as_query_engine()
response = query_engine.query("What is in my documents?")
print(response)
对话引擎
复制
# Create chat engine with memory
chat_engine = index.as_chat_engine(chat_mode="condense_question")
# Multi-turn conversation
response = chat_engine.chat("What is LemonData?")
print(response)
response = chat_engine.chat("How many models does it support?")
print(response)
异步用法
复制
import asyncio
async def main():
response = await llm.acomplete("Hello!")
print(response.text)
asyncio.run(main())
环境变量
为了使代码更整洁,请使用环境变量:复制
export OPENAI_API_KEY="sk-your-lemondata-key"
export OPENAI_API_BASE="https://api.lemondata.cc/v1"
复制
from llama_index.llms.openai import OpenAI
# Will automatically use environment variables
llm = OpenAI(model="gpt-4o")
最佳实践
选择合适的模型
选择合适的模型
对于嵌入和摘要任务,请使用速度更快的模型 (GPT-4o-mini);对于最终回答,请保留性能更强大的模型 (GPT-4o, Claude)。
优化分块大小
优化分块大小
根据您的文档类型调整分块大小。对于密集的技术文档,使用较小的分块;对于叙述性内容,使用较大的分块。
使用缓存
使用缓存
启用 LlamaIndex 缓存,以避免在开发过程中产生冗余的 API 调用。