(how to configure Yaml) ?
AutoRAG can use all LLMs supported by the LlamaIndex
vllm
moduleWe recommend using VLLM for fast inference!
We developed it parallelize, so you can experiment faster than using the llama_index_llm
module :)
Docs: vllm
Sample YAML
modules:
-module_type: vllm
llm: mistralai/Mistral-7B-Instruct-v0.2
temperature:[0.1, 1.0]
max_tokens: 512
llama_index_llm
moduleSample YAML
modules:
- module_type: llama_index_llm
llm: openailike
model: mistralai/Mistral-7B-Instruct-v0.2
api_base: your_api_base
api_key: your_api_key
ollama
, groq
)Ollama
ollama
with the code below in pythonimport autorag
from llama_index.llms.ollama import Ollama
autorag.generator_models["ollama"] = Ollama
nodes:
- node_line_name: node_line_1
nodes:
- node_type: generator
modules:
- module_type: llama_index_llm
llm: ollama
model: [llama3, qwen, mistral]
Additional parameters can be used directly in a YAML file