Query Mate provides intuitive controls for fine-tuning your RAG experience, combined with an unparalleled choice of Large Language Models (LLMs). This allows you to balance performance, creativity, and data privacy according to your specific needs.
These parameters allow you to control the behavior of the LLM and the RAG process:
Query Mate offers access to hundreds of LLM models, giving you the power to choose based on your data privacy concerns and performance needs. By default, private LLMs are preferred, ensuring your data never leaves your secure environment. However, you are in complete control:
The customer decides what is private and what is not, and how they trust big corporations. Here's an example of a private model available:
Version: | llama3.2:3b |
---|---|
Provider: | Ollama |
Host: | Ollama 87 nvidia (Private) Ollama fluxedge (Private) |
Capabilities: | Streaming |
Context size: | 128K tokens |
Number of parameters: | 3.21 billions parameters |
Quantization: | Q4_K_M |
This is just one example of the many private models you can deploy and manage with Query Mate.
This flexibility ensures that Query Mate can adapt to any enterprise security policy while providing powerful AI-driven insights.