RestAI is an AIaaS (AI as a Service) open-source platform. Built on top of LlamaIndex, Ollama and HF Pipelines. Supports any public LLM supported by LlamaIndex and any local LLM suported by Ollama. Precise embeddings usage and tuning.
This project is pretty great BUT we need more options to use different LLM's.You don't have to worry about creating a solution which supports 100+ LLM easily as LiteLLM is another foss project which is capable of doing this task for you.
Project LiteLLM link - https://github.com/BerriAI/litellm
You can study there project and see how it can be implemented in this project like you implemented support for ollama I believe similarly you can do it for LiteLLM which will be big win for the project as many will be easily able to use many more LLM easily which everyone wants and project will require 3 major parameters from user like base url,model name,api key that's all and with open ai api general structure it can query and give back result for the query.Many big projects have started adding support for this project in there project to make things advanced in easier way so study it and after that if you have any query you can ask them they are pretty responsive plus if u want to know more about my personal experience of using it with other great projects like flowise then I can tell you that too in detail.
A very good project.
I have a question, i see a docker-compose.yml located in the root directory. Is it means can directly launched it to run project without the need for other dependency operations.
Also, i think the document is worth improving, like ollama docs/ ๐
I see a lot of potential of this project. chroma and radis are added as database . it would be great if you kindly add at least one cloud database like pinecone. Thank you.
One of the growing and important discussions in the RAG Community now is Agent. It would be great to see Agent Implementation in this project and also some advanced Retrieval like Automerging Retrieval in llama index.
Could you please explain how this chat loading works in your code? Additionally, are you able to give the agent chat history when the agent is answering the next question?
The requirements.txt file is a text file used for Python projects, which lists all the required dependencies and their version numbers for the project.
llama_index-0.10.14 is wrong
from llama_index.embeddings.langchain import LangchainEmbedding
from llama_index.postprocessor.colbert_rerank import ColbertRerank
As a user of the project, I would like the ability to integrate my own API keys for inference and embedding services such as OpenAI, Cohere, and Anthropic, allowing for greater flexibility and customization in utilizing these services within the project.