Giter Club home page Giter Club logo

aios's Introduction

AIOS: LLM Agent Operating System

Code License

AIOS, a Large Language Model (LLM) Agent operating system, embeds large language model into Operating Systems (OS) as the brain of the OS, enabling an operating system "with soul" -- an important step towards AGI. AIOS is designed to optimize resource allocation, facilitate context switch across agents, enable concurrent execution of agents, provide tool service for agents, maintain access control for agents, and provide a rich set of toolkits for LLM Agent developers.

๐Ÿ  1. Architecture of AIOS

๐Ÿ“ฐ 2. News

  • [2024-05-01] ๐Ÿ› ๏ธ The agent creation in AIOS is refactored, which can be found in our OpenAGI package.
  • [2024-04-29] ๐Ÿ“Š The evaluation mode of AIOS is added, which supports customizable agent types and agent instance numbers in each agent type.
  • [2024-04-14] ๐Ÿš€ AIOS currently supports generation interrupt (for open-sourced llms from huggingface) and customized console loggers.
  • [2024-04-05] ๐Ÿ› ๏ธ AIOS codebase has been updated to add shell simulator, rapid API calls, and pre-commit test cases. Please see CONTRIBUTE for how to test your contributions and create pull requests.
  • [2024-04-02] ๐Ÿค AIOS Discord Community is up. Welcome to join the community for discussions, brainstorming, development, or just random chats!
  • [2024-03-25] โœˆ๏ธ Our paper AIOS: LLM Agent Operating System is released and AIOS repository is officially launched!
  • [2023-12-06] ๐Ÿ“‹ After several months of working, our perspective paper LLM as OS, Agents as Apps: Envisioning AIOS, Agents and the AIOS-Agent Ecosystem is officially released.

โœˆ๏ธ 3. Getting Started

3.1 Installation

To run AIOS, you will need to install our agent creation package, OpenAGI.

Git clone AIOS and OpenAGI

git clone https://github.com/agiresearch/AIOS.git
git clone https://github.com/agiresearch/OpenAGI.git

Make sure you have Python = 3.11 Install the required packages using pip

conda create -n AIOS python=3.11
source activate AIOS
cd AIOS
pip install -r requirements.txt

Allow your code to be able to see 'openagi'

cd ../OpenAGI
pip install -e .

3.2 Usage

If you use open-sourced models from huggingface, you need to setup your Hugging Face token and cache directory

export HUGGING_FACE_HUB_TOKEN=<YOUR READ TOKEN>
export HF_HOME=<YOUR CACHE DIRECTORY>

If you use LLM APIs, you need to setup your API key such as OpenAI API Key, Gemini API Key

export OPENAI_API_KEY=<YOUR OPENAI API KEY>
export GEMINI_API_KEY=<YOUR GEMINI API KEY>

If you use external tool APIs in your agents, you need to set up your API keys as follows:

export RAPID_API_KEY=""
export WOLFRAM_ALPHA_APPID=""

You can also create .env file from the .env.example file, and then use dotenv to load the environment variables using .env file into your application's environment at runtime.

cp .env.example .env

(1) Demonstration Mode

In the demonstration mode, we provide a toy example: we hardcode three agents and allow you to change the parameters. Then you can see the output of each step in running multiple agents For open-sourced LLMs, you need to setup the name of the LLM you would like to use the max gpu memory, the evaluation device and the maximum length of generated new tokens.

# For open-sourced LLMs
python main.py --llm_name <llm_name> --max_gpu_memory <max_gpu_memory> --eval_device <eval_device> --max_new_tokens <max_new_tokens>
## Use google/gemma-1.1-2b-it for example
python main.py --llm_name google/gemma-1.1-2b-it --max_gpu_memory '{"0": "24GB"}' --eval_device "cuda:0" --max_new_tokens 256

For close-sourced LLMs, you just need to setup the name of the LLM.

# For close-sourced LLMs
python main.py --llm_name <llm_name>
## Use gpt-4 for example
python main.py --llm_name gpt-4

You can use bash script to start the agent execution like this

bash scripts/run/gpt4.sh

You can use an open-source model on an Apple MacBook with Ollama. First, you will need to pull the model. Let's use llama3 as an example:

ollama pull llama3

Then, you can run the Python script with the input parameter to start using AIOS with Llama3 and Ollama on your MacBook:

python main.py --llm_name ollama/llama3

(2) Interactive Mode

In the deployment mode, the outputs of running agents are stored in files. And in this mode, you are provided with multiple commands to run agents and see resource usage of agents (e.g., run <xxxAgent>: <YOUR TASK>, print agent). Different from the interactive mode, you need to set all the default loggers as file loggers.

# For open-sourced LLMs
python simulator.py --llm_name <llm_name> --max_gpu_memory <max_gpu_memory> --eval_device <eval_device> --max_new_tokens <max_new_tokens> --scheduler_log_mode file --agent_log_mode file --llm_kernel_log_mode file
## Use google/gemma-1.1-2b-it for example
python simulator.py --llm_name google/gemma-1.1-2b-it --max_gpu_memory '{"0": "24GB"}' --eval_device "cuda:0" --max_new_tokens 256 --scheduler_log_mode file --agent_log_mode file --llm_kernel_log_mode file
# For close-sourced LLMs
python simulator.py --llm_name <llm_name> --scheduler_log_mode file --agent_log_mode file --llm_kernel_log_mode file
## Use gpt-4 for example
python simulator.py --llm_name gpt-4 --scheduler_log_mode file --agent_log_mode file --llm_kernel_log_mode file

You can use bash script to start the interactive simulation session like this

bash scripts/interactive/gpt4.sh

Instance of available commands

run NarrativeAgent: Craft a tale about a valiant warrior on a quest to uncover priceless treasures hidden within a mystical island.
print agent

(3) Evaluation Mode

In the evaluation mode, we allow you to configure different types of predefined agents (MathAgent, NarrativeAgent, RecAgent) with a configurable number of agents for each type. Additionally, you can evaluate the acceleration performance with or without AIOS by comparing the waiting time and turnaround time.

python eval.py --llm_name gpt-3.5-turbo --agents MathAgent:1,NarrativeAgent:1,RecAgent:1

You can use bash script to start the agent execution like this

bash scripts/eval/gpt4.sh

If you want to obtain metrics for either concurrent execution (with AIOS) or sequential execution (without AIOS), you can specify the mode parameter when running the eval.py file."

python eval.py --llm_name gpt-4 --agents MathAgent:1,NarrativeAgent:1,RecAgent:1 --mode concurrent-only
python eval.py --llm_name gpt-4 --agents MathAgent:1,NarrativeAgent:1,RecAgent:1 --mode sequential-only

3.3 Supported LLM backbones

  • gpt-3.5-turbo, gpt-4
  • gemini-pro
  • open-sourced LLM from Huggingface

๐Ÿ–‹๏ธ 4. References

@article{mei2024aios,
  title={AIOS: LLM Agent Operating System},
  author={Mei, Kai and Li, Zelong and Xu, Shuyuan and Ye, Ruosong and Ge, Yingqiang and Zhang, Yongfeng}
  journal={arXiv:2403.16971},
  year={2024}
}
@article{ge2023llm,
  title={LLM as OS, Agents as Apps: Envisioning AIOS, Agents and the AIOS-Agent Ecosystem},
  author={Ge, Yingqiang and Ren, Yujie and Hua, Wenyue and Xu, Shuyuan and Tan, Juntao and Zhang, Yongfeng},
  journal={arXiv:2312.03815},
  year={2023}
}

๐Ÿš€ 5. Contributions

AIOS is dedicated to facilitating the development and deployment of LLM agents in a systematic way, collaborators and contributions are always welcome to foster a cohesive, effective and efficient AIOS-Agent ecosystem!

For detailed information on how to contribute, see CONTRIBUTE. If you would like to contribute to the codebase, issues or pull requests are always welcome!

๐ŸŒ 6. AIOS Contributors

AIOS contributors

๐Ÿค 7. Discord Channel

If you would like to join the community, ask questions, chat with fellows, learn about or propose new features, and participate in future developments, join our Discord Community!

๐Ÿ“ช 8. Contact

For issues related to AIOS development, we encourage submitting issues, pull requests, or initiating discussions in the AIOS Discord Channel. For other issues please feel free to contact Kai Mei ([email protected]) and Yongfeng Zhang ([email protected]).

๐ŸŒŸ 9. Star History

Star History Chart

aios's People

Contributors

dongyuanjushi avatar agiresearch avatar evison avatar tata0703 avatar brama10 avatar eltociear avatar justsujay avatar itsthemoon avatar jhsuyu avatar lumiere-ml avatar arnoldioi avatar jgalego avatar peteryschneider avatar 1tylermitchell avatar ivanbelenky avatar shuyuan-x avatar zzfoo avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.