Start building LLM-empowered multi-agent applications in an easier way.
If you find our work helpful, please kindly cite our paper.
Visit our workstation to build multi-agent applications with dragging-and-dropping.
Discord | DingTalk |
---|---|
[2024-09-06] AgentScope version 0.1.0 is released now.
[2024-09-03] AgentScope supports Web Browser Control now! Refer to our example for more details.
[2024-07-15] AgentScope has implemented the Mixture-of-Agents algorithm. Refer to our MoA example for more details.
[2024-06-14] A new prompt tuning module is available in AgentScope to help developers generate and optimize the agents' system prompts! Refer to our tutorial for more details!
[2024-06-11] The RAG functionality is available for agents in AgentScope now! A quick introduction to RAG in AgentScope can help you equip your agent with external knowledge!
[2024-06-09] We release AgentScope v0.0.5 now! In this new version, AgentScope Workstation (the online version is running on agentscope.io) is open-sourced with the refactored AgentScope Studio!
AgentScope is an innovative multi-agent platform designed to empower developers to build multi-agent applications with large-scale models. It features three high-level capabilities:
🤝 Easy-to-Use: Designed for developers, with fruitful components, comprehensive documentation, and broad compatibility. Besides, AgentScope Workstation provides a drag-and-drop programming platform and a copilot for beginners of AgentScope!
✅ High Robustness: Supporting customized fault-tolerance controls and retry mechanisms to enhance application stability.
🚀 Actor-Based Distribution: Building distributed multi-agent applications in a centralized programming manner for streamlined development.
Supported Model Libraries
AgentScope provides a list of ModelWrapper
to support both local model
services and third-party model APIs.
API | Task | Model Wrapper | Configuration | Some Supported Models |
---|---|---|---|---|
OpenAI API | Chat | OpenAIChatWrapper |
guidance template |
gpt-4o, gpt-4, gpt-3.5-turbo, ... |
Embedding | OpenAIEmbeddingWrapper |
guidance template |
text-embedding-ada-002, ... | |
DALL·E | OpenAIDALLEWrapper |
guidance template |
dall-e-2, dall-e-3 | |
DashScope API | Chat | DashScopeChatWrapper |
guidance template |
qwen-plus, qwen-max, ... |
Image Synthesis | DashScopeImageSynthesisWrapper |
guidance template |
wanx-v1 | |
Text Embedding | DashScopeTextEmbeddingWrapper |
guidance template |
text-embedding-v1, text-embedding-v2, ... | |
Multimodal | DashScopeMultiModalWrapper |
guidance template |
qwen-vl-max, qwen-vl-chat-v1, qwen-audio-chat | |
Gemini API | Chat | GeminiChatWrapper |
guidance template |
gemini-pro, ... |
Embedding | GeminiEmbeddingWrapper |
guidance template |
models/embedding-001, ... | |
ZhipuAI API | Chat | ZhipuAIChatWrapper |
guidance template |
glm-4, ... |
Embedding | ZhipuAIEmbeddingWrapper |
guidance template |
embedding-2, ... | |
ollama | Chat | OllamaChatWrapper |
guidance template |
llama3, llama2, Mistral, ... |
Embedding | OllamaEmbeddingWrapper |
guidance template |
llama2, Mistral, ... | |
Generation | OllamaGenerationWrapper |
guidance template |
llama2, Mistral, ... | |
LiteLLM API | Chat | LiteLLMChatWrapper |
guidance template |
models supported by litellm... |
Yi API | Chat | YiChatWrapper |
guidance template |
yi-large, yi-medium, ... |
Post Request based API | - | PostAPIModelWrapper |
guidance template |
- |
Supported Local Model Deployment
AgentScope enables developers to rapidly deploy local model services using the following libraries.
Supported Services
Example Applications
Model
Conversation
Game
Distribution
More models, services and examples are coming soon!
AgentScope requires Python 3.9 or higher.
Note: This project is currently in active development, it's recommended to install AgentScope from source.
# Pull the source code from GitHub
git clone https://github.com/modelscope/agentscope.git
# Install the package in editable mode
cd agentscope
pip install -e .
pip install agentscope
To support different deployment scenarios, AgentScope provides several optional dependencies. Full list of optional dependencies refers to tutorial Taking distribution mode as an example, you can install its dependencies as follows:
# From source
pip install -e .[distribute]
# From pypi
pip install agentscope[distribute]
# From source
pip install -e .\[distribute\]
# From pypi
pip install agentscope\[distribute\]
In AgentScope, the model deployment and invocation are decoupled by
ModelWrapper
.
To use these model wrappers, you need to prepare a model config file as follows.
model_config = {
# The identifies of your config and used model wrapper
"config_name": "{your_config_name}", # The name to identify the config
"model_type": "{model_type}", # The type to identify the model wrapper
# Detailed parameters into initialize the model wrapper
# ...
}
Taking OpenAI Chat API as an example, the model configuration is as follows:
openai_model_config = {
"config_name": "my_openai_config", # The name to identify the config
"model_type": "openai_chat", # The type to identify the model wrapper
# Detailed parameters into initialize the model wrapper
"model_name": "gpt-4", # The used model in openai API, e.g. gpt-4, gpt-3.5-turbo, etc.
"api_key": "xxx", # The API key for OpenAI API. If not set, env
# variable OPENAI_API_KEY will be used.
"organization": "xxx", # The organization for OpenAI API. If not set, env
# variable OPENAI_ORGANIZATION will be used.
}
More details about how to set up local model services and prepare model configurations is in our tutorial.
Create built-in user and assistant agents as follows.
from agentscope.agents import DialogAgent, UserAgent
import agentscope
# Load model configs
agentscope.init(model_configs="./model_configs.json")
# Create a dialog agent and a user agent
dialog_agent = DialogAgent(name="assistant",
model_config_name="my_openai_config")
user_agent = UserAgent()
In AgentScope, message is the bridge among agents, which is a
dict that contains two necessary fields name
and content
and an
optional field url
to local files (image, video or audio) or website.
from agentscope.message import Msg
x = Msg(name="Alice", content="Hi!")
x = Msg("Bob", "What about this picture I took?", url="/path/to/picture.jpg")
Start a conversation between two agents (e.g. dialog_agent and user_agent) with the following code:
x = None
while True:
x = dialog_agent(x)
x = user_agent(x)
if x.content == "exit": # user input "exit" to exit the conversation_basic
break
AgentScope provides an easy-to-use runtime user interface capable of displaying multimodal output on the front end, including text, images, audio and video.
Refer to our tutorial for more details.
AgentScope is released under Apache License 2.0.
Contributions are always welcomed!
We provide a developer version with additional pre-commit hooks to perform checks compared to the official version:
# For windows
pip install -e .[dev]
# For mac
pip install -e .\[dev\]
# Install pre-commit hooks
pre-commit install
Please refer to our Contribution Guide for more details.
If you find our work helpful for your research or application, please cite our papers.
AgentScope: A Flexible yet Robust Multi-Agent Platform
@article{agentscope,
author = {Dawei Gao and
Zitao Li and
Xuchen Pan and
Weirui Kuang and
Zhijian Ma and
Bingchen Qian and
Fei Wei and
Wenhao Zhang and
Yuexiang Xie and
Daoyuan Chen and
Liuyi Yao and
Hongyi Peng and
Ze Yu Zhang and
Lin Zhu and
Chen Cheng and
Hongzhu Shi and
Yaliang Li and
Bolin Ding and
Jingren Zhou}
title = {AgentScope: A Flexible yet Robust Multi-Agent Platform},
journal = {CoRR},
volume = {abs/2402.14034},
year = {2024},
}