摘要(英) |
Recently, with the emergence of ChatGPT, the development of large language models has been flourishing. However, real-world applications are often more complex and cannot be han-
dled by a single agent alone. As a result, there has been growing interest in exploring how multiple agents can collaborate to solve more complex tasks, leading to the creation of various multi-agent conversation frameworks. Among them, AutoGen, developed by Microsoft, has garnered significant attention for allowing users to easily customize agents and tasks. It also supports Python coroutines, enabling concurrent handling of I/O-intensive tasks.
However, coroutines inherently run on a single thread. When tasks require substantial computational resources, they cannot leverage the performance of multi-core CPUs to achieve true parallel execution. For instance, in processing large-scale data, tasks are often divided into batches, but AutoGen’s architecture necessitates sequential execution of these tasks, resulting in increased processing time and limiting its performance and scalability in real-world scenarios.
Moreover, AutoGen’s execution model is constrained by its single-threaded architecture, preventing the system from scaling across multiple computing resources. In modern distributed systems, tasks can be distributed across multiple machines to maximize resource utilization.
However, AutoGen is unable to implement such processing, further restricting its ability to handle complex tasks and its range of applications. Additionally, the AutoGen system lacks fault tolerance. Since it operates on a single thread,
any errors or uncaptured exceptions during execution will cause the entire system to halt immediately. This results in interruptions, undermining its stability and reliability, and wasting the resources and costs invested in using LLM services.
To address these challenges, this study proposes AsyncGen, an improved multi-agent conversation framework based on distributed systems. By integrating AutoGen with the distributed system framework Ray, AsyncGen utilizes Ray’s resource scheduling to distribute computational tasks across multiple nodes for parallel processing. Experimental results show that the improved framework can operate simultaneously across multiple machines, significantly reducing execution time, enhancing fault tolerance, and improving scalability, demonstrating the potential and value of extending multi-agent frameworks using distributed systems. |
參考文獻 |
[1] T. Liang, Z. He, W. Jiao, X. Wang, Y. Wang, R. Wang, Y. Yang, S. Shi, and Z. Tu, “Encouraging divergent thinking in large language models through multi-agent debate,” arXiv preprint arXiv:2305.19118, 2023.
[2] G. Li, H. Hammoud, H. Itani, D. Khizbullin, and B. Ghanem, “Camel: Communicative agents for” mind” exploration of large language model society,” Advances in Neural Information Processing Systems, vol. 36, pp. 51991–52008, 2023.
[3] S. Hong, X. Zheng, J. Chen, Y. Cheng, J. Wang, C. Zhang, Z. Wang, S. K. S. Yau, Z. Lin, L. Zhou, et al., “Metagpt: Meta programming for multi-agent collaborative framework,” arXiv preprint arXiv:2308.00352, 2023.
[4] Q. Wu, G. Bansal, J. Zhang, Y. Wu, S. Zhang, E. Zhu, B. Li, L. Jiang, X. Zhang, and C. Wang, “AutoGen: Enabling next-gen llm applications via multi-agent conversation framework,” arXiv preprint arXiv:2308.08155, 2023.
[5] P. Moritz, R. Nishihara, S. Wang, A. Tumanov, R. Liaw, E. Liang, M. Elibol, Z. Yang, W. Paul, M. I. Jordan, et al., “Ray: A distributed framework for emerging {AI} applications,” in 13th USENIX symposium on operating systems design and implementation (OSDI 18), pp. 561–577, 2018.
[6] “LlamaIndex.” https://github.com/jerryjliu/llama_index, 2022.
[7] “LangChain.” https://github.com/langchain-ai/langchain, 2024.
[8] O. Khattab, A. Singhvi, P. Maheshwari, Z. Zhang, K. Santhanam, S. Vardhamanan, S. Haq, A. Sharma, T. T. Joshi, H. Moazam, et al., “Dspy: Compiling declarative language model calls into self-improving pipelines,” arXiv preprint arXiv:2310.03714, 2023.
[9] “CrewAI.” https://github.com/crewAIInc/crewAI, 2024.
[10] W. Chen, Z. You, R. Li, Y. Guan, C. Qian, C. Zhao, C. Yang, R. Xie, Z. Liu, and M. Sun, “Internet of agents: Weaving a web of heterogeneous agents for collaborative intelligence,” arXiv preprint arXiv:2407.07061, 2024.
[11] X. Tan, Y. Jiang, Y. Yang, and H. Xu, “Teola: Towards end-to-end optimization of llm-based applications,” arXiv preprint arXiv:2407.00326, 2024.
[12] L. Bonati, S. D’Oro, M. Polese, S. Basagni, and T. Melodia, “Intelligence and learning in o-ran for data-driven nextg cellular networks,” IEEE Communications Magazine, vol. 59, no. 10, pp. 21–27, 2021. |