Spaces:
Running
Running
Kunal Pai
commited on
Commit
·
1d27ba2
1
Parent(s):
c2ad08f
Add HASHIRU paper: Initial implementation of a hierarchical multi-agent system for hybrid intelligent resource utilization
Browse files- Introduced the main paper (conference_101719.tex) detailing the HASHIRU framework, its architecture, and core mechanisms.
- Added a PDF version of the paper (conference_101719.pdf).
- Created a references file (references.bib) with citations relevant to the research and framework.
- paper/HASHIRU.pdf +0 -0
- paper/IEEEtran.cls +0 -0
- paper/conference_101719.tex +248 -0
- paper/references.bib +368 -0
paper/HASHIRU.pdf
ADDED
Binary file (74.4 kB). View file
|
|
paper/IEEEtran.cls
ADDED
The diff for this file is too large to render.
See raw diff
|
|
paper/conference_101719.tex
ADDED
@@ -0,0 +1,248 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
\documentclass[conference]{IEEEtran}
|
2 |
+
\IEEEoverridecommandlockouts
|
3 |
+
% The preceding line is only needed to identify funding in the first footnote. If that is unneeded, please comment it out.
|
4 |
+
\usepackage{cite}
|
5 |
+
\usepackage{amsmath,amssymb,amsfonts}
|
6 |
+
\usepackage{algorithmic}
|
7 |
+
\usepackage{graphicx}
|
8 |
+
\usepackage{textcomp}
|
9 |
+
\usepackage{xcolor}
|
10 |
+
\usepackage{hyperref}
|
11 |
+
\def\BibTeX{{\rm B\kern-.05em{\sc i\kern-.025em b}\kern-.08em
|
12 |
+
T\kern-.1667em\lower.7ex\hbox{E}\kern-.125emX}}
|
13 |
+
\begin{document}
|
14 |
+
|
15 |
+
\title{HASHIRU: Hierarchical Agent System for Hybrid Intelligent Resource Utilization}
|
16 |
+
|
17 |
+
\author{\IEEEauthorblockN{Kunal Pai}
|
18 |
+
\IEEEauthorblockA{\textit{UC Davis} \\
|
19 |
+
kunpai@ucdavis.edu}
|
20 |
+
\and
|
21 |
+
\IEEEauthorblockN{Parth Shah}
|
22 |
+
\IEEEauthorblockA{\textit{Independent Researcher} \\
|
23 |
+
helloparthshah@gmail.com}
|
24 |
+
\and
|
25 |
+
\IEEEauthorblockN{Harshil Patel}
|
26 |
+
\IEEEauthorblockA{\textit{UC Davis} \\
|
27 |
+
hpppatel@ucdavis.edu}
|
28 |
+
\and
|
29 |
+
\IEEEauthorblockN{Saisha Shetty}
|
30 |
+
\IEEEauthorblockA{\textit{UC Davis} \\
|
31 |
+
spshetty@ucdavis.edu}
|
32 |
+
}
|
33 |
+
|
34 |
+
\maketitle
|
35 |
+
|
36 |
+
\section{Introduction}\label{sec:introduction}
|
37 |
+
|
38 |
+
The landscape of Artificial Intelligence (AI) is being reshaped by the rapid advancements in Large Language Models (LLMs), which exhibit profound capabilities in language understanding, generation, reasoning, and planning \cite{brown2020language, devlin2019bert, raffel2020exploring}. This progress has catalyzed the development of sophisticated AI agents capable of autonomous task execution. Increasingly, the focus is shifting from single-agent systems to Multi-Agent Systems (MAS), where collaborative teams of specialized agents address complex problems beyond the scope of individual agents \cite{dorri2018multi, wooldridge2009introduction}. Such collaborative approaches hold significant potential in diverse domains like scientific discovery \cite{boiko2023emergent}, software engineering \cite{qian2023communicative}, data analysis, and strategic decision-making \cite{wang2023decision}. The growing complexity of tasks, highlighted by benchmarks demanding advanced mathematical reasoning (e.g., GSM8K \cite{cobbe2021gsm8k}, SVAMP \cite{patel2021svamp}), coding (e.g., HumanEval \cite{chen2021codex}, CoDocBench \cite{pai2024codocbench}), and graduate-level technical knowledge and reasoning \cite{phan2025humanitysexam}, further underscores the need for agentic systems capable of effectively coordinating diverse cognitive resources \cite{wen2024benchmarkingcomplexinstructionfollowingmultiple}.
|
39 |
+
|
40 |
+
Despite this promise, contemporary agentic frameworks often encounter significant limitations. Many suffer from \textbf{rigidity}, relying on predefined agent roles and static team structures that hinder adaptation to dynamic task requirements \cite{zhang2023building}. Furthermore, \textbf{resource obliviousness} is prevalent; systems frequently lack mechanisms to monitor and optimize computational resources like API costs, memory usage, and CPU load, leading to inefficiency, particularly when scaling or deploying in resource-constrained environments \cite{park2023generative}. This is often exacerbated by a reliance on powerful, proprietary cloud-based LLMs, incurring substantial operational expenses. \textbf{Model homogeneity}, the default use of a single powerful LLM for all sub-tasks, neglects the potential efficiency gains from employing a diverse ecosystem of models, including smaller, specialized, or locally-run alternatives \cite{zhou2023agents}. Lastly, while \textbf{tool use} is fundamental \cite{yao2022react, parisi2022talm}, the ability for agents to autonomously \textbf{create and integrate new tools} during operation remains limited, restricting dynamic functional extension and long-term self-improvement without human intervention \cite{wang2023voyager}.
|
41 |
+
|
42 |
+
To address these challenges, we introduce \textbf{HASHIRU (Hierarchical Agent System for Hybrid Intelligent Resource Utilization)}, a novel MAS framework designed for enhanced flexibility, resource efficiency, and adaptability. HASHIRU employs a hierarchical structure led by a central ``CEO'' agent that dynamically manages a team of specialized ``employee'' agents, instantiated on demand for specific sub-tasks. A core tenet of HASHIRU is its \textbf{hybrid intelligence} approach, strategically prioritizing smaller (e.g., 3B--7B parameter), locally-run LLMs, often accessed via frameworks like Ollama \cite{ollama}, to promote cost-effectiveness and computational efficiency. While prioritizing local resources, the system retains the flexibility to integrate external APIs and potentially more powerful models when justified by task complexity and resource availability, under the CEO's management.
|
43 |
+
|
44 |
+
The primary contributions of this work are:
|
45 |
+
\begin{enumerate}
|
46 |
+
\item A novel MAS architecture combining a \textbf{hierarchical control structure} with \textbf{dynamic, resource-aware agent lifecycle management} (hiring/firing). This management is explicitly governed by computational budget constraints (cost, memory usage, concurrency) and incorporates an economic model with hiring/firing costs to discourage excessive churn.
|
47 |
+
\item A \textbf{hybrid intelligence model} that prioritizes cost-effective, local LLMs while adaptively incorporating external APIs and potentially larger models, optimizing the efficiency-capability trade-off.
|
48 |
+
\item An integrated mechanism enabling the \textbf{autonomous creation of API tools}, allowing the system to dynamically extend its functional repertoire in response to task demands.
|
49 |
+
\item The application of an \textbf{economic model} (hiring/firing fees) to agent management, promoting efficient resource allocation and team stability.
|
50 |
+
\end{enumerate}
|
51 |
+
|
52 |
+
This paper details the design and rationale behind HASHIRU. Section \ref{sec:background} discusses related work in agent architectures, dynamic management, resource allocation, model heterogeneity, and tool use. Section 3 elaborates on the HASHIRU architecture and its core mechanisms. Section 4 presents experimental results (or outlines planned experiments), followed by discussion and conclusion in Sections 5 and 6.
|
53 |
+
|
54 |
+
\section{Background and Related Work} \label{sec:background}
|
55 |
+
|
56 |
+
The concept of intelligent agents has evolved significantly from early work in symbolic AI and distributed problem-solving \cite{russell2010artificial, shoham1994agent} to the current era dominated by LLMs. Modern agentic frameworks leverage LLMs as their cognitive core, enabling sophisticated reasoning, planning, and interaction capabilities \cite{wang2023survey, xi2023rise}. HASHIRU builds upon this foundation while addressing specific limitations observed in the current state-of-the-art.
|
57 |
+
|
58 |
+
\subsection{Agent Architectures: Hierarchy and Dynamics}
|
59 |
+
MAS architectures vary widely, including flat, federated, and hierarchical structures \cite{dorri2018multi, horling2004survey}. Hierarchical models offer clear control flow and efficient task decomposition but risk bottlenecks and rigidity \cite{gaston2005agenta,gaston2005agentb}. HASHIRU utilizes a \textbf{CEO-Employee hierarchy} for centralized coordination but distinguishes itself through \textbf{dynamic team composition}. Unlike systems with static hierarchies or predefined roles (e.g., CrewAI \cite{crewai}, ChatDev \cite{qian2023communicative}), HASHIRU's CEO actively manages the employee pool based on runtime needs and resource constraints.
|
60 |
+
|
61 |
+
\subsection{Dynamic Agent Lifecycle Management}
|
62 |
+
The ability of an MAS to adapt its composition dynamically is crucial for complex environments \cite{valckenaers2005trends}. Various triggers for agent creation or deletion have been explored, often tied to task structure or environmental changes. HASHIRU introduces a specific mechanism where the CEO agent makes \textbf{hiring and firing decisions} based on a cost-benefit analysis considering agent performance, operational costs (API fees, inferred compute), memory footprint (tracked explicitly as a percentage of available resources), and agent concurrency limits. Furthermore, HASHIRU incorporates an \textbf{economic model} with explicit ``starting bonus'' (hiring) and ``invocation'' (usage) costs. This introduces economic friction, aiming to prevent excessive agent initialization or usage for marginal gains and promote team stability, a nuance often missing in simpler dynamic composition strategies.
|
63 |
+
|
64 |
+
\subsection{Resource Management and Agent Economies}
|
65 |
+
Resource awareness is critical for scalable and deployable MAS. Research areas in economics explore mechanisms like market-based auctions or contract nets for resource allocation \cite{clearwater1996market}. HASHIRU implements a more \textbf{centralized, budget-constrained resource management model}. The CEO operates within defined limits for financial cost, memory usage (as a percentage of total available memory), and concurrent agent count. This direct management, particularly the focus on memory percentage, suggests an orientation towards practical deployment, potentially on local hardware or edge devices with finite resources, contrasting with cloud-centric systems assuming elastic resources \cite{park2023generative}. Frameworks like AutoGen \cite{wu2023autogen} and LangGraph \cite{langgraph} typically rely on implicit cost tracking via API keys without such explicit multi-dimensional resource budgeting and control.
|
66 |
+
|
67 |
+
\subsection{Hybrid Intelligence and Heterogeneous Models}
|
68 |
+
Leveraging diverse LLMs with varying capabilities, costs, and latencies is an emerging trend \cite{zhou2023agents}. Techniques like model routing aim to select the optimal model for specific sub-tasks. HASHIRU embraces \textbf{model heterogeneity} with a specific strategic focus: \textbf{prioritizing smaller (3B--7B), locally-run models via Ollama integration} \cite{ollama}. This emphasizes cost-efficiency, low latency for simpler tasks, and potential privacy advantages over systems defaulting to large, proprietary cloud APIs (e.g., GPT-4 \cite{openai2023gpt4}, Claude 3 \cite{anthropic2024claude}). While capable of integrating external APIs (potentially invoking larger models), HASHIRU's default operational stance represents a distinct balance point in the capability vs. efficiency trade-off.
|
69 |
+
|
70 |
+
\subsection{Tool Use and Autonomous Tool Creation}
|
71 |
+
The ability to use external tools (APIs, functions, databases) is a cornerstone of modern agents, enabled by frameworks like ReAct \cite{yao2022react} and built-in function calling \cite{openai_func_calling}. Most systems rely on a predefined tool suite. HASHIRU advances this by incorporating a mechanism for \textbf{autonomous API tool creation}. When a required functionality is missing, the CEO can commission the generation (potentially via a specialized agent or code generation process) and deployment of a new API tool within the HASHIRU ecosystem. This capability for self-extension differentiates HASHIRU from systems limited to static toolsets and moves towards greater operational autonomy and adaptability \cite{wang2023voyager, park2023generative}.
|
72 |
+
|
73 |
+
In summary, HASHIRU integrates concepts from hierarchical control, dynamic MAS, resource management, and tool use, but its novelty lies in the synergistic combination of: (1) dynamic, resource-aware hierarchical management with (2) an economic model for stability, (3) a local-first hybrid intelligence strategy, and (4) integrated autonomous tool creation capabilities. This combination targets key limitations in current agentic systems concerning efficiency, adaptability, cost, and autonomy.
|
74 |
+
|
75 |
+
\section{HASHIRU System Architecture}
|
76 |
+
\label{sec:architecture}
|
77 |
+
|
78 |
+
The architecture of HASHIRU is designed to directly address the challenges of rigidity, resource obliviousness, and limited adaptability outlined in Section~\ref{sec:introduction}. It implements a hierarchical, dynamically managed multi-agent system optimized for hybrid resource utilization. This section details the core components and mechanisms underpinning HASHIRU's operation.
|
79 |
+
|
80 |
+
\subsection{Overview}
|
81 |
+
HASHIRU operates on a hierarchical model conceptually similar to a business organization, featuring a central coordinating agent (``CEO'') and specialized task-executing agents (``Employees''). Key architectural tenets include:
|
82 |
+
\begin{itemize}
|
83 |
+
\item \textbf{Centralized Coordination within a Dynamic Hierarchy:} A CEO agent manages overall strategy, task allocation, and team composition.
|
84 |
+
\item \textbf{Dynamic Lifecycle Management:} Employee agents are instantiated (hired) and terminated (fired) based on runtime requirements and resource constraints, governed by an economic model.
|
85 |
+
\item \textbf{Hybrid Intelligence:} Strategic preference for local, computationally cheaper LLMs, while retaining access to external APIs and potentially more powerful models.
|
86 |
+
\item \textbf{Explicit Resource Management:} Continuous monitoring and control of costs, memory usage, and agent concurrency against defined budgets.
|
87 |
+
\item \textbf{Adaptive Tooling:} Utilization of predefined tools alongside the capability for autonomous creation of new API tools.
|
88 |
+
\end{itemize}
|
89 |
+
Figure \ref{fig:arch} illustrates the overall structure and interaction flow.
|
90 |
+
|
91 |
+
\begin{figure}[ht]
|
92 |
+
\centering
|
93 |
+
\includegraphics[width=0.45\textwidth]{HASHIRU.pdf}
|
94 |
+
\caption{High-level architecture of the HASHIRU system, illustrating the CEO-Employee hierarchy.}
|
95 |
+
\label{fig:arch}
|
96 |
+
\end{figure}
|
97 |
+
|
98 |
+
\subsection{Hierarchical Structure: CEO and Employee Agents}
|
99 |
+
The system employs a two-tiered hierarchy:
|
100 |
+
|
101 |
+
\begin{itemize}
|
102 |
+
\item \textbf{CEO Agent:} This singleton agent serves as the central coordinator and entry point. Its primary responsibilities include:
|
103 |
+
\begin{itemize}
|
104 |
+
\item Receiving and interpreting the primary user query or task.
|
105 |
+
\item Decomposing the main task into smaller, manageable sub-tasks.
|
106 |
+
\item Identifying the capabilities required for each sub-task.
|
107 |
+
\item Managing the pool of Employee agents (see Section \ref{subsec:dynamic_mgmt}).
|
108 |
+
\item Assigning sub-tasks to suitable, active Employee agents.
|
109 |
+
\item Monitoring the progress and performance of Employee agents.
|
110 |
+
\item Synthesizing the results from Employee agents into a coherent final output or response.
|
111 |
+
\item Managing the system's overall resource budget (see Section \ref{subsec:resource_mgmt}).
|
112 |
+
\item Initiating the creation of new tools when required (see Section \ref{subsec:tooling}).
|
113 |
+
\end{itemize}
|
114 |
+
\item \textbf{Employee Agents:} These are specialized agents instantiated by the CEO to perform specific sub-tasks. Each Employee agent typically wraps an LLM (local via Ollama \cite{ollama} or external via API) or provides access to a specific tool/API. Key characteristics include:
|
115 |
+
\begin{itemize}
|
116 |
+
\item Specialization: Possessing capabilities tailored to certain types of sub-tasks (e.g., code generation, data analysis, information retrieval).
|
117 |
+
\item Dynamic Existence: Created and destroyed by the CEO based on need and performance.
|
118 |
+
\item Task Execution: Receives a sub-task description and context from the CEO, executes it, and returns the result.
|
119 |
+
\item Resource Consumption: Associated with specific resource costs (e.g., API call costs, memory footprint) tracked by the system.
|
120 |
+
\end{itemize}
|
121 |
+
\end{itemize}
|
122 |
+
This hierarchical structure facilitates organized task decomposition and result aggregation under centralized control, while the dynamic nature of the Employee pool provides flexibility.
|
123 |
+
|
124 |
+
\subsection{Dynamic Agent Lifecycle Management}
|
125 |
+
\label{subsec:dynamic_mgmt}
|
126 |
+
A core innovation in HASHIRU is the CEO's ability to dynamically manage the Employee agent team through ``hiring'' (instantiation) and ``firing'' (termination). This process is driven by a cost-benefit analysis aimed at optimizing task performance within resource constraints.
|
127 |
+
|
128 |
+
When a new sub-task requires capabilities not readily available or efficiently provided by the current pool of active Employee agents, the CEO may decide to hire a new agent. Conversely, if an agent is underperforming, consistently idle, excessively costly, or if resource limits are approached, the CEO may decide to fire it. This decision considers multiple factors:
|
129 |
+
\begin{itemize}
|
130 |
+
\item \textbf{Task Requirements:} The specific capabilities needed for pending sub-tasks.
|
131 |
+
\item \textbf{Agent Performance Metrics:} Historical success rate, quality of output, or efficiency of existing agents relevant to the task type.
|
132 |
+
\item \textbf{Operational Costs:} API costs, estimated computational load, or other costs associated with using the agent's underlying model or tools.
|
133 |
+
\item \textbf{Memory Footprint:} The amount of system memory the agent consumes, tracked as a percentage of the total available memory allocated to HASHIRU.
|
134 |
+
\item \textbf{Agent Concurrency:} The current number of active Employee agents relative to a predefined limit.
|
135 |
+
\end{itemize}
|
136 |
+
|
137 |
+
Crucially, HASHIRU incorporates an \textbf{economic model} for agent lifecycle events:
|
138 |
+
\begin{itemize}
|
139 |
+
\item \textbf{Hiring Cost (``Starting Bonus''):} A one-time cost incurred when a new agent is instantiated, representing setup overhead.
|
140 |
+
\item \textbf{Invocation Cost (``Salary''):} A multi-time cost incurred when an agent is used, representing load on the system, or on payment methods, for using an agent.
|
141 |
+
\end{itemize}
|
142 |
+
These explicit transaction costs discourage excessive agent churn, promoting stability. The CEO must evaluate if the anticipated long-term benefit of replacing an agent outweighs the immediate hiring and firing costs plus any difference in ongoing operational costs. This mechanism directly combats system rigidity and allows adaptation while actively managing computational budgets and preventing wasteful high-frequency agent turnover.
|
143 |
+
|
144 |
+
\subsection{Hybrid Intelligence and Model Management}
|
145 |
+
HASHIRU is designed for \textbf{hybrid intelligence}, leveraging a diverse set of cognitive resources. It strategically prioritizes the use of smaller (e.g., 3B--7B parameter), cost-effective LLMs that can be run locally via frameworks like Ollama \cite{ollama}. This approach enhances efficiency, reduces reliance on expensive external APIs, and potentially improves privacy and latency for certain tasks.
|
146 |
+
|
147 |
+
However, the system is not restricted to local models. The CEO agent can integrate and utilize:
|
148 |
+
\begin{itemize}
|
149 |
+
\item \textbf{External LLM APIs:} Access to powerful proprietary models (e.g., GPT-4 \cite{openai2023gpt4}, Claude 3 \cite{anthropic2024claude}) when deemed necessary for complex reasoning or specialized knowledge, subject to cost-benefit analysis.
|
150 |
+
\item \textbf{External Tool APIs:} Integration with third-party software or data sources.
|
151 |
+
\item \textbf{Self-Created APIs:} Tools generated by HASHIRU itself (see Section \ref{subsec:tooling}).
|
152 |
+
\end{itemize}
|
153 |
+
The CEO manages this heterogeneous pool, selecting the most appropriate resource (local model, external API, tool) for a given sub-task based on perceived difficulty, required capabilities, and the current resource budget status. This allows HASHIRU to balance cost-effectiveness and computational efficiency with the need for high capability when required.
|
154 |
+
|
155 |
+
\subsection{Resource Monitoring and Control}
|
156 |
+
\label{subsec:resource_mgmt}
|
157 |
+
Explicit resource management is central to HASHIRU's design, moving beyond simple API key cost tracking. The system, coordinated by the CEO, actively monitors:
|
158 |
+
\begin{itemize}
|
159 |
+
\item \textbf{Financial Costs:} Accumulating costs from external API calls.
|
160 |
+
\item \textbf{Memory Usage:} Tracking the memory footprint of active Employee agents, specifically as a percentage of a predefined total available memory resource.
|
161 |
+
\item \textbf{Agent Concurrency:} Maintaining a count of concurrently active Employee agents.
|
162 |
+
\end{itemize}
|
163 |
+
These metrics are monitored against predefined \textbf{budget limits} or hard caps. If initiating an action (like hiring a new agent) would exceed a budget limit (e.g., push memory usage over 90\% of allocated, or exceed the maximum concurrent agent count), the action is prevented. This mechanism ensures the system operates within defined operational constraints, crucial for deployment on devices with limited resources or under strict financial budgets.
|
164 |
+
|
165 |
+
\subsection{Tool Utilization and Autonomous Creation}
|
166 |
+
\label{subsec:tooling}
|
167 |
+
Like many modern agent systems, HASHIRU's agents can utilize predefined tools (functions, external APIs, databases) to interact with the environment and perform actions beyond pure text generation \cite{yao2022react, openai_func_calling}.
|
168 |
+
|
169 |
+
A distinctive feature of HASHIRU is its capability for \textbf{integrated, autonomous tool creation}. If the CEO agent determines, through task analysis or failure analysis of existing agents, that a specific functional capability is required but not available through existing Employee agents or tools, it can initiate a process to create a new tool. This typically involves:
|
170 |
+
\begin{enumerate}
|
171 |
+
\item Defining the specification for the required tool (inputs, outputs, functionality).
|
172 |
+
\item Commissioning the generation of the necessary logic (e.g., code implementing the functionality, potentially involving API calls to external services using provided credentials, possibly generated by a specialized code-generating Employee agent).
|
173 |
+
\item Deploying this logic as a new, callable API endpoint accessible within the HASHIRU ecosystem.
|
174 |
+
\item Potentially instantiating a new Employee agent dedicated to utilizing this newly created tool.
|
175 |
+
\end{enumerate}
|
176 |
+
This mechanism allows HASHIRU to dynamically extend its own functional repertoire over time, tailoring its capabilities to the tasks it encounters without requiring direct manual intervention for every new function, thereby enabling greater autonomy and long-term adaptation.
|
177 |
+
|
178 |
+
|
179 |
+
\section{Experimental Setup}
|
180 |
+
\label{sec:experiments}
|
181 |
+
|
182 |
+
To evaluate the performance, efficiency, and adaptability of HASHIRU, we designed a set of experiments targeting its core architectural features. Our evaluation focuses on assessing the benefits of dynamic resource-aware management, the hybrid intelligence model, and the autonomous tool creation capability compared to relevant baselines. Specifically, we investigate:
|
183 |
+
\begin{itemize}
|
184 |
+
\item The impact of dynamic agent management with economic constraints on resource utilization (cost, memory) and task performance compared to static configurations.
|
185 |
+
\item The effectiveness of the hybrid (local-first) model strategy versus homogeneous (cloud-only or local-only) approaches across tasks of varying complexity.
|
186 |
+
\item The system's ability to autonomously create and utilize necessary tools when faced with novel functional requirements within a task.
|
187 |
+
\end{itemize}
|
188 |
+
|
189 |
+
\subsection{Evaluation Tasks}
|
190 |
+
\label{subsec:tasks}
|
191 |
+
We selected tasks demanding complex reasoning, multi-perspective analysis, and interaction, suitable for exercising HASHIRU's hierarchical coordination and dynamic capabilities. The tasks fall into two main categories:
|
192 |
+
|
193 |
+
\subsubsection{Academic Paper Review}
|
194 |
+
This task evaluates HASHIRU's capacity to critically assess academic work by simulating the peer-review process. Given one or more scientific papers (e.g., in PDF format), the system must generate a review summary and ultimately recommend acceptance or rejection. This probes HASHIRU's ability to decompose evaluation criteria, delegate tasks to specialized agents (e.g., novelty assessment, methodological rigor, clarity), and manage resources effectively across long and complex documents.
|
195 |
+
|
196 |
+
\subsubsection{Reasoning and Problem-Solving Tasks}
|
197 |
+
To evaluate broader reasoning, knowledge retrieval, and problem-solving capabilities under different constraints, we employ a set of challenging benchmarks and puzzle-like tasks:
|
198 |
+
\begin{itemize}
|
199 |
+
\item \textbf{Humanity's Last Exam \cite{phan2025humanitysexam}:} A benchmark designed to test graduate-level technical knowledge and complex reasoning across multiple domains. Success requires deep understanding and sophisticated problem-solving, likely necessitating access to powerful external LLMs managed effectively within HASHIRU's hybrid framework.
|
200 |
+
\item \textbf{NYT Connections:} This popular puzzle requires identifying hidden semantic relationships or themes to categorize 16 words into four distinct groups. Solving this involves associative reasoning, broad world knowledge, and potentially hypothesis testing across different potential groupings, testing knowledge access and combinatorial reasoning coordination.
|
201 |
+
\item \textbf{Wordle:} The daily word puzzle requires deductive reasoning to identify a five-letter word within six guesses, using feedback on correct letters and positions. This tests logical deduction, constraint satisfaction, and vocabulary knowledge. It serves as a good test case for comparing the efficiency (speed, cost, number of guesses) of local versus external models for iterative reasoning. We assume interaction via a simulated game environment.
|
202 |
+
\item \textbf{Globle:} This geographic deduction game requires identifying a target country based on proximity feedback from guesses. It tests geographic knowledge retrieval, spatial reasoning, and iterative strategy refinement based on feedback (distance, direction). We assume interaction via a simulated game environment.
|
203 |
+
\end{itemize}
|
204 |
+
These diverse reasoning tasks challenge the system's ability to leverage appropriate cognitive resources (local vs. external models), potentially create simple tools, and coordinate problem-solving strategies effectively.
|
205 |
+
|
206 |
+
\subsection{Baselines for Comparison}
|
207 |
+
\label{subsec:baselines}
|
208 |
+
To quantify the benefits of HASHIRU's features, we will compare its performance against several baseline configurations:
|
209 |
+
\begin{itemize}
|
210 |
+
\item \textbf{Static-HASHIRU:} A version with a fixed, predefined set of Employee agents (e.g., one generalist agent per potential role identified in paper analysis), disabling dynamic hiring/firing.
|
211 |
+
\item \textbf{Cloud-Only HASHIRU:} HASHIRU operating exclusively with a powerful external LLM API and online function-calling for all agents, disabling the use of local models.
|
212 |
+
\item \textbf{Local-Only HASHIRU:} HASHIRU operating exclusively with smaller, local LLMs (e.g., selected models via Ollama) for all agents.
|
213 |
+
\item \textbf{HASHIRU (No-Economy):} HASHIRU with dynamic hiring/firing enabled but without the explicit costs, to isolate the impact of the economic model on agent churn and stability.
|
214 |
+
% \item \textbf{[Optional] Other Frameworks:} If feasible, configure AutoGen \cite{wu2023autogen} or CrewAI \cite{crewai} with a similar hierarchical structure to solve a subset of the tasks for comparison.
|
215 |
+
\end{itemize}
|
216 |
+
|
217 |
+
\subsection{Evaluation Metrics}
|
218 |
+
\label{subsec:metrics}
|
219 |
+
We will evaluate performance using a combination of quantitative and qualitative metrics:
|
220 |
+
\begin{itemize}
|
221 |
+
\item \textbf{Task Success Rate / Quality:}
|
222 |
+
\begin{itemize}
|
223 |
+
\item Percentage of tasks successfully completed (binary for games, potentially graded for analysis based on rubrics).
|
224 |
+
\item Quality of output for analysis tasks (human evaluation based on relevance, coherence, accuracy, completeness).
|
225 |
+
\item Accuracy for information extraction tasks.
|
226 |
+
\item Number of guesses/turns required for game tasks.
|
227 |
+
\end{itemize}
|
228 |
+
\item \textbf{Resource Consumption:}
|
229 |
+
\begin{itemize}
|
230 |
+
\item Total cost incurred from external API calls.
|
231 |
+
\item Peak and average memory usage (\% of allocated budget).
|
232 |
+
\item Wall-clock time per task.
|
233 |
+
\item Number and type (local/external) of LLM calls.
|
234 |
+
\end{itemize}
|
235 |
+
\item \textbf{System Dynamics and Adaptability:}
|
236 |
+
\begin{itemize}
|
237 |
+
\item Number of Employee agents hired and fired during tasks.
|
238 |
+
\item Frequency of agent churn (hires+fires / task duration or steps).
|
239 |
+
\item Number and utility of autonomously created tools (if applicable).
|
240 |
+
\end{itemize}
|
241 |
+
\end{itemize}
|
242 |
+
|
243 |
+
|
244 |
+
|
245 |
+
\bibliography{references}
|
246 |
+
\bibliographystyle{plain}
|
247 |
+
|
248 |
+
\end{document}
|
paper/references.bib
ADDED
@@ -0,0 +1,368 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
@article{shen2023hugginggpt,
|
2 |
+
title = {HuggingGPT: Solving AI Tasks with ChatGPT and its Friends in Hugging Face},
|
3 |
+
author = {Shen, Yongliang and Song, Kaitao and Tan, Xu and Li, Dongsheng and Lu, Weiming and Zhuang, Yueting},
|
4 |
+
journal = {arXiv preprint arXiv:2303.17580},
|
5 |
+
year = {2023}
|
6 |
+
}
|
7 |
+
|
8 |
+
@article{wu2023autogen,
|
9 |
+
title = {{AutoGen}: Enabling Next-Gen {LLM} Applications via Multi-Agent Conversation},
|
10 |
+
author = {Wu, Qingyun and Bansal, Gagan and Zhang, Jieyu and Wu, Yiran and Li, Beibin and Zhu, Erkang and Jiang, Li and Zhang, Xiaoyun and Zhang, Shaokun and Liu, Jiale and Awadallah, Ahmed H. and White, Ryen W. and Burger, Doug and Wang, Chi},
|
11 |
+
journal = {arXiv preprint arXiv:2308.08155},
|
12 |
+
year = {2023}
|
13 |
+
}
|
14 |
+
|
15 |
+
@inproceedings{yao2022react,
|
16 |
+
title = {{ReAct}: Synergizing Reasoning and Acting in Language Models},
|
17 |
+
author = {Yao, Shunyu and Zhao, Jeffrey and Yu, Dian and Du, Nan and Shafran, Izhak and Narasimhan, Karthik and Cao, Yuan},
|
18 |
+
booktitle = {International Conference on Learning Representations (ICLR)},
|
19 |
+
year = {2023},
|
20 |
+
note = {arXiv:2210.03629}
|
21 |
+
}
|
22 |
+
|
23 |
+
@article{schick2023toolformer,
|
24 |
+
title = {Toolformer: Language Models Can Teach Themselves to Use Tools},
|
25 |
+
author = {Schick, Timo and Dwivedi-Yu, Jane and Bitton, Yonatan and Yuan, Xi and Camburu, Oana-Maria and Houlsby, Neil},
|
26 |
+
journal = {arXiv preprint arXiv:2302.04761},
|
27 |
+
year = {2023}
|
28 |
+
}
|
29 |
+
|
30 |
+
@article{ong2024routellm,
|
31 |
+
title = {{RouteLLM}: Learning to Route {LLMs} with Preference Data},
|
32 |
+
author = {Ong, Isaac and Almahairi, Amjad and Wu, Vincent and Chiang, Wei-Lin and Wu, Tianhao and Gonzalez, Joseph E. and Kadous, M. Waleed and Stoica, Ion},
|
33 |
+
journal = {arXiv preprint arXiv:2406.18665},
|
34 |
+
year = {2024}
|
35 |
+
}
|
36 |
+
|
37 |
+
@article{fourney2024magentic,
|
38 |
+
title = {Magentic-One: A Generalist Multi-Agent System for Solving Complex Tasks},
|
39 |
+
author = {Fourney, Adam and Bansal, Gagan and Mozannar, Hussein and Tan, Cheng and et al.},
|
40 |
+
journal = {arXiv preprint arXiv:2411.04468},
|
41 |
+
year = {2024}
|
42 |
+
}
|
43 |
+
|
44 |
+
@inproceedings{cobbe2021gsm8k,
|
45 |
+
title = {Training Verifiers to Solve Math Word Problems},
|
46 |
+
author = {Cobbe, Karl and Kosaraju, Vineet and Bavarian, Mohammad and Chen, Mark and Jun, Heewoo and Kaiser, Lukasz and Plappert, Matthias and Tworek, Jerry and Hilton, Jacob and Nakano, Reiichiro and Hesse, Christopher and Schulman, John},
|
47 |
+
booktitle = {Advances in Neural Information Processing Systems (NeurIPS)},
|
48 |
+
year = {2021},
|
49 |
+
note = {Dataset introduced: GSM8K (Grade School Math 8K)}
|
50 |
+
}
|
51 |
+
|
52 |
+
@inproceedings{patel2021svamp,
|
53 |
+
title = {Are {NLP} Models really able to Solve Simple Math Word Problems?},
|
54 |
+
author = {Patel, Arkil and Bhattamishra, Satwik and Goyal, Navin},
|
55 |
+
booktitle = {Conference of the North American Chapter of the Association for Computational Linguistics (NAACL)},
|
56 |
+
year = {2021},
|
57 |
+
note = {Introduces the SVAMP challenge dataset}
|
58 |
+
}
|
59 |
+
|
60 |
+
@misc{phan2025humanitysexam,
|
61 |
+
title = {Humanity's Last Exam},
|
62 |
+
author = {Phan, Long and Gatti, Alice and Han, Ziwen and others},
|
63 |
+
year = {2025},
|
64 |
+
eprint = {2501.14249},
|
65 |
+
archivePrefix = {arXiv},
|
66 |
+
primaryClass = {cs.LG},
|
67 |
+
url = {https://arxiv.org/abs/2501.14249}
|
68 |
+
}
|
69 |
+
|
70 |
+
|
71 |
+
|
72 |
+
@article{chen2021codex,
|
73 |
+
title = {Evaluating Large Language Models Trained on Code},
|
74 |
+
author = {Chen, Mark and Tworek, Jerry and Jun, Heewoo and Yuan, Qiming and Ponde de Oliveira Pinto, Henrique and Kaplan, Jared and Edwards, Harri and Burda, Yuri and Joseph, Nicholas and Brockman, Greg and Ray, Alex and Puri, Raul and Krueger, Gretchen and Petrov, Michael and Khlaaf, Heidy and Sastry, Girish and Mishkin, Pamela and Chan, Brooke and Gray, Scott and Ryder, Nick and Pavlov, Mikhail and Power, Alethea and Kaiser, Lukasz and Bavarian, Mohammad and Winter, Clemens and Tillet, Philippe and Such, Felipe and Cummings, Dave and Plappert, Matthias and Chantzis, Fotios and Barnes, Elizabeth and Herbert-Voss, Ariel and Guss, William and Nichol, Alex and Paino, Alex and Tezak, Nikolas and Tang, Jie and Babuschkin, Igor and Balaji, Suchir and Jain, Shantanu and Saunders, William and Hesse, Christopher and Carr, Andrew N. and Leike, Jan and Achiam, Josh and Misra, Vedant and Morikawa, Evan and Radford, Alec and Knight, Matthew and Brundage, Miles and Murati, Mira and Mayer, Katie and Welinder, Peter and McGrew, Bob and Amodei, Dario and McCandlish, Sam and Sutskever, Ilya and Zaremba, Wojciech},
|
75 |
+
journal = {arXiv preprint arXiv:2107.03374},
|
76 |
+
year = {2021},
|
77 |
+
note = {OpenAI Codex paper; introduced HumanEval benchmark}
|
78 |
+
}
|
79 |
+
|
80 |
+
@article{pai2024codocbench,
|
81 |
+
title = {{CoDocBench}: A Dataset for Code-Documentation Alignment in Software Maintenance},
|
82 |
+
author = {Pai, Kunal and Devanbu, Premkumar and Ahmed, Toufique},
|
83 |
+
journal = {arXiv preprint arXiv:2502.00519},
|
84 |
+
year = {2024}
|
85 |
+
}
|
86 |
+
|
87 |
+
@inproceedings{kamienski2021pysstubs,
|
88 |
+
title = {{PySStuBs}: Characterizing Single-Statement Bugs in Popular Open-Source Python Projects},
|
89 |
+
author = {Kamienski, Arthur V. and Palechor, Luisa and Bezemer, Cor-Paul and Hindle, Abram},
|
90 |
+
booktitle = {IEEE/ACM International Conference on Mining Software Repositories (MSR)},
|
91 |
+
year = {2021}
|
92 |
+
}
|
93 |
+
|
94 |
+
@article{brown2020language,
|
95 |
+
title={Language models are few-shot learners},
|
96 |
+
author={Brown, Tom and Mann, Benjamin and Ryder, Nick and Subbiah, Melanie and Kaplan, Jared D and Dhariwal, Prafulla and Neelakantan, Arvind and Shyam, Pranav and Sastry, Girish and Askell, Amanda and others},
|
97 |
+
journal={Advances in neural information processing systems},
|
98 |
+
volume={33},
|
99 |
+
pages={1877--1901},
|
100 |
+
year={2020}
|
101 |
+
}
|
102 |
+
|
103 |
+
@inproceedings{devlin2019bert,
|
104 |
+
title={Bert: Pre-training of deep bidirectional transformers for language understanding},
|
105 |
+
author={Devlin, Jacob and Chang, Ming-Wei and Lee, Kenton and Toutanova, Kristina},
|
106 |
+
booktitle={Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: human language technologies, volume 1 (long and short papers)},
|
107 |
+
pages={4171--4186},
|
108 |
+
year={2019}
|
109 |
+
}
|
110 |
+
|
111 |
+
@article{raffel2020exploring,
|
112 |
+
title={Exploring the limits of transfer learning with a unified text-to-text transformer},
|
113 |
+
author={Raffel, Colin and Shazeer, Noam and Roberts, Adam and Lee, Katherine and Narang, Sharan and Matena, Michael and Zhou, Yanqi and Li, Wei and Liu, Peter J},
|
114 |
+
journal={Journal of machine learning research},
|
115 |
+
volume={21},
|
116 |
+
number={140},
|
117 |
+
pages={1--67},
|
118 |
+
year={2020}
|
119 |
+
}
|
120 |
+
|
121 |
+
@article{dorri2018multi,
|
122 |
+
title={Multi-agent systems: A survey},
|
123 |
+
author={Dorri, Ali and Kanhere, Salil S and Jurdak, Raja},
|
124 |
+
journal={Ieee Access},
|
125 |
+
volume={6},
|
126 |
+
pages={28573--28593},
|
127 |
+
year={2018},
|
128 |
+
publisher={IEEE}
|
129 |
+
}
|
130 |
+
|
131 |
+
@book{wooldridge2009introduction,
|
132 |
+
title={An introduction to multiagent systems},
|
133 |
+
author={Wooldridge, Michael},
|
134 |
+
year={2009},
|
135 |
+
publisher={John wiley \& sons}
|
136 |
+
}
|
137 |
+
|
138 |
+
@article{boiko2023emergent,
|
139 |
+
title={Emergent autonomous scientific research capabilities of large language models},
|
140 |
+
author={Boiko, Daniil A and MacKnight, Robert and Gomes, Gabe},
|
141 |
+
journal={arXiv preprint arXiv:2304.05332},
|
142 |
+
year={2023}
|
143 |
+
}
|
144 |
+
|
145 |
+
@inproceedings{gaston2005agenta,
|
146 |
+
title={Agent-organized networks for dynamic team formation},
|
147 |
+
author={Gaston, Matthew E and DesJardins, Marie},
|
148 |
+
booktitle={Proceedings of the fourth international joint conference on Autonomous agents and multiagent systems},
|
149 |
+
pages={230--237},
|
150 |
+
year={2005}
|
151 |
+
}
|
152 |
+
|
153 |
+
@misc{zhou2023agents,
|
154 |
+
title={Agents: An Open-source Framework for Large Language Model based Autonomous Agents},
|
155 |
+
author={Wangchunshu Zhou and Jianshu Chen and Jialong Wu and Yiheng Xu and Kexin Wang and Jintian Zhang and Yuan Gao and Zhiyong Wu and Kevin Tian and Yubo Feng and Linyi Yang and Bokai Quan and Cong Yu and Yuhang Wang and Shishen Lan and Yan Wang and Hong-Cheng Guo and Chaoyu Chen and Tianxiang Sun and Jin Xiong and Yi Lu and Peng Li and Lichao Sun and Lifan Yuan and Hang Li and Xiangang Li},
|
156 |
+
year={2023},
|
157 |
+
eprint={2309.07870},
|
158 |
+
archivePrefix={arXiv},
|
159 |
+
primaryClass={cs.AI},
|
160 |
+
url={https://arxiv.org/abs/2309.07870}
|
161 |
+
}
|
162 |
+
|
163 |
+
|
164 |
+
@misc{openai_func_calling,
|
165 |
+
title = {Function calling},
|
166 |
+
author = {{OpenAI}},
|
167 |
+
year = {2023},
|
168 |
+
howpublished = {OpenAI API Documentation},
|
169 |
+
url = {https://platform.openai.com/docs/guides/function-calling},
|
170 |
+
note = {Accessed: 2025-05-01}
|
171 |
+
}
|
172 |
+
|
173 |
+
@misc{wang2023voyager,
|
174 |
+
title={{Voyager}: An Open-Ended Embodied Agent with Large Language Models},
|
175 |
+
author={Guanzhi Wang and Yuqi Xie and Yunfan Jiang and Ajay Mandlekar and Chaowei Xiao and Yuke Zhu and Linxi Fan and Anima Anandkumar},
|
176 |
+
year={2023},
|
177 |
+
eprint={2305.16291},
|
178 |
+
archivePrefix={arXiv},
|
179 |
+
primaryClass={cs.AI},
|
180 |
+
url={https://arxiv.org/abs/2305.16291}
|
181 |
+
}
|
182 |
+
|
183 |
+
@book{russell2010artificial,
|
184 |
+
title={Artificial intelligence: a modern approach},
|
185 |
+
author={Russell, Stuart J. and Norvig, Peter},
|
186 |
+
year={2010},
|
187 |
+
edition={3rd},
|
188 |
+
publisher={Prentice Hall Press},
|
189 |
+
address={Upper Saddle River, NJ, USA}
|
190 |
+
}
|
191 |
+
|
192 |
+
@article{shoham1994agent,
|
193 |
+
author = {Yoav Shoham},
|
194 |
+
title = {Agent-oriented programming},
|
195 |
+
journal = {Artificial Intelligence},
|
196 |
+
volume = {60},
|
197 |
+
number = {1},
|
198 |
+
pages = {51--92},
|
199 |
+
year = {1993},
|
200 |
+
publisher = {Elsevier}
|
201 |
+
}
|
202 |
+
|
203 |
+
@misc{wang2023survey,
|
204 |
+
title={A Survey on Large Language Model based Autonomous Agents},
|
205 |
+
author={Lei Wang and Chen Ma and Xueyang Feng and Zeyu Zhang and Hao Yang and Jingsen Zhang and Zhiyuan Chen and Jiakai Tang and Xu Chen and Yankai Lin and Wayne Xin Zhao and Zhewei Wei and Ji-Rong Wen},
|
206 |
+
year={2023},
|
207 |
+
eprint={2308.11432},
|
208 |
+
archivePrefix={arXiv},
|
209 |
+
primaryClass={cs.AI}
|
210 |
+
}
|
211 |
+
|
212 |
+
@misc{xi2023rise,
|
213 |
+
title={The Rise and Potential of Large Language Model Based Agents: A Survey},
|
214 |
+
author={Zhiheng Xi and Wenxiang Chen and Xin Guo and Wei He and Yiwen Ding and Boyang Hong and Ming Zhang and Junzhe Wang and Senjie Jin and Enyu Zhou and Rui Zheng and Xiaoran Fan and Xiao Wang and Limao Xiong and Linyi Yang and Ting Ruan and Yongquan Yang and Peng Li and Yitao Chang and Yanlin Wang},
|
215 |
+
year={2023},
|
216 |
+
eprint={2309.07864},
|
217 |
+
archivePrefix={arXiv},
|
218 |
+
primaryClass={cs.AI}
|
219 |
+
}
|
220 |
+
|
221 |
+
|
222 |
+
@inproceedings{park2023generative,
|
223 |
+
author = {Park, Joon Sung and O'Brien, Joseph C. and Cai, Carrie J. and Morris, Meredith Ringel and Liang, Percy and Bernstein, Michael S.},
|
224 |
+
title = {Generative Agents: Interactive Simulacra of Human Behavior},
|
225 |
+
year = {2023},
|
226 |
+
isbn = {9798400701320},
|
227 |
+
publisher = {Association for Computing Machinery},
|
228 |
+
address = {New York, NY, USA},
|
229 |
+
url = {https://doi.org/10.1145/3586183.3606763},
|
230 |
+
doi = {10.1145/3586183.3606763},
|
231 |
+
booktitle = {The 36th Annual ACM Symposium on User Interface Software and Technology (UIST '23)},
|
232 |
+
pages = {1–22},
|
233 |
+
numpages = {22},
|
234 |
+
location = {San Francisco, CA, USA},
|
235 |
+
series = {UIST '23}
|
236 |
+
}
|
237 |
+
|
238 |
+
@misc{ollama,
|
239 |
+
title = {Ollama},
|
240 |
+
author = {{Ollama Team}},
|
241 |
+
howpublished = {\url{https://ollama.com/}},
|
242 |
+
year = {2023},
|
243 |
+
note = {Accessed: 2025-05-01}
|
244 |
+
}
|
245 |
+
|
246 |
+
@misc{anthropic2024claude,
|
247 |
+
title = {The {Claude 3} Model Family: {Opus, Sonnet, Haiku}},
|
248 |
+
author = {{Anthropic}},
|
249 |
+
year = {2024},
|
250 |
+
month = {March},
|
251 |
+
howpublished = {Model Card},
|
252 |
+
url = {https://www-cdn.anthropic.com/de8ba9b01c9ab7cbabf5c33b80b7bbc618857627/Model_Card_Claude_3.pdf},
|
253 |
+
note = {Accessed: 2025-05-01}
|
254 |
+
}
|
255 |
+
|
256 |
+
@misc{openai2023gpt4,
|
257 |
+
title={GPT-4 Technical Report},
|
258 |
+
author={OpenAI},
|
259 |
+
year={2023},
|
260 |
+
eprint={2303.08774},
|
261 |
+
archivePrefix={arXiv},
|
262 |
+
primaryClass={cs.CL},
|
263 |
+
url={https://arxiv.org/abs/2303.08774}
|
264 |
+
}
|
265 |
+
|
266 |
+
@misc{langgraph,
|
267 |
+
title = {LangGraph: A Framework for Agentic Workflows},
|
268 |
+
author = {LangChain},
|
269 |
+
year = {2024},
|
270 |
+
howpublished= {\url{https://www.langchain.com/langgraph}},
|
271 |
+
note = {Accessed: May 1, 2025}
|
272 |
+
}
|
273 |
+
|
274 |
+
@book{clearwater1996market,
|
275 |
+
title = {Market-Based Control: A Paradigm for Distributed Resource Allocation},
|
276 |
+
editor = {Scott H. Clearwater},
|
277 |
+
publisher = {World Scientific},
|
278 |
+
year = {1996}
|
279 |
+
}
|
280 |
+
|
281 |
+
@article{valckenaers2005trends,
|
282 |
+
title={Guest Editors' Introduction: Intelligent Control in the Manufacturing Supply Chain},
|
283 |
+
author={McFarlane, Duncan and Mar{\'\i}k, Vladim{\'\i}r and Valckenaers, Paul},
|
284 |
+
journal={IEEE Intelligent Systems},
|
285 |
+
volume={20},
|
286 |
+
number={1},
|
287 |
+
pages={24--26},
|
288 |
+
year={2005},
|
289 |
+
publisher={IEEE}
|
290 |
+
}
|
291 |
+
|
292 |
+
|
293 |
+
@article{horling2004survey,
|
294 |
+
title={A survey of multi-agent organizational paradigms},
|
295 |
+
author={Horling, Bryan and Lesser, Victor},
|
296 |
+
journal={The Knowledge engineering review},
|
297 |
+
volume={19},
|
298 |
+
number={4},
|
299 |
+
pages={281--316},
|
300 |
+
year={2004},
|
301 |
+
publisher={Cambridge University Press}
|
302 |
+
}
|
303 |
+
|
304 |
+
|
305 |
+
@inproceedings{gaston2005agentb,
|
306 |
+
title={Agent-organized networks for multi-agent production and exchange},
|
307 |
+
author={Gaston, Matthew E and DesJardins, Marie},
|
308 |
+
booktitle={Proceedings of the 20th national conference on Artificial intelligence-Volume 1},
|
309 |
+
pages={77--82},
|
310 |
+
year={2005}
|
311 |
+
}
|
312 |
+
|
313 |
+
@misc{zhang2023building,
|
314 |
+
title={Building Cooperative Embodied Agents Modularly with Large Language Models},
|
315 |
+
author={Hongxin Zhang and Weihua Du and Jiaming Shan and Qinhong Zhou and Yilun Du and Joshua B. Tenenbaum and Tianmin Shu and Chuang Gan},
|
316 |
+
year={2023},
|
317 |
+
eprint={2307.02485},
|
318 |
+
archivePrefix={arXiv},
|
319 |
+
primaryClass={cs.AI}
|
320 |
+
}
|
321 |
+
|
322 |
+
@misc{parisi2022talm,
|
323 |
+
title={TALM: Tool Augmented Language Models},
|
324 |
+
author={Aaron Parisi and Yao Zhao and Noah Fiedel},
|
325 |
+
year={2022},
|
326 |
+
eprint={2205.12255},
|
327 |
+
archivePrefix={arXiv},
|
328 |
+
primaryClass={cs.CL}
|
329 |
+
}
|
330 |
+
|
331 |
+
|
332 |
+
@misc{crewai,
|
333 |
+
title = {CrewAI},
|
334 |
+
author = {{CrewAI Inc.}},
|
335 |
+
year = {2025},
|
336 |
+
howpublished = {\url{https://www.crewai.com/}},
|
337 |
+
note = {Accessed: 2025-05-01}
|
338 |
+
}
|
339 |
+
|
340 |
+
|
341 |
+
@article{qian2023communicative,
|
342 |
+
title={Chatdev: Communicative agents for software development},
|
343 |
+
author={Qian, Chen and Liu, Wei and Liu, Hongzhang and Chen, Nuo and Dang, Yufan and Li, Jiahao and Yang, Cheng and Chen, Weize and Su, Yusheng and Cong, Xin and others},
|
344 |
+
journal={arXiv preprint arXiv:2307.07924},
|
345 |
+
year={2023}
|
346 |
+
}
|
347 |
+
|
348 |
+
@article{wang2023decision,
|
349 |
+
title={Decision-making driven by driver intelligence and environment reasoning for high-level autonomous vehicles: a survey},
|
350 |
+
author={Wang, Yuning and Jiang, Junkai and Li, Shangyi and Li, Ruochen and Xu, Shaobing and Wang, Jianqiang and Li, Keqiang},
|
351 |
+
journal={IEEE Transactions on Intelligent Transportation Systems},
|
352 |
+
volume={24},
|
353 |
+
number={10},
|
354 |
+
pages={10362--10381},
|
355 |
+
year={2023},
|
356 |
+
publisher={IEEE}
|
357 |
+
}
|
358 |
+
|
359 |
+
@misc{wen2024benchmarkingcomplexinstructionfollowingmultiple,
|
360 |
+
title={Benchmarking Complex Instruction-Following with Multiple Constraints Composition},
|
361 |
+
author={Bosi Wen and Pei Ke and Xiaotao Gu and Lindong Wu and Hao Huang and Jinfeng Zhou and Wenchuang Li and Binxin Hu and Wendy Gao and Jiaxin Xu and Yiming Liu and Jie Tang and Hongning Wang and Minlie Huang},
|
362 |
+
year={2024},
|
363 |
+
eprint={2407.03978},
|
364 |
+
archivePrefix={arXiv},
|
365 |
+
primaryClass={cs.CL},
|
366 |
+
url={https://arxiv.org/abs/2407.03978},
|
367 |
+
}
|
368 |
+
|