Governance in AI Agents Surveys and Experiments

Authors

  • Abhas Bali UG Student, Department of Computing Technologies, SRM Institute of Science and Technology, Chennai, India. Author
  • Neeraj Nair Department of Computing Technologies, SRM Institute of Science and Technology, Chennai, India. Author
  • N Sumukesh UG Student, Department of Computing Technologies, SRM Institute of Science and Technology, Chennai, India. Author
  • Sivakumar B Associate Professor, Department of Computing Technologies, SRM Institute of Science and Technology, Chennai, India. Author
  • Saswat Singh Senior Engineer, ZS Associates, Maharashtra, India. Author

DOI:

https://doi.org/10.62647/

Keywords:

Autonomous AI agents, Agent governance, Human-in-the-loop (HITL), Risk management, Ethical AI, Policy enforcement, Decision transparency, Compliance monitoring, Behavioral oversight, Sandboxed simulation

Abstract

The agent governance is an increasingly important domain of inquiry. Its main objective is to establish powerful checks and balances to make sure that autonomous AI agents act as planned, they comply with ethical standards, and they are in line with human and organizational interests. The field is faced with numerous complex problems: how to control these agents, how to monitor the behavior of the agents, how to ensure protection, and how to define their place in a larger social phenomenon. The people use numerous tools-strict rules, ongoing monitoring, and risk awareness to detect issues like bias or the non-adherence of the expected performance of the agents. What is the challenge? The abilities of the agents are advancing at a speed which is beyond the evolution of the current governance systems. To fill gaps in theory and the lack of sound empirical evidence, we are developing an extensive system architecture that makes governance take an active part in all stages of the AI lifecycle, and each evaluated using clear and specific criteria. The engine of our methodology includes a real-time compliance check with the ability to test in a sandbox and a bidirectional data flow, which facilitates the implementation of it on a large scale. We include multi-agent coordination, considering human and organization aspects, including organizational culture and stakeholder interest, and creating feedback loops, which start at creation and continue until system retirement. In this way we complete the gap between theory and practically viable solutions and offer the actionable solutions suitable to practice in a number of areas.

Downloads

Download data is not yet available.

Downloads

Published

25-03-2026

How to Cite

Governance in AI Agents Surveys and Experiments. (2026). International Journal of Information Technology and Computer Engineering, 14(1), 576-583. https://doi.org/10.62647/

Similar Articles

1-10 of 968

You may also start an advanced similarity search for this article.