| Abstract: |
AI agents are rapidly evolving from generating passive text to autonomously executing complex real-world tasks, such as booking travel, managing social media, and performing enterprise operations. Leading organizations including OpenAI, Google, Microsoft, and Salesforce are driving this transformation, positioning AI agents as the next major wave of generative AI technologies. As these systems gain autonomy and capability, effective security measurements and governance becomes essential to ensure their actions promote beneficial outcomes and prevent harm. This paper introduces CertAI, a certification framework designed to ensure the security, privacy, ethics, robustness, transparency and fairness of autonomous AI agents. CertAI generates verifiable certificates from structured metadata, embedding ethical, security and other parameters to include trust scores, risk levels, incident counts, and domain-specific compliance status. The framework is complemented by CertAI-Bench, a benchmark that systematically populates metadata through probing AI agents across multiple dimensions such as security, ethics, fairness, transparency, and others. Together, these components enable scalable certification, transparent evaluation, and continuous monitoring of autonomous AI systems, establishing a foundation for secure, responsible and verifiable AI governance. The certification process determines numerical values of the dimensions utilizing different models and calculates trust score. The results show that the agents based on larger models provide stronger safety guarantees, but fairness and transparency remain the weakest dimensions across all AI agents. |