Name: Towards AI Legal Name: Towards AI, Inc. Description: Towards AI is the world's leading artificial intelligence (AI) and technology publication. Read by thought-leaders and decision-makers around the world. Phone Number: +1-650-246-9381 Email: pub@towardsai.net
228 Park Avenue South New York, NY 10003 United States
Website: Publisher: https://towardsai.net/#publisher Diversity Policy: https://towardsai.net/about Ethics Policy: https://towardsai.net/about Masthead: https://towardsai.net/about
Name: Towards AI Legal Name: Towards AI, Inc. Description: Towards AI is the world's leading artificial intelligence (AI) and technology publication. Founders: Roberto Iriondo, , Job Title: Co-founder and Advisor Works for: Towards AI, Inc. Follow Roberto: X, LinkedIn, GitHub, Google Scholar, Towards AI Profile, Medium, ML@CMU, FreeCodeCamp, Crunchbase, Bloomberg, Roberto Iriondo, Generative AI Lab, Generative AI Lab VeloxTrend Ultrarix Capital Partners Denis Piffaretti, Job Title: Co-founder Works for: Towards AI, Inc. Louie Peters, Job Title: Co-founder Works for: Towards AI, Inc. Louis-François Bouchard, Job Title: Co-founder Works for: Towards AI, Inc. Cover:
Towards AI Cover
Logo:
Towards AI Logo
Areas Served: Worldwide Alternate Name: Towards AI, Inc. Alternate Name: Towards AI Co. Alternate Name: towards ai Alternate Name: towardsai Alternate Name: towards.ai Alternate Name: tai Alternate Name: toward ai Alternate Name: toward.ai Alternate Name: Towards AI, Inc. Alternate Name: towardsai.net Alternate Name: pub.towardsai.net
5 stars – based on 497 reviews

Frequently Used, Contextual References

TODO: Remember to copy unique IDs whenever it needs used. i.e., URL: 304b2e42315e

Resources

Our 15 AI experts built the most comprehensive, practical, 90+ lesson courses to master AI Engineering - we have pathways for any experience at Towards AI Academy. Cohorts still open - use COHORT10 for 10% off.

Publication

Why Your Software Development Life Cycle Will Not Work for Your AI Agents (And How to Change That)
Latest   Machine Learning

Why Your Software Development Life Cycle Will Not Work for Your AI Agents (And How to Change That)

Last Updated on October 13, 2025 by Editorial Team

Author(s): Gowtham Boyina

Originally published on Towards AI.

Classical software development is deterministic. You code, you test, you deploy, and the result — when provided the same input — is deterministic. The sequence of logic is predictable, and the failure modes.

AI-agents do not work under those rules. They do not reason procedurally but probabilistically. Two same prompts will have actions branching off in different ways. The same tool calls will have different results. “Works on my machine” is no longer valid where the machine itself learns, changes, and re-interprets context.

The Software Development Lifecycle (SDLC) falls apart under these circumstances since it presumes static behavior and fixed logic. Enterprise AI agents need an altogether different discipline — one where behavior and reasoning are treated as measurable, governable, and changing over time.

This is the Agent Development Lifecycle (ADLC).

Why SDLC Doesn’t Work for AI Agents

1. Deterministic vs. Probabilistic Behavior

SDLC assumes identical inputs yield identical outputs, enabling static testing and regression validation.
Agents operate probabilistically. The same input can lead to different reasoning paths, tool invocations, and conclusions. This unpredictability invalidates SDLC’s binary “pass/fail” testing model. ADLC replaces test cases with evaluation frameworks that track distributions of performance — measuring hallucination, groundedness, bias, and safety instead of correctness alone.

2. Static Logic vs. Adaptive Systems

Traditional software remains static until redeployed. Agents adapt dynamically through reinforcement, retraining, and continuous feedback. SDLC’s update cycle cannot accommodate continuous adaptation.
ADLC embeds Runtime Optimization Loops, allowing live tuning of prompts, models, and tools based on real-world telemetry.

3. Implementation-Centric vs. Evaluation-First

SDLC defines success by implementation fidelity.
ADLC defines success by behavioral outcomes — does the agent achieve enterprise KPIs under defined risk limits?
An agent with imperfect code but consistent, auditable performance passes ADLC’s standard; SDLC would incorrectly flag it as defective.

4. Closed Testing vs. Continuous Certification

SDLC treats testing as a phase. ADLC treats it as a perpetual discipline.
Agents must undergo ongoing behavioral certification: periodic red teaming, bias audits, compliance verification, and retraining. Governance catalogs replace static QA reports; each agent release carries lineage, eval metrics, and risk documentation.

5. Security by Code vs. Security by Behavior

SDLC security mitigates vulnerabilities in code.
ADLC security mitigates vulnerabilities in behavior — prompt injection, tool misuse, memory poisoning, and goal hijacking.
Agents demand sandboxed execution, cryptographic identity issuance, and MCP Gateway-level governance, enforcing runtime policy and throttling untrusted operations.

6. Linear Model vs. Continuous Loops

SDLC progresses linearly (Plan → Build → Test → Deploy → Maintain).
ADLC operates as a cyclic framework, embedding two continuous loops:

  • Experimentation Loop between Build and Test for evaluation-driven improvement.
  • Runtime Optimization Loop between Operate and Monitor for live adaptation.
    Agent systems evolve — therefore, the lifecycle must evolve continuously too.

7. Limited Governance vs. Total Accountability

SDLC documentation ends at deployment sign-off.
ADLC integrates governance, lineage, and risk management into every phase.
Every agent has an owner, version, authority boundary, and audit trail. Its decisions can be reconstructed and its performance linked to measurable enterprise value.

The ADLC Framework

ADLC extends DevSecOps through six interdependent phases, governed by experimentation and runtime optimization loops.

Why Your Software Development Life Cycle Will Not Work for Your AI Agents (And How to Change That)
from IBM’s Guide to Architecting Secure Enterprise AI Agents with MCP — Verified by Anthropic

1. Plan

Define use cases, measurable KPIs, and acceptable risk boundaries. Establish behavioral specifications in natural language and map them to evaluation criteria — accuracy, trust, compliance, latency, safety.
Generate or synthesize datasets for testing. Plan governance gates: leadership approval, ethical standards, and escalation procedures for agent autonomy.

2. Code & Build

Develop prompt logic, memory mechanisms, and orchestration frameworks. Integrate with enterprise APIs and MCP tools.

  • Implement hybrid model architectures (frontier + domain models).
  • Enforce secure-by-design practices: sandboxing, least privilege, and traceable agent identities.
  • Instrument observability hooks for reasoning traces, tool calls, and output lineage.

Deliverables include prompts-as-code, tool schemas, memory policies, and model manifests — all version-controlled.

3. Test, Optimize, Release

Validate not only function but behavior. Testing now includes:

  • Behavioral evals (groundedness, bias, hallucination).
  • Red teaming for adversarial robustness.
  • Compliance verification via LLM-as-a-Judge (LLM-aaJ) frameworks.
    Agents that pass all evaluation thresholds are certified into governed catalogs with metadata (version, owner, lineage, and policy adherence).

4. Deploy

Deploy into hybrid or multi-cloud environments under strict security and compliance controls.

  • Integrate MCP Gateways for policy enforcement and observability.
  • Enable multi-agent orchestration for collaborative workflows.
  • Implement progressive rollout, rollback, and kill-switch mechanisms.
    Deployment is a governance milestone, not a handoff.

5. Operate

Run agents as governed digital assets. Maintain catalogs of all active agents with their tools, permissions, and audit trails.
Perform continuous compliance audits, fairness analysis, and operational tuning.
Securely retire deprecated agents with full data lineage and evaluation history preserved.

6. Monitor & Runtime Optimization Loop

Establish real-time observability pipelines capturing telemetry across accuracy, cost, latency, and drift metrics.
Detect anomalies like goal manipulation, memory poisoning, or unauthorized tool calls.
Feed insights back into the optimization loop to refine prompts, retrain models, and rebalance resource allocation dynamically.

Monitoring’s core question shifts from “Is it running?” to “Is it correct, safe, and aligned?”.

Comparative Structure: SDLC vs. ADLC

Structural Outcome

SDLC builds applications that execute; ADLC builds agents that reason.
SDLC delivers stability. ADLC delivers alignment.
The two share DevSecOps roots, but ADLC introduces a new operational logic: every phase is observable, reversible, and empirically measured.

Traditional DevOps pipelines cannot govern agents that change behavior mid-run, use external tools autonomously, or make contextual decisions without explicit code paths.

ADLC integrates security, compliance, and evaluation into a self-correcting lifecycle — an adaptive feedback system that mirrors the adaptive nature of agents themselves.

The New Discipline

AI agents are not code — they are evolving systems that reason, act, and adapt. Treating them as software artifacts leads to operational failure, security gaps, and regulatory risk.

The Agent Development Lifecycle (ADLC) is not a framework trend — it’s the operational backbone for the next era of enterprise systems.

It acknowledges the probabilistic nature of intelligence, embeds evaluation into every stage, and extends DevSecOps into domains of reasoning and autonomy.

The organizations that adopt ADLC will not just build smarter software — they will operate governed, evolving intelligence that aligns with business intent.

Attribution:
Adapted from IBM’s “Guide to Architecting Secure Enterprise AI Agents with MCP,” verified by Anthropic, TechXchange 2025.

Access the Document here: Click on this

Join thousands of data leaders on the AI newsletter. Join over 80,000 subscribers and keep up to date with the latest developments in AI. From research to projects and ideas. If you are building an AI startup, an AI-related product, or a service, we invite you to consider becoming a sponsor.

Published via Towards AI


Take our 90+ lesson From Beginner to Advanced LLM Developer Certification: From choosing a project to deploying a working product this is the most comprehensive and practical LLM course out there!

Towards AI has published Building LLMs for Production—our 470+ page guide to mastering LLMs with practical projects and expert insights!


Discover Your Dream AI Career at Towards AI Jobs

Towards AI has built a jobs board tailored specifically to Machine Learning and Data Science Jobs and Skills. Our software searches for live AI jobs each hour, labels and categorises them and makes them easily searchable. Explore over 40,000 live jobs today with Towards AI Jobs!

Note: Content contains the views of the contributing authors and not Towards AI.