Last week, we explored AI Maturity Stage 1: Ad Hoc Use, where curiosity sparks innovation but also introduces risks like shadow IT and fragmented workflows. We discussed how leaders can observe grassroots experimentation and begin laying the groundwork for responsible growth.
This week, we pause our stage-by-stage journey to explore a critical leadership trait that underpins every phase of AI maturity: agility.
What Does Agility Look Like in AI Leadership?
Agility in AI leadership isn’t just about pivoting quickly or reacting to change, it’s about creating space for participation, curiosity, and collaboration. It’s about framing AI integration as a shared journey, not a top-down directive.
In many organizations, we’ve seen early AI initiatives unfold behind closed doors. Pilot projects are launched, but participation is limited to preselected teams. Tools like ChatGPT or Claude are made available, but without guidance, training, or clarity around use cases. Employees propose ideas, but they’re declined due to predefined priorities.
This isn’t just a missed opportunity—it’s a signal.
When innovation is restricted, engagement suffers. When strategy lacks transparency, adoption stalls. And when leaders frame AI as a technical rollout rather than a cultural shift, they risk building systems that no one trusts.
Follow Sakara Digital for weekly insights
Practical strategies for AI leadership, digital transformation, and fractional support.
Inclusive Leadership Is Agile Leadership
Agility means more than adapting strategy, it means remaining open to diverse input and cross-functional collaboration. It means:
- Welcoming volunteers from across the organization
- Creating channels for bottom-up ideas
- Structuring experimentation without stifling it
- Framing AI as a capability that grows with people, not apart from them
Professor Valentine’s model reminds us that GenAI is a horizontal technology, it touches every function. That means every function should have a voice in how it’s explored, deployed, and evaluated.
Framing, Structuring, and Evaluating in Real Time
Unlike predictive analytics, generative AI evolves rapidly and often unpredictably. Leaders must be willing to frame, structure, and evaluate AI integration as it unfolds—not just at the beginning.
This requires:
- Framing AI initiatives with clear intent and shared language
- Structuring participation to include diverse roles and perspectives
- Evaluating impact iteratively, with feedback loops and ethical oversight
Agility isn’t reactive, it’s relational. It’s about building trust, not just tools.
From Ad Hoc to Localized Projects
As organizations move from Stage 1 to Stage 2, Localized Projects, agility becomes even more critical. Teams begin building use-case-specific tools, and agentic systems start to appear in tightly scoped deployments. Leaders must track what’s being built, document patterns, and support both citizen developers and domain experts.
But without inclusive framing, even the most promising pilots can falter.
Next week, we’ll explore how to structure experimentation in Stage 2—balancing open-ended exploration with strategic focus, and laying the groundwork for scalable, ethical AI systems.
Teaser for Next Week:
Localized Projects are where GenAI begins to take root. But scaling success requires more than enthusiasm, it requires structure. We’ll explore how leaders can support pilots, build internal champions, and prepare for programmatic growth.
“When it comes to AI, early decisions shape long-term adoption. Leaders who embrace inclusive, cross-functional participation build more than tech—they build trust.”
This article was created in collaboration with GenAI and shaped by intentional human insight.
Further Reading
- Agentic AI Is Already Changing the Workforce. Harvard Business Review
- Rewired Pharma Companies Will Win in the Digital Age. McKinsey
#FractionalConsulting #LifeSciences #DigitalTransformation #AI #Leadership








Your perspective matters—join the conversation.