Product strategy, experimentation, and context-as-code for work AI
Product & Work AI Playbook
The Evolving Landscape of Work AI: Integrating Experimentation, Agentic Systems, and Context-as-Code
The enterprise AI landscape is rapidly transforming, driven by breakthroughs in autonomous, agentic AI systems, structured workflows, and a renewed emphasis on maintaining contextual integrity. For organizations like Glean, which focus on work-centric AI solutions, these developments present both significant opportunities and complex challenges. As the field advances, a key theme emerges: the necessity of combining rigorous experimentation practices with emerging AI capabilities—particularly agentic AI—to build reliable, trustworthy, and scalable products that truly serve organizational needs.
Building on a Foundation: Data, Context, and Systematic Prompt Engineering
At the core of effective enterprise AI is the principle of treating data and organizational context as a product. Glean’s success has been rooted in deep integrations with enterprise tools such as email, document repositories, Slack, and Teams. These integrations create high switching costs and foster trustworthy proprietary data, which feeds into AI models to deliver relevant and accurate insights.
A key evolution in this space is the concept of "context as code"—a disciplined approach to managing organizational knowledge through version-controlled, structured formats. This approach ensures that prompt engineering—the craft of designing AI prompts—is treated like software development, with templates, explicit variables, and iterative refinement. Such rigor not only enhances reproducibility and scalability but also builds trust in AI outputs, especially as models become more autonomous.
Complementing this is the rise of systematic prompt engineering, transforming ad hoc prompts into reusable, versioned templates. This discipline enables scaling AI capabilities across workflows while reducing variability and improving experiment reliability.
Furthermore, organizations are increasingly deploying Cultural Coding Index (CCI) metrics—tools designed to detect and quantify biases embedded within datasets and models. With AI systems influencing critical decisions, responsible AI practices demand awareness of cultural and societal biases. Incorporating CCI into the experimentation cycle allows teams to measure impact not just quantitatively but also from an inclusive and ethical perspective, fostering trustworthy AI.
The Rise of Autonomous, Agentic AI Systems in the Workplace
Recent innovations highlight a surge in autonomous, agentic AI systems capable of multi-step, goal-oriented behaviors that emulate human decision-making in complex workflows. Notable examples include:
- Perplexity’s Perplexity Computer, functioning as a digital worker that orchestrates tasks across 19 different AI models to execute sophisticated workflows.
- Zavi AI, a voice-driven OS that enables users to initiate, modify, and execute actions within applications via natural language, exemplifying fluid, natural interaction.
- Microsoft’s CORPGEN architecture, which introduces hierarchical planning and persistent memory, empowering AI to manage long-term, multi-horizon tasks—a crucial capability for strategic enterprise automation.
These advancements imply that AI agents will increasingly operate autonomously, managing multi-step workflows, long-term objectives, and complex decision pathways. For Glean and similar organizations, this underscores the importance of structured context pipelines—the backbone for multi-model orchestration—to manage, propagate, and adapt organizational knowledge reliably and securely.
Reimagining Experimentation, Observability, and Governance
The integration of agentic AI into enterprise workflows necessitates new approaches to experimentation and control:
- Experimentation frameworks must evolve beyond traditional A/B tests, accommodating autonomous behaviors and multi-step interactions. This could involve simulation environments or scenario-based testing to evaluate AI decision pathways.
- Observability tools need to trace decision pathways, detect failures, and measure biases at every step, ensuring transparency and control over AI actions.
- Governance protocols are critical to oversee automation boundaries, enforce ethical standards, and ensure regulatory compliance. As AI agents operate with increasing independence, defining control points and audit trails becomes essential.
A key concept emerging here is the "context pipeline"—a structured, programmable workflow that propagates organizational knowledge through multi-model orchestration. These pipelines enable enterprise control and security, while also serving as bias mitigation tools, ensuring AI systems remain aligned with organizational values.
Transitioning to an AI-Native Product Operating Model
To effectively harness these technological shifts, organizations must adopt an AI-native product operation model that emphasizes fast validation, iterative development, and scalable deployment. This means moving beyond pilots toward continuous integration of AI features, supported by versioned prompt templates, automated testing, and impact assessments that incorporate bias metrics like CCI.
This approach enables teams to rapidly validate hypotheses, refine models, and deploy autonomous workflows confidently, reducing time-to-market and improving product reliability.
UX and Adoption: Lowering Barriers with Natural, Non-Coder Workflows
The rise of natural language interfaces—such as voice commands and conversational AI—lowers barriers for non-technical users. As exemplified by tools like Zavi AI, users can initiate complex workflows without coding, enabling more inclusive adoption and fostering new patterns of collaboration in the workplace.
This shift not only democratizes AI access but also changes user expectations, emphasizing seamless, intuitive interactions and context-aware automation that align with daily work routines.
Tactical Roadmap for the Future
Building on these insights, enterprise AI teams should prioritize:
- Developing version-controlled, modular prompt templates to ensure reproducibility and scalability.
- Integrating cultural bias metrics like CCI into experiment design and impact evaluations.
- Building context pipelines capable of managing multi-step, multi-model workflows that support hierarchical planning and persistent memory.
- Enhancing observability and governance tools to trace decision pathways, detect biases, and enforce ethical standards.
- Promoting cross-disciplinary collaboration with cultural experts, ethicists, and product managers to embed inclusive, responsible AI practices from the ground up.
Implications and Current Status
The confluence of systematic prompt engineering, cultural bias quantification, and autonomous AI infrastructure signals a paradigm shift in enterprise work AI. Organizations like Glean are well-positioned to lead this evolution by adapting their product strategies to prioritize structured context management, multi-model orchestration, and robust governance.
As AI systems become more autonomous and integrated into daily workflows, maintaining trust, transparency, and ethical standards is paramount. The future of work AI hinges on balancing automation with oversight, mitigating biases, and building scalable, responsible solutions that enhance productivity while respecting societal values.
In summary, the current landscape demands a holistic approach—one that combines technical rigor, ethical foresight, and user-centric design—to realize the full potential of AI in transforming the digital workplace.