June 17, 2025

From R&D to Production: Our GenAI Journey at AWS Summit

My AWS Summit talk has been published! It was great to reflect on Lexer's journey from doing early R&D on GenAI back in 2023, to deploying into production in early 2024, to optimizing and leveraging the latest and greatest in 2024. Here's what we learned along the way.

The Journey Overview

It was great to reflect on Lexer's journey from doing early R&D on GenAI back in 2023 ➡️ Deploying into production early 2024 ➡️ Optimizing and leveraging the latest and greatest in 2024.

This timeline represents more than just a technology adoption story—it's a case study in how to thoughtfully integrate emerging AI capabilities into production systems while maintaining the reliability and performance that customers depend on.

Each phase of this journey taught us different lessons about the practical realities of building with GenAI, from the experimental excitement of early research to the sobering challenges of production deployment to the ongoing work of optimization and improvement.

Phase 1: Early R&D (2023)

Looking back at 2023, the GenAI landscape was both incredibly exciting and wildly uncertain. We were among the early adopters exploring how large language models could enhance our customer data platform.

What we explored:

  • Natural language querying of customer data
  • Automated insight generation from behavioral patterns
  • Content personalization using AI-driven segmentation
  • Conversational interfaces for data exploration

Key lessons from the R&D phase:

Start with clear use cases: The most promising applications were those that solved specific, well-defined problems rather than trying to "add AI" generally.

Prototype extensively: We built numerous small prototypes to understand what worked, what didn't, and where the technology's sweet spots were.

Focus on user experience: The most technically impressive AI features were useless if they didn't improve the actual user experience.

Plan for variability: Early GenAI outputs were inconsistent, so we had to design systems that could handle and improve upon that variability.

Phase 2: Production Deployment (Early 2024)

Moving from R&D prototypes to production systems in early 2024 required solving an entirely different set of challenges. This phase was about reliability, scalability, and maintaining the quality standards our customers expected.

Production challenges we solved:

Latency and performance: Research prototypes could take 30+ seconds to generate insights. Production systems needed sub-5-second response times.

Cost management: API costs that were negligible during R&D became significant at scale. We had to implement intelligent caching, batching, and prompt optimization.

Quality assurance: We developed robust testing and validation systems to ensure AI-generated content met our quality standards.

Error handling: Production systems needed graceful degradation when AI services were unavailable or returned unexpected results.

Security and privacy: Customer data handling required careful attention to data residency, encryption, and audit trails.

Technical Architecture Decisions

Some key architectural decisions that enabled our successful production deployment:

Hybrid approach: We combined multiple AI models rather than relying on a single solution. Different models excel at different tasks.

Intelligent caching: We implemented sophisticated caching strategies to reduce API costs and improve response times while maintaining result freshness.

Fallback systems: Every AI-powered feature had a non-AI fallback to ensure system reliability.

Gradual rollout: We used feature flags and gradual rollouts to monitor performance and user adoption before full deployment.

Monitoring and observability: We built comprehensive monitoring for AI service performance, costs, and quality metrics.

Phase 3: Optimization and Latest Capabilities (2024)

With production systems stable, 2024 became about optimization and leveraging the rapidly evolving AI landscape to deliver even better capabilities.

Areas of optimization:

Model efficiency: We continuously evaluated new models and techniques to improve performance while reducing costs.

User experience refinement: Based on real usage data, we refined interfaces and workflows to better match how customers actually wanted to interact with AI features.

Advanced capabilities: We integrated newer capabilities like multimodal processing and more sophisticated reasoning as they became available.

Custom model training: We began training custom models on specific use cases where generic models weren't optimal.

What We Got Right

Reflecting on the journey, several strategic decisions proved particularly valuable:

Early investment in R&D: Starting exploration in 2023 gave us a significant head start in understanding the technology's practical applications.

Customer-centric approach: We focused on solving real customer problems rather than implementing AI for its own sake.

Incremental deployment: Rather than betting everything on AI, we integrated capabilities gradually where they added the most value.

Investment in infrastructure: Building robust monitoring, testing, and deployment infrastructure paid dividends as we scaled.

Team education: We invested heavily in upskilling our team to understand both the capabilities and limitations of GenAI.

Challenges and Lessons Learned

Not everything went smoothly. Here are some key challenges and what we learned from them:

Managing expectations: Early demos created unrealistic expectations. We learned to be more transparent about AI limitations and set appropriate expectations.

Cost management: Initial cost projections were way off. We had to rapidly develop sophisticated cost optimization strategies.

Quality consistency: Maintaining consistent output quality required much more engineering effort than anticipated.

Integration complexity: Integrating AI capabilities into existing systems was more complex than building greenfield applications.

Regulatory considerations: As regulations evolved, we had to adapt our implementations to maintain compliance.

The Competitive Advantage

Moving quickly through this journey created several competitive advantages:

Customer insights: Real-world usage data helped us understand what customers actually wanted from AI-powered features.

Technical expertise: Our team developed deep practical knowledge of building production AI systems.

Market positioning: Being early to market with reliable AI features helped differentiate us from competitors.

Continuous improvement: The feedback loops from production usage enabled rapid iteration and improvement.

Looking Forward

As I reflected during the AWS Summit talk, this journey is far from over. The AI landscape continues to evolve rapidly, and new capabilities are emerging constantly.

What's next for us:

  • Deeper personalization using advanced reasoning capabilities
  • Multimodal processing for richer customer insights
  • More sophisticated automation of routine tasks
  • Better integration between AI and traditional analytics
  • Continued focus on cost optimization and performance

Advice for Others Starting the Journey

For companies beginning their own GenAI journey, here's what I'd recommend based on our experience:

Start with experimentation: Build small prototypes to understand the technology before committing to large implementations.

Focus on specific problems: Don't try to "add AI everywhere." Identify specific use cases where AI can provide clear value.

Invest in infrastructure early: Good monitoring, testing, and deployment systems are essential for production success.

Plan for costs: AI services can be expensive at scale. Factor in cost optimization from the beginning.

Educate your team: Success requires understanding both the capabilities and limitations of the technology.

Stay customer-focused: The goal is to solve customer problems better, not to implement the latest technology.

The Broader Impact

This journey has fundamentally changed how we think about product development and customer experience. AI isn't just a feature we've added—it's become a lens through which we evaluate all potential improvements.

The combination of early research, careful production deployment, and continuous optimization has positioned us well in an increasingly AI-driven market. But more importantly, it's enabled us to deliver genuinely better experiences for our customers.

The AWS Summit talk was a great opportunity to share these learnings with the broader community and reflect on how far we've come in just over a year. The journey from R&D to production optimization has been challenging, exciting, and ultimately rewarding.

For anyone considering their own GenAI journey, the message is clear: start now, start small, but start. The learning curve is steep, but the potential rewards for both your business and your customers are significant.

Aaron Wallis is a founder and technology leader who positions himself as a "Technical-Creative Bridge." With over two decades of experience building technology that bridges creative vision with technical excellence.

Share this article

More Articles

AI Reveals What Actually Requires Thinking: Reflections on the Future of Development

Exploring Tim O'Reilly's insights on how AI is transforming programming - not killing it, but revealing what truly requires human thinking.

Read More

When AI Becomes Your Best Referral Partner: The New Era of Organic Discovery

How AI recommendations are reshaping business discovery and what this means for the future of SEO and organic search.

Read More

From R&D to Production: Our GenAI Journey at AWS Summit

Reflecting on Lexer's journey from early GenAI research in 2023 to production deployment and optimization, as presented at AWS Summit.

Read More

When Conversation Becomes the Interface: The GenAI UX Revolution

How GenAI chat agents are fundamentally changing our UI/UX expectations and why traditional forms are starting to feel obsolete.

Read More

GEM: The New SEO for the GenAI Era

How Generative Engine Marketing is reshaping digital discovery, with insights from Vercel's success getting 10% of signups from ChatGPT.

Read More

Vibe Coding: The Self-Driving Car for Development

How LLMs are transforming development by handling the tedious work, freeing developers to focus on architecture and decisions that actually matter.

Read More

Authentic Selves: Reflections on Pride, Community, and the Retail Industry

A personal reflection on diversity, inclusion, and the privilege of showing up authentically in professional spaces during Pride Month.

Read More

Being a Verb: The Power of Action in a World of Overthinking

Explore the concept of 'being a verb' and why prioritizing action over overthinking leads to better outcomes.

Read More

AI Strategy: Balancing Innovation with Implementation

Practical considerations for implementing AI in your organization without getting lost in the hype.

Read More

From COVID to AI: The Next Phase of Social Distancing

How AI might create a more profound shift in human interaction than COVID social distancing, further abstracting our connections.

Read More

On Avalanches and Snowballs

A reflection on balancing short-term and long-term projects in product teams, using debt payment strategies as an analogy.

Read More

A Recipe for Resilience in Pride Month and Beyond

A personal reflection on maintaining pride and resilience during difficult times, with a meaningful connection to Pride Month celebrations.

Read More

War and Peace

An exploration of how businesses operate either 'at war' or 'at peace' and how this mindset affects their strategies and operations.

Read More