October 2025 Showcase: LangGraph 1.0, Agent Protocols, and AI Cancer Research
Our October 2025 monthly showcase was packed with cutting-edge presentations on LangGraph 1.0 middleware, agent-to-agent communication protocols, groundbreaking cancer detection research, and local LLM inference strategies. With over 60 attendees and 6 first-time visitors, the energy was electric as our community explored the future of AI agent architectures.
LangGraph 1.0 Alpha: Middleware Revolutionβ
Colin McNamara kicked off the evening with an exploration of LangGraph 1.0's new middleware architecture, demonstrating how three simple hooks are transforming agent development:
The Three Middleware Hooksβ
-
before_model - Control what enters your agent
- Manage context and token limits automatically
- Implement guardrails with Nemo Guardrails
- Jump to different nodes based on input
-
model - Optimize model execution
- Swap models on-the-fly based on query complexity
- Implement caching layers for cost reduction
- Control tool availability and output formats
-
after_model - Validate and route outputs
- Implement circuit breakers for unsafe outputs
- Add human-in-the-loop at critical decision points
- Collect metrics for OpenTelemetry
Real-World Impactβ
Colin shared his production use case in food manufacturing, where LangGraph middleware enables:
- Automatic processing of EDI orders into QuickBooks
- Human approval for capacity-exceeding orders
- State persistence with rollback capabilities
- Native OpenTelemetry logging for observability
With companies like JPMorgan Chase, NuvoBank, and LinkedIn using LangGraph in production, the 1.0 stable release marks a significant milestone for enterprise AI applications.
Read the full LangGraph middleware deep dive β
A2A and AP2: The Future of Agent Commerceβ
Ryan Booth delivered a fascinating exploration of Google's Agent-to-Agent (A2A) protocol and the emerging Agent Payment Protocol (AP2), complete with a live demo of his Corvara marketplace platform.
Key Insightsβ
A2A enables agent discovery: Just as SEO helps humans find content, A2A allows AI agents to discover and negotiate with service providers. Businesses must now optimize for "Agent Engine Optimization" (AEO) to remain discoverable in an AI-first world.
AP2 enables automated transactions: The new payment protocol allows agents to complete financial transactions within chat interfaces, bringing us closer to fully automated service marketplaces.
The Corvara Demoβ
Ryan's prototype demonstrated:
- Automatic agent discovery based on capabilities
- Contract negotiation and pricing comparison
- Integration with Codex and Claude orchestrators
- Human-in-the-loop approval points for payments
The vision? A marketplace where agents bid on software development projects, negotiate contracts, complete work, and process paymentsβall with minimal human intervention.
Explore the A2A/AP2 protocols in detail β
AI-Powered Cancer Detection: High School Research at Stanfordβ
In one of the evening's most inspiring presentations, high school student Venika Kakarla shared her Stanford research on hepatocellular carcinoma (liver cancer) using spatial data analysis and machine learning.
Research Highlightsβ
Working with 1.4 million cells across 124+ tumor regions, Venika employed:
- Direct and indirect interaction studies to map cell distributions around blood vessels
- Microvessel density (MVD) analysis using 8Γ8 grid-based spatial partitioning
- Voronoi tessellation clustering to identify dispersion patterns
- Python-based statistical analysis with Matplotlib, Pandas, and SciPy
Future: Graph Attention Networksβ
Venika is now developing a Graph Attention Network (GAT) that combines:
- Spatial clustering data
- Her previous thermal stability research
- Clinical patient information
- Treatment history
The goal: Predict cancer recurrence probability and progression risk with unprecedented accuracy.
Her journey from Google Colab experiments (reducing 2,500 hours of compute to 5 minutes) to Stanford's School of Medicine demonstrates the democratizing power of AI tools in scientific research.
Read Venika's full research presentation β
LLM Inference: Choosing the Right Providerβ
Dmitri Iourovitski concluded the evening with a comprehensive comparison of local LLM inference providers, cutting through the confusion with practical guidance.
The Three Main Optionsβ
Ollama/Llama.cpp: Best for local development and quick demos
- β Easiest setup, no GPU required
- β No batch processing, opaque memory management
- Use when: Prototyping on macOS or CPU-only systems
Hugging Face Transformers: Best for flexibility and latest models
- β Fine-grained control, Flash Attention support since 2023
- β Day-one access to new model architectures
- β Steep learning curve, manual memory management
- Use when: Production with customization needs
VLLM: Best for high-throughput production
- β Paged attention, automatic batching, enterprise-grade
- β Complex setup, requires high-end GPUs (12GB+ VRAM)
- Use when: Scaling to multiple users and concurrent requests
Performance Numbersβ
Dmitri's OCR benchmark (20 pages, Q1 2.5 VL 7B model):
- Ollama on 3090: 35 minutes
- Transformers on 3090: 8 minutes (despite more aggressive quantization!)
The difference? Flash Attention and proper optimization.
Dive into the full inference provider comparison β
Community Highlightsβ
- 60+ attendees with 6 first-time visitors joining the AIMUG family
- Student mentorship showcasing world-class research from high school to Stanford
- Production deployments sharing real-world use cases in manufacturing, finance, and healthcare
- Open knowledge sharing with all presentations documented and available to the community
What's Nextβ
Upcoming Eventsβ
- Tuesday Office Hours: Every Tuesday at 5 PM Central on Google Meet
- Hacky Hour: Check Discord/Meetup for the next social coding session
- November Showcase: First Wednesday of November - mark your calendars!
Get Involvedβ
- πΊ Watch the full recording
- π Explore the documentation
- π¬ Join our Discord
- π RSVP on Meetup
Final Thoughtsβ
October's showcase demonstrated the rapid maturation of AI agent architectures. From production-ready middleware patterns to agent marketplaces to life-saving medical research, our community is at the forefront of practical AI innovation.
Whether you're building production agents with LangGraph, exploring agent protocols, conducting cutting-edge research, or optimizing inference performance, there's a place for you in the AIMUG community.
See you at the next showcase!
Want to present at a future event? Reach out on Discord or email us. We're always looking for community members to share their learning and innovations.