The LLM 'Ops' Stack: Prompt Management, Evaluation, and Observability in 2023
This article maps the emerging LLMOps ecosystem in late 2023, examining specialized tools for prompt management, observability, evaluation, and safety testing that address unique challenges of operating production LLM systems. It explores platforms like PromptLayer, Helicone, and Braintrust while analyzing how teams are adapting DevOps practices for non-deterministic AI systems, highlighting the skills gap and architectural patterns forming as LLM applications scale.
8/28/202313 min read


When software moved from desktop applications to web services, an entire ecosystem of operational tools emerged: monitoring, logging, error tracking, performance analysis, A/B testing. DevOps and SaaS operations became distinct disciplines with specialized tools—Datadog, New Relic, Sentry, PagerDuty, LaunchDarkly—addressing challenges unique to running production services at scale.
We're watching a parallel evolution for large language models. As LLMs move from research curiosities and demos to production applications serving millions of users, teams are discovering that operating these systems at scale requires an entirely new operational toolkit. The "LLMOps" or "AI Ops" stack is crystallizing rapidly, and by late August 2023, the contours are becoming clear.
Why Traditional Ops Tools Don't Suffice
Traditional software operates deterministically. The same input produces the same output. Bugs are reproducible. Performance is measurable with clear metrics—latency, throughput, error rates. Testing involves checking that outputs match expectations.
LLMs break these assumptions fundamentally. The same prompt can produce different outputs across runs (with temperature > 0). "Correctness" is often subjective or context-dependent. Performance includes dimensions like relevance, coherence, and helpfulness that defy simple metrics. Edge cases are infinite—you can't enumerate all possible prompts.
This creates operational challenges that traditional tools don't address:
Prompt versioning and management: In traditional software, code is versioned. With LLMs, prompts are code—they define system behavior. But prompts are natural language strings scattered across codebases, often modified by non-engineers (product managers, domain experts). How do you version, test, and deploy prompt changes systematically?
Quality evaluation: How do you know if a prompt change improved output quality? Unlike code changes where test suites verify correctness, LLM outputs require human judgment for many tasks. Scaling this evaluation is non-trivial.
Cost and latency monitoring: LLM API calls cost money per token and have variable latency. Applications need to track costs per user, per feature, and identify expensive prompts. Traditional APM tools don't understand LLM-specific costs.
Debugging and observability: When an LLM produces unexpected output, understanding why requires seeing the full prompt (including any dynamic context), model parameters, and output. Traditional logs don't capture this structured information.
Safety and content moderation: LLMs can generate harmful content. Production systems need monitoring for policy violations, automated content filtering, and human review workflows.
Model comparison and selection: With multiple models available (GPT-4, Claude, open-source alternatives) at different price points, teams need frameworks for comparing performance and choosing appropriately.
These challenges have spawned an emerging ecosystem of specialized tools.
Prompt Management and Versioning
The simplest operational challenge is managing prompts themselves. Early LLM applications hard-coded prompts in application code—scattered string literals modified by developers. This quickly becomes unmaintainable.
PromptLayer was among the first dedicated prompt management tools. It provides a centralized repository for prompts, version history, and the ability to test prompts against multiple models simultaneously. Non-technical team members can modify prompts in a web interface, with changes tracked and deployable through proper release processes.
The value becomes clear at scale. One startup reported having 47 different prompts across their application, modified by product managers, engineers, and domain experts. Without centralized management, tracking which prompt version was live, who changed what, and why became impossible.
Humanloop takes a similar approach, positioning itself as "version control and collaboration for prompts." The platform provides a prompt IDE where teams can develop, test, and deploy prompts. It includes evaluation tools for comparing prompt versions and tracking performance over time.
LangChain provides programmatic prompt management. While primarily an orchestration framework, LangChain includes abstractions for prompt templates, few-shot examples, and dynamic prompt construction. For developers comfortable with code, it offers more flexibility than GUI-based tools.
The emerging pattern resembles content management systems (CMS) for traditional software. Just as CMSs separated content from code, prompt management tools separate prompts from application logic. This enables faster iteration, better governance, and clearer ownership.
Key features emerging across these tools:
Version control: Track every prompt change with diffs and rollback capability
A/B testing: Deploy multiple prompt variants and measure performance
Templating: Define prompts with variables that get populated dynamically
Multi-model testing: Test the same prompt across GPT-4, Claude, and other models
Collaboration: Enable non-engineers to modify prompts with appropriate review processes
Analytics: Track which prompts are used most, cost per prompt, and performance metrics
Several teams report that adopting prompt management tools reduced time-to-deploy prompt improvements from days (requiring engineer time) to hours (product managers updating prompts directly).
Observability and Logging
Understanding LLM behavior in production requires specialized observability. Traditional logs capture events but don't structure LLM-specific information—full prompts, model responses, token counts, costs, latency.
Helicone focuses on OpenAI API observability. It acts as a proxy between your application and OpenAI, capturing every request and response with rich metadata. The platform provides dashboards showing costs over time, latency distributions, common prompts, and the ability to drill into specific requests.
The debugging value is substantial. When users report unexpected behavior, developers can search Helicone for that user's requests, see the exact prompt sent (including any dynamic context), view the model response, and understand what happened. This visibility dramatically reduces debugging time.
Weights & Biases (W&B), known for ML experiment tracking, has expanded into LLM observability. Their platform captures prompts, completions, embeddings, and metadata, organizing them into traces that show multi-step LLM workflows. For complex applications involving chains of LLM calls, W&B helps visualize the flow and identify bottlenecks.
LangSmith, LangChain's observability companion, provides tracing specifically designed for LangChain applications. Since LangChain orchestrates complex workflows (chains, agents, tools), LangSmith traces show how data flows through these components, which LLM calls succeed or fail, and where latency accumulates.
Portkey offers a unified observability layer across multiple LLM providers. If your application uses both OpenAI and Anthropic, Portkey provides consistent monitoring and analytics regardless of provider. This is valuable for teams exploring multiple models or implementing fallback strategies.
The observability pattern emerging resembles distributed tracing in microservices. Complex LLM applications involve multiple calls—retrieval, summarization, generation, evaluation—and understanding the full workflow requires seeing how these steps connect.
Key capabilities across observability platforms:
Request/response logging: Capture full prompts and completions with timestamps
Cost tracking: Calculate spend per user, per feature, per model with granular attribution
Latency monitoring: Measure and alert on response times
Error tracking: Identify failed API calls, rate limits, content policy violations
User session replay: Reconstruct user interactions to debug reported issues
Search and filtering: Find requests by user, prompt content, time range, or other dimensions
One engineering lead shared that observability tools revealed their application was making redundant LLM calls due to a caching bug—costing them $3,000 monthly that simple logging hadn't surfaced. The structured visibility into LLM operations paid for itself immediately.
Evaluation and Testing
Perhaps the hardest LLMOps challenge is evaluating whether outputs are good. Traditional unit tests check exact output matches. LLM outputs vary, and "correctness" often requires judgment.
LangChain evaluation modules provide frameworks for systematic testing. You define test cases (input prompts and expected characteristics), run your LLM application, and evaluate outputs against criteria. The framework supports both automated metrics (e.g., checking that outputs contain specific keywords) and human evaluation workflows.
PromptLayer's A/B testing enables comparing prompt variants systematically. Deploy two prompts to random user subsets, collect outputs, and evaluate which performs better. The platform facilitates human evaluation—showing raters outputs from both variants without revealing which is which.
Humanloop's evaluation suite emphasizes human-in-the-loop evaluation. The platform presents outputs to human evaluators, collects ratings, and calculates statistical significance of differences between prompt or model variants. This systematic approach replaces ad-hoc testing with rigorous methodology.
OpenAI Evals (open-sourced by OpenAI) provides a framework for creating and running evaluations. Teams define evaluation datasets with input/output pairs and scoring functions. The framework runs prompts against these datasets and reports performance metrics. It's become a de facto standard for sharing evaluation benchmarks.
Scale AI offers human evaluation as a service. For teams needing evaluation at scale, Scale provides access to trained evaluators who assess LLM outputs against defined criteria. This is expensive but higher quality than crowdsourced evaluation for complex tasks.
The evaluation challenge has multiple dimensions:
Automated metrics work for some tasks. If you're extracting structured data, you can check accuracy automatically. If generating code, you can check if it compiles and passes tests. But for open-ended generation—writing, conversation, creative content—automated metrics are limited.
Human evaluation provides ground truth but doesn't scale. Having humans rate every output is prohibitively expensive. Teams use sampling strategies—evaluate a representative subset and extrapolate.
LLM-as-judge is an emerging pattern. Use a capable LLM (often GPT-4) to evaluate outputs from your application LLM. Research suggests GPT-4 evaluations correlate reasonably with human judgments for many tasks. This provides scalable, consistent evaluation at lower cost than human raters.
Regression testing prevents degradation. As you modify prompts or change models, regression tests ensure quality doesn't decrease. Teams are building evaluation datasets from production examples—real user inputs and human-approved outputs—using them as regression test suites.
Several teams report that systematic evaluation revealed prompt improvements they assumed were beneficial actually decreased output quality. Without measurement, "improvements" were based on intuition rather than evidence.
A/B Testing and Experimentation
Once you can evaluate LLM outputs, you need frameworks for running experiments—testing whether prompt changes, model switches, or parameter adjustments improve performance.
Statsig and LaunchDarkly, traditional feature flagging platforms, now support LLM experimentation. Statsig's LLM metrics track costs, latency, and can integrate custom quality metrics. Teams use feature flags to gradually roll out prompt changes, monitor impact, and rollback if metrics degrade.
Braintrust positions itself specifically as an LLM evaluation and experimentation platform. It provides datasets for evaluation, tools for running experiments comparing prompt or model variants, and statistical analysis determining which variant performs better with confidence intervals.
The experimentation workflow mirrors traditional A/B testing but with LLM-specific considerations:
Define metrics: Cost, latency, and quality measures (automated or human-evaluated)
Create variants: Different prompts, models, or parameters
Deploy experiment: Route users to variants randomly
Collect data: Gather outputs and metric measurements
Evaluate: Determine which variant performs better statistically
Deploy winner: Roll out the better variant to all users
The challenge is sample size. Quality evaluation often requires human rating, which is expensive. Teams typically run experiments with smaller samples (100-500 examples) using human evaluation, then deploy winners and monitor production metrics to confirm improvements hold at scale.
Prompt optimization tools automate parts of this process. Some platforms use reinforcement learning or evolutionary algorithms to generate and test prompt variants automatically, searching for versions that maximize defined metrics. This is nascent but promising for structured tasks with clear success criteria.
One product team shared that systematic experimentation revealed counter-intuitive findings. Shorter prompts they assumed would save costs actually decreased quality enough to increase support tickets, making them more expensive overall. Without measurement, they'd have optimized the wrong metric.
Red Teaming and Safety Testing
As discussed in our previous red teaming article, systematically testing LLMs for safety issues, policy violations, and adversarial attacks is critical for production deployment.
HuggingFace Red Team provides a platform for adversarial testing. Teams can run automated attacks attempting to elicit harmful outputs, test against known jailbreak techniques, and document findings. The platform helps teams proactively identify vulnerabilities before they're exploited in production.
Arthur AI focuses on model monitoring including bias detection and fairness testing. The platform can test LLM outputs across demographic dimensions, identifying potential bias in model responses that could create legal or reputational risk.
Anthropic's Model Diff (not a commercial product but a pattern) involves systematically comparing new model versions against previous versions for capability changes, including potentially dangerous capabilities. This regression testing for safety helps ensure model updates don't introduce new risks.
Robust Intelligence offers adversarial testing specifically for AI systems. Their platform attempts prompt injections, tests for data leakage, and validates that models behave safely under adversarial conditions. This security-first approach helps teams deploy with confidence.
The red teaming workflow typically involves:
Threat modeling: Identify potential harms (misinformation, toxic content, privacy violations, etc.)
Attack generation: Create prompts designed to elicit harmful behaviors
Automated testing: Run attacks at scale using test frameworks
Human red teaming: Expert testers attempt creative attacks automation might miss
Mitigation: Develop defenses (prompt engineering, output filtering, fine-tuning)
Continuous monitoring: Watch for novel attacks in production
Several startups have hired dedicated red team roles—security researchers, domain experts, and creative adversarial thinkers who probe for weaknesses full-time. This investment reflects the reputational and regulatory risks of deploying unsafe LLM systems.
Model Selection and Management
With multiple LLM providers and models available at different capabilities and price points, teams need frameworks for choosing and managing models.
OpenRouter provides a unified API across 20+ LLM providers. Instead of integrating with OpenAI, Anthropic, Cohere, and others separately, teams call OpenRouter's API and specify which model to use. This abstraction enables easy model switching and A/B testing across providers.
Martian offers similar multi-provider access with added intelligence—automatically routing requests to the best model based on the task, falling back to alternatives if primary models are unavailable, and optimizing for cost or latency based on your preferences.
Model registries are emerging in larger organizations. Teams maintain catalogs of approved models with documentation of capabilities, costs, latency characteristics, and appropriate use cases. This helps prevent every team evaluating models independently.
The model selection decision involves multiple dimensions:
Capability: Can the model handle your task's complexity?
Cost: What's the price per 1,000 tokens?
Latency: What's the typical response time?
Context window: How much context does your application require?
Availability: What's the uptime and rate limit?
Privacy: Does the provider use data for training?
Safety: How well does it handle adversarial inputs?
Teams are developing sophisticated routing strategies. Use GPT-3.5 for simple tasks, GPT-4 for complex ones. Try Claude for very long contexts. Fall back to open-source models if API providers are down. This requires operational infrastructure to manage model selection programmatically.
The Emerging LLMOps Stack Architecture
By late August 2023, a reference architecture is emerging:
Development layer: Prompt management tools (PromptLayer, Humanloop) where teams develop and version prompts. Integration with orchestration frameworks (LangChain, LlamaIndex) for building application logic.
Evaluation layer: Testing frameworks (OpenAI Evals, Braintrust) for systematic quality assessment. Human evaluation services (Scale AI) for ground truth. LLM-as-judge systems for scalable evaluation.
Deployment layer: Feature flags (Statsig, LaunchDarkly) for gradual rollout. Model routing (OpenRouter, Martian) for provider abstraction and fallbacks.
Observability layer: Request/response logging (Helicone, LangSmith) for visibility. Cost and latency monitoring dashboards. Error tracking and alerting.
Safety layer: Red teaming platforms (HuggingFace) for adversarial testing. Content moderation and filtering. Bias and fairness monitoring (Arthur AI).
Analytics layer: Usage analytics, cost attribution, performance tracking. A/B test analysis and statistical significance testing.
This stack resembles mature SaaS operations with LLM-specific adaptations. The tools are newer and less mature than traditional DevOps tools, but the patterns are familiar to anyone who's operated production services.
Adoption Patterns and Challenges
Despite available tools, LLMOps adoption lags application development. Many teams are building LLM applications without proper operational infrastructure, accumulating technical debt that will become painful as systems scale.
Cost is a barrier for startups. Many LLMOps tools charge monthly fees or per-request pricing. For early-stage startups already paying for expensive LLM APIs, additional operational tooling can feel like luxury rather than necessity. This changes when systems reach scale and operational issues become acute.
Complexity creates hesitation. The LLMOps stack involves multiple tools—prompt management, observability, evaluation—each requiring setup and integration. Teams struggling to ship products deprioritize operational excellence until problems force attention.
Maturity concerns are valid. These tools are mostly less than a year old. APIs change, companies pivot, and long-term viability is uncertain. Teams hesitate to build dependencies on young startups.
Build versus buy decisions favor building initially. Engineers often implement basic logging and evaluation internally before adopting third-party tools. This works at small scale but becomes unmaintainable as systems grow.
The teams that have adopted LLMOps tools report high value:
Faster debugging: Structured logging reduces debugging time from hours to minutes
Cost savings: Visibility into spending identifies expensive prompts and optimization opportunities
Quality improvements: Systematic evaluation enables confident prompt improvements
Risk reduction: Red teaming and safety monitoring prevent embarrassing public incidents
Velocity: Proper tooling enables faster iteration and deployment
One engineering director estimated that LLMOps tools improved their team's productivity by 30%—saving more in engineering time than the tools cost.
Open Source vs. Commercial
The LLMOps ecosystem includes both open-source and commercial offerings:
Open source: LangChain, LangSmith (freemium), OpenAI Evals, LiteLLM, various logging libraries. These provide foundational capabilities and community-driven development.
Commercial: PromptLayer, Humanloop, Helicone (freemium), Braintrust, Scale AI, Arthur AI. These offer polished UIs, managed infrastructure, and enterprise features.
The pattern mirrors broader software tooling. Open source provides base functionality and community-driven standards. Commercial tools add convenience, scalability, and support.
Many teams start with open source, adding commercial tools as operational needs grow. LangChain for orchestration, then LangSmith for observability. OpenAI Evals for basic testing, then Braintrust for sophisticated experimentation.
What Enterprise Customers Need
Enterprise adoption requires capabilities that current LLMOps tools often lack:
On-premise deployment: Regulated industries can't send data to third-party observability platforms. They need deployable solutions running in their infrastructure.
Governance and compliance: Audit trails, access controls, data retention policies, and compliance certifications (SOC 2, HIPAA, etc.).
Integration with existing tooling: Enterprises have established DevOps stacks. LLMOps tools must integrate with existing monitoring (Datadog), incident management (PagerDuty), and workflow tools (Jira).
Multi-tenancy: Large organizations need to support multiple teams, projects, and environments (dev, staging, production) with appropriate isolation and resource allocation.
Advanced analytics: Custom metrics, complex queries, and integration with business intelligence platforms.
Several LLMOps vendors are building enterprise offerings, but the market is early. Expect significant development in 2024 as enterprises move beyond pilots to production LLM systems.
The Skills Gap
Operating LLMs requires new skills that bridge traditional software engineering and AI/ML understanding:
Prompt engineering: Understanding how to construct effective prompts, debug prompt issues, and optimize for quality and cost.
LLM evaluation: Designing evaluation frameworks, conducting human evaluation studies, and interpreting results.
Cost optimization: Understanding token costs, implementing caching strategies, and optimizing prompt length without sacrificing quality.
Safety and red teaming: Thinking adversarially about potential misuse and implementing appropriate safeguards.
Observability: Instrumenting applications for LLM-specific monitoring and debugging production issues.
These skills are rare. Most software engineers lack ML background. Most ML engineers lack production operations experience. The people who combine both are highly sought and well-compensated.
Job postings for "LLM Engineer," "Prompt Engineer," and "AI Ops Engineer" have proliferated over the past months, with salaries ranging from $150K to $400K+ depending on experience and company.
Looking Forward
The LLMOps stack will mature rapidly over the next 12-18 months:
Consolidation: The current landscape has dozens of point solutions. Expect consolidation as successful tools expand scope or get acquired. Winners will offer broader platforms rather than single-feature tools.
Standards emergence: Common formats for prompt storage, evaluation datasets, and observability data will reduce fragmentation and enable tool interoperability.
Deeper provider integration: OpenAI, Anthropic, and others will build more operational features into their platforms—native A/B testing, built-in evaluation, better observability. This may commoditize some third-party tools.
Better automation: Manual evaluation and prompt engineering will become increasingly automated. Reinforcement learning from human feedback (RLHF) applied to prompt optimization, automated red teaming, and AI-powered observability analysis.
Enterprise focus: As the market matures beyond early adopters, tools will need enterprise features—security, compliance, sophisticated access controls, and integration with existing enterprise infrastructure.
Vertical specialization: Generic LLMOps tools will face competition from vertical-specific solutions. Healthcare LLMOps with HIPAA compliance, financial services LLMOps with regulatory focus, etc.
For teams building LLM applications today, the strategic question isn't whether to adopt LLMOps practices but when. Early in development, manual processes and basic logging suffice. As systems scale and stakes rise, proper operational infrastructure becomes essential.
The pattern we've seen repeatedly in technology: early adopters build custom tooling, best practices crystallize, commercial tools mature, and operational excellence becomes table stakes for competitive products.
We're early in this evolution. The LLMOps stack of late 2023 will look primitive compared to what emerges over the next few years. But the fundamental needs—managing prompts, evaluating quality, monitoring costs, ensuring safety—will persist and grow more critical as LLMs become infrastructure that powers mission-critical applications.
The era of LLM applications running without proper operational infrastructure is ending. The era of treating LLM operations as a proper engineering discipline is beginning. The tools are here. The practices are emerging. The question is whether teams will adopt them proactively or wait until operational crises force their hand.

