If you're looking for a replacement for Langfuse, Humanloop is a good option. It's a collaborative environment for building and optimizing LLM applications, with a prompt management system with version control, an evaluation and monitoring tool, and model fine-tuning tools. Humanloop integrates with major LLM providers and offers Python and TypeScript SDKs for integration. It also offers a free tier for prototyping and an enterprise tier for more advanced use cases.
Another contender is Langtail. Langtail is a suite of tools for debugging, testing and deploying LLM prompts. It includes features like prompt fine-tuning, testing, deploying prompts as API endpoints and monitoring production performance. Langtail also offers a no-code playground for writing and running prompts, adjustable parameters and logging. The service is available in three pricing tiers, including a free tier for small businesses and solopreneurs.
If you're looking for a more powerful service with serious security controls, Vellum is worth a look. Vellum offers tools for prompt engineering, semantic search, prompt chaining, evaluation and monitoring. It's geared for enterprise-scale operations and offers SOC 2 Type II and HIPAA compliance. Vellum is designed to be used for a variety of use cases, including document analysis, chatbots and workflow automation.
Last is LastMile AI, a full-stack developer platform designed to help you productionize generative AI applications. It offers features like Auto-Eval for prompt optimization, an RAG Debugger to improve performance and a notebook-like environment for prototyping. LastMile AI supports multiple AI models and has a range of integration options, making it a good choice for those who want to deploy production-grade generative AI applications.