Home

Awesome

Langtrace

Open Source Observability for LLM applications

🌐 Open Source & Open Telemetry(OTEL) Observability for LLM applications

Static Badge NPM Typescript SDK PyPI Python SDK Downloads NPM Trace Attributes PyPI Trace Attributes Static Badge Deploy on Railway


Langtrace is an open source observability software which lets you capture, debug and analyze traces and metrics from all your applications that leverages LLM APIs, Vector Databases and LLM based Frameworks.

image

πŸ“Š Open Telemetry Support

The traces generated by Langtrace adhere to Open Telemetry Standards(OTEL). We are developing semantic conventions for the traces generated by this project. You can checkout the current definitions in this repository. Note: This is an ongoing development and we encourage you to get involved and welcome your feedback.


πŸ“¦ SDK Repositories


πŸš€ Getting Started

Langtrace Cloud ☁️

To use the managed SaaS version of Langtrace, follow the steps below:

  1. Sign up by going to this link.
  2. Create a new Project after signing up. Projects are containers for storing traces and metrics generated by your application. If you have only one application, creating 1 project will do.
  3. Generate an API key by going inside the project.
  4. In your application, install the Langtrace SDK and initialize it with the API key you generated in the step 3.
  5. The code for installing and setting up the SDK is shown below:

If your application is built using typescript/javascript

npm i @langtrase/typescript-sdk
import * as Langtrace from '@langtrase/typescript-sdk' // Must precede any llm module imports
Langtrace.init({ api_key: <your_api_key> })

OR

import * as Langtrace from "@langtrase/typescript-sdk"; // Must precede any llm module imports
LangTrace.init(); // LANGTRACE_API_KEY as an ENVIRONMENT variable

If your application is built using python

pip install langtrace-python-sdk
from langtrace_python_sdk import langtrace
langtrace.init(api_key=<your_api_key>)

OR

from langtrace_python_sdk import langtrace
langtrace.init() # LANGTRACE_API_KEY as an ENVIRONMENT variable

🏠 Langtrace self hosted

To run the Langtrace locally, you have to run three services:

[!IMPORTANT] Checkout our documentation for various deployment options and configurations.

Requirements:

The .env file

Feel free to modify the .env file to suit your needs.

Starting the servers

docker compose up

The application will be available at http://localhost:3000.

Take down the setup

To delete containers and volumes

docker compose down -v

-v flag is used to delete volumes

Telemetry

Langtrace collects basic, non-sensitive usage data from self-hosted instances by default, which is sent to a central server (via PostHog).

The following telemetry data is collected by us:

This data helps us to:

No sensitive information is gathered, and the data is not shared with third parties.

If you prefer to disable telemetry, you can do so by setting TELEMETRY_ENABLED=false in your configuration.


πŸ”— Supported integrations

Langtrace automatically captures traces from the following vendors:

VendorTypeTypescript SDKPython SDK
OpenAILLM:white_check_mark::white_check_mark:
AnthropicLLM:white_check_mark::white_check_mark:
Azure OpenAILLM:white_check_mark::white_check_mark:
CohereLLM:white_check_mark::white_check_mark:
xAILLM:white_check_mark::white_check_mark:
GroqLLM:white_check_mark::white_check_mark:
PerplexityLLM:white_check_mark::white_check_mark:
GeminiLLM:white_check_mark::white_check_mark:
AWS BedrockLLM:white_check_mark::x:
MistralLLM:x::white_check_mark:
LangchainFramework:x::white_check_mark:
LlamaIndexFramework:white_check_mark::white_check_mark:
LanggraphFramework:x::white_check_mark:
LiteLLMFramework:x::white_check_mark:
DSPyFramework:x::white_check_mark:
CrewAIFramework:x::white_check_mark:
OllamaFramework:x::white_check_mark:
VertexAIFramework:white_check_mark::white_check_mark:
Vercel AIFramework:white_check_mark::x:
PineconeVector Database:white_check_mark::white_check_mark:
ChromaDBVector Database:white_check_mark::white_check_mark:
QDrantVector Database:white_check_mark::white_check_mark:
WeaviateVector Database:white_check_mark::white_check_mark:
PGVectorVector Database:white_check_mark::white_check_mark: (SQLAlchemy)

πŸ“ Langtrace System Architecture

image


πŸ’‘ Feature Requests and Issues


🀝 Contributions

We welcome contributions to this project. To get started, fork this repository and start developing. To get involved, join our Slack workspace.


🌟 Langtrace Star History

Langtrace Star History Chart


πŸ”’Security

To report security vulnerabilities, email us at security@scale3labs.com. You can read more on security here.


πŸ“œ License


❓Frequently Asked Questions

1. Can I self host and run Langtrace in my own cloud? Yes, you can absolutely do that. Follow the self hosting setup instructions in our documentation.

2. What is the pricing for Langtrace cloud? Currently, we are not charging anything for Langtrace cloud and we are primarily looking for feedback so we can continue to improve the project. We will inform our users when we decide to monetize it.

3. What is the tech stack of Langtrace? Langtrace uses NextJS for the frontend and APIs. It uses PostgresDB as a metadata store and Clickhouse DB for storing spans, metrics, logs and traces.

4. Can I contribute to this project? Absolutely! We love developers and welcome contributions. Get involved early by joining our Discord Community.

5. What skillset is required to contribute to this project? Programming Languages: Typescript and Python. Framework knowledge: NextJS. Database: Postgres and Prisma ORM. Nice to haves: Opentelemetry instrumentation framework, experience with distributed tracing.