LangFuse
LangFuse is an LLM engineering platform that helps teams collaboratively develop, monitor, evaluate, and debug AI applications. This guide demonstrates how to integrate Vercel AI Gateway with LangFuse to access various AI models and providers.
First, create a new directory for your project and initialize it:
terminalmkdir langfuse-ai-gateway cd langfuse-ai-gateway pnpm dlx init -yInstall the required LangFuse packages along with the
dotenvand@types/nodepackages:pnpm i langfuse openai dotenv @types/nodeCreate a
.envfile with your Vercel AI Gateway API key and LangFuse API keys:.envAI_GATEWAY_API_KEY=your-api-key-here LANGFUSE_PUBLIC_KEY=your_langfuse_public_key LANGFUSE_SECRET_KEY=your_langfuse_secret_key LANGFUSE_HOST=https://cloud.langfuse.comIf you're using the AI Gateway from within a Vercel deployment, you can also use the
VERCEL_OIDC_TOKENenvironment variable which will be automatically provided.Create a new file called
index.tswith the following code:index.tsimport { observeOpenAI } from 'langfuse'; import OpenAI from 'openai'; const openaiClient = new OpenAI({ apiKey: process.env.AI_GATEWAY_API_KEY, baseURL: 'https://ai-gateway.vercel.sh/v1', }); const client = observeOpenAI(openaiClient, { generationName: 'fun-fact-request', // Optional: Name of the generation in Langfuse }); const response = await client.chat.completions.create({ model: 'moonshotai/kimi-k2', messages: [ { role: 'system', content: 'You are a helpful assistant.' }, { role: 'user', content: 'Tell me about the food scene in San Francisco.' }, ], }); console.log(response.choices[0].message.content);The following code:
- Creates an OpenAI client configured to use the Vercel AI Gateway
- Uses
observeOpenAIto wrap the client for automatic tracing and logging - Makes a chat completion request through the AI Gateway
- Automatically captures request/response data, token usage, and metrics
Run your application using Node.js:
pnpm dlx tsx index.tsYou should see a response from the AI model in your console.
Was this helpful?