This package provides an easy way to use the Langflow API to run flows from within server-side JavaScript applications.
npm install @datastax/langflow-client
yarn add @datastax/langflow-client
pnpm add @datastax/langflow-client
To use this Langflow client you will need a DataStax account with which you can use DataStax Langflow. You will need a flow before you can call its API. If you don't already have one, you can get started with the Basic Prompting Flow.
You can also use this with an installation of open-source Langflow.
You will need a Langflow API key. You provide the API key when you create a new client. You will also need your Langflow ID.
You can generate API keys and find your Langflow ID on the API modal in the Langflow flow editor.
You will need an installation of Langflow. You may also require an API key if you have set up authentication on Langflow.
You can import the client like so:
import { LangflowClient } from "@datastax/langflow-client";
You can then create a new client object with the following options:
baseURL
: This is set tohttps://api.langflow.astra.datastax.com
by default, if you are running your own instance of Langflow you will need to provide the URLlangflowId
: The ID of your organisation, which can be found in the API modal of the flow editor. This is not required for your own instance of LangflowapiKey
: A Langflow API key that can be generated within your DataStax account or in the settings of open-source Langflow
// using DataStax Langflow, you do not need to provide the baseUrl
const dsLangflowClient = new LangflowClient({ langflowId, apiKey });
// for open-source langflow, you will need to provide a baseUrl and optionally an apiKey
const baseUrl = "http://localhost:7860";
const apiKey = "sk-...";
const osLangflowClient = new LangflowClient({ baseUrl, apiKey });
Once you have a client, you can create a reference to a flow using the flowID
. This can be found in the API modal in Langflow.
const flow = client.flow(flowId);
You can run a flow by calling run
with the text input to the flow:
const response = await client.flow(flowId).run(input);
You can add tweaks to a flow like so:
const response = await client
.flow(flowId)
.tweak(tweakName, tweakOptions)
.run(input);
Or you can pass all tweaks as an object:
const response = await client.flow(flowId).run(input, { tweaks });
You can also pass input and output options, as well as a session ID.
import { InputTypes, OutputTypes } from "@datastax/langflow-client/consts";
const response = await client.flow(flowId).run(input, {
input_type: InputTypes.CHAT,
output_type: OutputTypes.CHAT,
session_id,
tweaks,
});
The available input types are "chat", "text" and "any". The available output types are "chat", "text", "any" and "debug". The default for both is "chat".
Langflow is very flexible in its output. So the FlowResponse
object gives you raw access to the sessionId
and the outputs
.
const response = await client.flow(flowId).run(input);
console.log(response.outputs);
There is one convenience function that will return you the first chat output message text. If you only have one chat output component in your flow, this is a useful shortcut to get to that response.
const response = client.flow(flowId).run(input);
console.log(response.chatOutputText());
The Langflow API supports streaming responses. Instead of calling run
on a Flow
object, you can call stream
with the same arguments and the response will be a ReadableStream
of objects.
const response = await client.flow(flowId).stream(input);
for await (const event of response) {
console.log(event);
}
There are three different events: add_message
, token
, and end
. The events mean:
add_message
: a message being added to the chat and can refer to a human input message or a response from an AItoken
: a token that is emitted as part of a message being generated by the flowend
: all tokens have been returned, this message will also contain a fullFlowResponse
Event objects have the format:
{
event: "add_message" | "token" | "end",
data: object
}
The event.data
is different per event type. The token
event type is the simplest and looks like this:
{
"event": "token",
"data": {
"chunk": "hello ",
"id": "6686ff20-0c95-40bb-8879-fd90ed3d634e",
"timestamp": "2025-02-12 22:18:04 UTC"
}
}
There's more documentation and examples of a streaming response in the Langflow docs.
Chat input components support files as input as well as text. You need to upload your file first, using the file upload function, then provide the file path to the flow as a tweak.
const flow = client.flow(flowId)
const file = await flow.uploadFile(pathToFile);
console.log(file);
// => { flowId: "XXX", filePath: "YYY" }
const response = await flow.tweak("ChatInput-abcd": { files: file.filePath }).run("What can you see in this image?");
Warning
DataStax Langflow doesn't make file upload available, you will receive a 501 Not Implemented error.
You can use the standard AbortController
to cancel requests by passing a signal
to the run
or uploadFile
functions. The functions will reject with a DOMException
error with the name AbortError
or, if you use AbortSignal.timeout
, TimeoutError
.
For example, when running the following code, if the entire request takes longer than 500ms, then the promise will reject and the error message will be, "The operation was aborted due to timeout".
const signal = AbortSignal.timeout(500);
try {
const response = await client.flow(flowId).run(input, { signal });
} catch (error) {
console.error(error.message);
}
To run and contribute to this library you can clone the repository:
git clone [email protected]:datastax/langflow-client-ts.git
cd langflow-client-ts
Install the dependencies:
npm install
Run the tests:
npm test
Transpile the TypeScript to JavaScript:
npm run build
Transpile the TypeScript to JavaScript in watch mode:
npm run build:watch
Lint the code:
npm run lint
Format the code:
npm run format
Check the formatting of the code:
npm run format:check