Programming language for building LLM workflows.
and its runtime implementation in Python.
pip install flowity
Hello, who are you?
$resp = $query()
$end($resp)
Any statements not starting with $
are considered as prompt, and will be sent to the model while calling $query()
, the response will be stored in the variable $resp
.
Prompt will be cached until the next $query()
call.
Syntax details can be found here.
import os
# === Setup Langchain ===
from langchain_openai import ChatOpenAI
os.environ["OPENAI_API_KEY"] = "sk-xxxx" # Your OpenAI api key
# Set OPENAI_API_BASE if you're using a reverse proxy
# os.environ["OPENAI_API_BASE"] = "https://api.openai.com/v1"
model = ChatOpenAI()
# === Write Flowity Code ===
code = """
Hello, who are you?
$str = $query()
$end($str)
"""
# === Run Flowity Code ===
from flowity.rt import rtime
rt = rtime.FlowityRuntime()
ret = rt.run(
code=code,
model=model
)
print(ret)