Quickstart / 快速开始
Install the SDK, create an app, deploy your first GPU function in minutes.
pip install duanflow
Read guide
像调用本地函数一样调用云端 GPU。DuanFlow 自动匹配实时最优算力,把 Python 代码部署成可扩缩的 AI Endpoint。
Call cloud GPUs from Python. No clusters, no YAML, no idle machines.
import duanflow as df
app = df.App("my-ai-app")
@app.function(
gpu="H100",
memory=80,
optimize="price"
)
def run_inference(prompt):
# 这里写你的 AI 业务代码
return model.generate(prompt)
run_inference.remote("解释端流是什么")
Mock 演示 DuanFlow 如何读取代码需求、选择实时最优 GPU,并完成云端部署。
import duanflow as df
app = df.App("my-ai-app")
@app.function(gpu="H100", memory=80)
def run_inference(prompt):
# 这里写你的 AI 业务代码
return "生成结果..."
Mock 文档入口采用双语结构,方便演示 SDK、部署、Endpoint 和 GPU 调度能力。
Install the SDK, create an app, deploy your first GPU function in minutes.
pip install duanflow
Read guide
Use decorators to define CPU, memory, GPU type, timeout, and autoscaling policy.
@app.function(gpu="H100")
Read guide
Turn Python functions into secure HTTPS APIs with logs and versioned releases.
duanflow deploy app.py
Read guide
Route workloads by price, latency, region, or GPU availability across resource pools.
optimize="price"
Read guide
每个示例都用同一种 SDK 心智模型:写函数、声明资源、远程调用。
@app.endpoint(gpu="H100", memory="80Gi")
def chat(prompt):
# Load model once, autoscale on demand
return qwen.generate(prompt)
View full example
@app.function(gpu="A100", concurrency=128)
def embed(doc):
# 自动扩容到多容器执行
return encoder.encode(doc)
vectors = embed.map(documents)
View full example
@app.job(gpu="H100", gpu_count=4)
def finetune(dataset):
# 训练完成后自动释放算力
trainer.train(dataset)
return trainer.metrics
View full example