RunCue

I developed Runcue over December of 2025 to while working on various LLM pipelines and knowledge management projects, and found it made development much faster and easier to reason about.

Run interdependent background tasks in parallel from Python without all the infrastructure overhead.

Runcue is a Python library for coordinating work across rate-limited services. Define your tasks, tell runcue when they're ready to run, and let it handle the scheduling and throttling. Dependencies are handled via artifacts, not task completion - this makes it more like a Makefile than a task queue - and therefore easier to build and scale.

What runcue Is

An embedded coordinator for rate-limited work

runcue handles the WHEN. You handle the WHAT.

Example

import runcue
from pathlib import Path

cue = runcue.Cue()
cue.service("api", rate="60/min", concurrent=5)

@cue.task("process", uses="api")
def process(work):
    # Your logic here
    output = do_work(work.params["input"])
    Path(work.params["output"]).write_text(output)
    return {"done": True}

@cue.is_ready
def is_ready(work):
    return Path(work.params["input"]).exists()

@cue.is_stale
def is_stale(work):
    return not Path(work.params["output"]).exists()

async def main():
    cue.start()
    await cue.submit("process", params={
        "input": "data.txt",
        "output": "result.txt"
    })
    await cue.stop()

Simulator

Runcue comes with a sim that includes various scenarios, to stress-test situations like deadlocks and timeouts, and find the max parallel limits of the host.

Mandlebrot Example

The repo has an example that can generate animated gifs using rucue to generate fractals in parallel and then assemble them at the end.