llm.rb v4.11.0 released
llm.rb v4.11.0 introduces streaming tool execution—tools can start while the model is still responding, overlapping latency with output. It adds MCP support over both stdio and HTTP (with connection pooling), OpenAI’s Responses API, and a complete concurrency model with threads, fibers, and async tasks.
The release includes a local model registry for cost tracking, JSON Schema unions, and production fixes across providers, making it ready for real systems where control and performance matter.
Post a comment