·llm-streaming-response-handler
!

llm-streaming-response-handler

Build production LLM streaming UIs with Server-Sent Events, real-time token display, cancellation, error recovery. Handles OpenAI/Anthropic/Claude streaming APIs. Use for chatbots, AI assistants, real-time text generation. Activate on "LLM streaming", "SSE", "token stream", "chat UI", "real-time AI". NOT for batch processing, non-streaming APIs, or WebSocket bidirectional chat.

12Installs·1Trend·@curiositech

Installation

$npx skills add https://github.com/curiositech/some_claude_skills --skill llm-streaming-response-handler

How to Install llm-streaming-response-handler

Quickly install llm-streaming-response-handler AI skill to your development environment via command line

  1. Open Terminal: Open your terminal or command line tool (Terminal, iTerm, Windows Terminal, etc.)
  2. Run Installation Command: Copy and run this command: npx skills add https://github.com/curiositech/some_claude_skills --skill llm-streaming-response-handler
  3. Verify Installation: Once installed, the skill will be automatically configured in your AI coding environment and ready to use in Claude Code, Cursor, or OpenClaw

Source: curiositech/some_claude_skills.

SKILL.md

View raw

Expert in building production-grade streaming interfaces for LLM responses that feel instant and responsive.

| OpenAI | SSE | data: {"choices":[{"delta":{"content":"token"}}]} | | Anthropic | SSE | data: {"type":"contentblockdelta","delta":{"text":"token"}} | | Claude (API) | SSE | data: {"delta":{"text":"token"}} | | Vercel AI SDK | SSE | Normalized across providers |

Problem: Stream fails mid-response, user sees partial text with no indication of failure.

Build production LLM streaming UIs with Server-Sent Events, real-time token display, cancellation, error recovery. Handles OpenAI/Anthropic/Claude streaming APIs. Use for chatbots, AI assistants, real-time text generation. Activate on "LLM streaming", "SSE", "token stream", "chat UI", "real-time AI". NOT for batch processing, non-streaming APIs, or WebSocket bidirectional chat. Source: curiositech/some_claude_skills.

Facts (cite-ready)

Stable fields and commands for AI/search citations.

Install command
npx skills add https://github.com/curiositech/some_claude_skills --skill llm-streaming-response-handler
Category
!Security
Verified
First Seen
2026-03-09
Updated
2026-03-10

Browse more skills from curiositech/some_claude_skills

Quick answers

What is llm-streaming-response-handler?

Build production LLM streaming UIs with Server-Sent Events, real-time token display, cancellation, error recovery. Handles OpenAI/Anthropic/Claude streaming APIs. Use for chatbots, AI assistants, real-time text generation. Activate on "LLM streaming", "SSE", "token stream", "chat UI", "real-time AI". NOT for batch processing, non-streaming APIs, or WebSocket bidirectional chat. Source: curiositech/some_claude_skills.

How do I install llm-streaming-response-handler?

Open your terminal or command line tool (Terminal, iTerm, Windows Terminal, etc.) Copy and run this command: npx skills add https://github.com/curiositech/some_claude_skills --skill llm-streaming-response-handler Once installed, the skill will be automatically configured in your AI coding environment and ready to use in Claude Code, Cursor, or OpenClaw

Where is the source repository?

https://github.com/curiositech/some_claude_skills