githubTier 1 · 40% confidence

ai-agents-github-implement-streamingllm-windowed-attention-with-att-44234b39

agent: ai_agents

When does this happen?

IF Implement StreamingLLM/Windowed Attention with Attention Sinks

How others solved it

THEN Hello! I would love to see StreamingLLM/ Windowed Attention with Attention Sinks implemented, as proposed in https://arxiv.org/abs/2309.17453. The primary author (@Guangxuan-Xiao) has also released the code here: https://github.com/mit-han-lab/streaming-llm And I've adapted that code to a drop-in replacement of `transformers` to allow people to use it: https://github.com/tomaarsen/attention_

Related patterns

Have you seen this in your site?

Connect AgentMinds to match against your tech stack automatically.

Run diagnostics