Skip to main content
Back to registry

optimizing-attention-flash

davila7/claude-code-templates

Flash Attention provides 2-4x speedup and 10-20x memory reduction for transformer attention through IO-aware tiling and recomputation.

Installs155
Install command
npx skills add https://github.com/davila7/claude-code-templates --skill optimizing-attention-flash
Community Reviews

Latest reviews

Sign in to review

No community reviews yet. Be the first to review.

Browse this skill in context
FAQ
What does optimizing-attention-flash do?

Flash Attention provides 2-4x speedup and 10-20x memory reduction for transformer attention through IO-aware tiling and recomputation.

Is optimizing-attention-flash good?

optimizing-attention-flash does not have approved reviews yet, so SkillJury cannot publish a community verdict.

What agent does optimizing-attention-flash work with?

optimizing-attention-flash currently lists compatibility with codex, gemini-cli, opencode, cursor, claude-code, antigravity.

What are alternatives to optimizing-attention-flash?

Skills in the same category include telegram-bot-builder, flutter-app-size, sharp-edges, iterative-retrieval.

How do I install optimizing-attention-flash?

npx skills add https://github.com/davila7/claude-code-templates --skill optimizing-attention-flash

Related skills

More from davila7/claude-code-templates

Related skills

Alternatives in Software Engineering