From 2d5f20684a9e4574120a4e632b711d4037301da8 Mon Sep 17 00:00:00 2001 From: WadydX <65117428+WadydX@users.noreply.github.com> Date: Thu, 23 Apr 2026 14:47:08 +0100 Subject: [PATCH] docs: remove dead reference links in flash-attention skill --- optional-skills/mlops/flash-attention/SKILL.md | 4 ---- 1 file changed, 4 deletions(-) diff --git a/optional-skills/mlops/flash-attention/SKILL.md b/optional-skills/mlops/flash-attention/SKILL.md index 6a3839bf78..89a860e67d 100644 --- a/optional-skills/mlops/flash-attention/SKILL.md +++ b/optional-skills/mlops/flash-attention/SKILL.md @@ -345,10 +345,6 @@ Flash Attention uses float16/bfloat16 for speed. Float32 not supported. **Performance benchmarks**: See [references/benchmarks.md](references/benchmarks.md) for detailed speed and memory comparisons across GPUs and sequence lengths. -**Algorithm details**: See [references/algorithm.md](references/algorithm.md) for tiling strategy, recomputation, and IO complexity analysis. - -**Advanced features**: See [references/advanced-features.md](references/advanced-features.md) for rotary embeddings, ALiBi, paged KV cache, and custom attention masks. - ## Hardware requirements - **GPU**: NVIDIA Ampere+ (A100, A10, A30) or AMD MI200+