Naifan Li
Tags Categories Archives About
Naifan Li
Cancel
TagsCategoriesArchivesAbout

Contents

Summary: Sequence Packing

Published on Jul 29, 2021 Updated on Jan 15, 2026 LLMs LLMs, Sequence Packing One minute
Contents
  • References

. Instead of padding, we concatenate multiple shorter sequences into a single, longer sequence. This minimizes wasted compute (through padding tokens)

References

  • Efficient Sequence Packing without Cross-contamination: Accelerating Large Language Models without Impacting Performance
  • https://huggingface.co/blog/sirluk/llm-sequence-packing
  • https://docs.nvidia.com/nemo-framework/user-guide/24.12/nemotoolkit/features/optimizations/sequence_packing.html
  • https://docs.pytorch.org/docs/stable/generated/torch.nn.utils.rnn.pack_sequence.html
CLIP: Learning Transferable Visual Models From Natural Language Supervision FLAN: Finetuned Language Models Are Zero-Shot Learners

Related Posts

  • Qwen3-VL Technical Report 2025-11-27
  • DeepSeek R1 2025-06-10
  • Qwen Series: Technical Summary 2025-05-14
  • Qwen3 Technical Report 2025-05-14
  • Qwen2.5-VL Technical Report 2025-02-25
2018 - 2026 Naifan Li