Skip to content

[Feature Request] Support RoPE Scaling (YaRN) for Long Context Extension Training #276

@dawson-chen

Description

@dawson-chen

We are using the Qwen3-32B model for RL training. This model natively supports 32K context length and can extend the context window to 64K or even 128K using the YaRN method.

During the previous SFT stage, we enabled YaRN for long context training.

We hope ROLL can support rope_scaling configuration, specifically including:

Training Stage (Trainer): Support passing rope_scaling configuration when loading the model
Inference Stage (Rollout/vLLM/SGLang): Support corresponding RoPE scaling configuration during generation

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions