qgallouedec/huggingface-static-fc6272-bucket
423 kB
Qwen/Qwen3.6-27B, Qwen/Qwen3.6-35B-A3B) reuses the Qwen3.5-MoE architecture but ships a slightly different chat template, so we updated the stack end-to-end: new training template with {% generation %} markers, tool-call response schema routing, tiny test models for the VLM matrix.from trl import SFTConfig, SFTTrainer
trainer = SFTTrainer(
model="Qwen/Qwen3.6-27B",
args=SFTConfig(assistant_only_loss=True),
train_dataset=dataset,
)
trainer.train()tools=[...] to GRPOTrainer.trl vllm-serve (Qwen3 MTP / Eagle3 drafts), 12 more KTO ↔ DPO alignment PRs (KTO promotion to stable is now in reach), three more {% generation %} chat templates (Gemma/Gemma 2, Phi-3, GLM-4-MoE), and a chunky SFT entropy bug fix.Compare two code files side‑by‑side
View project metrics with an interactive dashboard
View and monitor your data with an interactive dashboard
Track and visualize your project data on an interactive dashboard
Track and visualize your data with a real‑time dashboard
View and manage your tracking data with an interactive dashboard