Hugging Face
Models
Datasets
Spaces
Buckets
new
Docs
Enterprise
Pricing
Log In
Sign Up
Open to Collab
85
17
53
Jun Zhang
PRO
jzhang533
Follow
jinghuiPP's profile picture
branikita's profile picture
xianbao's profile picture
55 followers
Β·
72 following
jzhang533
AI & ML interests
None yet
Recent Activity
reacted
to
qgallouedec
's
post
with π
3 days ago
TRL v1.3 ships day-one training support for Qwen 3.6 π The new Qwen 3.6 family (`Qwen/Qwen3.6-27B`, `Qwen/Qwen3.6-35B-A3B`) reuses the Qwen3.5-MoE architecture but ships a slightly different chat template, so we updated the stack end-to-end: new training template with `{% generation %}` markers, tool-call response schema routing, tiny test models for the VLM matrix. SFT with assistant-only loss works out of the box: ```python from trl import SFTConfig, SFTTrainer trainer = SFTTrainer( model="Qwen/Qwen3.6-27B", args=SFTConfig(assistant_only_loss=True), train_dataset=dataset, ) trainer.train() ``` So does GRPO tool-calling β just hand `tools=[...]` to `GRPOTrainer`. v1.3 also brings a new experimental TPO trainer (Triple Preference Optimization), speculative decoding in `trl vllm-serve` (Qwen3 MTP / Eagle3 drafts), 12 more KTO β DPO alignment PRs (KTO promotion to stable is now in reach), three more `{% generation %}` chat templates (Gemma/Gemma 2, Phi-3, GLM-4-MoE), and a chunky SFT entropy bug fix. Full release notes: https://github.com/huggingface/trl/releases/tag/v1.3.0
updated
a Space
16 days ago
baidu/README
liked
a Space
16 days ago
baidu/ERNIE-Image-Turbo
View all activity
Organizations
jzhang533
's activity
All
Models
Datasets
Spaces
Buckets
Papers
Collections
Community
Posts
Upvotes
Likes
Articles
published
an
article
7 months ago
view article
Article
Unleashing the Full Potential of ERNIE4.5 using FastDeploy
Sep 19, 2025
β’
11
published
an
article
8 months ago
view article
Article
PP-OCRv5 on Hugging Face: A Specialized Approach to OCR
Sep 10, 2025
β’
111