AI & ML interests

None defined yet.

Recent Activity

pagezyhfΒ 
posted an update 16 days ago
view post
Post
2719
πŸš€ Big news for AI builders!

We’re thrilled to announce that the Qwen3-VL family of vision-language models is now available on Azure AI Foundry, thanks to our collaboration with Microsoft.

We bring open-source innovation to enterprise-grade AI infrastructure, making it easier than ever for enterprise to deploy and scale the latest and greatest from models from hugging Face securely within Azure.

πŸ” Highlights:

- Deploy Qwen3-VL instantly via managed endpoints
- Built-in governance, telemetry, and lifecycle management
- True multimodal reasoning β€” vision, language, and code understanding
- State-of-the-art performance, outperforming closed-source models like Gemini 2.5 Pro and GPT-5
- Available in both *Instruct* and *Thinking* modes, across 24 model sizes

πŸ‘‰ Get started today: search for Qwen3-VL in the Hugging Face Collection on Azure AI Foundry.
  • 1 reply
Β·

Update cache format

#3 opened about 1 month ago by
cyrilvallez

Fix formatting.

3
#4 opened about 1 month ago by
ariG23498
MolbapΒ 
posted an update about 1 month ago
view post
Post
3079
πŸš€ New blog: Maintain the unmaintainable – 1M+ Python LOC, 400+ models

How do you stop a million-line library built by thousands of contributors from collapsing under its own weight?
At πŸ€— Transformers, we do it with explicit software-engineering tenets, principles that make the codebase hackable at scale.

πŸ” Inside the post:
– One Model, One File: readability first β€” you can still open a modeling file and see the full logic, top to bottom.
– Modular Transformers: visible inheritance that cuts maintenance cost by ~15Γ— while keeping models readable.
– Config-Driven Performance: FlashAttention, tensor parallelism, and attention scheduling are config-level features, not rewrites.

Written with @lysandre ,@pcuenq and @yonigozlan , this is a deep dive into how Transformers stays fast, open, and maintainable.

Read it here β†’ transformers-community/Transformers-tenets

Integrate first returns

1
#6 opened about 2 months ago by
ror
pagezyhfΒ 
posted an update about 2 months ago
view post
Post
839
What’s your biggest headache deploying Hugging Face models to the cloudβ€”and how can we fix it for you?
Β·

Fix a bug

#1 opened about 2 months ago by
RaushanTurganbay

Fix a bug

#1 opened about 2 months ago by
RaushanTurganbay
lysandreΒ 
posted an update 2 months ago
view post
Post
6585
We're kick-starting the process of Transformers v5, with @ArthurZ and @cyrilvallez !

v5 should be significant: we're using it as a milestone for performance optimizations, saner defaults, and a much cleaner code base worthy of 2025.

Fun fact: v4.0.0-rc-1 came out on Nov 19, 2020, nearly five years ago!
  • 6 replies
Β·
pagezyhfΒ 
posted an update 2 months ago
pagezyhfΒ 
posted an update 2 months ago
view post
Post
3893
🀝 Collaborating with AMD to ensure Hugging Face Transformers runs smoothly on AMD GPUs!

We run daily CI on AMD MI325 to track the health of the most important model architectures and we’ve just made our internal dashboard public.

By making this easily accessible, we hope to spark community contributions and improve support for everyone!
  • 2 replies
Β·