Join the conversation

Join the community of Machine Learners and AI enthusiasts.

Sign Up

All HF Hub posts

danielhanchenย 
posted an update about 14 hours ago
view post
Post
600
You can now fine-tune embedding models in our free Unsloth notebook! ๐Ÿค—

Fine-tuning embedding models improves retrieval & RAG by aligning vectors to your domain-specific notion of similarity, improving search, clustering, and recommendations on your data.

โญ Blog + Notebooks: https://unsloth.ai/docs/new/embedding-finetuning

Unsloth trains embedding models 1.8-3.3x faster with 20% less VRAM, 2x longer context & no accuracy loss vs. FA2 setups.

We'd like to thank Hugging Face and Unsloth contributor: electroglyph for making this possible!
hassenhamdiย 
posted an update 1 day ago
view post
Post
1194
Google published the paper. I shipped the code. ๐Ÿš€

DeepMind just released PACEvolve (Progress-Aware Consistent Evolution), a massive overhaul of the AlphaEvolve framework. It solves the critical issues of "Context Pollution" and "Mode Collapse" that have historically crippled evolutionary coding agents.

But there was no public implementation. So I built one.

Introducing OpenPACEvolve: A fully open-source, production-grade implementation of the PACEvolve framework.

๐Ÿ›  I engineered this framework solo, but I wasn't working alone. I orchestrated a custom coding agents powered by Claude Opus 4.5 as Engineer and Gemini Pro 3 Preview ensuring fiedelity and quallty.

By leveraging these SOTA models, I was able to translate complex theoretical research into functional, modular Python architecture in record time. This is what the future of AI engineering looks like: Human architectural oversight + AI velocity.

๐Ÿง  What OpenPACEvolve Solves: Unlike standard agents that get "stuck" in loops, this framework implements the paper's full recipe for long-horizon stability: โœ… Hierarchical Context Management (HCM): Bi-level pruning to keep the agent's memory clean. โœ… Momentum-Based Backtracking (MBB): Uses "power-law backtracking" to detect stagnation and force pivots. โœ… Self-Adaptive Crossover: Intelligent code-sharing between parallel "islands."

๐Ÿ‘จโ€๐Ÿ’ป This project is more than a repo; it's a demonstration of rapid research-to-production cycles using next-gen AI workflows.

๐Ÿ“Ž Link of the paper : https://arxiv.org/abs/2601.10657

The code is live. The agents are ready. Check out the repository below. ๐Ÿ‘‡
https://github.com/hassenhamdi/OpenPACEvolve
Star the repo ๐ŸŒŸ.
branikitaย 
posted an update 2 days ago
view post
Post
1234
Our engineer Alan from https://robonine.com/ (Educational Robotics) integrated Feetech STS3250 and STS3215 servo motors into the prototype and completed the first test run of a 6-DOF semi-SCARA manipulator.

During motion, the structure demonstrates high stiffness with no visible backlash or mechanical play. The kinematic chain remains stable throughout the test trajectory, confirming the rigidity of the mechanical design and joint assembly.

The next stage includes full assembly with all actuators operating in backlash compensation mode, followed by quantitative measurement of positioning accuracy and repeatability.
  • 1 reply
ยท
mitkoxย 
posted an update 2 days ago
view post
Post
964
GLM-4.7-Flash is fast, good and cheap.
3,074 tokens/sec peak at 200k tokens context window on my desktop PC.
Works with Claude Code and opencode for hours. No errors, drop-in replacement of the Anthropic cloud AI.
MIT licensed, open weights, free for commercial use and modifications.
Supports speculative decoding using MTP, which is highly effective in mitigating latency.
Great for on device AI coding as AWQ 4bit at 18.5 GB. Hybrid inference on a single consumer GPU + CPU RAM.
  • 2 replies
ยท
mahimairajaย 
posted an update about 8 hours ago
IlyasMoutawwakilย 
posted an update about 16 hours ago
view post
Post
438
After 2 months of refinement, I'm happy to announce that a lot of Transformers' modeling code is now significantly more torch-compile & export-friendly ๐Ÿ”ฅ

Why it had to be done ๐Ÿ‘‡
PyTorch's Dynamo compiler is increasingly becoming the default interoperability layer for ML systems. Anything that relies on torch.export or torch.compile, from model optimization to cross-framework integrations, benefits directly when models can be captured as a single dynamo-traced graph !

Transformers models are now easier to:
โš™๏ธ Compile end-to-end with torch.compile backends
๐Ÿ“ฆ Export reliably via torch.export and torch.onnx.export
๐Ÿš€ Deploy to ONNX / ONNX Runtime, Intel Corporation's OpenVINO, NVIDIA AutoDeploy (TRT-LLM), AMD's Quark, Meta's Executorch and more hardware-specific runtimes.

This work aims at unblocking entire TorchDynamo-based toolchains that rely on exporting Transformers across runtimes and accelerators.

We are doubling down on Transformers commitment to be a first-class citizen of the PyTorch ecosystem, more exportable, more optimizable, and easier to deploy everywhere.

There are definitely some edge-cases that we still haven't addressed so don't hesitate to try compiling / exporting your favorite transformers and to open issues / PRs.

PR in the comments ! More updates coming coming soon !
  • 1 reply
ยท
Reubencfย 
posted an update 1 day ago
view post
Post
764
Now Live: The Reubencf/Nano_Banana_Editor now includes 10 free requests/day! ๐ŸŒ I'm personally sponsoring these credits to help make open AI accessible to all.
(Note: Limits are subject to change based on funding).

Enjoy !
mahimairajaย 
posted an update 1 day ago
view post
Post
760
Lacking vllm support for Transformers v5, frustrating only me?
danielhanchenย 
posted an update 3 days ago
Ujjwal-Tyagiย 
posted an update 4 days ago
view post
Post
2714
So, Koreans are also doing great progress behind Chinese,
Their two open source ai models that are actually good in coding. upstage/Solar-Open-100B skt/A.X-K1
  • 1 reply
ยท