Join the conversation

Join the community of Machine Learners and AI enthusiasts.

Sign Up

All HF Hub posts

XenovaΒ 
posted an update 1 day ago
view post
Post
818
NEW: Real-time conversational AI models can now run 100% locally in your browser! 🀯

πŸ” Privacy by design (no data leaves your device)
πŸ’° Completely free... forever
πŸ“¦ Zero installation required, just visit a website
⚑️ Blazingly-fast WebGPU-accelerated inference

Try it out: webml-community/conversational-webgpu

For those interested, here's how it works:
- Silero VAD for voice activity detection
- Whisper for speech recognition
- SmolLM2-1.7B for text generation
- Kokoro for text to speech

Powered by Transformers.js and ONNX Runtime Web! πŸ€— I hope you like it!
  • 1 reply
Β·
danaaubakirovaΒ 
posted an update 2 days ago
yeonseok-zeticaiΒ 
posted an update 3 days ago
view post
Post
1920
πŸš€ NEW DROP: run your own on-device LLMβ€”in minutes, on any phone
Today we’re open-sourcing everything you need to put Qwen3-0.6B straight into a production-ready mobile app:

πŸŽ₯ Watch Qwen3-0.6B chat in real time on any smartphones!

πŸ“Š TPS benchmarks – slides comparing token-per-second across heterogeneous mobile devices

πŸ’» Plug-and-play source – Just Copy & Run the source to your project for Android (Kotlin & Java) and iOS (Swift).

🀞 Cross-platform, one pipeline – ZETIC.MLange auto-tunes kernels for every different devices, we’ve tested.

πŸ‘¨β€πŸ’» Ready for production – swap in your own model, re-benchmark with one command, publish.

Get started
Just Sign-up and check the playground project, QWEN-0.6B
- https://mlange.zetic.ai/p/zetic-example/Qwen3-0.6B

We built this to show that cloud-free LLMs are ready today. Dive in, fork it, and tag ZETIC.ai when you launch your own on-device assistant, game NPC, or offline content generatorβ€”we’ll spotlight the best projects.
danieldkΒ 
posted an update 1 day ago
view post
Post
698
We have been working on a project called kernels. kernels makes it possible to load compute kernels directly from the Hub! πŸš€

We plan to give kernels a more proper introduction soon. But for those who have been following along, we are happy to announce a new release:

- New layer API with torch.compile support.
- Experimental support for loading Apple Silicon Metal 🀘 Kernels.
- Generate wheels from Hub kernels for legacy deployments.

Full release notes here: https://github.com/huggingface/kernels/releases/tag/v0.6.0
VirtualOasisΒ 
posted an update 3 days ago
view post
Post
2261
I built an AI Website: ai-garden.netlify.app
It is a curated garden of AI resources.
It's my database for writing and research, organized by category and designed for quick access. Whether you're looking for learning materials, development tools, research papers, or industry news, everything's laid out in a clean, searchable format.

Feel free to suggest new resources or improvements - this garden grows better with community input. 🍻
Β·
merveΒ 
posted an update 1 day ago
view post
Post
910
Past week was insanely packed for open AI! 😱
Luckily we picked some highlights for you ❀️ lfg!

πŸ’¬ LLMs/VLMs
> Deepseek 🐳 released deepseek-ai/DeepSeek-R1-0528, 38B model, only 0.2 and 1.4 points behind o3 in AIME 24/25 🀯 they also released an 8B distilled version based on Qwen3 (OS) deepseek-ai/deepseek-r1-678e1e131c0169c0bc89728d
> Xiaomi released MiMo-7B-RL (LLM for code and math) and MiMo-VL-7B-RL (VLM for visual reasoning, GUI agentic task and general use) (OS) 😍 XiaomiMiMo/mimo-vl-68382ccacc7c2875500cd212
> NVIDIA released , new reasoning model nvidia/Nemotron-Research-Reasoning-Qwen-1.5B
> DS: MiniMax released https://huggingface.co/MiniMaxAI/SynLogic, new 49k logical reasoning examples across 35 tasks including solving cipher, sudoku and more!

πŸ–ΌοΈ Image/Video Generation
> tencent released tencent/HunyuanPortrait, a new model for consistent portrait generation with SVD Research license. They also released tencent/HunyuanVideo-Avatar, audio driven avatar generation (OS)
> showlab released showlab/OmniConsistency, consistent stylization model (OS)
> Rapidata/text-2-video-human-preferences-veo3 is a new T2V preference dataset based on videos from Veo3 with 46k examples (OS)

AudioπŸ—£οΈ
> https://huggingface.co/ResembleAI/Chatterbox is a new 500M text-to-speech model preferred more than ElevenLabs (OS) 😍
> PlayHT/PlayDiffusion is a new speech editing model (OS)

Other
> https://huggingface.co/NX-AI/TiReX is a new time series foundation model
> Yandex released a huge (4.79B examples!) video recommendation dataset https://huggingface.co/yandex/yambda

OS ones have Apache2.0 or MIT licenses, find more models and datasets here merve/releases-30-may-6840097345e0b1e915bff843
merveΒ 
posted an update 1 day ago
view post
Post
867
Yesterday was the day of vision language action models (VLAs)!

> SmolVLA: open-source small VLA for robotics by Hugging Face LeRobot team πŸ€–
Blog: https://huggingface.co/blog/smolvla
Model: lerobot/smolvla_base

> Holo-1: 3B & 7B web/computer use agentic VLAs by H Company πŸ’»
Model family: Hcompany/holo1-683dd1eece7eb077b96d0cbd
Demo: https://huggingface.co/spaces/multimodalart/Holo1
Blog: https://huggingface.co/blog/Hcompany/holo1
super exciting times!!
JawardΒ 
posted an update 1 day ago
view post
Post
907
bumped into one of the OG reads today!! handwriting generation & synthesis is still my favorite application of RNNs - supper amazed at how such a small model (3.6M params), trained overnight on cpu could reach such peak performance. Huge credit to the data (IAM-OnDBπŸ”₯) which was meticulously curated using an infra-red device to track pen position.
Try demo here: https://www.calligrapher.ai/
Code: https://github.com/sjvasquez/handwriting-synthesis
azettlΒ 
posted an update 2 days ago
view post
Post
895
Agents & MCP Hackathon Day 2

Again, a short night, but here are some updates from my Hackathon projects before starting night #3.

I managed to get the first version of both submissions (custom Gradio component and MCP server) online!Β 

You can check the roundtable MCP where multiple AIs discuss your question and try to reach consensus: azettl/consilium_mcp.

The Gradio component is here: azettl/gradio_consilium_roundtable.

I placed my API keys in the env variables, so you can test without needing your own keys, but I will remove them soon as I did not find a limit setting in Sambanova. Still, you can check them by adding your own keys in the config tab.

Looking forward to your feedback, there are still many days I can and will improve this.
  • 1 reply
Β·
danielhanchenΒ 
posted an update 3 days ago