AI & ML interests

None defined yet.

Recent Activity

TonicΒ 
posted an update 3 days ago
AbhaykoulΒ 
posted an update 3 days ago
view post
Post
3569
πŸš€ Dhanishtha-2.0-preview-0825 Is Here

The Intermediate Thinking Model just leveled up again.

With sharper reasoning, better tool use, and expanded capabilities, Dhanishtha-2.0-preview-0825 is now live and ready to impress.

🧠 What Makes Dhanishtha Special?
Unlike typical CoT models that only thinks one time, Dhanishtha thinks iteratively:

> Think β†’ Answer β†’ Rethink β†’ Improve β†’ Rethink again if needed.

πŸ”— Try it now: HelpingAI/Dhanishtha-2.0-preview-0825

πŸ”ž Dhanishtha NSFW Preview

For those exploring more expressive and immersive roleplay scenarios, we’re also releasing:

HelpingAI/Dhanishtha-nsfw
A specialized version tuned for adult-themed interactions and character-driven roleplay.

πŸ”— Explore it here: HelpingAI/Dhanishtha-nsfw

πŸ’¬ You can also try all of these live at chat.helpingai.co
Β·

FEATURE REQUEST : Add Tmux

#25 opened 5 days ago by
Tonic
IlyasMoutawwakilΒ 
posted an update 5 days ago
view post
Post
3247
πŸš€ Optimum: The Last v1 Release πŸš€
Optimum v1.27 marks the final major release in the v1 series. As we close this chapter, we're laying the groundwork for a more modular and community-driven future:
- Optimum v2: A lightweight core package for porting Transformers, Diffusers, or Sentence-Transformers to specialized AI hardware/software/accelerators..
- Optimum‑ONNX: A dedicated package where the ONNX/ONNX Runtime ecosystem lives and evolves, faster-moving and decoupled from the Optimum core.

🎯 Why this matters:
- A clearer governance path for ONNX, fostering stronger community collaboration and improved developer experience..
- Enable innovation at a faster pace in a more modular, open-source environment.

πŸ’‘ What this means:
- More transparency, broader participation, and faster development driven by the community and key actors in the ONNX ecosystem (PyTorch, Microsoft, Joshua Lochner πŸ‘€, ...)
- A cleaner, more maintainable core Optimum, focused on extending HF libraries to special AI hardware/software/accelerators tooling and used by our partners (Intel Corporation, Amazon Web Services (AWS), AMD, NVIDIA, FuriosaAI, ...)

πŸ› οΈ Major updates I worked on in this release:
βœ… Added support for Transformers v4.53 and SmolLM3 in ONNX/ONNXRuntime.
βœ… Solved batched inference/generation for all supported decoder model architectures (LLMs).

✨ Big shoutout to @echarlaix for leading the refactoring work that cleanly separated ONNX exporter logic and enabled the creation of Optimum‑ONNX.

πŸ“ Release Notes: https://lnkd.in/gXtE_qji
πŸ“¦ Optimum : https://lnkd.in/ecAezNT6
🎁 Optimum-ONNX: https://lnkd.in/gzjyAjSi
#Optimum #ONNX #OpenSource #HuggingFace #Transformers #Diffusers
nroggendorffΒ 
posted an update 10 days ago
view post
Post
3662
Is it possible to apply for a resources grant for a whole organization, or do you need to apply for each repo individually? I think it'd be pretty cool to have something like the discord-community org for None-yet in terms of resource allocation (multiple spaces running on cpu upgrade.

I realize the scale of the community is just a tiny bit different, and that having this for a public org (one where anyone can join) isn't super fiscally responsible, but we'll be good. I promise we will! Right, guys?
  • 1 reply
Β·
BrigitteTousiΒ 
posted an update 11 days ago
view post
Post
472
This is what Hugging Face is all about. We want everyone, hobbyists, researchers and industry alike, to be able to contribute to AI because everyone is affected by it. Kudos to HF's @irenesolaiman for spreading the word!πŸ”₯πŸ€—
AtAndDevΒ 
posted an update 13 days ago
view post
Post
297
Qwen 3 Coder is a personal attack to k2, and I love it.
It achieves near SOTA on LCB while not having reasoning.
Finally people are understanding that reasoning isnt necessary for high benches...

Qwen ftw!

DECENTRALIZE DECENTRALIZE DECENTRALIZE
TonicΒ 
posted an update 16 days ago
view post
Post
682
πŸ‘‹ Hey there folks,

just submitted my plugin idea to the G-Assist Plugin Hackathon by @nvidia . Check it out, it's a great way to use a local SLA model on a windows machine to easily and locally get things done ! https://github.com/NVIDIA/G-Assist
TonicΒ 
posted an update 17 days ago
view post
Post
531
πŸ™‹πŸ»β€β™‚οΈ Hey there folks ,

Yesterday , Nvidia released a reasoning model that beats o3 on science, math and coding !

Today you can try it out here : Tonic/Nvidia-OpenReasoning

hope you like it !
AbhaykoulΒ 
posted an update 20 days ago
view post
Post
3005
πŸŽ‰ Dhanishtha-2.0-preview-0725 is Now Live

The Intermediate Thinking Model just got even better.
With the new update, Dhanishtha is now sharper, smarter, and trained further on tool use

🧠 What Makes Dhanishtha Different?
Unlike standard COT models that give one-shot responses, Dhanishtha thinks in layers:

> Think β†’ Answer β†’ Rethink β†’ Improve β†’ Rethink again if needed.

HelpingAI/Dhanishtha-2.0-preview-0725
macadelicccΒ 
posted an update 22 days ago
view post
Post
225
I was messing around with the HF api trying to get some stats on all time downloads for my models, and then I made it into a space so that anyone can use it.

macadeliccc/hf_downloads_dashboard

Let me know if you think it needs any changes or if you find it useful.
TonicΒ 
posted an update 24 days ago
view post
Post
3269
πŸ™‹πŸ»β€β™‚οΈ Normalize adding compute & runtime traces to your model cards
  • 2 replies
Β·
nroggendorffΒ 
posted an update 26 days ago
view post
Post
2990
Since when are H200s on ZeroGPU?
Β·
louisbrulenaudetΒ 
posted an update 27 days ago
view post
Post
2771
Because hackathons are often the starting point for many AI projects, I've created a Python-backend template incorporating my feedback to streamline collaboration and urgent deployments 🏎️

Within a year, I had the opportunity to participate in hackathons organized by Mistral, OpenAI, and DeepMind and this GitHub template is structured around several fundamental building blocks and recommendations I offer developers eager to participate in their first hackathon, whether as part of a team or individually. Its emphasis is on rapid setup and deployment through:
- uv as a package manager, simplifying usage via a series of pre-configured make commands.
- FastAPI for API management, structured in a modular architecture designed to minimize branch conflicts during merges to main branches (using minimal health-check and ping routes to verify Docker’s proper execution and backend accessibility on the local network).
- Pydantic for validation and type handling, which simplifies debugging and enhances understanding of data objects.
- A set of custom instructions tailored for agents (Cline and GitHub Copilot), aimed at improving overall comprehension of the application and optimizing the vibe-coding experience.

This template includes unit tests with a 100% success rate and test coverage, as well as a minimal CI file ensuring that the FastAPI application runs correctly. Thus, merging code that breaks the server into production becomes impossible ⛔️

In general, I would reiterate an essential piece of advice: your two main adversaries are branch conflictsβ€”particularly when the same file is modified concurrently within a brief period, especially if your architecture isn’t built for scalabilityβ€”and deployment issues under urgent circumstances ⏱️

Link to GitHub: https://github.com/louisbrulenaudet/hackathon-backend

Simply issue these commands and you can ship your code at the speed of light:
make init
make dev
TonicΒ 
posted an update 29 days ago
view post
Post
489
Who's going to Raise Summit in Paris Tomorrow ?

If you're around , I would love to meet you :-)
AbhaykoulΒ 
posted an update about 1 month ago
view post
Post
2984
πŸŽ‰ Dhanishtha 2.0 Preview is Now Open Source!

The world's first Intermediate Thinking Model is now available to everyone!

Dhanishtha 2.0 Preview brings revolutionary intermediate thinking capabilities to the open-source community. Unlike traditional reasoning models that think once, Dhanishtha can think, answer, rethink, answer again, and continue rethinking as needed using multiple blocks between responses.

πŸš€ Key Features
- Intermediate thinking: Think β†’ Answer β†’ Rethink β†’ Answer β†’ Rethink if needed...
- Token efficient: Uses up to 79% fewer tokens than DeepSeek R1 on similar queries
- Transparent thinking: See the model's reasoning process in real-time
- Open source: Freely available for research and development


HelpingAI/Dhanishtha-2.0-preview
https://helpingai.co/chat
  • 1 reply
Β·
frascuchonΒ 
posted an update about 1 month ago
freddyaboultonΒ 
posted an update about 1 month ago
AbhaykoulΒ 
posted an update about 1 month ago
view post
Post
4387
Introducing Dhanishtha 2.0: World's first Intermediate Thinking Model

Dhanishtha 2.0 is the world's first LLM designed to think between the responses. Unlike other Reasoning LLMs, which think just once.

Dhanishtha can think, rethink, self-evaluate, and refine in between responses using multiple <think> blocks.
This technique makes it Hinghlt Token efficient it Uses up to 79% fewer tokens than DeepSeek R1
---

You can try our model from: https://helpingai.co/chat
Also, we're gonna Open-Source Dhanistha on July 1st.

---
For Devs:
πŸ”‘ Get your API key at https://helpingai.co/dashboard
from HelpingAI import HAI  # pip install HelpingAI==1.1.1
from rich import print

hai = HAI(api_key="hl-***********************")

response = hai.chat.completions.create(
    model="Dhanishtha-2.0-preview",
    messages=[{"role": "user", "content": "What is the value of ∫0∞π‘₯3/π‘₯βˆ’1𝑑π‘₯ ?"}],
    stream=True,
    hide_think=False # Hide or show models thinking
)

for chunk in response:
    print(chunk.choices[0].delta.content, end="", flush=True)
  • 2 replies
Β·