Florent Gbelidji's picture

Florent Gbelidji

florentgbelidji

AI & ML interests

None yet

Recent Activity

Organizations

Hugging Face's profile picture Orange's profile picture Data For Good's profile picture Demo Corp's profile picture Atmos Bank's profile picture Hugging Test Lab's profile picture Deutsche Gesellschaft fรผr internationale Zusammenarbeit's profile picture RTLAI's profile picture Florent's profile picture Gradio-Blocks-Party's profile picture Sphere Spring 2023 Class's profile picture HF Canonical Model Maintainers's profile picture Carmaker Quality Datalab's profile picture Witty Works's profile picture Periplus AG's profile picture Besedo's profile picture Whoopsies's profile picture Amazon SageMaker's profile picture Adevinta's profile picture Shopify's profile picture Vanguard Investment Management Fintech Strategies's profile picture Vanguard's profile picture H&H Communication Lab GmbH's profile picture FactSet's profile picture Capital Fund Management's profile picture FactSet's profile picture Gorgias's profile picture Journalists on Hugging Face's profile picture Deutsche Telekom - Hugging Face's profile picture Social Post Explorers's profile picture Dev Mode Explorers's profile picture Aodocs's profile picture Paris AI Running Club's profile picture Hugging Face Party @ PyTorch Conference's profile picture Nerdy Face's profile picture Open R1's profile picture

florentgbelidji's activity

posted an update 20 days ago
view post
Post
1435
๐—ฃ๐—น๐—ฎ๐—ป๐—ป๐—ถ๐—ป๐—ด ๐—ฌ๐—ผ๐˜‚๐—ฟ ๐—ก๐—ฒ๐˜…๐˜ ๐—ฆ๐—ธ๐—ถ ๐—”๐—ฑ๐˜ƒ๐—ฒ๐—ป๐˜๐˜‚๐—ฟ๐—ฒ ๐—๐˜‚๐˜€๐˜ ๐—š๐—ผ๐˜ ๐—ฆ๐—บ๐—ฎ๐—ฟ๐˜๐—ฒ๐—ฟ: ๐—œ๐—ป๐˜๐—ฟ๐—ผ๐—ฑ๐˜‚๐—ฐ๐—ถ๐—ป๐—ด ๐—”๐—น๐—ฝ๐—ถ๐—ป๐—ฒ ๐—”๐—ด๐—ฒ๐—ป๐˜!๐Ÿ”๏ธโ›ท๏ธ

With the big hype around AI agents these days, I couldnโ€™t stop thinking about how AI agents could truly enhance real-world activities.
What sort of applications could we build with those AI agents: agentic RAG? self-correcting text-to-sql? Nah, boringโ€ฆ

Passionate about outdoors, Iโ€™ve always dreamed of a tool that could simplify planning mountain trips while accounting for all potential risks. Thatโ€™s why I built ๐—”๐—น๐—ฝ๐—ถ๐—ป๐—ฒ ๐—”๐—ด๐—ฒ๐—ป๐˜, a smart assistant designed to help you plan safe and enjoyable itineraries in the French Alps and Pyrenees.

Built using Hugging Face's ๐˜€๐—บ๐—ผ๐—น๐—ฎ๐—ด๐—ฒ๐—ป๐˜๐˜€ library, Alpine Agent combines the power of AI with trusted resources like ๐˜š๐˜ฌ๐˜ช๐˜ต๐˜ฐ๐˜ถ๐˜ณ.๐˜ง๐˜ณ (https://skitour.fr/) and METEO FRANCE. Whether itโ€™s suggesting a route with moderate difficulty or analyzing avalanche risks and weather conditions, this agent dynamically integrates data to deliver personalized recommendations.

In my latest blog post, I share how I developed this projectโ€”from defining tools and integrating APIs to selecting the best LLMs like ๐˜˜๐˜ธ๐˜ฆ๐˜ฏ2.5-๐˜Š๐˜ฐ๐˜ฅ๐˜ฆ๐˜ณ-32๐˜‰-๐˜๐˜ฏ๐˜ด๐˜ต๐˜ณ๐˜ถ๐˜ค๐˜ต, ๐˜“๐˜ญ๐˜ข๐˜ฎ๐˜ข-3.3-70๐˜‰-๐˜๐˜ฏ๐˜ด๐˜ต๐˜ณ๐˜ถ๐˜ค๐˜ต, or ๐˜Ž๐˜—๐˜›-4.

โ›ท๏ธ Curious how AI can enhance adventure planning?โ€จTry the app and share your thoughts: florentgbelidji/alpine-agent

๐Ÿ‘‰ Want to build your own agents? Whether for cooking, sports training, or other passions, the possibilities are endless. Check out the blog post to learn more: https://huggingface.co/blog/florentgbelidji/alpine-agent

Many thanks to @m-ric for helping on building this tool with smolagents!
  • 1 reply
ยท
reacted to MoritzLaurer's post with ๐Ÿ”ฅ 30 days ago
view post
Post
2213
๐Ÿš€ Releasing a new zeroshot-classifier based on ModernBERT! Some key takeaways:

- โšก Speed & efficiency: It's multiple times faster and uses significantly less memory than DeBERTav3. You can use larger batch sizes and enabling bf16 (instead of fp16) gave me a ~2x speed boost as well
- ๐Ÿ“‰ Performance tradeoff: It performs slightly worse than DeBERTav3 on average across my zeroshot classification task collection
- ๐Ÿง  Use cases: I recommend using it for scenarios requiring speed and a larger context window (8k).
- ๐Ÿ’ก Whatโ€™s next? Iโ€™m preparing a newer version trained on better + longer synthetic data to fully leverage the 8k context window and improve upon the training mix of my older zeroshot-v2.0 models. I also hope that there will be a multilingual variant in the future.

Great work by https://huggingface.co/answerdotai !

If youโ€™re looking for a high-speed zeroshot classifier, give it a try!

๐Ÿ“„ Resources below: ๐Ÿ‘‡
Base model: MoritzLaurer/ModernBERT-base-zeroshot-v2.0
Large model: MoritzLaurer/ModernBERT-large-zeroshot-v2.0
Updated zeroshot collection: MoritzLaurer/zeroshot-classifiers-6548b4ff407bb19ff5c3ad6f
ModernBERT collection with paper: answerdotai/modernbert-67627ad707a4acbf33c41deb
reacted to m-ric's post with ๐Ÿ”ฅ 30 days ago
view post
Post
5091
Since I published it on GitHub a few days ago,
Hugging Face's new agentic library ๐˜€๐—บ๐—ผ๐—น๐—ฎ๐—ด๐—ฒ๐—ป๐˜๐˜€ has gathered nearly 4k stars ๐Ÿคฏ

โžก๏ธ But we are just getting started on agents: so we are hiring an ML Engineer to join me and double down on this effort!

The plan is to build GUI agents: agents that can act on your computer with mouse & keyboard, like Claude Computer Use.

We will make it work better, and fully open. โœจ

Sounds like something you'd like to do? Apply here ๐Ÿ‘‰ https://apply.workable.com/huggingface/j/AF1D4E3FEB/
ยท
reacted to m-ric's post with ๐Ÿ”ฅ 3 months ago
view post
Post
1387
Great feature alert: ๐—ฌ๐—ผ๐˜‚ ๐—ฐ๐—ฎ๐—ป ๐—ป๐—ผ๐˜„ ๐˜‚๐˜€๐—ฒ ๐—ฎ๐—ป๐˜† ๐—ฆ๐—ฝ๐—ฎ๐—ฐ๐—ฒ ๐—ฎ๐˜€ ๐—ฎ ๐˜๐—ผ๐—ผ๐—น ๐—ณ๐—ผ๐—ฟ ๐˜†๐—ผ๐˜‚๐—ฟ ๐˜๐—ฟ๐—ฎ๐—ป๐˜€๐—ณ๐—ผ๐—ฟ๐—บ๐—ฒ๐—ฟ๐˜€.๐—ฎ๐—ด๐—ฒ๐—ป๐˜! ๐Ÿ› ๏ธ๐Ÿ”ฅ๐Ÿ”ฅ

This lets you take the coolest spaces, like FLUX.1-dev, and use them in agentic workflows with a few lines of code! ๐Ÿง‘โ€๐Ÿ’ป

On the video below, I set up my fake vacation pictures where I'm awesome at surfing (I'm really not) ๐Ÿ„

Head to the doc to learn this magic ๐Ÿ‘‰ https://huggingface.co/docs/transformers/main/en/agents_advanced#import-a-space-as-a-tool-
reacted to joaogante's post with ๐Ÿค— 9 months ago
view post
Post
3161
New sampling strategy dropped in ๐Ÿค— transformers -- Min P sampling ๐Ÿ”ฅ

Are you tired of having top_k arbitrarily discarding high-quality continuations? Or top_p forgetting to exclude low-probability tokens, derailing your generation? Try out the new min_p flag in generate, fresh from a PR merged today! ๐Ÿฅฌ

Min P consists of a dynamic token filter -- as opposed to Top K, which keeps the K most likely tokens, and Top P, which keeps the most likely tokens up to a fixed cumulative probability, both static filters. Min P takes a base probability (defined in the min_p flag) and multiplies it by the probability of the most likely token in the distribution for the next token. All tokens less likely than the resulting value are filtered. What happens with this strategy?
๐Ÿ‘‰ High probability token present -> aggressive filter (we don't want to miss on that high-probability case and risk derailing generation)
๐Ÿ‘‰ No high probability token present -> relaxed filter (there are many continuation possibilities that the model finds plausible)

You should set min_p to a low value, between 0.05 and 0.1. It behaves particularly well for creative text generation when paired up with temperature > 1.

Kudos to @kalomaze and @menhguin for creating this technique ๐Ÿ”ฅ Read their discussion in the original issue for benchmarks (https://github.com/huggingface/transformers/issues/27670)

Copy-pasteable version of the example in the image below here: https://pastebin.com/VqXNtuxd

Have fun experimenting! ๐Ÿ˜Ž
reacted to m-ric's post with ๐Ÿ”ฅ 11 months ago
view post
Post
Interesting paper: ๐†๐š๐‹๐จ๐ซ๐ž: ๐ญ๐ซ๐š๐ข๐ง ๐Ÿ•๐ ๐ฆ๐จ๐๐ž๐ฅ๐ฌ ๐จ๐ง ๐œ๐จ๐ง๐ฌ๐ฎ๐ฆ๐ž๐ซ-๐ ๐ซ๐š๐๐ž ๐†๐๐”๐ฌ ๐Ÿ’ช
It's now possible to ๐™›๐™ช๐™ก๐™ก๐™ฎ ๐™ฅ๐™ง๐™š-๐™ฉ๐™ง๐™–๐™ž๐™ฃ a 7B model on a consumer-grade GPU of 24Gb RAM, without any performance loss!

The memory usage of training models has always been an acute issue. For instance full pre-training of a 7B model used to eat ~50Gb of RAM!

The common workarounds to reduce memory load are:
- separate models on multiple GPUs ("sharding")
- quantize models: encode weights on fewer bits

Another technique is to ๐™ฅ๐™ง๐™ค๐™Ÿ๐™š๐™˜๐™ฉ ๐™ฉ๐™๐™š ๐™ฌ๐™š๐™ž๐™œ๐™๐™ฉ ๐™ข๐™–๐™ฉ๐™ง๐™ž๐™ญ ๐™ฉ๐™ค ๐™ก๐™ค๐™ฌ๐™š๐™ง-๐™ง๐™–๐™ฃ๐™  ๐™จ๐™ฅ๐™–๐™˜๐™š๐™จ, (since sometimes the weights do not really vary on all dimensions): this can save a lot of space!
This low-rank projection can be done on adapters to preserve the original weights (go check out LoRA), but it still generally hurts the performance too much for pre-training.

โžก๏ธ Enter the authors of ๐˜Ž๐˜ข๐˜“๐˜ฐ๐˜ณ๐˜ฆ: ๐˜”๐˜ฆ๐˜ฎ๐˜ฐ๐˜ณ๐˜บ-๐˜Œ๐˜ง๐˜ง๐˜ช๐˜ค๐˜ช๐˜ฆ๐˜ฏ๐˜ต ๐˜“๐˜“๐˜” ๐˜›๐˜ณ๐˜ข๐˜ช๐˜ฏ๐˜ช๐˜ฏ๐˜จ ๐˜ฃ๐˜บ ๐˜Ž๐˜ณ๐˜ข๐˜ฅ๐˜ช๐˜ฆ๐˜ฏ๐˜ต ๐˜“๐˜ฐ๐˜ธ-๐˜™๐˜ข๐˜ฏ๐˜ฌ ๐˜—๐˜ณ๐˜ฐ๐˜ซ๐˜ฆ๐˜ค๐˜ต๐˜ช๐˜ฐ๐˜ฏ. They gather (and prove) interesting insights:
โ›” The weight matrix does not reliably converge to lower ranks during training.
โœ… But the gradient matrix does!

Based on these insights, ๐˜๐—ต๐—ฒ๐˜† ๐—ฏ๐˜‚๐—ถ๐—น๐—ฑ ๐—š๐—ฎ๐—Ÿ๐—ผ๐—ฟ๐—ฒ, that projects the gradient to lower ranks.
๐Ÿ—บ๏ธ ๐—š๐—ฟ๐—ฒ๐—ฎ๐˜ ๐—ถ๐—ฑ๐—ฒ๐—ฎ: to leave the optimization free to explore more space, they periodically re-build the low-rank projection throughout the training (a nice illustration is in the paper).

๐Ÿค This method can even be combined with previous ones such as 8-bit Adam (quantizing the optimizer states to 8-bit).

โžก๏ธ ๐‘๐ž๐ฌ๐ฎ๐ฅ๐ญ๐ฌ:
๐Ÿ“‰ Of course, huge reduction in memory footprint allowing the training on consumer-grade GPU (cf figure).
๐Ÿ’ช No reduction in performance: this scales well up to 7B parameters (and was independently confirmed since) โ‡’ this is essential, it confirms that the method is viable!

Read the full paper here: GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection (2403.03507)
reacted to clem's post with ๐Ÿค— about 1 year ago
view post
Post
Is synthetic data the future of AI? ๐Ÿ”ฅ๐Ÿ”ฅ๐Ÿ”ฅ

@HugoLaurencon @Leyo & @VictorSanh are introducing HuggingFaceM4/WebSight , a multimodal dataset featuring 823,000 pairs of synthetically generated HTML/CSS codes along with screenshots of the corresponding rendered websites to train GPT4-V-like models ๐ŸŒ๐Ÿ’ป

While crafting their upcoming foundation vision language model, they faced the challenge of converting website screenshots into usable HTML/CSS codes. Most VLMs suck at this and there was no public dataset available for this specific task, so they decided to create their own.

They prompted existing LLMs to generate 823k HTML/CSS codes of very simple websites. Through supervised fine-tuning of a vision language model on WebSight, they were able to generate the code to reproduce a website component, given a screenshot.

You can explore the dataset here: HuggingFaceM4/WebSight

What do you think?
ยท