Skip to content

Commit

Permalink
Smolagents can see (#2611)
Browse files Browse the repository at this point in the history

---------

Co-authored-by: Merve Noyan <[email protected]>
Co-authored-by: Albert Villanova del Moral <[email protected]>
  • Loading branch information
3 people authored Jan 24, 2025
1 parent 4d76287 commit 6851e20
Show file tree
Hide file tree
Showing 6 changed files with 353 additions and 1 deletion.
15 changes: 15 additions & 0 deletions _blog.yml
Original file line number Diff line number Diff line change
Expand Up @@ -3945,6 +3945,7 @@
- nlp
- LLM
- agents
- smolagents
- transformers
- gpt
- mixtral
Expand Down Expand Up @@ -4266,6 +4267,7 @@
date: July 1, 2024
tags:
- agents
- smolagents
- nlp
- community
- research
Expand Down Expand Up @@ -5267,6 +5269,7 @@
date: December 31, 2024
tags:
- agents
- smolagents
- open-source
- nlp
- tools
Expand Down Expand Up @@ -5374,3 +5377,15 @@
- on-device
- smolvlm
- community

- local: smolagents-can-see
title: "We now support VLMs in smolagents!"
author: m-ric
thumbnail: /blog/assets/smolagents-can-see/thumbnail.png
date: January 24, 2025
tags:
- agents
- smolagents
- multimodal
- vlm
- vision
10 changes: 9 additions & 1 deletion agents.md
Original file line number Diff line number Diff line change
Expand Up @@ -23,6 +23,10 @@ We are releasing Transformers Agents 2.0!

🚀 Go try it out and climb ever higher on the GAIA leaderboard!

> [!WARNING]
> `transformers.agents` has now been upgraded to the stand-alone library [smolagents](https://github.com/huggingface/smolagents)! The two libraries have very similar APIs, so switching is easy.
> Go checkout the `smolagents` introduction blog [here](https://huggingface.co/blog/smolagents).
## Table of Contents

- [What is an agent?](#what-is-an-agent)
Expand Down Expand Up @@ -450,4 +454,8 @@ We will keep improving this package in the coming months. We have already identi

👉 **Go try out transformers agents!** We’re looking forward to receiving your feedback and your ideas.

Let’s fill the top of the leaderboard with more open-source models! 🚀
Let’s fill the top of the leaderboard with more open-source models! 🚀

> [!WARNING]
> `transformers.agents` has now been upgraded to the stand-alone library [smolagents](https://github.com/huggingface/smolagents)! The two libraries have very similar APIs, so switching is easy.
> Go checkout the `smolagents` introduction blog [here](https://huggingface.co/blog/smolagents).
Binary file added assets/smolagents-can-see/thumbnail.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
3 changes: 3 additions & 0 deletions beating-gaia.md
Original file line number Diff line number Diff line change
Expand Up @@ -10,6 +10,9 @@ authors:

After some experiments, we were impressed by the performance of Transformers Agents to build agentic systems, so we wanted to see how good it was! We tested using a [Code Agent built with the library](https://github.com/aymeric-roucher/GAIA) on the GAIA benchmark, arguably the most difficult and comprehensive agent benchmark… and ended up on top!

> [!WARNING]
> The framework `transformers.agents` used in this blog post has now been upgraded to the stand-alone library [smolagents](https://github.com/huggingface/smolagents)! The two libraries have very similar APIs, so switching is easy.
> Go checkout the `smolagents` introduction blog [here](https://huggingface.co/blog/smolagents).
## GAIA: a tough benchmark for Agents

Expand Down
4 changes: 4 additions & 0 deletions open-source-llms-as-agents.md
Original file line number Diff line number Diff line change
Expand Up @@ -12,6 +12,10 @@ authors:

Open-source LLMs have now reached a performance level that makes them suitable reasoning engines for powering agent workflows: [Mixtral](https://huggingface.co/blog/mixtral) even [surpasses GPT-3.5](#results) on our benchmark, and its performance could easily be further enhanced with fine-tuning.

> [!NOTE]
> We've released the simplest agentic library out there: [smolagents](https://github.com/huggingface/smolagents)!
> Go checkout the `smolagents` introduction blog [here](https://huggingface.co/blog/smolagents).
## Introduction

Large Language Models (LLMs) trained for [causal language modeling](https://huggingface.co/docs/transformers/tasks/language_modeling) can tackle a wide range of tasks, but they often struggle with basic tasks like logic, calculation, and search. The worst scenario is when they perform poorly in a domain, such as math, yet still attempt to handle all the calculations themselves.
Expand Down
Loading

0 comments on commit 6851e20

Please sign in to comment.