Remove explicit model arg in HfApiModel - which wouldn't run without Pro (#373)
* Remove explicit model arg in HfApiModel - which wouldn't run without Pro
This commit is contained in:
parent
023419bf92
commit
aa55f137e5
|
@ -127,20 +127,15 @@ The agent will need these arguments upon initialization:
|
||||||
- `model`: the LLM that powers the agent.
|
- `model`: the LLM that powers the agent.
|
||||||
Our `model` must be a callable that takes as input a list of messages and returns text. It also needs to accept a stop_sequences argument that indicates when to stop its generation. For convenience, we directly use the HfEngine class provided in the package to get a LLM engine that calls Hugging Face's Inference API.
|
Our `model` must be a callable that takes as input a list of messages and returns text. It also needs to accept a stop_sequences argument that indicates when to stop its generation. For convenience, we directly use the HfEngine class provided in the package to get a LLM engine that calls Hugging Face's Inference API.
|
||||||
|
|
||||||
And we use [meta-llama/Llama-3.3-70B-Instruct](meta-llama/Llama-3.3-70B-Instruct) as the llm engine because:
|
>[!NOTE] To use a specific model, pass it like this: `HfApiModel("meta-llama/Llama-3.3-70B-Instruct")`. The Inference API hosts models based on various criteria, and deployed models may be updated or replaced without prior notice. Learn more about it [here](https://huggingface.co/docs/api-inference/supported-models).
|
||||||
- It has a long 128k context, which is helpful for processing long source documents
|
|
||||||
- It is served for free at all times on HF's Inference API!
|
|
||||||
|
|
||||||
_Note:_ The Inference API hosts models based on various criteria, and deployed models may be updated or replaced without prior notice. Learn more about it [here](https://huggingface.co/docs/api-inference/supported-models).
|
|
||||||
|
|
||||||
```py
|
```py
|
||||||
from smolagents import HfApiModel, CodeAgent
|
from smolagents import HfApiModel, CodeAgent
|
||||||
|
|
||||||
agent = CodeAgent(
|
agent = CodeAgent(
|
||||||
tools=[retriever_tool], model=HfApiModel("meta-llama/Llama-3.3-70B-Instruct"), max_steps=4, verbosity_level=2
|
tools=[retriever_tool], model=HfApiModel(), max_steps=4, verbosity_level=2
|
||||||
)
|
)
|
||||||
```
|
```
|
||||||
|
|
||||||
Upon initializing the CodeAgent, it has been automatically given a default system prompt that tells the LLM engine to process step-by-step and generate tool calls as code snippets, but you could replace this prompt template with your own as needed.
|
Upon initializing the CodeAgent, it has been automatically given a default system prompt that tells the LLM engine to process step-by-step and generate tool calls as code snippets, but you could replace this prompt template with your own as needed.
|
||||||
|
|
||||||
Then when its `.run()` method is launched, the agent takes care of calling the LLM engine, and executing the tool calls, all in a loop that ends only when tool `final_answer` is called with the final answer as its argument.
|
Then when its `.run()` method is launched, the agent takes care of calling the LLM engine, and executing the tool calls, all in a loop that ends only when tool `final_answer` is called with the final answer as its argument.
|
||||||
|
|
Loading…
Reference in New Issue