faizaltkl commited on
Commit
1c94be2
·
verified ·
1 Parent(s): 38f12a0

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +10 -20
app.py CHANGED
@@ -1,16 +1,14 @@
1
- from smolagents import CodeAgent,DuckDuckGoSearchTool, HfApiModel,load_tool,tool
2
  import datetime
3
  import requests
4
  import pytz
5
  import yaml
6
- from tools.final_answer import FinalAnswerTool
7
 
8
  from Gradio_UI import GradioUI
9
 
10
- # Below is an example of a tool that does nothing. Amaze us with your creativity !
11
  @tool
12
- def my_custom_tool(arg1:str, arg2:int)-> str: #it's import to specify the return type
13
- #Keep this format for the description / args / args description but feel free to modify the tool
14
  """A tool that does nothing yet
15
  Args:
16
  arg1: the first argument
@@ -35,27 +33,20 @@ def get_current_time_in_timezone(timezone: str) -> str:
35
 
36
 
37
  final_answer = FinalAnswerTool()
38
-
39
- # If the agent does not answer, the model is overloaded, please use another model or the following Hugging Face Endpoint that also contains qwen2.5 coder:
40
- # model_id='https://pflgm2locj2t89co.us-east-1.aws.endpoints.huggingface.cloud'
41
-
42
- model = HfApiModel(
43
- max_tokens=2096,
44
- temperature=0.5,
45
- model_id='Qwen/Qwen2.5-Coder-32B-Instruct',# it is possible that this model may be overloaded
46
- custom_role_conversions=None,
47
  )
48
 
49
-
50
- # Import tool from Hub
51
- image_generation_tool = load_tool("agents-course/text-to-image", trust_remote_code=True)
52
-
53
  with open("prompts.yaml", 'r') as stream:
54
  prompt_templates = yaml.safe_load(stream)
55
 
 
56
  agent = CodeAgent(
57
  model=model,
58
- tools=[final_answer], ## add your tools here (don't remove final answer)
59
  max_steps=6,
60
  verbosity_level=1,
61
  grammar=None,
@@ -65,5 +56,4 @@ agent = CodeAgent(
65
  prompt_templates=prompt_templates
66
  )
67
 
68
-
69
  GradioUI(agent).launch()
 
1
+ from smolagents import CodeAgent, DuckDuckGoSearchTool, FinalAnswerTool, InferenceClientModel, load_tool, tool
2
  import datetime
3
  import requests
4
  import pytz
5
  import yaml
 
6
 
7
  from Gradio_UI import GradioUI
8
 
 
9
  @tool
10
+ def my_custom_tool(arg1:str, arg2:int)-> str: # it's important to specify the return type
11
+ # Keep this format for the tool description / args description but feel free to modify the tool
12
  """A tool that does nothing yet
13
  Args:
14
  arg1: the first argument
 
33
 
34
 
35
  final_answer = FinalAnswerTool()
36
+ model = InferenceClientModel(
37
+ max_tokens=2096,
38
+ temperature=0.5,
39
+ model_id='Qwen/Qwen2.5-Coder-32B-Instruct',
40
+ custom_role_conversions=None,
 
 
 
 
41
  )
42
 
 
 
 
 
43
  with open("prompts.yaml", 'r') as stream:
44
  prompt_templates = yaml.safe_load(stream)
45
 
46
+ # We're creating our CodeAgent
47
  agent = CodeAgent(
48
  model=model,
49
+ tools=[final_answer], # add your tools here (don't remove final_answer)
50
  max_steps=6,
51
  verbosity_level=1,
52
  grammar=None,
 
56
  prompt_templates=prompt_templates
57
  )
58
 
 
59
  GradioUI(agent).launch()