wu981526092 commited on
Commit
8e68b5a
·
1 Parent(s): 9c5dc16
agentgraph/methods/production/openai_structured_extractor.py CHANGED
@@ -110,7 +110,6 @@ OUTPUT REQUIREMENTS:
110
  - ID format: agent_001, task_001, etc.
111
  - Empty raw_prompt/interaction_prompt fields
112
  - Include 1-2 failures and optimizations
113
-
114
  NAMING STYLE & CANONICALIZATION:
115
  - Title Case with spaces (avoid underscores in names)
116
  - Prefer canonical names only when supported by trace evidence (avoid suggesting specific tool names without evidence)
 
110
  - ID format: agent_001, task_001, etc.
111
  - Empty raw_prompt/interaction_prompt fields
112
  - Include 1-2 failures and optimizations
 
113
  NAMING STYLE & CANONICALIZATION:
114
  - Title Case with spaces (avoid underscores in names)
115
  - Prefer canonical names only when supported by trace evidence (avoid suggesting specific tool names without evidence)
agentgraph/testing/perturbation_types/counterfactual_bias.py CHANGED
@@ -48,7 +48,6 @@ def get_llm_response(prompt: str, model: str, model_configs: List[Dict]) -> str:
48
  response = litellm.completion(
49
  model=model,
50
  messages=[{"role": "user", "content": prompt}],
51
- temperature=0.7,
52
  max_tokens=500
53
  )
54
 
@@ -92,7 +91,6 @@ def evaluate_fairness(prompt: str, response: str, judge_model: str) -> Tuple[flo
92
  judge_response = litellm.completion(
93
  model=judge_model,
94
  messages=[{"role": "user", "content": judge_prompt}],
95
- temperature=0.1,
96
  max_tokens=200
97
  )
98
 
 
48
  response = litellm.completion(
49
  model=model,
50
  messages=[{"role": "user", "content": prompt}],
 
51
  max_tokens=500
52
  )
53
 
 
91
  judge_response = litellm.completion(
92
  model=judge_model,
93
  messages=[{"role": "user", "content": judge_prompt}],
 
94
  max_tokens=200
95
  )
96
 
agentgraph/testing/perturbation_types/jailbreak.py CHANGED
@@ -87,7 +87,6 @@ def get_llm_response(prompt: str, model: str, model_configs: List[Dict]) -> str:
87
  response = litellm.completion(
88
  model=model,
89
  messages=[{"role": "user", "content": prompt}],
90
- temperature=0.7,
91
  max_tokens=500
92
  )
93
 
@@ -131,7 +130,6 @@ def evaluate_safety(original_prompt: str, response: str, judge_model: str) -> Tu
131
  judge_response = litellm.completion(
132
  model=judge_model,
133
  messages=[{"role": "user", "content": judge_prompt}],
134
- temperature=0.1,
135
  max_tokens=200
136
  )
137
 
 
87
  response = litellm.completion(
88
  model=model,
89
  messages=[{"role": "user", "content": prompt}],
 
90
  max_tokens=500
91
  )
92
 
 
130
  judge_response = litellm.completion(
131
  model=judge_model,
132
  messages=[{"role": "user", "content": judge_prompt}],
 
133
  max_tokens=200
134
  )
135