|
11 | 11 | for Tool Success evaluator using inline dataset content. |
12 | 12 |
|
13 | 13 | USAGE: |
14 | | - python sample_tool_success.py |
| 14 | + python sample_tool_call_success.py |
15 | 15 |
|
16 | 16 | Before running the sample: |
17 | 17 |
|
@@ -71,18 +71,18 @@ def main() -> None: |
71 | 71 | testing_criteria = [ |
72 | 72 | { |
73 | 73 | "type": "azure_ai_evaluator", |
74 | | - "name": "tool_success", |
75 | | - "evaluator_name": "builtin.tool_success", |
| 74 | + "name": "tool_call_success", |
| 75 | + "evaluator_name": "builtin.tool_call_success", |
76 | 76 | "initialization_parameters": {"deployment_name": f"{model_deployment_name}"}, |
77 | 77 | "data_mapping": {"tool_definitions": "{{item.tool_definitions}}", "response": "{{item.response}}"}, |
78 | 78 | } |
79 | 79 | ] |
80 | 80 |
|
81 | 81 | print("Creating Eval Group") |
82 | 82 | eval_object = client.evals.create( |
83 | | - name="Test Tool Success Evaluator with inline data", |
84 | | - data_source_config=data_source_config, # type: ignore |
85 | | - testing_criteria=testing_criteria, # type: ignore |
| 83 | + name="Test Tool Call Success Evaluator with inline data", |
| 84 | + data_source_config=data_source_config, # type: ignore |
| 85 | + testing_criteria=testing_criteria, # type: ignore |
86 | 86 | ) |
87 | 87 | print(f"Eval Group created") |
88 | 88 |
|
|
0 commit comments