diff --git a/src/google/adk/evaluation/response_evaluator.py b/src/google/adk/evaluation/response_evaluator.py index 463f21f..c5da692 100644 --- a/src/google/adk/evaluation/response_evaluator.py +++ b/src/google/adk/evaluation/response_evaluator.py @@ -42,7 +42,7 @@ class ResponseEvaluator: A note on evaluation_criteria: `response_match_score`: This metric compares the agents final natural - language reponse with the expected final response, stored in the + language response with the expected final response, stored in the "reference" field in test/eval files. We use Rouge metric to compare the two responses. diff --git a/src/google/adk/flows/llm_flows/contents.py b/src/google/adk/flows/llm_flows/contents.py index e473963..f284755 100644 --- a/src/google/adk/flows/llm_flows/contents.py +++ b/src/google/adk/flows/llm_flows/contents.py @@ -310,7 +310,7 @@ def _merge_function_response_events( function_response_events: A list of function_response events. NOTE: function_response_events must fulfill these requirements: 1. The list is in increasing order of timestamp; 2. the first event is the - initial function_reponse event; 3. all later events should contain at + initial function_response event; 3. all later events should contain at least one function_response part that related to the function_call event. (Note, 3. may not be true when aync function return some intermediate response, there could also be some intermediate model diff --git a/src/google/adk/tests/unittests/utils.py b/src/google/adk/tests/unittests/utils.py index 592ff15..2e74db9 100644 --- a/src/google/adk/tests/unittests/utils.py +++ b/src/google/adk/tests/unittests/utils.py @@ -231,7 +231,7 @@ class MockModel(BaseLlm): if not responses: return cls(responses=[]) elif isinstance(responses[0], LlmResponse): - # reponses is list[LlmResponse] + # responses is list[LlmResponse] return cls(responses=responses) else: responses = [