Add initial working implementation using search grounding #17
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
This pull request introduces a new agent evaluation framework for knowledge-grounded question answering (QA) using Gemini's Google Search grounding, along with rich display utilities for evaluation outputs. It adds well-structured modules for agent implementation, configuration, evaluation, and user-friendly display, and updates environment and linting configurations to support the new features.
Agent and Evaluation Framework:
aieng-eval-agents/aieng/agent_evals/knowledge_agent/agent.py)aieng-eval-agents/aieng/agent_evals/knowledge_agent/agent.py)aieng-eval-agents/aieng/agent_evals/knowledge_agent/__init__.py)__init__.pyto expose display utilities and document the agent evaluation tools. (aieng-eval-agents/aieng/agent_evals/__init__.py)Display and Usability Enhancements:
richlibrary for improved readability and user experience. (aieng-eval-agents/aieng/agent_evals/display.py)Configuration and Tooling:
.env.example).pre-commit-config.yaml)