-
Notifications
You must be signed in to change notification settings - Fork 3
notebook: Document an agentic AI system #472
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
notebook: Document an agentic AI system #472
Conversation
AnilSorathiya
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
minor comment:
We have remove the StepEfficiency scorer from the code due to bug. Main branch doesn't have it.
Otherwise it's looks good to me. Thanks 👍
5ae6576 to
8ea7f00
Compare
|
@AnilSorathiya It works and passes the code quality test in the current version:
|
PR SummaryThis pull request introduces a new YAML template (agentic_ai_template.yaml) that defines comprehensive guidelines for documenting agentic AI systems. The template is structured into multiple sections including conceptual soundness, data evaluation, model evaluation, and observability and monitoring. Each section comprises detailed guidelines (with examples and hierarchical parent section references) aimed at enabling users to document features such as autonomy, reasoning, memory, risk management, regulatory compliance, and more. Additionally, a new Jupyter Notebook (document_agentic_ai.ipynb) has been added. This notebook provides step‐by‐step instructions to build and document an agentic AI system using the ValidMind Library. It includes detailed markdown explanations, code cells for installing dependencies, initializing the ValidMind environment, building and testing agent workflows, and running validation tests. The notebook guides users to verify LLM API access via environment variable configuration, integrate banking tools, bind the tools to the agent, and finally to capture test results including AI evaluation metrics. Significantly, a legacy notebook (langgraph_agent_simple_banking_demo.ipynb) has been removed, likely because its functionality is being replaced or superseded by the new, more comprehensive documentation template and notebook. Overall, the PR refactors the documentation and testing approach for agentic AI systems by providing a structured template and modernizing the developer guides. Test Suggestions
|


Pull Request Description
What and why?
Agentic AItemplate as a notebook artifact as it isn't a default out-of-the-box template: agentic_ai_template.yamlHow to test
gh pr checkout 472StepEfficiencyin your library environment by registering a new Python kernelnotebooks/code_samples/agents/document_agentic_ai.ipynbWhat needs special review?
The notebook runs end-to-end without issues in my environment, but you should check that everything looks fine to them as well.
General topic and test wording
I understand how to use the library functions, how to run tests, etc. very well at this point but I don't necessarily know why we're choosing to run the tests that we do, so please make sure that my descriptions are accurate and relevant.
Assigning AI evaluation metric scores section
Important
Please check that the following are all correct and do what we want them to do for the reasons outlined in the notebook:
Dependencies, breaking changes, and deployment notes
Refer to the line about
StepEfficiency.pyabove.Release notes
Learn how to build and document an agentic AI system with the ValidMind Library with our new notebook. Construct a LangGraph-based banking agent that selects and invokes tools in response to user requests. You'll assign AI evaluation scores to your agent, run accuracy, RAGAS, and safety tests, and log the results of your tests to the ValidMind Platform.
Document an agentic AI system
Checklist