-
Notifications
You must be signed in to change notification settings - Fork 943
RAG Performance & Fairness Evaluation Toolkit (OpenVINO + LangChain) #3114
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: latest
Are you sure you want to change the base?
Conversation
…. The toolkit computes standard metrics (BERT, BLEU, ROUGE, perplexity score) and a racial-bias indicator, and it is implemented using Optimum-Intel’s OVModelForCausalLM with the OpenVINO backend and LangChain for orchestration.
|
Check out this pull request on See visual diffs & provide feedback on Jupyter Notebooks. Powered by ReviewNB |
|
Hi @pkhara31 , the notebook https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/llm-rag-langchain/llm-rag-langchain.ipynb seems to cover the same tasks you described in comment. Could you please check ? Does this match your idea of the notebook you wanted to add ? |
|
Hi @sbalandi , This NB covers the methodology to evaluate the performance of RAG pipeline by computing the BERT, BLEU, ROGUE, perplexity, racial bias scores. |
|
Thanks for your contribution @pkhara31 Do you think if we can add this NB to RAG notebook, as a separate |
|
Another idea is to change this |
|
The model loading redundancy can be avoided, but I think the evaluation and the ability to apply RAG pipeline on web docs (scraping) should still remain in the helper. |
|
May I know the latest update on this? |
Hi, sorry for the delay, I'm in the process of reviewing notebook, I'll write comments within a week |
This toolkit enables developers to build, evaluate, and optimize Retrieval-Augmented Generation (RAG) applications with comprehensive quality metrics including accuracy, bias detection, and perplexity analysis plus a racial-bias indicator. This uses RAG pipeline optimized with Intel OpenVINO for enhanced performance on CPU, GPU, and NPU. The pipeline leverages:
Optimum-Intel’s OVModelForCausalLM with the OpenVINO backend for efficient inference.
LangChain for orchestration of document loading, chunking, embedding, retrieval, reranking, and generation.
Goal: Provide a portable notebook-driven workflow for rapid experimentation, model comparison, and validation of RAG systems on custom/private corpora.