I’m currently working on evaluating LLMs and conversational AI solutions, focusing primarily on testing their domain-specific performance and multilingual capabilities.☔🌦️🪐❄️
-
The Mojolab Foundation
- Delhi, India
-
13:13
(UTC +05:30) - https://orcid.org/0009-0006-7881-5826
- in/palak-garg-0a3937267
- palakgarg21_
Pinned Loading
-
-
aditisinghh17/PubEvalAI
aditisinghh17/PubEvalAI PublicDeveloping an Evaluation Framework for Conversational AI in Public Procurement: Enhancing Decision-Making in AI Product Selection
Jupyter Notebook
-
aditisinghh17/LegalRAG-Evaluation
aditisinghh17/LegalRAG-Evaluation PublicThis repository contains evaluations of LLM-based Retrieval-Augmented Generation (RAG) pipelines tailored for the legal domain. It currently features RAG with three different LLMs: Mistral, OpenAI,…
Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
If the problem persists, check the GitHub status page or contact support.