-
./technical-report- It contains
Technical_Report.pdf - Kindly download for better readability.
- It contains
-
./queries-generation- It generate queries and answer, with three methods.
- method-1: transcript agnostic, using domain, intent, mece-category and sub-category.
- method-2: transcript agnostic, using Self-Instruct and Auto-Evol instruct.
- method-3: with raw transcripts.
-
./simulated-real-query-system- Curated dataset of more than 70 queries.
- task-1-queries.xlsx, and task-2-queries.xlsx.
- Kindly download both the files for better readability.
-
./recor-rag(dockerized)- Dockerized Jupyter Notebook for
final query answer generationfor task 1 and task 2. - Below are the step to re-produce this.
- Dockerized Jupyter Notebook for
-
Create account in https://lightning.ai.
- it has all the requirements needed to run the docker file.
- you need an email, and phone number (no credit/debit card needed).
-
After creating an account, open https://lightning.ai and click on
+ new studiobutton on the top right corner. -
In the machine type select
GPU, thenL40S, also make sureInterruptibleis toggled off. (If you've already started the studio withCPU, then change toGPUfrom top right corner) -
Now confirm and start.
-
Upload the zip-file provided with this readme (or run the below command if the zip file does not work) on lightning.ai instance just created.
# or if you prefer to download using wget wget https://storage.googleapis.com/recor-rag/team_65.zip -
Unzip the file by running the following commands
sudo apt update sudo apt install -y unzip unzip team_65.zip
-
Run the followings commands to download the models
# 1. dowload openai/gpt-oss-20b model (~41 GB) gsutil -m cp -r gs://recor-rag/model-gptoss . # 2. dowload colbert, and embeddings model (~3 GB) gsutil -m cp -r gs://recor-rag/huggingface .
-
Current directory should look like this
. ├── .gitignore ├── README.md ├── docker-compose.yml ├── images/* ├── queries-generation │ ├── method-1 │ │ ├── query-evaluation │ │ │ ├── README.md │ │ │ ├── llm-as-a-judge-task-1.ipynb │ │ │ ├── llm-as-a-judge-task-2.ipynb │ │ │ ├── llm-eval-stage-1.ipynb │ │ │ ├── non-llm-eval-task-1.ipynb │ │ │ ├── non-llm-eval-task-2.ipynb │ │ │ ├── test1.json │ │ │ └── test2.json │ │ └── query-generation │ │ ├── README.md │ │ └── query-generator-task-1-2.ipynb │ ├── method-2 │ │ ├── README.md │ │ ├── auto_evol.py │ │ ├── config.py │ │ ├── main.py │ │ ├── requirements.txt │ │ ├── seed_generator.py │ │ ├── self_instruct.py │ │ └── utils │ │ ├── failure_detector.py │ │ ├── init.py │ │ ├── llm_client.py │ │ └── prompts.py │ └── method-3 │ ├── README.md │ ├── dataset-generation-part-1.ipynb │ ├── dataset-generation-pipeline.ipynb │ ├── final-transcripts-domain-corrected.json │ ├── images/* │ ├── requirements.txt │ └── summary_clusters.ipynb ├── recor-rag │ ├── Dockerfile │ ├── answer-generation.ipynb │ ├── const │ │ ├── faiss_index │ │ │ ├── index.faiss │ │ │ └── index.pkl │ │ └── summaries-20k.json │ └── requirements.txt ├── simulated-real-query-system │ ├── task-1-queries.xlsx │ └── task-2-queries.xlsx └── technical-report └── Technical_Report.pdf
-
Start the container.
# to build the `vllm-server`, and `recor-rag` container docker compose build # to load those container, `recor-rag` will start quickly # but `vllm-server` one will take around 3-4 minutes, as it needs to load the weights in the GPU, and caputre CUDA Graphs docker compose up
-
Once the container starts,
- The vLLM server will be live at
http://localhost:8000 - Recor-Rag sever will be live at
http://localhost:8888
- The vLLM server will be live at
-
To open the jupyter notebook,
-
Open jupyter notebook (in the public link)
- All the images above are from
https://lightning.ai.





