An AI project for processing and training models on question datasets.
This project requires Python 3.12 or higher. Check your Python version:
python --versionCreate a virtual environment using Python 3.12:
# Create virtual environment
py -3.12 -m venv venv312
# Activate the environment
# Windows
venv312\Scripts\activate
# Linux/macOS
source venv312/bin/activateAfter activating the virtual environment, install the required packages:
pip install -r requirement.txtNote: The requirements.txt has been updated to support Python 3.12 with the latest compatible package versions including:
- numpy 2.1.2 (latest stable version)
- scipy 1.14.1 (latest stable version)
- torch 2.5.1+cu121 (with CUDA 12.1 support)
To access Hugging Face models and datasets, you need to authenticate with your Hugging Face token.
- Install the Hugging Face CLI (already included in requirements):
pip install huggingface_hub- Login with your Hugging Face token:
huggingface-cli login-
Enter your token when prompted. You can get your token from:
- Go to https://huggingface.co/settings/tokens
- Click "New token"
- Choose appropriate permissions (Read/Write)
- Copy the generated token
-
The token will be automatically saved to
~/.cache/huggingface/tokenfor future use.
Alternatively, you can set an environment variable:
# Windows (PowerShell)
$env:HUGGINGFACE_HUB_TOKEN="your_token_here"
# Windows (Command Prompt)
set HUGGINGFACE_HUB_TOKEN=your_token_here
# Linux/macOS
export HUGGINGFACE_HUB_TOKEN="your_token_here"Create a .env file in the project root:
HUGGINGFACE_HUB_TOKEN=your_token_here.env to .gitignore.
Test your setup by running:
from huggingface_hub import whoami
print(whoami())[Add your usage instructions here]
questions-ai/
├── cleanData/ # Data cleaning utilities
├── data/ # Raw datasets
├── cleaned_data/ # Processed datasets
├── requirement.txt # Python dependencies
└── README.md # This file