Authors: Soham Bhattacharya1,3, Ali Reza Alaei2 & Umapada Pal3
Affiliated institutions: 1Ramakrishna Mission Vivekananda Educational and Research Institute Belur, India; 2Southern Cross University, Australia; 3Indian Statistical Institute Kolkata, India;
- The development of an enhanced MSA framework that demonstrates superior performance over the existing baselines
- The integration of a novel Fault-Tolerance (FT) mechanism that ensures architectural robustness and handles out-of-distribution (OOD) data which are faulty inputs
- Extending and a comprehensive curation and refinement of the dataset DocImsent introduced by Ahuja et. al (2024) to improve data quality and volume of the dataset.
How to get the models:
Please contact the author directly to request access to the trained model files. Once received, place them in the src/trained/ folder before running the application.
-
Clone the repository
git clone [https://github.com/soham-b-github/msa.git](https://github.com/soham-b-github/msa.git) cd msa -
Install dependencies
pip install -r requirements.txt
Key requirements:
torch,transformers,google-cloud-vision,clip(OpenAI),Pillow,flask/streamlit(for app.py). -
Setup Google Cloud credentials
- Export your service account key:
export GOOGLE_APPLICATION_CREDENTIALS="path/to/your/service-account-file.json"
- Export your service account key:
Once you have obtained the trained model files from the author:
- Ensure the model file is in the correct directory.
- Run the frontend application:
streamlit run app.py
- Upload an image (e.g., a meme or poster) to receive the predicted sentiment (Positive, Negative, or Neutral).
To evaluate the framework, you can either use the samples provided in this repository or integrate your own dataset. The system expects a specific directory hierarchy within the data/ folder:
- Using provided samples: The repository includes sample images located in
data/docimsentv1-samples/anddata/external-samples/. These are ready for immediate testing. - Using custom datasets: If you wish to evaluate the framework on a full dataset, create a new subdirectory within the
data/folder and place your images there. - Run the following code in the terminal:
python3 evaluation.py --DATASET_PATH=./../data/my_new_datasetRecommended directory structure for data/ folder:
mevitsa/
├── data/
│ ├── docimsentv1-samples/ # Existing sample data
│ ├── external-samples/ # Existing sample data
│ └── [your-dataset-name]/ # <--- PLACE CUSTOM DATASETS HERE
│ ├── image_001.jpg
│ └── image_002.png
This repository is associated with a research paper currently under communication with ICPR. Citation information will be made available after the paper is accepted and published.