Update README.md
Browse files
README.md
CHANGED
|
@@ -12,7 +12,9 @@ tags:
|
|
| 12 |
- EmoSet
|
| 13 |
---
|
| 14 |
|
| 15 |
-
|
|
|
|
|
|
|
| 16 |
|
| 17 |
## Training Details
|
| 18 |
|
|
@@ -23,7 +25,8 @@ PreceptCLIP-Emotions is a model designed to predict the emotions that an image e
|
|
| 23 |
- *Learning Rate*: 0.0001
|
| 24 |
- *Batch Size*: 32
|
| 25 |
|
| 26 |
-
## Requirements
|
|
|
|
| 27 |
- python=3.9.15
|
| 28 |
- cudatoolkit=11.7
|
| 29 |
- torchvision=0.14.0
|
|
|
|
| 12 |
- EmoSet
|
| 13 |
---
|
| 14 |
|
| 15 |
+
**PerceptCLIP-Emotions** is a model designed to predict the **emotions** that an image evokes in users. This is the official model from the paper:
|
| 16 |
+
📄 **["Don't Judge Before You CLIP: A Unified Approach for Perceptual Tasks"](https://arxiv.org/abs/2503.13260)**
|
| 17 |
+
We apply **LoRA adaptation** on the **CLIP visual encoder** and add an **MLP head** for emotion classification. Our model achieves **state-of-the-art results**.
|
| 18 |
|
| 19 |
## Training Details
|
| 20 |
|
|
|
|
| 25 |
- *Learning Rate*: 0.0001
|
| 26 |
- *Batch Size*: 32
|
| 27 |
|
| 28 |
+
## Installation & Requirements
|
| 29 |
+
You can set up the environment using environment.yml or manually install dependencies:
|
| 30 |
- python=3.9.15
|
| 31 |
- cudatoolkit=11.7
|
| 32 |
- torchvision=0.14.0
|