Spaces:
Sleeping
Sleeping
title: Submission Oriaz | |
emoji: 🔥 | |
colorFrom: yellow | |
colorTo: green | |
sdk: docker | |
pinned: True | |
# Benchmarkusing different techniques | |
## Global Informations : | |
#### Intended Use | |
- **Primary intended uses**: Baseline comparison for climate disinformation classification models | |
- **Primary intended users**: Researchers and developers participating in the Frugal AI Challenge | |
- **Out-of-scope use cases**: Not intended for production use or real-world classification tasks | |
### Training Data | |
The model uses the QuotaClimat/frugalaichallenge-text-train dataset: | |
- Size: ~6000 examples | |
- Split: 80% train, 20% test | |
- 8 categories of climate disinformation claims | |
#### Labels | |
0. No relevant claim detected | |
1. Global warming is not happening | |
2. Not caused by humans | |
3. Not bad or beneficial | |
4. Solutions harmful/unnecessary | |
5. Science is unreliable | |
6. Proponents are biased | |
7. Fossil fuels are needed | |
### Environmental Impact | |
Environmental impact is tracked using CodeCarbon, measuring: | |
- Carbon emissions during inference | |
- Energy consumption during inference | |
This tracking helps establish a baseline for the environmental impact of model deployment and inference. | |
### Ethical Considerations | |
- Dataset contains sensitive topics related to climate disinformation | |
- Environmental impact is tracked to promote awareness of AI's carbon footprint | |
## ML model for Climate Disinformation Classification | |
### Model Description | |
Find the best ML model to process vectorized quotes to detect climate change disinformation. | |
### Performance | |
#### Metrics (I used NVIDIA T4 small GPU) | |
- **Accuracy**: ~69-72% | |
- **Environmental Impact**: | |
- Emissions tracked in gCO2eq (~0,7g) | |
- Energy consumption tracked in Wh (~1,8wh) | |
#### Model Architecture | |
ML models prefers numeric values so we need to embed our quotes. I used *MTEB Leaderboard* on HuggingFace to find the model with the best trade-off between performance and the number of parameters. | |
I then chosed "dunzhang/stella_en_400M_v5" model as embedder. It has the 7th best performance score with only 400M parameters. | |
Once the quote are embedded, I have 6091 values x 1024 features. After that, train-test split (70%, 30%). | |
Using TPOT Classifier, I found that the best model on my data was a Logistic Regressor. | |
Then here is the Confusion Matrix : | |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/66169e1ce557753f30eab31b/tfAcfFu3Cnc9XJ00ixrWB.png) | |
### Limitations | |
- Embedding phase take ~30 secondes for 1800 quotes. It can be optimised and can have a real influence on carbon emissions. | |
- Hard to go over 70% accuracy with "simple" ML. | |
- Textual data have some interpretations limitations that little models can't find. | |
## Bert model for Climate Disinformation Classification | |
### Model Description | |
Fine tune model for model classification. | |
### Performance | |
#### Metrics (I used NVIDIA T4 small GPU) | |
- **Accuracy**: ~90% | |
- **Environmental Impact**: | |
- Emissions tracked in gCO2eq (~0,25g) | |
- Energy consumption tracked in Wh (~0.7wh) | |
#### Model Architecture | |
Fine tuning of "bert-uncased" model with 70% train, 15% eval, 15% test datasets. | |
### Limitations | |
- Not optimized. I need to try to run it on CPU | |
- Little models have limitations. Regularly between 70-80% accuracy. Hard to go over just by changing params. | |
# Contacts : | |
*LinkedIn* : Mattéo GIRARDEAU | |
*email* : [email protected] | |
``` | |