Spaces:
Sleeping
Sleeping
Update README.md
Browse files
README.md
CHANGED
|
@@ -1,8 +1,6 @@
|
|
| 1 |
---
|
| 2 |
title: Image App Space
|
| 3 |
emoji: 🔥
|
| 4 |
-
colorFrom: red
|
| 5 |
-
colorTo: yellow
|
| 6 |
sdk: gradio
|
| 7 |
sdk_version: 5.47.0
|
| 8 |
app_file: app.py
|
|
@@ -11,4 +9,29 @@ license: mit
|
|
| 11 |
short_description: A simple user-facing interface for an image model
|
| 12 |
---
|
| 13 |
|
| 14 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
| 2 |
title: Image App Space
|
| 3 |
emoji: 🔥
|
|
|
|
|
|
|
| 4 |
sdk: gradio
|
| 5 |
sdk_version: 5.47.0
|
| 6 |
app_file: app.py
|
|
|
|
| 9 |
short_description: A simple user-facing interface for an image model
|
| 10 |
---
|
| 11 |
|
| 12 |
+
# AutoGluon Sign Identification Predictor
|
| 13 |
+
|
| 14 |
+
This repository contains a trained MultiModalPredictor from the AutoGluon library, which was trained to identify signs from images. Which can also be found in the files and versions section under AutoML_for_Neural_Networks
|
| 15 |
+
|
| 16 |
+
# Dataset
|
| 17 |
+
The model was trained on the ecopus/sign_identification dataset. The augmented split was used for training and validation, while the original split was used for the final evaluation of the model's performance.
|
| 18 |
+
|
| 19 |
+
# Model
|
| 20 |
+
The model is borrowed from its-zion-18/sign-image-autogluon-predictor model.
|
| 21 |
+
|
| 22 |
+
# Evaluation Results
|
| 23 |
+
|
| 24 |
+
The final performance of the best model on the original dataset is as follows:
|
| 25 |
+
|
| 26 |
+
- **Accuracy**: `1.0000`
|
| 27 |
+
- **Weighted F1**: `1.0000`
|
| 28 |
+
|
| 29 |
+
# Potential Errors
|
| 30 |
+
The augmented split in the ecopus/sign_identification dataset is specifically designed to be an artificially expanded version of the original split.
|
| 31 |
+
The images in the augmented set are simple transformations—like rotations, flips, or slight color changes—of the images in the original set.
|
| 32 |
+
The code then trains the model on a portion of the augmented data (df_aug_train) and evaluates it on the original data (df_orig).
|
| 33 |
+
Because the model was trained on data that is derived directly from the evaluation data, it's not actually seeing truly "new" information during the final test.
|
| 34 |
+
Which could be leading to data leakage and overfitting
|
| 35 |
+
|
| 36 |
+
# Contact
|
| 37 |
+
Feel free to contact me for any questions or concerns: aslann@cmu.edu
|