Update README.md
Browse files
README.md
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
# RTDETR Model on COCO8 Dataset
|
2 |
|
3 |
-
This model is a **Vision Transformer** (ViT) based object detection and tracking model, trained on the **COCO8** dataset
|
4 |
|
5 |
## Model Details
|
6 |
|
@@ -8,17 +8,8 @@ This model is a **Vision Transformer** (ViT) based object detection and tracking
|
|
8 |
- **Trained On**: COCO8 dataset (people with and without coats)
|
9 |
- **Training Epochs**: 100 epochs
|
10 |
- **Input Size**: 640x640 pixels
|
11 |
-
- **Output**:
|
12 |
-
- `1`: People wearing coats
|
13 |
-
- `0`: People not wearing coats
|
14 |
|
15 |
## How to Use
|
16 |
|
17 |
You can use this model directly from the Hugging Face Hub. Below is an example of how to use it for inference on your images.
|
18 |
-
|
19 |
-
### Install Dependencies
|
20 |
-
|
21 |
-
First, ensure you have the `transformers` and `torch` libraries installed:
|
22 |
-
|
23 |
-
```bash
|
24 |
-
pip install transformers torch
|
|
|
1 |
# RTDETR Model on COCO8 Dataset
|
2 |
|
3 |
+
This model is a **Vision Transformer** (ViT) based object detection and tracking model, trained on the **COCO8** dataset.
|
4 |
|
5 |
## Model Details
|
6 |
|
|
|
8 |
- **Trained On**: COCO8 dataset (people with and without coats)
|
9 |
- **Training Epochs**: 100 epochs
|
10 |
- **Input Size**: 640x640 pixels
|
11 |
+
- **Output**: Detects and tracks objects through the frames in any input video
|
|
|
|
|
12 |
|
13 |
## How to Use
|
14 |
|
15 |
You can use this model directly from the Hugging Face Hub. Below is an example of how to use it for inference on your images.
|
|
|
|
|
|
|
|
|
|
|
|
|
|