HemaAM commited on
Commit
04cadd7
·
1 Parent(s): 1071e1e

Updated Readme

Browse files
Files changed (1) hide show
  1. README.md +29 -0
README.md CHANGED
@@ -10,4 +10,33 @@ pinned: false
10
  license: mit
11
  ---
12
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
  Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
 
10
  license: mit
11
  ---
12
 
13
+ # Object Detection App for Yolov3 model using Pascal VOC dataset
14
+ ## How to Use the App
15
+ 1. The app has one tab of images:
16
+ - **Examples**: In this tab, the app displays a gallery of images from Pascal VOC dataset. We can upload any one of the images provided in the list shown below to the app widget screen. We can choose the IoU threshold, threshold using the given sliders. Then we click on submit for the inference to show the objects detected in the image and the corresponding top classes along with their confidence levels. We visualize the class activation maps using GradCAM, show/hide the GradCAM overlay by controling the transparency of the overlay, second to the last layer in the Yolo v3 network architecture is used for displaying the GradCam in this application.
17
+
18
+ 2. **Examples Tab - Description of the options to choose**:
19
+ - **Input Image**: Select one of the example images from the given list.
20
+ - **IoU threshold**: Move the Slider to the float value between 0 to 1 to get the best one bounding box that covers the object the most. The default value for this is different for each of the images in the gallery. And the default is set in a way that redundant bounding boxes are not cluttering the image.
21
+ - **Threshold**: Move the Slider to the float value between 0 to 1 to set the best confidence threshold for the object to be detected. The default value for this is different for each of the images in the gallery.
22
+ - **Enable GradCAM**: Check this box to display the GradCAM overlay on the input image. Uncheck it to view only the original image.
23
+ - **Transparency**: Control the transparency of the GradCAM overlay. The default value is 0.6.
24
+
25
+ 4. After adjusting the settings, click the "Submit" button to see the results.
26
+
27
+ ## Source code of training the model
28
+ The main code using which training was performed can be viewed at below location:
29
+
30
+ [https://github.com/mHemaAP/S13](https://github.com/mHemaAP/S13)
31
+
32
+ ## Credits
33
+
34
+ - This app is built using the Gradio library ([https://www.gradio.app/](https://www.gradio.app/)) for interactive model interfaces.
35
+ - The PyTorch library ([https://pytorch.org/](https://pytorch.org/)) is used for the deep learning model and GradCAM visualization.
36
+ - The CIFAR-10 dataset ([https://www.kaggle.com/datasets/aladdinpersson/pascal-voc-dataset-used-in-yolov3-video](https://www.kaggle.com/datasets/aladdinpersson/pascal-voc-dataset-used-in-yolov3-video)) is used for training and evaluation.
37
+
38
+ ## License
39
+
40
+ This project is licensed under the MIT License - see the [LICENSE](LICENSE) file for details.
41
+
42
  Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference