Commit
·
df530f5
1
Parent(s):
5da3e0f
Update README.md
Browse files
README.md
CHANGED
|
@@ -1,16 +1,15 @@
|
|
| 1 |
---
|
| 2 |
-
|
| 3 |
tags:
|
| 4 |
- Keras
|
| 5 |
-
|
| 6 |
license: mit
|
| 7 |
-
|
| 8 |
metrics:
|
| 9 |
- PSNR
|
| 10 |
-
|
|
|
|
|
|
|
|
|
|
| 11 |
---
|
| 12 |
|
| 13 |
-
|
| 14 |
Here is a fully trained model of EDSR (Enhanced Deep Residual Networks for Single Image Super-Resolution) model. This model surpassed the performance of the current available SOTA models.
|
| 15 |
|
| 16 |
Spaces link - https://huggingface.co/spaces/keras-io/EDSR
|
|
@@ -21,7 +20,27 @@ Keras Example link - https://keras.io/examples/vision/edsr/
|
|
| 21 |
|
| 22 |
It was trained for 500 epochs with 200 steps each.
|
| 23 |
|
| 24 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 25 |
|
| 26 |
Hack to make this work for any image size. Currently the model takes input of image size 150 x 150.
|
| 27 |
We pad the input image with transparant pixels so that it is a square image, which is a multiple of 150 x 150
|
|
@@ -29,4 +48,4 @@ Then we chop the image into multiple 150 x 150 sub images
|
|
| 29 |
Upscale it and stich it together.
|
| 30 |
|
| 31 |
The output image might look a bit off, because each sub-image dosent have data about other sub-images.
|
| 32 |
-
This approach assumes that the
|
|
|
|
| 1 |
---
|
|
|
|
| 2 |
tags:
|
| 3 |
- Keras
|
|
|
|
| 4 |
license: mit
|
|
|
|
| 5 |
metrics:
|
| 6 |
- PSNR
|
| 7 |
+
datasets:
|
| 8 |
+
- eugenesiow/Div2k
|
| 9 |
+
library_name: keras
|
| 10 |
+
pipeline_tag: image-to-image
|
| 11 |
---
|
| 12 |
|
|
|
|
| 13 |
Here is a fully trained model of EDSR (Enhanced Deep Residual Networks for Single Image Super-Resolution) model. This model surpassed the performance of the current available SOTA models.
|
| 14 |
|
| 15 |
Spaces link - https://huggingface.co/spaces/keras-io/EDSR
|
|
|
|
| 20 |
|
| 21 |
It was trained for 500 epochs with 200 steps each.
|
| 22 |
|
| 23 |
+
|
| 24 |
+
|
| 25 |
+
# Enhanced Deep Residual Networks for Single Image Super-Resolution
|
| 26 |
+
|
| 27 |
+
## Introduction
|
| 28 |
+
|
| 29 |
+
This repository contains a trained model based on the Enhanced Deep Residual Networks for Single Image Super-Resolution paper. The model was trained for 500 epochs with 200 steps each, resulting in a high-quality super-resolution model.
|
| 30 |
+
|
| 31 |
+
## Dataset Used
|
| 32 |
+
|
| 33 |
+
The model was trained on the DIV2K dataset, which is a newly proposed high-quality (2K resolution) image dataset for image restoration tasks. The DIV2K dataset consists of 800 training images, 100 validation images, and 100 test images.
|
| 34 |
+
|
| 35 |
+
## Architecture
|
| 36 |
+
|
| 37 |
+
The Enhanced Deep Residual Networks for Single Image Super-Resolution paper presents an enhanced deep super-resolution network (EDSR) and a new multi-scale deep super-resolution system (MDSR) that outperform current state-of-the-art SR methods. The EDSR model optimizes performance by analyzing and removing unnecessary modules to simplify the network architecture. The MDSR system is a multi-scale architecture that shares most of the parameters across different scales, using significantly fewer parameters compared with multiple single-scale models but showing comparable performance.
|
| 38 |
+
|
| 39 |
+
## Metrics
|
| 40 |
+
|
| 41 |
+
The model was evaluated using the PSNR (Peak Signal-to-Noise Ratio) metric, which measures the quality of the reconstructed image compared to the original image. The model achieved a PSNR of approximately 31, which is a high-quality result.
|
| 42 |
+
|
| 43 |
+
## TODO:
|
| 44 |
|
| 45 |
Hack to make this work for any image size. Currently the model takes input of image size 150 x 150.
|
| 46 |
We pad the input image with transparant pixels so that it is a square image, which is a multiple of 150 x 150
|
|
|
|
| 48 |
Upscale it and stich it together.
|
| 49 |
|
| 50 |
The output image might look a bit off, because each sub-image dosent have data about other sub-images.
|
| 51 |
+
This approach assumes that the sub-image has enough data about its surroundings
|