Publication:

Context-Aware Image Super-Resolution Using Deep Neural Networks

cris.lastimport.scopus

2024-08-07T09:30:27Z

cris.lastimport.wos

2024-08-05T03:24:36Z

cris.legacyId

286804

cris.virtual.author-scopus

35554798200

cris.virtual.department

LTS5

cris.virtual.parent-organization

STI

cris.virtual.parent-organization

EPFL

cris.virtual.parent-organization

EPFL

cris.virtual.parent-organization

EDOC

cris.virtual.parent-organization

ETU

cris.virtual.parent-organization

EPFL

cris.virtual.parent-organization

IEM

cris.virtual.parent-organization

STI

cris.virtual.parent-organization

EPFL

cris.virtual.sciperId

115534

cris.virtual.sciperId

194306

cris.virtual.unitId

10954

cris.virtual.unitManager

Thiran, Jean-Philippe

cris.virtualsource.author-scopus

18c34af8-ff12-466f-97c1-52e627259417

cris.virtualsource.author-scopus

dde778f1-5db3-4325-aee5-e926f9f969f7

cris.virtualsource.department

18c34af8-ff12-466f-97c1-52e627259417

cris.virtualsource.department

dde778f1-5db3-4325-aee5-e926f9f969f7

cris.virtualsource.orcid

18c34af8-ff12-466f-97c1-52e627259417

cris.virtualsource.orcid

dde778f1-5db3-4325-aee5-e926f9f969f7

cris.virtualsource.parent-organization

b90b43a5-ca1d-4299-9a3f-53ece71373f9

cris.virtualsource.parent-organization

b90b43a5-ca1d-4299-9a3f-53ece71373f9

cris.virtualsource.parent-organization

b90b43a5-ca1d-4299-9a3f-53ece71373f9

cris.virtualsource.parent-organization

e241245b-0e63-4d9e-806e-b766e62006ef

cris.virtualsource.parent-organization

e241245b-0e63-4d9e-806e-b766e62006ef

cris.virtualsource.parent-organization

5520d273-fb5f-458c-93a9-f6a9eee8961b

cris.virtualsource.parent-organization

5520d273-fb5f-458c-93a9-f6a9eee8961b

cris.virtualsource.parent-organization

5520d273-fb5f-458c-93a9-f6a9eee8961b

cris.virtualsource.parent-organization

5520d273-fb5f-458c-93a9-f6a9eee8961b

cris.virtualsource.parent-organization

24256c18-3294-40df-91a5-eefdca985e4a

cris.virtualsource.parent-organization

24256c18-3294-40df-91a5-eefdca985e4a

cris.virtualsource.parent-organization

24256c18-3294-40df-91a5-eefdca985e4a

cris.virtualsource.parent-organization

24256c18-3294-40df-91a5-eefdca985e4a

cris.virtualsource.rid

18c34af8-ff12-466f-97c1-52e627259417

cris.virtualsource.rid

dde778f1-5db3-4325-aee5-e926f9f969f7

cris.virtualsource.sciperId

18c34af8-ff12-466f-97c1-52e627259417

cris.virtualsource.sciperId

dde778f1-5db3-4325-aee5-e926f9f969f7

cris.virtualsource.unitId

24256c18-3294-40df-91a5-eefdca985e4a

cris.virtualsource.unitManager

24256c18-3294-40df-91a5-eefdca985e4a

datacite.rights

openaccess

dc.contributor.advisor

Thiran, Jean-Philippe

dc.contributor.author

Rad, Mohammad Saeed

dc.date.accepted

2021

dc.date.accessioned

2021-06-21T11:35:07

dc.date.available

2021-06-21T11:35:07

dc.date.created

2021-06-21

dc.date.issued

2021

dc.date.modified

2025-02-19T13:43:54.234864Z

dc.description.abstract

Image super-resolution is a classic ill-posed computer vision and image processing problem, addressing the question of how to reconstruct a high-resolution image from its low-resolution counterpart. Current state-of-the-art methods have improved the performance of the single image super-resolution task significantly by benefiting from machine learning and AI-powered algorithms, and more specifically, with the advent of Deep Learning-based approaches. Although these advances allow a machine to learn and have better exploitation of an image and its content, recent methods are still unable to constrain the plausible solution space based on the available contextual information within an image. This limitation mostly results in poor reconstructions, even for well-known types of objects and textures easily recognizable for humans. In this thesis, we aim at proving that the categorical prior, which characterizes the semantic class of a region in an image (e.g., sky, building, plant), is crucial in super-resolution task for reaching a higher reconstruction quality. In particular, we propose several approaches to improve the perceived image quality and generalization capability of deep learning-based methods by exploiting the context and semantic meaning of images. To prove the effectiveness of this categorical information, we first propose a convolutional neural network-based framework that is able to extract and use semantic information to super-resolve a given image by using multitask learning, simultaneously for learning image super-resolution and semantic segmentation. The proposed decoder is forced to explore categorical information during training, as this setting employs only one shared deep network for both semantic segmentation and super-resolution tasks. We further investigate the possibility of using semantic information by a novel objective function to introduce additional spatial control over the training process. We propose penalizing images at different semantic levels using appropriate loss terms by benefiting from our new OBB (Object, Background, and Boundary) labels generated from segmentation labels. Then, we introduce a new test time adaptation-based technique to leverage high-resolution images with perceptually similar context to a given test image to improve the reconstruction quality. We further validate this approach's effectiveness by using a novel numerical experiment analyzing the correlation between filters learned by our network and what we define as `ideal' filters. Finally, we present a generic solution to enable adapting all our previous contributions in this thesis, as well as other recent super-resolution works trained on synthetic datasets, to real-world super-resolution problem. Real-world super-resolution refers to super-resolving images with real degradations caused by physical imaging systems, instead of low-resolution images from simulated datasets assuming a simple and uniform degradation model (i.e., bicubic downsampling). We study and develop an image-to-image translator to map the distribution of real low-resolution images to the well-understood distribution of bicubically downsampled images. This translator is used as a plug-in to integrate real inputs into any super-resolution framework trained on simulated datasets. We carry out extensive qualitative and quantitative experiments for each mentioned contribution, including user studies, to compare our proposed approaches to state-of-the-art method.

dc.description.sponsorship

LTS5

dc.identifier.doi

10.5075/epfl-thesis-9549

dc.identifier.uri

https://infoscience.epfl.ch/handle/20.500.14299/179454

dc.language.iso

en

dc.publisher

EPFL

dc.publisher.place

Lausanne

dc.relation

https://infoscience.epfl.ch/record/286804/files/EPFL_TH9549.pdf

dc.size

148

dc.subject

super-resolution

dc.subject

neural network

dc.subject

deep learning

dc.subject

generative adversarial networks

dc.subject

computer vision

dc.subject

image processing

dc.title

Context-Aware Image Super-Resolution Using Deep Neural Networks

dc.type

thesis::doctoral thesis

dspace.entity.type

Publication

dspace.file.type

n/a

dspace.legacy.oai-identifier

oai:infoscience.epfl.ch:286804

epfl.legacy.itemtype

Theses

epfl.legacy.submissionform

THESIS

epfl.oai.currentset

fulltext

epfl.oai.currentset

DOI

epfl.oai.currentset

thesis-public

epfl.oai.currentset

STI

epfl.oai.currentset

thesis

epfl.oai.currentset

thesis-bn

epfl.oai.currentset

OpenAIREv4

epfl.publication.version

http://purl.org/coar/version/c_970fb48d4fbd8a85

epfl.thesis.doctoralSchool

EDEE

epfl.thesis.faculty

STI

epfl.thesis.institute

IEL

epfl.thesis.jury

Prof. Pascal Frossard (président) ; Prof. Jean-Philippe Thiran (directeur de thèse) ; Prof. Sabine Süsstrunk, Dr. Claudiu Musat, Dr. Christopher Schroers (rapporteurs)

epfl.thesis.number

9549

epfl.thesis.originalUnit

LTS5

epfl.thesis.publicDefenseYear

2021-07-02

epfl.writtenAt

EPFL

oaire.licenseCondition

Copyright

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
EPFL_TH9549.pdf
Size:
86.9 MB
Format:
Adobe Portable Document Format

License bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
license.txt
Size:
1.71 KB
Format:
Item-specific license agreed to upon submission
Description:

Collections