"""
Overview:
Restore the images using `NafNet <https://github.com/megvii-research/NAFNet>`_.
.. image:: nafnet_demo.plot.py.svg
:align: center
This is an overall benchmark of all the NafNet models:
.. image:: nafnet_benchmark.plot.py.svg
:align: center
.. warning::
Currently, we've identified a significant issue with NafNet when images contain gaussian noise.
To ensure your code functions correctly, please ensure the credibility of
your image source or preprocess them using SCUNet.
.. note::
New in version v0.4.4, **images with alpha channel supported**.
If you use an image with alpha channel (e.g. RGBA images),
it will return a RGBA image, otherwise return RGG image.
"""
from typing import Literal
import numpy as np
from PIL import Image
from huggingface_hub import hf_hub_download
from ..data import ImageTyping
from ..generic import ImageEnhancer
from ..utils import open_onnx_model, area_batch_run, ts_lru_cache
NafNetModelTyping = Literal['REDS', 'GoPro', 'SIDD']
@ts_lru_cache()
def _open_nafnet_model(model: NafNetModelTyping):
"""
Open the NAFNet model for image restoration.
:param model: The NAFNet model type ('REDS', 'GoPro', 'SIDD').
:type model: NafNetModelTyping
:return: The NAFNet ONNX model.
"""
return open_onnx_model(hf_hub_download(
f'deepghs/image_restoration',
f'NAFNet-{model}-width64.onnx',
))
class _Enhancer(ImageEnhancer):
def __init__(self, model: NafNetModelTyping = 'REDS', tile_size: int = 256, tile_overlap: int = 16,
batch_size: int = 4, silent: bool = False):
self.model = model
self.tile_size = tile_size
self.tile_overlap = tile_overlap
self.batch_size = batch_size
self.silent = silent
def _process_rgb(self, rgb_array: np.ndarray):
input_ = rgb_array[None, ...]
def _method(ix):
ox, = _open_nafnet_model(self.model).run(['output'], {'input': ix})
return ox
output_ = area_batch_run(
input_, _method,
tile_size=self.tile_size, tile_overlap=self.tile_overlap, batch_size=self.batch_size,
silent=self.silent, process_title='NafNet Restore',
)
output_ = np.clip(output_, a_min=0.0, a_max=1.0)
return output_[0]
@ts_lru_cache()
def _get_enhancer(model: NafNetModelTyping = 'REDS', tile_size: int = 256, tile_overlap: int = 16,
batch_size: int = 4, silent: bool = False) -> _Enhancer:
return _Enhancer(model, tile_size, tile_overlap, batch_size, silent)
[docs]def restore_with_nafnet(image: ImageTyping, model: NafNetModelTyping = 'REDS',
tile_size: int = 256, tile_overlap: int = 16, batch_size: int = 4,
silent: bool = False) -> Image.Image:
"""
Restore an image using the NAFNet model.
:param image: The input image.
:type image: ImageTyping
:param model: The NAFNet model type ('REDS', 'GoPro', 'SIDD'). Default is 'REDS'.
:type model: NafNetModelTyping
:param tile_size: The size of processing tiles. Default is 256.
:type tile_size: int
:param tile_overlap: The overlap between tiles. Default is 16.
:type tile_overlap: int
:param batch_size: The batch size of inference. Default is 4.
:type batch_size: int
:param silent: If True, the progress will not be displayed. Default is False.
:type silent: bool
:return: The restored image.
:rtype: Image.Image
"""
return _get_enhancer(model, tile_size, tile_overlap, batch_size, silent).process(image)