Skip to content

Releases: Phhofm/models

2xBHI_small_esrgan_pretrain

14 Oct 12:09
c16e557

Choose a tag to compare

2xBHI_small_esrgan_pretrain

Scale: 2x
Network type: ESRGAN
Author: Philip Hofmann
License: CC-BY-4.0
Purpose: 2x esrgan pretrain model with l1&mssim loss only.
Training iterations: 300'000
Description: A 2x esrgan pretrain model.

Slowpic Example
https://slow.pics/s/92cYa0w3?image-fit=cover

2xBHI_small_drct-xl_pretrain

14 Oct 12:09
c16e557

Choose a tag to compare

2xBHI_small_drct-xl_pretrain

Scale: 2x
Network type: DRCT-XL
Author: Philip Hofmann
License: CC-BY-4.0
Purpose: 2x drct-xl pretrain model with l1&mssim loss only.
Training iterations: 180'000
Description: A 2x drct-xl pretrain model

Slowpic Example
https://slow.pics/s/PkqzlPH1?image-fit=cover

image

1xgaterv3_r_sharpen

14 Oct 12:38
c16e557

Choose a tag to compare

1xgaterv3_r_sharpen

Scale:
Network Type: GaterV3
Author: Philip Hofmann
Iterations: 90,000
License: CC BY 4.0
Purpose: A 1× enhancement model designed to subtly sharpen image outputs.
It can be used as a standalone model or chained directly after 1xgaterv3_r_restore for improved sharpness and clarity.


Usage

This model was trained on CC0 images using the GaterV3 architecture (MIT).
It is released under the CC BY 4.0 license, permitting personal, commercial, and open use with proper attribution or prior agreement.
ONNX conversions are included for easy inference.


🖼️ Visual Results

example1 example2 example3 example4 example5 example6 example7

Full Changelog: 2xPublic_realplksr_dysample_layernorm_gan...1xgaterv3_r_sharpen

1xgaterv3_r_restore

14 Oct 12:38
c16e557

Choose a tag to compare

1xgaterv3_r_restore

Scale:
Network Type: GaterV3
Author: Philip Hofmann
Iterations: 180,000
License: CC BY 4.0
Purpose: A 1× restoration model for improving image quality — handles noise, resizing, JPEG compression, and mild blur.
If outputs appear too soft, you can chain 1xgaterv3_r_sharpen afterward to slightly enhance sharpness.


Usage

This model was trained on CC0 images using the GaterV3 architecture (MIT).
It is released under the CC BY 4.0 license, permitting personal, commercial, and open use with proper attribution or prior agreement.
ONNX conversions are included for easy inference.


🖼️ Visual Results

restore_example example1 example2 example3 example4 example5 example6 example7

2xPublic_realplksr_dysample_layernorm_real_nn

23 Jun 13:35
90313fd

Choose a tag to compare

🚀 2xPublic_realplksr_dysample_layernorm_real_nn

Scale: 2x
Network Type: RealPLKSR Dysample LayerNorm
Author: Philip Hofmann
Iterations: 120'000
License: Apache 2.0
Purpose: 2× upscaling of images with some degradation handling (blur, JPEG and sharpening artifacts), but no denoising, which can help the preservation of subtle textures


🔍 Overview

2× super-resolution model combining:

  • RealPLKSR backbone (GAN-optimized architecture)
  • Dysample dynamic upsampling
  • LayerNorm stabilization
    Trained on rigorously curated public domain imagery for robust real-world performance. Specifically no noise has been added to the low resolution training data counterpart. This might result in keeping more details, specifically preservation of subtle textures, since denoising can lead to smooth outputs, washing out of subtle, high-frequency textures.

🖼️ Visual Results

1
2
3
4
5
002
007
009
010
12
100
136
144
184
195


🧠 Model Details

🏋️ Training Dataset

  • Input: 56K+ public domain images (CC0/public domain marked)
  • Processing: On average 6 hours/10K images (processing alone)
  • Final Training Dataset: 4,297 512×512 tiles after:
    • Multi-scale strategy
    • Visual de-duplication
    • Strict quality thresholds of 10 different IQA metric scores

⚙️ LR Degradation Pipeline

Probabilistic degradation stack pre-applied on the training set:

  1. Optical Blur
    📌 Gaussian kernels (3×3 to 7×7)
    📌 σ = 0.8-2.5 (randomized per sample)

  2. Resampling Cycles
    📌 Downsampling methods = INTER_AREA, INTER_LINEAR, INTER_CUBIC
    📌 Upsampling methods = INTER_CUBIC, INTER_LINEAR, INTER_NEAREST, INTER_LANCZOS4
    📌 Asymmetric x/y scaling factors

  3. JPEG Compression
    📌 1-3 compression cycles
    📌 Chroma subsampling (420/422)
    📌 Quality: 30-90 (variable)

  4. Sharpening Artifacts
    📌 Laplacian edge detection
    📌 Intensity difference masking

  5. Final Downscaling
    📌 DPID (Deep Parametric Image Downscaling)
    📌 Kernel width: 0.4-0.6 (randomized)

🧩 Architecture

Component Origin License Contribution
PLKSR dslisleedh/PLKSR MIT Base architecture
RealPLKSR neosr-project MIT GAN stabilization
Dysample tiny-smart/dynsample MIT Dynamic upsampling
LayerNorm trainner-redux Apache 2.0 Training stability

🔮 Future Plans

I plan to enlarge the training dataset from the current 4,297 512x512 tiles. This expansion should significantly improve model quality by:

  • Increasing the amount of information and contextual understanding available during training
  • Enhancing the distribution coverage of degradation strengths
  • Improving generalization capabilities across diverse real-world scenarios
  • Reducing potential overfitting to specific artifact patterns

Also adjusting degradations&strengths applied to LR creation, and maybe adjust losses, all from learnings taken from inspecting the validation images


📜 License (Apache 2.0)

✅ Permitted Usage

  • Commercial applications (SaaS, apps, APIs)
  • Non-commercial/research projects
  • Model modification & fine-tuning
  • Redistribution of original/derived versions
  • Patent implementation of included techniques

🚫 Restrictions

  • Removing original attribution
  • Implying creator endorsement
  • Patent litigation against included techniques

📎 Attribution Requirements

Required when:

  • Public product/service usage
  • Research publications
  • Model redistribution

Attribution Template:

Image enhancement powered by 2xPublic_realplksr_dysample_layernorm_real 
by Philip Hofmann. [GitHub Release](https://github.com/Phhofm/models)

⚠️ Legal Note: Full license terms govern all usage. This summary doesn't constitute legal advice.
📄 Complete license: Apache 2.0 TEXT


🔗 References

2xPublic_realplksr_dysample_layernorm_real

23 Jun 13:32
90313fd

Choose a tag to compare

🚀 2xPublic_realplksr_dysample_layernorm_real

Scale: 2x
Network Type: RealPLKSR Dysample LayerNorm
Author: Philip Hofmann
Iterations: 250'000
License: Apache 2.0
Purpose: 2× upscaling of images with some degradation handling (blur, noise, JPEG and sharpening artifacts)


🔍 Overview

2× super-resolution model combining:

  • RealPLKSR backbone (GAN-optimized architecture)
  • Dysample dynamic upsampling
  • LayerNorm stabilization
    Trained on rigorously curated public domain imagery for robust real-world performance.

🖼️ Visual Results

1
2
3
4
5
002
007
009
010
12
100
136
144
184
195


🧠 Model Details

🏋️ Training Dataset

  • Input: 56K+ public domain images (CC0/public domain marked)
  • Processing: On average 6 hours/10K images (processing alone)
  • Final Training Dataset: 4,297 512×512 tiles after:
    • Multi-scale strategy
    • Visual de-duplication
    • Strict quality thresholds of 10 different IQA metric scores

⚙️ LR Degradation Pipeline

Probabilistic degradation stack pre-applied on the training set:

  1. Optical Blur
    📌 Gaussian kernels (3×3 to 7×7)
    📌 σ = 0.8-2.5 (randomized per sample)

  2. Sensor Noise
    📌 Physics-based Poisson-Gaussian model
    📌 Channel-dependent sensitivity parameters

  3. Resampling Cycles
    📌 Downsampling methods = INTER_AREA, INTER_LINEAR, INTER_CUBIC
    📌 Upsampling methods = INTER_CUBIC, INTER_LINEAR, INTER_NEAREST, INTER_LANCZOS4
    📌 Asymmetric x/y scaling factors

  4. JPEG Compression
    📌 1-3 compression cycles
    📌 Chroma subsampling (420/422)
    📌 Quality: 30-90 (variable)

  5. Sharpening Artifacts
    📌 Laplacian edge detection
    📌 Intensity difference masking

  6. Final Downscaling
    📌 DPID (Deep Parametric Image Downscaling)
    📌 Kernel width: 0.4-0.6 (randomized)

🧩 Architecture

Component Origin License Contribution
PLKSR dslisleedh/PLKSR MIT Base architecture
RealPLKSR neosr-project MIT GAN stabilization
Dysample tiny-smart/dynsample MIT Dynamic upsampling
LayerNorm trainner-redux Apache 2.0 Training stability

🔮 Future Plans

I plan to enlarge the training dataset from the current 4,297 512x512 tiles. This expansion should significantly improve model quality by:

  • Increasing the amount of information and contextual understanding available during training
  • Enhancing the distribution coverage of degradation strengths
  • Improving generalization capabilities across diverse real-world scenarios
  • Reducing potential overfitting to specific artifact patterns

Also adjusting degradations&strengths applied to LR creation, and maybe adjust losses, all from learnings taken from inspecting the validation images


📜 License (Apache 2.0)

✅ Permitted Usage

  • Commercial applications (SaaS, apps, APIs)
  • Non-commercial/research projects
  • Model modification & fine-tuning
  • Redistribution of original/derived versions
  • Patent implementation of included techniques

🚫 Restrictions

  • Removing original attribution
  • Implying creator endorsement
  • Patent litigation against included techniques

📎 Attribution Requirements

Required when:

  • Public product/service usage
  • Research publications
  • Model redistribution

Attribution Template:

Image enhancement powered by 2xPublic_realplksr_dysample_layernorm_real 
by Philip Hofmann. [GitHub Release](https://github.com/Phhofm/models/releases/tag/2xPublic_realplksr_dysample_layernorm_real)

⚠️ Legal Note: Full license terms govern all usage. This summary doesn't constitute legal advice.
📄 Complete license: Apache 2.0 TEXT


🔗 References

2xPublic_realplksr_dysample_layernorm_pretrain

23 Jun 13:42
90313fd

Choose a tag to compare

🚀 2xPublic_realplksr_dysample_layernorm_pretrain

Scale: 2x
Network Type: RealPLKSR Dysample LayerNorm
Author: Philip Hofmann
Iterations: 100'000
License: Apache 2.0
Purpose: 2× pretrain


🔍 Overview

2× super-resolution pretraincombining:

  • RealPLKSR backbone (GAN-optimized architecture)
  • Dysample dynamic upsampling
  • LayerNorm stabilization
    Trained on rigorously curated public domain imagery, this one has been trained on dpid downscaled content only

🖼️ Visual Results

1
2
3
4
5
002
007
009
010
12
100
136
144
184
195


🧠 Model Details

🏋️ Training Dataset

  • Input: 56K+ public domain images (CC0/public domain marked)
  • Processing: On average 6 hours/10K images (processing alone)
  • Final Training Dataset: 4,297 512×512 tiles after:
    • Multi-scale strategy
    • Visual de-duplication
    • Strict quality thresholds of 10 different IQA metric scores

⚙️ LR Degradation Pipeline

DPID downscaled only

🧩 Architecture

Component Origin License Contribution
PLKSR dslisleedh/PLKSR MIT Base architecture
RealPLKSR neosr-project MIT GAN stabilization
Dysample tiny-smart/dynsample MIT Dynamic upsampling
LayerNorm trainner-redux Apache 2.0 Training stability

🔮 Future Plans

I plan to enlarge the training dataset from the current 4,297 512x512 tiles. This expansion should significantly improve model quality by:

  • Increasing the amount of information and contextual understanding available during training
  • Enhancing the distribution coverage of degradation strengths
  • Improving generalization capabilities across diverse real-world scenarios
  • Reducing potential overfitting to specific artifact patterns

📜 License (Apache 2.0)

✅ Permitted Usage

  • Commercial applications (SaaS, apps, APIs)
  • Non-commercial/research projects
  • Model modification & fine-tuning
  • Redistribution of original/derived versions
  • Patent implementation of included techniques

🚫 Restrictions

  • Removing original attribution
  • Implying creator endorsement
  • Patent litigation against included techniques

📎 Attribution Requirements

Required when:

  • Public product/service usage
  • Research publications
  • Model redistribution

Attribution Template:

Image enhancement powered by 2xPublic_realplksr_dysample_layernorm_real 
by Philip Hofmann. [GitHub Release](https://github.com/Phhofm/models)

⚠️ Legal Note: Full license terms govern all usage. This summary doesn't constitute legal advice.
📄 Complete license: Apache 2.0 TEXT


🔗 References

2xPublic_realplksr_dysample_layernorm_gan

23 Jun 13:39
90313fd

Choose a tag to compare

🚀 2xPublic_realplksr_dysample_layernorm_gan

Scale: 2x
Network Type: RealPLKSR Dysample LayerNorm
Author: Philip Hofmann
Iterations: 125'000
License: Apache 2.0
Purpose: 2× upscaling of images, no degradation handling


🔍 Overview

2× super-resolution model combining:

  • RealPLKSR backbone (GAN-optimized architecture)
  • Dysample dynamic upsampling
  • LayerNorm stabilization
    Trained on rigorously curated public domain imagery, this one has been trained on dpid downscaled content only

🖼️ Visual Results

1
2
3
4
5
002
007
009
010
12
100
136
144
184
195


🧠 Model Details

🏋️ Training Dataset

  • Input: 56K+ public domain images (CC0/public domain marked)
  • Processing: On average 6 hours/10K images (processing alone)
  • Final Training Dataset: 4,297 512×512 tiles after:
    • Multi-scale strategy
    • Visual de-duplication
    • Strict quality thresholds of 10 different IQA metric scores

⚙️ LR Degradation Pipeline

DPID downscaled only

🧩 Architecture

Component Origin License Contribution
PLKSR dslisleedh/PLKSR MIT Base architecture
RealPLKSR neosr-project MIT GAN stabilization
Dysample tiny-smart/dynsample MIT Dynamic upsampling
LayerNorm trainner-redux Apache 2.0 Training stability

🔮 Future Plans

I plan to enlarge the training dataset from the current 4,297 512x512 tiles. This expansion should significantly improve model quality by:

  • Increasing the amount of information and contextual understanding available during training
  • Enhancing the distribution coverage of degradation strengths
  • Improving generalization capabilities across diverse real-world scenarios
  • Reducing potential overfitting to specific artifact patterns

Also adjusting degradations&strengths applied to LR creation, and maybe adjust losses, all from learnings taken from inspecting the validation images


📜 License (Apache 2.0)

✅ Permitted Usage

  • Commercial applications (SaaS, apps, APIs)
  • Non-commercial/research projects
  • Model modification & fine-tuning
  • Redistribution of original/derived versions
  • Patent implementation of included techniques

🚫 Restrictions

  • Removing original attribution
  • Implying creator endorsement
  • Patent litigation against included techniques

📎 Attribution Requirements

Required when:

  • Public product/service usage
  • Research publications
  • Model redistribution

Attribution Template:

Image enhancement powered by 2xPublic_realplksr_dysample_layernorm_real 
by Philip Hofmann. [GitHub Release](https://github.com/Phhofm/models)

⚠️ Legal Note: Full license terms govern all usage. This summary doesn't constitute legal advice.
📄 Complete license: Apache 2.0 TEXT


🔗 References

2xBHI_small_realplksr_small_pretrain

21 May 12:47
c815325

Choose a tag to compare

2xBHI_small_realplksr_small_pretrain

Scale: 2x
Network type: realplksr_small
Author: Philip Hofmann
License: CC-BY-4.0
Release: 21.05.2025
Purpose: 2x realplksr_small pretrain model with l1&mssim loss only.
Training iterations: 100'000
Description: A 2x realplksr_small pretrain model.

Visual Examples

11
14
23
99
100

Tensorboard Validation Graphs on BHI100

image

2xBHI_small_realplksr_small_dysample_pretrain

21 May 12:48
c815325

Choose a tag to compare

2xBHI_small_realplksr_small_dysample_pretrain

Scale: 2x
Network type: realplksr_small dysample
Author: Philip Hofmann
License: CC-BY-4.0
Release: 21.05.2025
Purpose: 2x realplksr_small dysample pretrain model with l1&mssim loss only.
Training iterations: 100'000
Description: A 2x realplksr_small dysample pretrain model.

Visual Examples

11
14
23
99
100

Tensorboard Validation Graphs on BHI100

image