Releases: Phhofm/models
2xBHI_small_esrgan_pretrain
2xBHI_small_esrgan_pretrain
Scale: 2x
Network type: ESRGAN
Author: Philip Hofmann
License: CC-BY-4.0
Purpose: 2x esrgan pretrain model with l1&mssim loss only.
Training iterations: 300'000
Description: A 2x esrgan pretrain model.
Slowpic Example
https://slow.pics/s/92cYa0w3?image-fit=cover
2xBHI_small_drct-xl_pretrain
2xBHI_small_drct-xl_pretrain
Scale: 2x
Network type: DRCT-XL
Author: Philip Hofmann
License: CC-BY-4.0
Purpose: 2x drct-xl pretrain model with l1&mssim loss only.
Training iterations: 180'000
Description: A 2x drct-xl pretrain model
Slowpic Example
https://slow.pics/s/PkqzlPH1?image-fit=cover

1xgaterv3_r_sharpen
1xgaterv3_r_sharpen
Scale: 1×
Network Type: GaterV3
Author: Philip Hofmann
Iterations: 90,000
License: CC BY 4.0
Purpose: A 1× enhancement model designed to subtly sharpen image outputs.
It can be used as a standalone model or chained directly after 1xgaterv3_r_restore for improved sharpness and clarity.
Usage
This model was trained on CC0 images using the GaterV3 architecture (MIT).
It is released under the CC BY 4.0 license, permitting personal, commercial, and open use with proper attribution or prior agreement.
ONNX conversions are included for easy inference.
🖼️ Visual Results
Full Changelog: 2xPublic_realplksr_dysample_layernorm_gan...1xgaterv3_r_sharpen
1xgaterv3_r_restore
1xgaterv3_r_restore
Scale: 1×
Network Type: GaterV3
Author: Philip Hofmann
Iterations: 180,000
License: CC BY 4.0
Purpose: A 1× restoration model for improving image quality — handles noise, resizing, JPEG compression, and mild blur.
If outputs appear too soft, you can chain 1xgaterv3_r_sharpen afterward to slightly enhance sharpness.
Usage
This model was trained on CC0 images using the GaterV3 architecture (MIT).
It is released under the CC BY 4.0 license, permitting personal, commercial, and open use with proper attribution or prior agreement.
ONNX conversions are included for easy inference.
🖼️ Visual Results

2xPublic_realplksr_dysample_layernorm_real_nn
🚀 2xPublic_realplksr_dysample_layernorm_real_nn
Scale: 2x
Network Type: RealPLKSR Dysample LayerNorm
Author: Philip Hofmann
Iterations: 120'000
License: Apache 2.0
Purpose: 2× upscaling of images with some degradation handling (blur, JPEG and sharpening artifacts), but no denoising, which can help the preservation of subtle textures
🔍 Overview
2× super-resolution model combining:
- RealPLKSR backbone (GAN-optimized architecture)
- Dysample dynamic upsampling
- LayerNorm stabilization
Trained on rigorously curated public domain imagery for robust real-world performance. Specifically no noise has been added to the low resolution training data counterpart. This might result in keeping more details, specifically preservation of subtle textures, since denoising can lead to smooth outputs, washing out of subtle, high-frequency textures.
🖼️ Visual Results
🧠 Model Details
🏋️ Training Dataset
- Input: 56K+ public domain images (CC0/public domain marked)
- Processing: On average 6 hours/10K images (processing alone)
- Final Training Dataset: 4,297 512×512 tiles after:
- Multi-scale strategy
- Visual de-duplication
- Strict quality thresholds of 10 different IQA metric scores
⚙️ LR Degradation Pipeline
Probabilistic degradation stack pre-applied on the training set:
-
Optical Blur
📌 Gaussian kernels (3×3 to 7×7)
📌 σ = 0.8-2.5 (randomized per sample) -
Resampling Cycles
📌 Downsampling methods = INTER_AREA, INTER_LINEAR, INTER_CUBIC
📌 Upsampling methods = INTER_CUBIC, INTER_LINEAR, INTER_NEAREST, INTER_LANCZOS4
📌 Asymmetric x/y scaling factors -
JPEG Compression
📌 1-3 compression cycles
📌 Chroma subsampling (420/422)
📌 Quality: 30-90 (variable) -
Sharpening Artifacts
📌 Laplacian edge detection
📌 Intensity difference masking -
Final Downscaling
📌 DPID (Deep Parametric Image Downscaling)
📌 Kernel width: 0.4-0.6 (randomized)
🧩 Architecture
| Component | Origin | License | Contribution |
|---|---|---|---|
| PLKSR | dslisleedh/PLKSR | MIT | Base architecture |
| RealPLKSR | neosr-project | MIT | GAN stabilization |
| Dysample | tiny-smart/dynsample | MIT | Dynamic upsampling |
| LayerNorm | trainner-redux | Apache 2.0 | Training stability |
🔮 Future Plans
I plan to enlarge the training dataset from the current 4,297 512x512 tiles. This expansion should significantly improve model quality by:
- Increasing the amount of information and contextual understanding available during training
- Enhancing the distribution coverage of degradation strengths
- Improving generalization capabilities across diverse real-world scenarios
- Reducing potential overfitting to specific artifact patterns
Also adjusting degradations&strengths applied to LR creation, and maybe adjust losses, all from learnings taken from inspecting the validation images
📜 License (Apache 2.0)
✅ Permitted Usage
- Commercial applications (SaaS, apps, APIs)
- Non-commercial/research projects
- Model modification & fine-tuning
- Redistribution of original/derived versions
- Patent implementation of included techniques
🚫 Restrictions
- Removing original attribution
- Implying creator endorsement
- Patent litigation against included techniques
📎 Attribution Requirements
Required when:
- Public product/service usage
- Research publications
- Model redistribution
Attribution Template:
Image enhancement powered by 2xPublic_realplksr_dysample_layernorm_real
by Philip Hofmann. [GitHub Release](https://github.com/Phhofm/models)
⚠️ Legal Note: Full license terms govern all usage. This summary doesn't constitute legal advice.
📄 Complete license: Apache 2.0 TEXT
🔗 References
2xPublic_realplksr_dysample_layernorm_real
🚀 2xPublic_realplksr_dysample_layernorm_real
Scale: 2x
Network Type: RealPLKSR Dysample LayerNorm
Author: Philip Hofmann
Iterations: 250'000
License: Apache 2.0
Purpose: 2× upscaling of images with some degradation handling (blur, noise, JPEG and sharpening artifacts)
🔍 Overview
2× super-resolution model combining:
- RealPLKSR backbone (GAN-optimized architecture)
- Dysample dynamic upsampling
- LayerNorm stabilization
Trained on rigorously curated public domain imagery for robust real-world performance.
🖼️ Visual Results
🧠 Model Details
🏋️ Training Dataset
- Input: 56K+ public domain images (CC0/public domain marked)
- Processing: On average 6 hours/10K images (processing alone)
- Final Training Dataset: 4,297 512×512 tiles after:
- Multi-scale strategy
- Visual de-duplication
- Strict quality thresholds of 10 different IQA metric scores
⚙️ LR Degradation Pipeline
Probabilistic degradation stack pre-applied on the training set:
-
Optical Blur
📌 Gaussian kernels (3×3 to 7×7)
📌 σ = 0.8-2.5 (randomized per sample) -
Sensor Noise
📌 Physics-based Poisson-Gaussian model
📌 Channel-dependent sensitivity parameters -
Resampling Cycles
📌 Downsampling methods = INTER_AREA, INTER_LINEAR, INTER_CUBIC
📌 Upsampling methods = INTER_CUBIC, INTER_LINEAR, INTER_NEAREST, INTER_LANCZOS4
📌 Asymmetric x/y scaling factors -
JPEG Compression
📌 1-3 compression cycles
📌 Chroma subsampling (420/422)
📌 Quality: 30-90 (variable) -
Sharpening Artifacts
📌 Laplacian edge detection
📌 Intensity difference masking -
Final Downscaling
📌 DPID (Deep Parametric Image Downscaling)
📌 Kernel width: 0.4-0.6 (randomized)
🧩 Architecture
| Component | Origin | License | Contribution |
|---|---|---|---|
| PLKSR | dslisleedh/PLKSR | MIT | Base architecture |
| RealPLKSR | neosr-project | MIT | GAN stabilization |
| Dysample | tiny-smart/dynsample | MIT | Dynamic upsampling |
| LayerNorm | trainner-redux | Apache 2.0 | Training stability |
🔮 Future Plans
I plan to enlarge the training dataset from the current 4,297 512x512 tiles. This expansion should significantly improve model quality by:
- Increasing the amount of information and contextual understanding available during training
- Enhancing the distribution coverage of degradation strengths
- Improving generalization capabilities across diverse real-world scenarios
- Reducing potential overfitting to specific artifact patterns
Also adjusting degradations&strengths applied to LR creation, and maybe adjust losses, all from learnings taken from inspecting the validation images
📜 License (Apache 2.0)
✅ Permitted Usage
- Commercial applications (SaaS, apps, APIs)
- Non-commercial/research projects
- Model modification & fine-tuning
- Redistribution of original/derived versions
- Patent implementation of included techniques
🚫 Restrictions
- Removing original attribution
- Implying creator endorsement
- Patent litigation against included techniques
📎 Attribution Requirements
Required when:
- Public product/service usage
- Research publications
- Model redistribution
Attribution Template:
Image enhancement powered by 2xPublic_realplksr_dysample_layernorm_real
by Philip Hofmann. [GitHub Release](https://github.com/Phhofm/models/releases/tag/2xPublic_realplksr_dysample_layernorm_real)
⚠️ Legal Note: Full license terms govern all usage. This summary doesn't constitute legal advice.
📄 Complete license: Apache 2.0 TEXT
🔗 References
2xPublic_realplksr_dysample_layernorm_pretrain
🚀 2xPublic_realplksr_dysample_layernorm_pretrain
Scale: 2x
Network Type: RealPLKSR Dysample LayerNorm
Author: Philip Hofmann
Iterations: 100'000
License: Apache 2.0
Purpose: 2× pretrain
🔍 Overview
2× super-resolution pretraincombining:
- RealPLKSR backbone (GAN-optimized architecture)
- Dysample dynamic upsampling
- LayerNorm stabilization
Trained on rigorously curated public domain imagery, this one has been trained on dpid downscaled content only
🖼️ Visual Results
🧠 Model Details
🏋️ Training Dataset
- Input: 56K+ public domain images (CC0/public domain marked)
- Processing: On average 6 hours/10K images (processing alone)
- Final Training Dataset: 4,297 512×512 tiles after:
- Multi-scale strategy
- Visual de-duplication
- Strict quality thresholds of 10 different IQA metric scores
⚙️ LR Degradation Pipeline
DPID downscaled only
🧩 Architecture
| Component | Origin | License | Contribution |
|---|---|---|---|
| PLKSR | dslisleedh/PLKSR | MIT | Base architecture |
| RealPLKSR | neosr-project | MIT | GAN stabilization |
| Dysample | tiny-smart/dynsample | MIT | Dynamic upsampling |
| LayerNorm | trainner-redux | Apache 2.0 | Training stability |
🔮 Future Plans
I plan to enlarge the training dataset from the current 4,297 512x512 tiles. This expansion should significantly improve model quality by:
- Increasing the amount of information and contextual understanding available during training
- Enhancing the distribution coverage of degradation strengths
- Improving generalization capabilities across diverse real-world scenarios
- Reducing potential overfitting to specific artifact patterns
📜 License (Apache 2.0)
✅ Permitted Usage
- Commercial applications (SaaS, apps, APIs)
- Non-commercial/research projects
- Model modification & fine-tuning
- Redistribution of original/derived versions
- Patent implementation of included techniques
🚫 Restrictions
- Removing original attribution
- Implying creator endorsement
- Patent litigation against included techniques
📎 Attribution Requirements
Required when:
- Public product/service usage
- Research publications
- Model redistribution
Attribution Template:
Image enhancement powered by 2xPublic_realplksr_dysample_layernorm_real
by Philip Hofmann. [GitHub Release](https://github.com/Phhofm/models)
⚠️ Legal Note: Full license terms govern all usage. This summary doesn't constitute legal advice.
📄 Complete license: Apache 2.0 TEXT
🔗 References
2xPublic_realplksr_dysample_layernorm_gan
🚀 2xPublic_realplksr_dysample_layernorm_gan
Scale: 2x
Network Type: RealPLKSR Dysample LayerNorm
Author: Philip Hofmann
Iterations: 125'000
License: Apache 2.0
Purpose: 2× upscaling of images, no degradation handling
🔍 Overview
2× super-resolution model combining:
- RealPLKSR backbone (GAN-optimized architecture)
- Dysample dynamic upsampling
- LayerNorm stabilization
Trained on rigorously curated public domain imagery, this one has been trained on dpid downscaled content only
🖼️ Visual Results
🧠 Model Details
🏋️ Training Dataset
- Input: 56K+ public domain images (CC0/public domain marked)
- Processing: On average 6 hours/10K images (processing alone)
- Final Training Dataset: 4,297 512×512 tiles after:
- Multi-scale strategy
- Visual de-duplication
- Strict quality thresholds of 10 different IQA metric scores
⚙️ LR Degradation Pipeline
DPID downscaled only
🧩 Architecture
| Component | Origin | License | Contribution |
|---|---|---|---|
| PLKSR | dslisleedh/PLKSR | MIT | Base architecture |
| RealPLKSR | neosr-project | MIT | GAN stabilization |
| Dysample | tiny-smart/dynsample | MIT | Dynamic upsampling |
| LayerNorm | trainner-redux | Apache 2.0 | Training stability |
🔮 Future Plans
I plan to enlarge the training dataset from the current 4,297 512x512 tiles. This expansion should significantly improve model quality by:
- Increasing the amount of information and contextual understanding available during training
- Enhancing the distribution coverage of degradation strengths
- Improving generalization capabilities across diverse real-world scenarios
- Reducing potential overfitting to specific artifact patterns
Also adjusting degradations&strengths applied to LR creation, and maybe adjust losses, all from learnings taken from inspecting the validation images
📜 License (Apache 2.0)
✅ Permitted Usage
- Commercial applications (SaaS, apps, APIs)
- Non-commercial/research projects
- Model modification & fine-tuning
- Redistribution of original/derived versions
- Patent implementation of included techniques
🚫 Restrictions
- Removing original attribution
- Implying creator endorsement
- Patent litigation against included techniques
📎 Attribution Requirements
Required when:
- Public product/service usage
- Research publications
- Model redistribution
Attribution Template:
Image enhancement powered by 2xPublic_realplksr_dysample_layernorm_real
by Philip Hofmann. [GitHub Release](https://github.com/Phhofm/models)
⚠️ Legal Note: Full license terms govern all usage. This summary doesn't constitute legal advice.
📄 Complete license: Apache 2.0 TEXT
🔗 References
2xBHI_small_realplksr_small_pretrain
2xBHI_small_realplksr_small_pretrain
Scale: 2x
Network type: realplksr_small
Author: Philip Hofmann
License: CC-BY-4.0
Release: 21.05.2025
Purpose: 2x realplksr_small pretrain model with l1&mssim loss only.
Training iterations: 100'000
Description: A 2x realplksr_small pretrain model.
Visual Examples
Tensorboard Validation Graphs on BHI100
2xBHI_small_realplksr_small_dysample_pretrain
2xBHI_small_realplksr_small_dysample_pretrain
Scale: 2x
Network type: realplksr_small dysample
Author: Philip Hofmann
License: CC-BY-4.0
Release: 21.05.2025
Purpose: 2x realplksr_small dysample pretrain model with l1&mssim loss only.
Training iterations: 100'000
Description: A 2x realplksr_small dysample pretrain model.




















