Welcome to Doom9's Forum, THE in-place to be for everyone interested in DVD conversion. Before you start posting please read the forum rules. By posting to this forum you agree to abide by the rules. |
11th May 2020, 12:35 | #21 | Link |
Registered User
Join Date: Oct 2001
Posts: 454
|
Yes, Waifu2x is fine for anime and stuff - but in no way potent for real-life stuff..
As mentioned above, take a look at "hybrid" from selur - he already has a working implementation of the vapoursynth ports with vulkan etc... No need to re-invent the wheel for avisynth with CPU-based OPEN CL .... |
11th May 2020, 14:45 | #23 | Link | ||
Acid fr0g
Join Date: May 2002
Location: Italy
Posts: 2,580
|
Quote:
Quote:
CUDA has many many tools and resources to help programmers. AMD has great hardware but really poor support and drivers.
__________________
@turment on Telegram |
||
11th May 2020, 18:45 | #26 | Link |
I'm Siri
Join Date: Oct 2012
Location: void
Posts: 2,633
|
note that ESRGAN is already available to vaporsynth thru vs_mxnet, you can turn your SD masters to native 2k anytime you like, the quality is levels above waifu2x or nnedi3, I therefore find all this waifu2x discussion totally meaningless
|
11th May 2020, 18:56 | #27 | Link | |
Registered User
Join Date: Sep 2007
Posts: 5,377
|
Quote:
https://github.com/rlaPHOENiX/VSGAN But that image looks nothing like a "native 2K image". It's a huge overstatement But waifu2x is meant for anime sources, and was trained on anime sources - how is it "meaningless ? " Or, do you have an ESRGAN model that gives similar or better results than waifu2x on typical anime sources ? It's all about the training used and the specific model. I tried some public distributed anime trained models for ESRGAN , and the results are significantly worse |
|
11th May 2020, 19:06 | #28 | Link |
I'm Siri
Join Date: Oct 2012
Location: void
Posts: 2,633
|
the problem is not about anime or photos, waifu2x is a really "early" project, its models are too small and do not have enough capacity compared to state-of-the-art models in the academia. if someone's gonna port neural net based resizers to avisynth, I think it makes more sense to directly go for the modern models.
|
11th May 2020, 19:12 | #29 | Link | |
I'm Siri
Join Date: Oct 2012
Location: void
Posts: 2,633
|
Quote:
I am 99.9% sure that, given the same training set, anime or not, ESRGAN will produce much better results |
|
11th May 2020, 19:25 | #30 | Link | |||
Registered User
Join Date: Sep 2007
Posts: 5,377
|
Quote:
Quote:
Quote:
Waifu2x has many different GUI's for it, several types of pretrained models available, some with lua scripting in other programs (aviutl) , many different implementations supporting AMD, Nvidia) - it just seems so mature and that it should be easier to port to avisynth. Many of the newer research projects are python based, pytorch, etc...so they should be able to be work in vapoursynth much more easily, than something like avisynth |
|||
11th May 2020, 19:37 | #32 | Link | |
Registered User
Join Date: Sep 2007
Posts: 5,377
|
Quote:
Which ESRGAN model did you use above? |
|
11th May 2020, 19:59 | #33 | Link |
I'm Siri
Join Date: Oct 2012
Location: void
Posts: 2,633
|
the official one
also the results above pass for 2k to me, it surely ain't high quality 2k, it looks a bit like the quality of 16mm film. native 2k could be u know, slightly out of focus and stuff and not necessarily very sharp or with extremely delicate details. screenshot took directly from a 2k master (filmed in 2018), does it also look "fake" to you? |
11th May 2020, 20:39 | #34 | Link |
Registered User
Join Date: Mar 2018
Posts: 447
|
As good as ESRGAN is, it's still meant for still images. When applied to video it may not be temporally coherent. Has anyone thought about porting GANs which are meant for video? Such as TecoGAN (and I'm sure there are many more).
|
11th May 2020, 22:11 | #35 | Link | ||
Registered User
Join Date: Sep 2007
Posts: 5,377
|
Quote:
Quote:
There is no way if you had a comparable shot up close that you would have the coarse detail characteristics, image artifacts like that Britney upscale |
||
11th May 2020, 22:44 | #36 | Link | |
I'm Siri
Join Date: Oct 2012
Location: void
Posts: 2,633
|
Quote:
have you tried scaling the original image with, say, spline64? that's what upscale normally looks like, then there're nnedi3 and waifu2x which produce sharper edges but the result becomes "oil-painted", neither looks like the result of ESRGAN. The "eye-lash" or the whole "eye" region looks very native hi-res to me. |
|
11th May 2020, 23:13 | #37 | Link | ||
Registered User
Join Date: Sep 2007
Posts: 5,377
|
Quote:
Nobody would think this was an upscale, almost everyone (that has any experience) would think Britney was. You might be able to pass Britney off to the uninitiated or lay person, not to anyone that works in image or video industry. Quote:
ESRGAN is "less bad", and it does a generally better job with finer details. But there are telltale signs that make it look like an upscale and not native 2K. If you used the original models very much, certain texture patterns are prone to serious errors as well . Certain frames are unusable because of this The flicker or temporal inconsistency was mentioned earlier too; it's not that great on video because of those issues |
||
12th May 2020, 14:19 | #38 | Link |
I'm Siri
Join Date: Oct 2012
Location: void
Posts: 2,633
|
I'm not sure what you mean by "edge clumps", I found a native HD close-up shot of britney in another video and the hair also looks "clumpy", probably just the texture of her hair.
all these were shot in circa 2010, when music video production began shifting towards 1080p. a lot of HD videos shot around that time or earlier have this "coarse" look, you can tell it's HD, but it's not high quality HD, it definitely does not squeeze out the full potential of 2k resolution. you just know the image could still be much "finer" and more "delicate" at 2k. the gap between SD and HD wasn't all that huge since SD videos at that time usually look much "finer" at their native resolution. more recent shots have managed to get rid of this "coarse" look, they look more "HD" than them shots in 2010, the image looks much more "delicate" and you can tell instantly that it has exploited every last bit of 2k resolution potential. this is the "real" HD quality I'm talking about. I'm not quite sure what changed in the production process of recent years that led to such quality boost. Last edited by feisty2; 12th May 2020 at 14:31. |
12th May 2020, 15:57 | #39 | Link | |
Broadcast Encoder
Join Date: Nov 2013
Location: Royal Borough of Kensington & Chelsea, UK
Posts: 2,904
|
Quote:
Not that cameras of the past were bad, but we made huge step forwards compared to years ago as technology became so much better. And this was all years before HDR was even a thing! You can see how cameras became much more advanced to record frequencies than what the official standardized specs were allowing broadcasters to air... We squeezed everything we could for years out of the FULL HD standard... |
|
12th May 2020, 16:16 | #40 | Link |
I'm Siri
Join Date: Oct 2012
Location: void
Posts: 2,633
|
I think it probably has more to do with lens and cameras. older cameras can record 1080p videos but that's highest resolution they can handle, while modern cameras can record 6k or 8k videos and when downscaled to 2k, it should look much "finer" than footage shot by 2k cameras.
|
Thread Tools | Search this Thread |
Display Modes | |
|
|