-
Notifications
You must be signed in to change notification settings - Fork 93
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Investigate Windows 10 1607-1703 subpar d3d9 performance #164
Comments
Ok, so... a few other oddities. Secondly, at least when testing my favourite HDR_FP16x2.. I figured even W7 seemed to be somewhat suboptimal (220fps, vs 1511 doing 330fps and 1703 barely touching 170fps). Last but not least, I tried to swap 1703's |
Ok never mind, mystery unravelled for W7 falling behind the highest expectations FWIW I also tested this natively on my 9600k+2080S desktop (with the legacy 473 branch versus 531, but still) and I could report 900fps in W7 vs 550-600 in W10 22H2. Funnily enough, not even dxvk (640) could compete. |
You could try to record an ETW trace if it's really a problem on the latest version of the OS. |
Ok so, uh.. There's a lot to unpack here. I couldn't find (or at least I couldn't be bothered to have play nice) the exact same original build environment of the day, but the newer one works just as good. In fact.. it seems even too good? Pick up the fps numbers of the last post (from whatever ~2004 dx sdk and VS .NET 2003 exe they give you), and now multiply them by 2.5. But I also gave a run to the old build while I was there, and wtf? Linux can pull off 1700fps even with that. And given how fairly quirky the virtual 3d device can be, I wonder if that couldn't also be responsible for the biggest ass imbalances that I have measured inside my vm. |
Follows #111.
It turns out that newer Windows did far more than just gimping VP.
I couldn't get 1607 to experience more than a ~15% handicap (and boy wasn't it hard to find the right scenes), but with 1703 you'd have to be blind not to see a difference any time you are CPU limited.
I tested (in a 640x480 window at minimum detail for games, default size and settings for the rest):
I understand performance isn't exactly the kind of "qualitative" issue that the project usually addresses, but the effect is consistent and just as annoying if your framerates weren't ludicrous to begin with (in CSGO the performance uplift of multicore rendering is practically nullified, while the worst case microbenchmark scenario that I could scavenge was not even ONE THIRD of the W7/1511 speed).
There are of course other caveats, but none that should lessen the main point.
Yes, I did all of this testing inside VMWare workstation, but while the virtual SVGA device isn't exactly comparable to native, the provided d3d9 driver should be pretty legit to compare with itself (unlike normal gpus I believe it even has the same codepaths for every Windows).
Secondly this was done on my i7-6500U+950M old laptop, which isn't exactly a workhorse. But either through downclocking or core limiting (which should come especially easy if you use a VM), I see no reason anybody couldn't get down to the same level if it turned out it was actually required for reproducing the problem.
Last but not least, I also took care of excluding any possible spectre and meltdown consideration (<2018 Windows should know nothing about them, and my host has mitigations disabled anyway).
The text was updated successfully, but these errors were encountered: