This is getting to be ridiculous. We're weeks away from having a huge flock of new players coming over to PC. Now granted it's great that D2 can run at 30-60fps off basically a potato...that's nice. But, my huge but, is those of us who've invested a lot of money into a gaming rig, specifically graphics cards to get a higher FPS out of destiny, well....nothing's been done for us since launch. I'm running a 2080ti now, and getting the same FPS at max graphics as I do when running my 980ti. Now this is great for someone who only has a 980ti, but for those of us with newer, more expensive graphics cards, not being able to have the option to sacrifice graphics for FPS is ridiculous. I have switched from 1440p, max graphics settings, and an FOV of 105, all the way down to 720p, lowest graphics with the lowest FOV (I think it's like 55 or something) and for some reason, my FPS still won't change (FPS number completely map dependant, what's going on, how many enemies have to be rendered, yada, yada, yada. My point is, no matter the situation and FPS in that location, the difference in resolution and graphics quality doesn't change my FPS at all, which is so strange) Now before anyone asks me anything about my monitors refresh rates, drivers, etc.... etc.... understand that the last 4 months of my life have been dedicated to researching and testing every possible method to increase my FPS, run every test imaginable, even bought 3 different graphics cards.... It's 100% destiny 2 and it's incredibly poor optimization for PC. For whatever reason, D2 doesn't ask your graphics card for more resources when outputting FPS. When I switch from 720p low graphics to 1440p at max graphics I can see with HWiNFO64 that my GPU core load drops dramatically, as does my Total GPU Power (normalized) [% of TDP] when playing at lower resolutions, which means D2 should be asking my GPU for more resources to output higher FPS when lowering graphical settings...right? Wrong, for whatever reason my % of usage on both can be below 50, or above 90 and no matter what, my FPS stays the same. That being said my FPS is also never a steady 118- or 90, 150, etc... the FPS itself varies constantly based on what it has to render, number of enemies, what planet I'm on, activity, etc.... When I say my FPS stays exactly the same, I'm saying in a public event on the EDZ if my FPS is between 105-112, I can then switch between max graphics 1440p fullscreen, to lowest graphics 720p fullscreen and the FPS stays in that 105-112 range. Same thing in the crucible, gambit, tower, etc... I never get a boost in FPS by lowering graphics or resolution. Why? Because Bungie hasn't optimized D2 for PC since year 1 D2. [spoiler]Moderator edit: This thread has been updated with tags that are more appropriate. Feel free to private message the moderator who moved your post, link to topic, for further clarification about why this topic was moved.[/spoiler]
*begins jacking off* Oh yes, my frames. MY FRAMES! *bursts onto my keyboard at 200 frames per second*
Edited by Yuno: 9/19/2019 2:10:57 AMIts probably because destiny is pretty CPU heavy. If you just go to a better graphics card then there will be improvement but probably not like you would expect in most games. The performance upgrade you are looking for in the 2080ti from the 980ti will probably come from something like upgrading to an i9 from an i7. I’m not the most educated in the world about pc stuff but I do know that this game cares more about your CPU.
Edited by KingRosso1996: 9/18/2019 3:25:54 PMI agree, i have fps spikes like crazy aswell. on the top end its 200+ and at the low end its 50. And its pretty frustrating. I have a great setup aswell. And sure if u dont want fps spikes you should set the fps limiter on. But honestly if it gets below 60 while having 8gb of vram. Is pretty ridicilouse. Destiny now is capped at 2,5gb vram at max graphic i think so basicly with 3gb of vram u get the same result as with you're 2080ti.