nobody pulled your arm, Black person
good job being a consumer goy
the GPU market has been dead since cryptoBlack folk broke the supply/demand curve with their homosexualry
>turn on RT >lose all control over your character lighting when it comes time to block for dialogue or in-engine cutscenes >all at the cost of half your frames
No wonder devs only put this shit in when they're paid by Nvidia to do so.
Yes. Nvidia sponsors games to put in raytracing. I never said it was Nvidia exclusive. AMD cards just shit the bed so bad with raytracing and no FSR3 that it might as well be Nvidia exclusive.
https://www.nvidia.com/en-us/geforce/news/computex-2021-rtx-dlss-game-announcements/
The ray tracing can look good in certain situations, but it breaks a lot of the lightning composition in cutscenes. Look at how the lighting is more flat in the next gen update here for example.
>soul removed
Developers remastering the visuall style out of their old games, and calling it "improved lighting". Has there been a single good remaster that respected the source material?
That's not even an rtx thing, that's actual lighting compared to global illumination with no light source, the rtx is minimal the big difference was that the actual lighting system was changed, Skyrim 2011 could pull off similar to this in 2012 with a simple mod that removes global illumination and adds proper light sources.
>removes global illumination and adds proper light sources.
This requires actual work, better to just drop a weekend project into a release build and call it RTX ON.
Are you moronic? Skyrim has separate cells for each room that can have whatever light parameters. Witcher is fully open world and you literally can't just remove the global light source which is the sun from the equation in a classically rasterized environment.
Back to the original question which you’ve avoided because you’re a smarmy homosexual, why is reducing visual clarity in a visual medium a good thing. NOT in a horror game 😉
9 months ago
Anonymous
Cause it makes the important details pop out more, keep crying glass wearer
>60+fps
That's not enough for anyone with functioning eyeballs or someone who is used to high framerate gaming (I'm talking about 144+). Try going back to 60-80 fps after playing on 144+ for like a week.
The truth is that even a beast like the 4090 is not ready for 4K or raytracing.
For you, and it's okay. Having high framerates but then a crappy screen for me is a mortal sin, I care much more about perfect blacks than reaching an arbitrary framerate because that's what r/masterrace says. In my case it's 80-120fps with Witcher3 Ultra RTX on an OLED screen, and I couldn't ask for more
I would agree with you anon if the ray tracing was implemented in a way that compliments the original artistic intent. As it stands the RT ruins almost all cutscenes and thats not acceptable in a cinematic game.
as a turbo autist who bought a crt in 2010 for high refresh rates, the first tft 120hz monitor, 144 and 240 hz.. 100fps is the sweet spot but 75-90 is doable on slow ass rpgs like this
You don't even need rtx to do most of that it's just global illumination on with no light source, old games could do similar but they probably cut it from the original because of performance issues like they gutted most of the game back in 2014-15.
So comparing a gutted area to one they did properly is a terrible comparison.
I get 60-100 FPS everywhere apart from novigrad central, unless I turn frame generation on and then it's right back up there. If you have a GPU that supports it, it's definitely nice. Is it worth the upgrade? No. But when you get a new card because you need to? It's worth it. DLSS is great. Now if only nvidia would stop releasing shitty drivers 90% of the time.
Cool but there are so few places where RT makes any significant difference. Until we get fully path traced games I can wait. Cybertroony 77 is almost there but hardware is still not fast enough.
It looks good in motion while exploring but it sucks for cutscenes. Colored light bounces screw with the composition and washes out shadows. It's really bad in kaer morhen where the light bounces off the white brick castle and removes all shadows from faces.
cope. most of ya'll frickers don't have a capable PC to make use of it. it's going to keep evolving and adaptation is going to keep increasing. rasterization is going to be killed off very slowly
>play portal RTX in 1440p with 4070 >game looks cool I guess >but runs like dogshit unless you turn on performance mode
it's a gimmick that's years away from being optimized. upscaling is a nice technology (in a vacuum) but the rt push is so desperate and annoying
who here is complaining about graphical stagnation? I mostly see people that are sick of 4k horse balls textures and want devs to prioritize smooth performance
You can just switch to 1080p, but you've been conditioned to think 4k is the new standard now because consoles pushed this meme shit on you.
1080p now seems extremely low.
I play at 1440p. I just want devs to optimize stuff instead of relying on upscaling
the resolution isn't the problem, gpu prices and lazy devs are. look at remnant 2 for example, they literally said the game is made for upscaling while it looks like shit
Ah, so you're just lying then for the sake of lying.
9 months ago
Anonymous
what?
post specs
9 months ago
Anonymous
You're saying it looks like shit just because.
9 months ago
Anonymous
it looks like shit relative to performance, there's no reason for it to need upscaling like the devs claim other than incompetence
9 months ago
Anonymous
>performance bad
Let me guess you can't run it at 200 fps so it's bad.
It runs fine.
9 months ago
Anonymous
not in native, no. the devs *literally* admitted that they rely on upscaling, it's genuinely mind-boggling that you defend this when the devs don't
9 months ago
Anonymous
>I don't get 200fps in native that means performance bad >can't even tell the difference between native and quality upscaling unless specifically told
You're just b***hing about shit for the sake of it.
9 months ago
Anonymous
you're pulling my chain aren't you
9 months ago
Anonymous
Yes, I'm calling you out.
As long as you can run it at 60 fps, that means it runs great.
9 months ago
Anonymous
post specs
9 months ago
Anonymous
post frames
9 months ago
Anonymous
6:50
do you really think 30-40 fps on a 4070 is acceptable?
9 months ago
Anonymous
>30-40 fps
more like 10fps with 30 generated
9 months ago
Anonymous
Frame Gen is off though moron?
9 months ago
Anonymous
>literally 100 frames with no visual difference at all if you turn on DLSS >I won't turn it on and instead choose to b***h about it because I just will, ok?
Way to destroy your own argument.
9 months ago
Anonymous
can you post your specs already? I have a feeling you don't even have an rtx card
9 months ago
Anonymous
This game doesn't use ray tracing.
9 months ago
Anonymous
but it uses dlss which you specifically talked about, now post your specs console-kun
9 months ago
Anonymous
Any upscaler is just as good.
You don't need an rtx card.
9 months ago
Anonymous
why did you specifically mention dlss if you're using a different upscaler? you don't own a pc you poor third worlder
9 months ago
Anonymous
Because you said you have an RTX card and it's also the best upscaler, but the rest are getting just as cdecent.
Your problem was with upscaling in general, which is entirely your fault.
9 months ago
Anonymous
>my upscaler is better than yours
And these are the people who yell poorgay at every opportunity?
9 months ago
Anonymous
Those are facts because dlss uses specifically designed cores, while other upscalers just leech off of the cards normal performance.
The quality of DLSS is much better as a result.
9 months ago
Anonymous
except when it breaks all your post processing effects of course. Which it absolutely does in witcher 3 at the moment.
9 months ago
Anonymous
DLSS fricks DOF in the witcher 3 so bad, I was wondering why the game looked better on xbox series x compared to my PC, turned off DLSS and cutscenes looked good again. Absolute meme technology.
9 months ago
Anonymous
I have zero issues with upscaling itself, it's a great technology. I do however have an issue with lazy devs and people who defend them for free. remnant 2 isn't even the standard, it's an especially bad example so I don't understand why you pretend like it's normal
so what gpu do you have?
RTGI is.
of course that doesn't count because you're too poor to afford a new card.
Also yes, upscaling and frame generation are great.
I can run games with my 7 year old card at a point where I absolutely should not be able to run them without completely tanking the resolution and everything looking pixelated.
9 months ago
Anonymous
>Also yes, upscaling and frame generation are great.
I have a zero tolerance policy for fake frames in my household.
9 months ago
Anonymous
Everything you see on your screen is fake.
Also you clearly haven't tried it, you're just sperging out because someone told you you're supposed to hate it.
9 months ago
Anonymous
>Everything you see on your screen is fake.
I don't know what kind of screen you are operating anon, but you should maybe invest in a better one before ending yourself on a overpriced gfx card
>waaa graphical and technological stagnation waaaa >ok here are some advancements >noooooo not like this waaaaaah
this thread summarized
>technological stagnation
shoving unoptimized RT into a game is your idea of progress? How about respecting the source material for once and do a proper remaster instead of releasing a highschool level tech demo.
>"advancements" are making games look like fricking shit
I too clapped and did the Nintendo Switch face when motion blur and DOF were introduced. I specially like it when those "advancements" are forced in consoleshit games and also offer better FPS when you disable them via hacking
tw3 looks like shit either way. Without RT it has stupidly bright colors, with RT on it looks like a fan mod "THIS IS WHAT THE WITCHER 3 WOULD LOOK LIKE IN 2035"
More like you paid 1600 dollars for Cyberpunk Overdrive RT. That shit is game changing. It's a night and day difference between regular ray tracing and path tracing.
Because they are out of touch with reality.
if you actually look at reviews through the years, the top card of the day often couldn't run certain games maxed out at decent frames as well.
Peak resolution was like 1600*1200 and they still only got like 48 frames with all the eye candy turned on.
>not resting and vesting in some boomer company like Google
stay mad, maybe you'll get that girlfriend you missed out at age 65 while looking like a balding shriveled walnut from bangladesh kek
comfy 2070 Super user here. Looks like i bought the correct pc parts which will last 10 years again. Nothing really tests it on high settings unless you're a moron and go ultra and raytracing with literally no discernable difference during gameplay. I'll just wait for this rtx dlss amd whatever thing to blow over and then buy another nvidia card.
My dad says buying a card now is like buying a card back when dx9 first came out. Look up the 5200 card.
>rtx dlss amd whatever thing to blow over
Never going to happen.
Real time Ray tracing, or specifically path tracing is the future, because it was literally the goal from the start and we are now able to achieve it.
Upscaling will always be used to ush 8k and 12k.
by blow over i mean I'll wait for 1. games to release that were built with ray tracing as standard not an option
and 2. for the rtx cards to be actually capable of providing decent fps on these games
for now the difference simply isnt noticable enough to warrant shit sub 100 fps
>games to release that were built with ray tracing as standard not an option
Has been done years ago.
Metro Exodus.
Nobody talks about it because you never got any performance issues so there was nothing to complain about.
Point of raytracing is general accuracy
Rasterization can create realistic lighting scenarios, but is too much handiwork to be applied EVERYWHERE
Open world games have light bleed everywhere
You want to have a dark cave? Can be done, has been done since 90s. Dark cave + flashlight that works like REAL FLASHLIGHT? Can't be done rasterized, you end up with all the dogshit solutions that make it look like the cave is made of black hole darkness swallowing the light from the flashlight.
Like you fire up your maglite in a cave and the whole PLACE is now full of light because it bounces off the rocks
Those homosexuals already are using UE with shit results because every game has to be open world for no reason.
I shit on Ubishit all the time but their games don't suffer those performance issues nearly as much.
Still I want to see decent UE5 game with both lumen and nanite.
Immortals of Aveum will be the first aside from Fortnite
It's funny because people complain all the time about the hacky workarounds used in traditional rasterized pipelines to achieve certain effects but then turn their nose up at raytracing because muh frames. Like color grading when you approach a cave or fake mirrors. Workarounds that aren't needed with raytracing.
Like if all you care about is framerate why not just play everything on the lowest setting.
Only the most absolute morons buy the top cards.
Always have been the case.
That said, it's literally the best card this gen has to offer because nvidia fricked up the 4XXX series this badly with it's moronic 4080 12GB fiasco.
keep going bud. 4080 is good but I probably could have waited a generation to be honest as even though I have a 1440p monitor I often just use my 1080p for the smoothness and higher refresh. It's a tft so its automatically better for darker games. IPS is a meme I wont fall for again thats for sure
GPUs peaked with Pascal. No one cares anymore. So to sell new cards nvidia have to shill bullshit tech that cripples your performance. Now you *need* a new card, or you can't use their stupid new lighting tech. Enjoy getting 60 fps unless you have a 1000 dollar gpu... what are you, poor? And people buy into it because like with anything people cannot be behind the curve
nobody pulled your arm, Black person
good job being a consumer goy
the GPU market has been dead since cryptoBlack folk broke the supply/demand curve with their homosexualry
/thread
Kek at you
worth every nickel
which one is supposed to be better?
Left is ray tracing on, with hideous checkerboard rendering artifacts.
you're both moronic lmao
compare textures
>compare textures
The frick you mean? textures are worse when you enable ray tracing. This is a screenshot from PS5. Non RT mode has higher res
Look at the chest hair, the hair is missing. Fricking moron. And PS5 doesnt have actual raytracing lmfao you dumb shit
LMFAO HAHAHAHAHAH WTFFFFF.
Stfu you creature.
why dont you post the video this is from
(you wont)
>why dont you post the video this is from
It's PS4 vs PS5 KEK
So its base game vs remaster
And no the remaster does not have raytracing on PS5. It has at best, shadows, not actual raytracing.
So you're moving the goalposts.
Nextgen looks very good with RT off. The fact that all cinematics are now done in engine is a massive improvement on it's own
>turn on RT
>lose all control over your character lighting when it comes time to block for dialogue or in-engine cutscenes
>all at the cost of half your frames
No wonder devs only put this shit in when they're paid by Nvidia to do so.
>RT is an Nvidia exclusive tech
Can you stop talking about shit you don't understand?
AMD cards can now do RT just as well.
Whomst are you quoting because it isn't me.
>paid by nvidia
>not me
Yes, it's you.
Yes. Nvidia sponsors games to put in raytracing. I never said it was Nvidia exclusive. AMD cards just shit the bed so bad with raytracing and no FSR3 that it might as well be Nvidia exclusive.
https://www.nvidia.com/en-us/geforce/news/computex-2021-rtx-dlss-game-announcements/
Except AMD cards can to RT just fine now.
https://twitter.com/HardwareUnboxed/status/1626194150210277377?cxt=HHwWgsDQsda9s5EtAAAA
uhoh stinky
Better comparison here
The ray tracing can look good in certain situations, but it breaks a lot of the lightning composition in cutscenes. Look at how the lighting is more flat in the next gen update here for example.
soul removed
>soul removed
Developers remastering the visuall style out of their old games, and calling it "improved lighting". Has there been a single good remaster that respected the source material?
>Has there been a single good remaster that respected the source material?
that CD-i Zelda fan remaster, out of all things
Fan remasters are a different category
No
>removed moon light
Why?
That's not even an rtx thing, that's actual lighting compared to global illumination with no light source, the rtx is minimal the big difference was that the actual lighting system was changed, Skyrim 2011 could pull off similar to this in 2012 with a simple mod that removes global illumination and adds proper light sources.
>removes global illumination and adds proper light sources.
This requires actual work, better to just drop a weekend project into a release build and call it RTX ON.
Are you moronic? Skyrim has separate cells for each room that can have whatever light parameters. Witcher is fully open world and you literally can't just remove the global light source which is the sun from the equation in a classically rasterized environment.
This just looks like if someone turned on fullbright with ambient occlusion on versus off.
the rtx update downgraded the SSAO and HBAO so they could spoof these comparisons
and then they patched HBAO back in after the media coverage was over. The polaks know how to spin them
i don't remember any area looking like these. where are they?
Frick off, I played on a 1070 and it looked better than the first image.
now post the same comparison with last gen where they didnt remove ambient occlusion
both looks like shit
oh good now witcher 3 can run at 10 fps instead of 40 fps inside
This isnt for poorgays like yourself
But the guy at gamestop said I was the coolest Black person on the block when I bought witcher 3 for the ps4
>Witcher 3
>PS4
I think he said that just to be polite
Rt on is the right and off is the left side you lying amd troony
RT on is left you tech illiterate moron, you can see the RT light bounces color Geralts white hair to be more golden in the left image.
huehuehue
>testing games
What is this supposed to prove? The low 1% and 0.1% are all much worse.
The 4080 and 7900 XTX are in two different price tiers anon. Would be like if the 4080 performed within 2% of the 4090. It's bad for AMD.
I thought that screenshot was supposed to prove the XTX was better so nvm
looks like a fan mod
I seen shit like this being done whithout ray tracing, it just seems like developers delegating their work to costly tech
Same with optimization
>I seen shit like this being done whithout ray tracing
Ok post a open world game with dynamic lighting and day night cycles that does this
Skyrim.
Incorrect
Now 90% of the screen is black and you can see less shit! Wow!
>poorly lit area is dark
Wow..........
Thanks for clarifying with that big brain of yours. Now explain why reducing visual clarity in a visual medium is a good thing.
Yeah horror games are shit cause I cant see anything!
The Witcher is a horror game?
>only horror can have dark areas
True true
Back to the original question which you’ve avoided because you’re a smarmy homosexual, why is reducing visual clarity in a visual medium a good thing. NOT in a horror game 😉
Cause it makes the important details pop out more, keep crying glass wearer
nice
>Playing the b***her 3 in 2015+8
Oof.
cope tendie
DUDE! LE RAIN! OMG REFLECTIONS!
lipstick for a pig, cp2077 looks awful like a hd mod for ps3 era game
ok so 90% performance decrease for some reflections, awesome
Nice but still not worth cutting framerate by 70%.
For you, and rightly so if you can't afford more powerful hardware that still allows you to get 60+fps
It's clearly borked turning DX12 alone will tank performance even without enabling RT.
>60+fps
That's not enough for anyone with functioning eyeballs or someone who is used to high framerate gaming (I'm talking about 144+). Try going back to 60-80 fps after playing on 144+ for like a week.
The truth is that even a beast like the 4090 is not ready for 4K or raytracing.
For you, and it's okay. Having high framerates but then a crappy screen for me is a mortal sin, I care much more about perfect blacks than reaching an arbitrary framerate because that's what r/masterrace says. In my case it's 80-120fps with Witcher3 Ultra RTX on an OLED screen, and I couldn't ask for more
I would agree with you anon if the ray tracing was implemented in a way that compliments the original artistic intent. As it stands the RT ruins almost all cutscenes and thats not acceptable in a cinematic game.
as a turbo autist who bought a crt in 2010 for high refresh rates, the first tft 120hz monitor, 144 and 240 hz.. 100fps is the sweet spot but 75-90 is doable on slow ass rpgs like this
>60+fps
I would rather have 90-120 fps than only 60 and this dumb RT shit.
You don't even need rtx to do most of that it's just global illumination on with no light source, old games could do similar but they probably cut it from the original because of performance issues like they gutted most of the game back in 2014-15.
So comparing a gutted area to one they did properly is a terrible comparison.
I get 60-100 FPS everywhere apart from novigrad central, unless I turn frame generation on and then it's right back up there. If you have a GPU that supports it, it's definitely nice. Is it worth the upgrade? No. But when you get a new card because you need to? It's worth it. DLSS is great. Now if only nvidia would stop releasing shitty drivers 90% of the time.
Cool but there are so few places where RT makes any significant difference. Until we get fully path traced games I can wait. Cybertroony 77 is almost there but hardware is still not fast enough.
4090 can run cyberpunk at 4k with 60-100 fps everything maxed including overdrive enabled. Just don't be poor.
Yea, with DLSS + FG. Frick that.
It looks good in motion while exploring but it sucks for cutscenes. Colored light bounces screw with the composition and washes out shadows. It's really bad in kaer morhen where the light bounces off the white brick castle and removes all shadows from faces.
RT is the biggest meme of all time
My PC turns into a heater with RT on, can't play like this in the summer.
cope. most of ya'll frickers don't have a capable PC to make use of it. it's going to keep evolving and adaptation is going to keep increasing. rasterization is going to be killed off very slowly
>play portal RTX in 1440p with 4070
>game looks cool I guess
>but runs like dogshit unless you turn on performance mode
it's a gimmick that's years away from being optimized. upscaling is a nice technology (in a vacuum) but the rt push is so desperate and annoying
Portal with RTX has less soul than the original
>soul
buzzword
its been years now and Ganker still doesnt kno what RT actually goes besides reflections
that's because it's barely noticeable and a waste of gpu resources
With the updated version of TW3 are there any mods worth looking at?
>mods
Brothers in Arms
4K Armor Textures (fixes the oversaturated color armor got after BW DLC)
>waaa graphical and technological stagnation waaaa
>ok here are some advancements
>noooooo not like this waaaaaah
this thread summarized
who here is complaining about graphical stagnation? I mostly see people that are sick of 4k horse balls textures and want devs to prioritize smooth performance
You can just switch to 1080p, but you've been conditioned to think 4k is the new standard now because consoles pushed this meme shit on you.
1080p now seems extremely low.
I thought 4k was a meme for a long time, but now that I have gotten used to it 1080p looks like garbage.
I play at 1440p. I just want devs to optimize stuff instead of relying on upscaling
the resolution isn't the problem, gpu prices and lazy devs are. look at remnant 2 for example, they literally said the game is made for upscaling while it looks like shit
>it looks like shit because I have a shit card
fixed
4070
post specs
Ah, so you're just lying then for the sake of lying.
what?
post specs
You're saying it looks like shit just because.
it looks like shit relative to performance, there's no reason for it to need upscaling like the devs claim other than incompetence
>performance bad
Let me guess you can't run it at 200 fps so it's bad.
It runs fine.
not in native, no. the devs *literally* admitted that they rely on upscaling, it's genuinely mind-boggling that you defend this when the devs don't
>I don't get 200fps in native that means performance bad
>can't even tell the difference between native and quality upscaling unless specifically told
You're just b***hing about shit for the sake of it.
you're pulling my chain aren't you
Yes, I'm calling you out.
As long as you can run it at 60 fps, that means it runs great.
post specs
post frames
6:50
do you really think 30-40 fps on a 4070 is acceptable?
>30-40 fps
more like 10fps with 30 generated
Frame Gen is off though moron?
>literally 100 frames with no visual difference at all if you turn on DLSS
>I won't turn it on and instead choose to b***h about it because I just will, ok?
Way to destroy your own argument.
can you post your specs already? I have a feeling you don't even have an rtx card
This game doesn't use ray tracing.
but it uses dlss which you specifically talked about, now post your specs console-kun
Any upscaler is just as good.
You don't need an rtx card.
why did you specifically mention dlss if you're using a different upscaler? you don't own a pc you poor third worlder
Because you said you have an RTX card and it's also the best upscaler, but the rest are getting just as cdecent.
Your problem was with upscaling in general, which is entirely your fault.
>my upscaler is better than yours
And these are the people who yell poorgay at every opportunity?
Those are facts because dlss uses specifically designed cores, while other upscalers just leech off of the cards normal performance.
The quality of DLSS is much better as a result.
except when it breaks all your post processing effects of course. Which it absolutely does in witcher 3 at the moment.
DLSS fricks DOF in the witcher 3 so bad, I was wondering why the game looked better on xbox series x compared to my PC, turned off DLSS and cutscenes looked good again. Absolute meme technology.
I have zero issues with upscaling itself, it's a great technology. I do however have an issue with lazy devs and people who defend them for free. remnant 2 isn't even the standard, it's an especially bad example so I don't understand why you pretend like it's normal
so what gpu do you have?
>4070
You mean an overpriced 4060?
>4060
you mean an overpriced 4050?
>muh optimization
Literally nothing wrong with upscaling. it's a form of optimization.
>upscaling
>frame generation
This is what Ganker view as technological progress.
480p upscaled to 8k, 15fps with 100 fake frames inbetween. This is the future of gaming.
RTGI is.
of course that doesn't count because you're too poor to afford a new card.
Also yes, upscaling and frame generation are great.
I can run games with my 7 year old card at a point where I absolutely should not be able to run them without completely tanking the resolution and everything looking pixelated.
>Also yes, upscaling and frame generation are great.
I have a zero tolerance policy for fake frames in my household.
Everything you see on your screen is fake.
Also you clearly haven't tried it, you're just sperging out because someone told you you're supposed to hate it.
>Everything you see on your screen is fake.
I don't know what kind of screen you are operating anon, but you should maybe invest in a better one before ending yourself on a overpriced gfx card
>technological stagnation
shoving unoptimized RT into a game is your idea of progress? How about respecting the source material for once and do a proper remaster instead of releasing a highschool level tech demo.
It's been fine for a while now. Stop fricking b***hing about everything when you have a decade old card.
>ok here are some advancements
doko?
>noooooo not like this waaaaaah
>"advancements" are making games look like fricking shit
I too clapped and did the Nintendo Switch face when motion blur and DOF were introduced. I specially like it when those "advancements" are forced in consoleshit games and also offer better FPS when you disable them via hacking
tw3 looks like shit either way. Without RT it has stupidly bright colors, with RT on it looks like a fan mod "THIS IS WHAT THE WITCHER 3 WOULD LOOK LIKE IN 2035"
>gaytracing
More like you paid 1600 dollars for Cyberpunk Overdrive RT. That shit is game changing. It's a night and day difference between regular ray tracing and path tracing.
>It's a night and day difference
The open world is still boring and the gameplay is shit. But now it is lighted in a nice way.
Yes, good thing we were talking about graphics, goalpost moving homosexual.
>poorgay cope thread
well if your moronic ass did a single bit of research you'd know that raytracing is dogshit and has a terrible visuals/performance ratio
Why do these threads always turn into poorgay coping.
Because they are out of touch with reality.
if you actually look at reviews through the years, the top card of the day often couldn't run certain games maxed out at decent frames as well.
Peak resolution was like 1600*1200 and they still only got like 48 frames with all the eye candy turned on.
how much latency does it add? i saw some Gankertard post screenshot bragging with rts and dlss and he had 60ms LOL
>60ms
Jesus christ wtf, how is this acceptable? Not even bloodborn on the PS4 has this much latency.
>thread full of nvidia employees
frick off pajeets, dont you have an 80 hour work week to get back to
stay poor, troony lover
>not resting and vesting in some boomer company like Google
stay mad, maybe you'll get that girlfriend you missed out at age 65 while looking like a balding shriveled walnut from bangladesh kek
Nope. I'm sitting on my bitcoin wealth made off all the normies and poor gays that bought in during thr pandemic lol.
Nice larp, it's OK to get scammed you know :^)
The best part of having shitty eyesight is that anything beyond 1080p looks basically the same to me.
comfy 2070 Super user here. Looks like i bought the correct pc parts which will last 10 years again. Nothing really tests it on high settings unless you're a moron and go ultra and raytracing with literally no discernable difference during gameplay. I'll just wait for this rtx dlss amd whatever thing to blow over and then buy another nvidia card.
My dad says buying a card now is like buying a card back when dx9 first came out. Look up the 5200 card.
>rtx dlss amd whatever thing to blow over
Never going to happen.
Real time Ray tracing, or specifically path tracing is the future, because it was literally the goal from the start and we are now able to achieve it.
Upscaling will always be used to ush 8k and 12k.
by blow over i mean I'll wait for 1. games to release that were built with ray tracing as standard not an option
and 2. for the rtx cards to be actually capable of providing decent fps on these games
for now the difference simply isnt noticable enough to warrant shit sub 100 fps
>games to release that were built with ray tracing as standard not an option
Has been done years ago.
Metro Exodus.
Nobody talks about it because you never got any performance issues so there was nothing to complain about.
Point of raytracing is general accuracy
Rasterization can create realistic lighting scenarios, but is too much handiwork to be applied EVERYWHERE
Open world games have light bleed everywhere
You want to have a dark cave? Can be done, has been done since 90s. Dark cave + flashlight that works like REAL FLASHLIGHT? Can't be done rasterized, you end up with all the dogshit solutions that make it look like the cave is made of black hole darkness swallowing the light from the flashlight.
Like you fire up your maglite in a cave and the whole PLACE is now full of light because it bounces off the rocks
left looks better.
homosexuals who b***h about RT are just moronic.
It's not ready for mass market but it's the future.
meh, its best selling point is that it will make games easier produce. But the hardware is not there for the consumer to actually enjoy it
Those homosexuals already are using UE with shit results because every game has to be open world for no reason.
I shit on Ubishit all the time but their games don't suffer those performance issues nearly as much.
Still I want to see decent UE5 game with both lumen and nanite.
Immortals of Aveum will be the first aside from Fortnite
It's funny because people complain all the time about the hacky workarounds used in traditional rasterized pipelines to achieve certain effects but then turn their nose up at raytracing because muh frames. Like color grading when you approach a cave or fake mirrors. Workarounds that aren't needed with raytracing.
Like if all you care about is framerate why not just play everything on the lowest setting.
ray-tracing, and even path-tracing, is a complete fricking meme
Hope you enjoy all the fake frames you paid for
Only the most absolute morons buy the top cards.
Always have been the case.
That said, it's literally the best card this gen has to offer because nvidia fricked up the 4XXX series this badly with it's moronic 4080 12GB fiasco.
I bought that card to play games at 1440p at 165fps. it's pretty overkill for that and Raytracing is an absolute meme.
>I bought that card to play games at 1440p at 165fps
moron
>Raytracing is an absolute meme.
double moron
I get a lot of coil whine when I play games with it at 200+ fps.
moron here. I bought a 1080 on release and it lasted me 6 years and probably still plays games well. got a 4080 now
Still using my 1080, ironic that it ended up being the peak for 1080p gaming
keep going bud. 4080 is good but I probably could have waited a generation to be honest as even though I have a 1440p monitor I often just use my 1080p for the smoothness and higher refresh. It's a tft so its automatically better for darker games. IPS is a meme I wont fall for again thats for sure
Metro Exodus converted me into raytracing believer
GPUs peaked with Pascal. No one cares anymore. So to sell new cards nvidia have to shill bullshit tech that cripples your performance. Now you *need* a new card, or you can't use their stupid new lighting tech. Enjoy getting 60 fps unless you have a 1000 dollar gpu... what are you, poor? And people buy into it because like with anything people cannot be behind the curve
Witcher 3 sucks