I'm not against Ray Tracing in principle, but Capcom's implementation so far has been shitty and pointless.for the most part.
Mostly reflections on marble, which DD2 won't even have for 99.9% of environments, too.
tbh most RT implementations in most games are shit, and that will never change because consoles are the primary market and consoles will never play well with RT
there's like a maximum of 5 games in total where you can look at RT and say "yeah that's worth turning on"
Full on pathtracing looks good in Metro and Cyberpunk, and Control also had good enough reflections, simply because most of the game was glass or shiny surfaces.
Obviously it only makes sense if you have a top of the line GPU and you have extra FPS to spare anyway.
if they werent you wouldn't be able to run shit on that old ass garbage GPU, and you'd still whine and moan, and the devs would have to optmize it for people like you.
>games are being held back by consoles because devs have to make the games run on weak hardware >but my weaker than a PS5 gpu should run it!
You make zero sense, you pick one or the other, if they go to town with the graphics and dont hold themselves back because of a PS5, your shit GPU wouldnt be able to handle it either, if you could, the ps5 could, and the games would have been held back.
5 months ago
Anonymous
Except no one would cry about a game pushing graphical boundries not being able to be played on 7 year old hardware.
5 months ago
Anonymous
You obviously missed the cyberpunk threads.
5 months ago
Anonymous
you can't fool me. that's driver.
5 months ago
Anonymous
No it isn't. It is cyberpunk.
5 months ago
Anonymous
cyberpunk looks very SOVLFVL
5 months ago
Anonymous
5 months ago
Anonymous
no, thats midtown madness 2 you fricking moron
5 months ago
Anonymous
No it isn't. It is cyberpunk.
Holy shit. Cyberpunk at max settings looks like THAT?
PCbros, what's your excuse for that?
5 months ago
Anonymous
Read the filename, this is clearly PS4 capture :^)
5 months ago
Anonymous
They did with Alan Wake 2 also, recommended specs are a 4090, and the game looks horrible on the PS5 so they clearly don't care.
and you can also categorically frick off
the 1070 was a shit card even at release, its trash throw it away and then throw yourself away because you're an idiot clinging to hardware like a console tard
The 1080Ti came out a year later. The 1080 was garbage for the price, but the 1060 and 1070 were pretty good... at MSRP.
The RX 480 was a 1060 competitor, either one was fine.
zoomers on this board weren't even old enough to buy PC hardware in 2016 let alone 2010. They don't remember we were even having shortages back then to miners in 2016 because they'd buy all the GPUs the second they became available. Because of the miners I was able to sell my R9 280x in late 2017 for 20 dollars MORE than I bought it new in 2014. It looks like that train is long gone now though.
According to Gaymer Nexus, PS5 equivalent GPU is in between GTX 1060 to GTX 1080 while Digital Foundry claims that it's RTX 2070 S.
I trust the former so you should do fine.
Nope, they just tested shit port of DMC5 which ran at some stupid internal resolution on PS5 even if you set the output to 1080p.
This is where the >PS5 GPU is just GTX 1060
meme come from.
>According to Gaymer Nexus, PS5 equivalent GPU is in between GTX 1060 to GTX 1080 while Digital Foundry claims that it's RTX 2070 S.
Direct hardware comparisons don't really paint the full picture because the set hardware of consoles allows for greater optimization since the programmers know exactly what the game is running on and playtest it on that hardware. It's similar to how you can't really compare phone specs to PC specs due to phones having to to deal with battery and heat more
It's a 6600XT-equivalent, so 2070-ish is about right.
But also what
>According to Gaymer Nexus, PS5 equivalent GPU is in between GTX 1060 to GTX 1080 while Digital Foundry claims that it's RTX 2070 S.
Direct hardware comparisons don't really paint the full picture because the set hardware of consoles allows for greater optimization since the programmers know exactly what the game is running on and playtest it on that hardware. It's similar to how you can't really compare phone specs to PC specs due to phones having to to deal with battery and heat more
says is true, in some games you need a 6700XT (2080Ti-ish) to match how the game runs on PS5. It really depends on how garbage the PC port is.
Nope, they just tested shit port of DMC5 which ran at some stupid internal resolution on PS5 even if you set the output to 1080p.
This is where the >PS5 GPU is just GTX 1060
meme come from.
Its a 6700 non xt this isnt a debate since thats literally what it is, Gaymers Nexus is dogshit and genuinely more deceptive than that cuck Linus
It’s a Navi 12 GPU with RT cores bolted on. It has the cores, TMUs and rops of a RX 5700 with a frequency boost. It’s closer to a 5700 XT in performance.
Not him. I installed Anal Wake 2 but I didn't bother to play it, game without any AA looks bad. There is just too much foliage and everything keeps shimmering all the time.
I bet hair will simply look like shit without blur coming from TAA.
>Is FSR the one that makes everything look blurry when the camera moves?
That would be TAA, FSR (and DLSS) try to make things better and sharper, but they're not that good at it, specially FSR.
Doesn't matter if you use FSR, TAA, DLSS textures will get blurred in the process. Those techniques hate sharp contrast.
MSAA would not work with amount of post processing effects modern games use. Static images don't show how blurry the image becomes when you move the camera, but the majority of homosexuals got used to awful image quality.
I could tweak the sharpness of the image with driver or reshade but it would only help a little.
based
the only reason to upgrade would be for RTX and raytracing is a meme so there really is no excuse for the shit tier optimization games have since as you say, games look like shit these days
you've got some mouth breathing morons clowning on you, because they miss the point
games today don't look good enough to warrant the hardware they require, and with most triple A games being absolute trash fires you don't really miss out either way
>Ganker is the normalgay board >bunch of normalgays pretending that consumers having 5 year old hardware is why the industry is behind
You're not only based for not being a moron who spends 200+ on games every year, but also making morons who thinks games would be good if everyone could run Starfield seethe
It's Capcom's first 9th gen only game
Gameplay footage showed a lot of physics, like foliage being affected by shockwaves
Dragon's Dogma ran extremely shit on target consoles
Dragon's Dogma was 30 fps even on PS4, when plenty of 7th gen remasters ran at 1080p 60 fps on PS4.
Capcom only shows requirements for 30 fps. Even for recommended settings, they use 2160i 30 fps as a baseline, instead of 1080p 60 fps like for Resident Evil games, including recent 4.
Monster Hunter World ran like absolute shit, which was Capcom's the only other open world game on "high end platforms".
Ryzen 3600, which is for 30 fps in this game, is similar in performance to consoles and it's still a relatively good CPU.
>Dragon's Dogma ran extremely shit on target consoles
I think people who haven't played it on either PS3 or 360 won't get just how bad it was. PC port gave DD second wind and is the version to play without a doubt.
>I think people who haven't played it on either PS3 or 360 won't get just how bad it was.
It ran like shit by today's standards but keep in mind this was a time when 24fps was considered acceptable
i mean shit i played all my 360 games on a crt via component and i had run that game through 3 times perfectly fine. People on console only just didn't care about fps standards back then
DD ran like shit. The PS3 version dipped as low as 20fps regularly. 360 didn't dip as low, but had screen tearing and pop-in. People argued over which was worse.
5 months ago
Anonymous
>20fps
Using magic cannon in bbi on ps3 would get you all the way down below 10 fps
It was 2 years before the game was announced and 1 year before capcom got hacked >I am in a position that allows me to know certain details regarding the next Dragon's Dogma.
>I am pleased to notee that some of you have kept track of what I've wrote in the past, but one post seems to have slipped through the cracks (maybe deleted or not posted perhaps in error). Anyway, I've updated it and am posting it here for you for better clarity
>Almost all of what I wrote last time is confirmed except for the end-game coop, it's still happening but downgraded to a lower priority (most likely as update post-launch). >Let's start from some technical info: there have been some hiccups in the last months. The team is more than satisfied with what has been achieved with the RE Engine, especially in the dungeons (magics are amazing but for now there is a very limited ray tracing, some sources are baked in a smart way, especially a new kind of fake tone mapping).However, problems have occurred in large open areas with a lot of trees/vegetation (some fights have literally brought dev-kits to their knees, even using extra power from upgraded pc stations). Currently the team is forced to choose whether to have more than 10+ enemies on screen or use sorceries with very complex animations like the tidal wave or 溶岩の海(?). The developers are still consulting with the RE guys occasionally, working to optimize the performance as much as possible.
i'm very worried the english won't have the charm of the first game going by what we've heard so far
capcoms insistence on movie type quality in games might really hurt the game
Capcom has had Denuvo and removing them a few months later to all their PC games for years. The fact that people still act shocked and appalled every single time is amazing. >Bro, the sun rose again today! This is fricking HORSE shit >Oh my God, it's setting a few hours later!
Because not everyone buy games on release? The last CAPCOM game i bought was RE2 and MH:W before that. None of those had Denuvo. I don't buy games with Denuvo. If people are buying it while it still has it then they should stop so we can erase this bloat software out of existence.
>RE2 and MH:W before that. None of those had Denuvo
they did though, they removed it later like they always do
you bought denuvo games and you didn't even realize proving it's irrelevant
Are you genuinely moronic?
I buy games i really want without Denuvo at full price. Examples being BG3 and Rogue Trader.
Capcom does not get the full price purchase. And I'm trying to be revolutionary here, i don't support Denuvo. And people that do are a blight.
5 months ago
Anonymous
you literally just said you bought RE2 and MHW, both games that released with denuvo
5 months ago
Anonymous
I didn't bought them when they had Denuvo. They were games without Denuvo and at discount when i bought them. Not hard to grasp.
5 months ago
Anonymous
you were talking about buying games new and blanket statement said RE2 and MHW didn't have denuvo as if they never did
5 months ago
Anonymous
homie, you bought games that had Denuvo at one point, meaning you fricking caved in after how much you talked up. You also act like if Capcom is devastated at the fact that you bought their games at a reduced price. They always put their games on sale, they don't care. Total units sold is what they're ultimately after.
The fact that you morons scream DENUVO!!! after everyone of their PC release is stupid. You know they support it. It has never been a secret.
Not really.
Games are essentially thousand and thousands of lines of code constantly executing. Denuvo is like a few extra lines of code. People can point at this and say SO IT AFFECTS PERFORMANCE, SEE?! but it's the same as throwing a stone in a river and saying it's affected the river's flow.
There were rare cases when it did hamper performance or loading times a few years ago, but 90% of the time whining about denuvo performance is just sour grapes cope from people who don't have the money.
7 years is basically fricking nothing, we've been sinking into diminishing returns for like a decade now. Only gaytracing makes any real difference (if implemented well), and DD2 doesn't use it.
>Minimum: >Requires a 64-bit processor and operating system >OS: Windows 10 (64 bit)/Windows 11 (64 bit)
REEE WHY WON'T IT RUN ON MY 15 YEAR OLD OS THIS IS BULLSHIT >Processor: Intel Core i5 10600 / AMD Ryzen 5 3600
REEEE MY 2500K IS PLENTY ENOUGH AVX IS A FRICKING HACK >Memory: 16 GB RAM
REEEE NOBODY NEEDS 16GB IN CURRENT YEAR. 4GB IS FINE FRICKING moronic CAPCOM >Graphics: NVIDIA GeForce GTX 1070 / AMD Radeon RX 5500 XT with 8GB VRAM
REEEEE VRAM IS A MEME MY 2014 GTX 970 STILL MAXES OUT EVERYTHING I SWEAR ON JENSEN THIS IS BULLSHIT
>people are still excited after the denuvo reveal
lol
lmao
there has not been a single game with denuvo that was good
you are as delusional as the people who thought starfield was going to be good
Yeah but this is Capcom, they always spoil everything about their games before release, especially something as big as playable classes
That'd be like having a secret weapon class in Monster Hunter or a secret character in Street Fighter, that's just not something they do
>itsuno has been adamant about wanting to make all the classes more distinct >to the point of removing knives and staves from magik archers, separating strider into thief and archer, etc.
taking this into account explain to me why he would decide to give maces(the only exclusive weapon to mystic knights) to fighters. you know that the only answer that makes sense is that mystic knight is gone.
What if MK gets something like magick swords and maces instead?
5 months ago
Anonymous
>what if it gets a new weapon
at that point it would be a different class entirety
5 months ago
Anonymous
Not really, magick shields and skills were his bread and butter keeping him special.
Having a special category of swords or maces that fit him better on the magic side of things would only make sense to differentiate him more.
Now if we see his cannon on some other class, that would be true confirmation in my eyes.
Because vocations share weapons that were previously vocation-distinct in DD1
"archer" is the bow class, it gets shortbows and longbows, and strider & ranger skills
It's not impossible that "fighter" is the sword + board class, and mystic knight returns as just "fighter" with a magick shield >I do think this is unlikely, because otherwise magick archer could have come under archer, but maybe you only progress to magick archer after levelling archer for a while or something
I want to be 100% unspoiled, but god I want to know how many vocations there are. I don't believe in the lmao9 posters, but I just need to know, truly.
I will feel so vindicated in just a few months. Patience.
I got a 7900xt for 750 upgraded from a 2070. Nvidia brand loyalty has really damaged the GPU market IMO, that and AMD drivers being shit until the 6000 series.
i dont play new releases but not very optimism toward capcom but my gtx 970 & i5 4690k yet to trip on any game i play vsync on ultra mod skyrim, destiny 2 etc 60-55fps mwh max setting 60-55 fps any optimized game worth one salt no perfomance issue
These threads always make me so happy.
It feels great to not be poor and listen to the desperate cries of the plebs, knowing that I will never have to deal with their struggles.
The recommended specs are still set for 30fps. Why the frick would someone go for 2160p before targeting 60fps?!
I'd rather go down one step to 720p before looking to play at 30fps. These specs tell me nothing.
>Why the frick would someone go for 2160p before targeting 60fps?!
Because consoles advertise as 4k capable, therefore that's the new "standard"
Why do you think all this "huur duur modern cards can't run shit" even started? It's because everyone now targets 4k 120fps.
>Why the frick would someone go for 2160p before targeting 60fps?! >These specs tell me nothing.
Because the CPU is an issue here, not GPU.
You people got too spoiled by 8th gen that you can't comprehend CPU bottleneck, now that consoles actually have competitive CPU.
>CPU is the bottleneck, so we won't list higher CPU requirements for the recommended
It doesn't make any fricking sense. Same for their low GPU requirement
>But muh 4k
You know exactly why. Cucksoles will run at those settings more or less. But it's strange because every RE Engine game I've tried ran rather well. And this game isn't even using the new engine.
They had no issues listing 1080p 60 fps as recommended target for Resident Evil 4.
It's clear that CPU is the bottleneck. >But it's strange because every RE Engine game I've tried ran rather well.
All of them were PS4 games. Even Resident Evil 4, despite what that one guy claims.
Dragon's Dogma 2 is Capcom's first 9th gen only game and it's clear they aim at 30 fps.
But it is a cpu limitation. Why would you guys think Dragon's Dogma 2 would not thrash the cpu then gpu when its trying to be an extremely dynamic combat sandbox type-ass game.
Ryzen 3600 owner here, Ready or Not ran like shit on the CPU side, even on the lowest settings, when GPU was running at like 50% utilization.
Any milsim wannabe Squad clone (not that I like them, to the contrary, I'd rather Red Orchestra clone than Squad clone) is guaranteed to run like shit with a lot of players.
I'm planning on building new PC specifically on the CPU side, hopefully 8800X3D, while waiting for proper RTX 5000 series.
5 months ago
Anonymous
I didn't play it but it sounds like bad CPU utilization with reliance on single thread.
There is difference between high requirements and just trash optimization. But for high player count shooters CPU performance was always bad for sure.
Dragons dogma is also open world and theyve gotten rid of traversal loading screens. Stop ignoring the March of time and progress and upgrade
5 months ago
Anonymous
R5 3600 is plenty new for a cpu
5 months ago
Anonymous
I'm on 5600 and I'm waiting for Zen5X3D, I don't want to upgrade every gen for 25% uplift.
I would rather have loading screens maybe not as many as in scatfield lol than constant traversal stutter.
JS's CPU hunger is insane but the #800x3d chips run that game very well. I stuttered like hell on my r3600 but on a 5800x3d it's smooth sailing.
5 months ago
Anonymous
It stutters even on 7800X3D, there are certain levels where performance drops for no reason. This is mostly the case if you enable RT.
5 months ago
Anonymous
Shader compilation stutters and no amount of hardware can be thrown at it to overcome the problem.
5 months ago
Anonymous
Feels good to not play or buy EA published garbage.
5 months ago
Anonymous
I'm on 5800x3d + 4070 and the only place that kind of stutters to me is the dust planet's village and the droid camp near it. >RT
It's not worth using. The reflections and RTGI is very barebones and ununoptimized. It looks like 3% better, the baseline graphics are already good enough.
No it's not, it's a sort of checkerboard rendering method. It renders something like every second line of 2 frames and blends them together. It's basically 2k.
Who the frick even uses interlaced these days!?
It's Capcom's own checkerboard rendering tech for RE Engine. I don't mean they invented interlacing or whatever, but it's one of their display options, not unlike a DLSS/FSR toggle.
>Do people don't look up game settings nowadays?
I wouldn't expect a modern game from a reasonable sized developer to use a display method not used for over a decade.
Dude, how is it not higher CPU load to run 60FPS than to run 30FPS?
CPU load doesn't change much if you change the resolution and keep the same framerate.
5 months ago
Anonymous
>how is it not higher CPU load to run 60FPS than to run 30FPS?
Because the CPU can already deliver those frames, but the GPU is limiting it from doing so.
Yes, technically you will se a slightly larger load on the CPU, but the point is that the 3600 is more than enough for the task, which is why only the GPU requirement changes.
5 months ago
Anonymous
You don't know that.
This game can do stable 30FPS on Cucksole CPU but it can't hold 60FPS for example.
R5 3600 is nearly equal to current gen CPUs.
If they targeted 30FPS from the start they didn't bother to optimize CPU performance.
If you don't have something over a 1070 or a 9th gen console that's kind of on you at that point. Probably gonna opt for console anyways since PC releases seem to be more shit than ever at launch now.
?t=23
Basically the 2160i is only 1080p pixel count, but half the framerate, which explains the steeper GPU requirement while having no real affect on the CPU.
Wow, rude, homosexual.
Could've just said that 2160p is 4 times 1080p, but due to it only drawing half the pixels on the Y axis, you still need to draw the full X.
Anyway, point still stands.
Higher GPU to do that, same CPU because the 3600 can handle it just fine.
>Higher GPU to do that, same CPU because the 3600 can handle it just fine.
This is just your assumption without any evidence.
You don't know how that game runs, it might be awfully optimized right now.
Before game launches it's all speculation.
5 months ago
Anonymous
>This is just your assumption without any evidence.
The evidence being the CPU is unchanged between minimum and recommended.
You keep panicking if you want though, not my problem.
Still should be double the 1080p pixel count, since a true 2160p is 4 times that.
5 months ago
Anonymous
Also illustrates how much of a meme 4K is.
1440p already looks way better than 1080p, but it's only 1.6 times the pixel count.
You start hitting massive diminishing returns on 4k.
5 months ago
Anonymous
This is assuming you can stomach no anti-aliasing at lower resolutions than 4k. Have you played any AAA game in the past 6 years? They all use TAA which will blur the frick out of 1080p and to a lesser degree 1440p. At 4k TAA is only a boon to clean up remaining edges
5 months ago
Anonymous
Is it though?
Isn’t it just two alternating 1920x1080p checkerboard patterns?
5 months ago
Anonymous
You need to remember that 4k is basically 4 1920x1080p screens.
So it's 4 screens alternating, so 2 normal 1080p screens.
The 10gb 3080 sucks. Get the 12gb if anything. That said, they're really expensive now compared to a little over a year ago, so you're better off getting a 4070 or 4070ti.
>singleplayer MMO
I will be skipping and hoping that anons get rused by marketers into buying the game day 1. Nothing more amusing than seeing hundreds of cope and sneed posts over some easily-visible slop.
7800xt or a rtx 4070? I've never used AMD before but those are about as much as I'm willing to spend. No I won't just wait until 2025 for nvidia do further price gouging, my 1070 is already outdated.
Black person, wait until February nGreedia is refreshing some of the cards.
4070 might get slightly cheaper or maybe 4070Super won't be any more expensive doubt that
You don't have to buy right now.
Whatever you can get cheaper honestly.
Do you want RT on? 4070
What resolution are you going to play at? Are you willing to risk it that 12GB will be enough for 1440p for the next 5-6 years, judging by you sitting on the 1070 still. Probably going to be enough for 1080p, probably.
I got a 7800XT for 7700XT money, so I don't think I could've gotten a better deal.
Anons don';t play games, this is a known fact. Being positive about a game means either potentially playing it (and as established, that is verboten) or being a corpo shill (likely unpaid). Now NOT talking about vidya is inconceivable so that leaves being a raging doom poster.
we've seen bridges collpsing, boulders rolling down hills, bombs blowing open creeks causing a rushing river to knock over a large foe etc.
besdies, Dragon's Dogma 2 is the first RE Engine title that isn't releasing on last gen consoles, so obviously the hardware floor requirment will be increased on PC
>speculation
You're a moron. They've said it and shown it multiple times. Look at this webm. The fricking table shakes when he lands near it. There are even shockwaves/impacts from physics objects colliding near each other.
It'll run fine on modern hardware. This game is not targeting a 2013 console OR your PC hardware from like 6 years ago because the PS5 didn't exist yet
damn, that's crazy given how heaps of PS5 games have gaytracing which would utterly cripple a 1070
truly fascinating >posts single slide from a single Gamers Nexus video in regard to a single game
wow!
>buy 7800x3d for this >didn't check cpu socket prior to install >install in the same manner as all other systems I've built (extremely slowly and gently lower the cpu into the socket in the keyed direction) >gently confirm it's in place by lightly moving it in place to confirm it's in propertly >latch it >no display out >rebuild like twice >still nothing >clean it all up and put parts in box >notice like 3 bent pins on oppositing corners of the motherboard
There's no way I bent them installing it in the correct orientation in the same manner as I have the other half a dozen systems I've built, right?
Left side 1/4 up-ish and top right-ish and 1/4 up from the bottom in the middleish on the right side. Nah I'm Australian, but I'm still fricked and it's an out of warranty repair. I was quoted 50-60 bucks or 110 if they replace the entire socket. Beat a $450 motherboard but it's shit.
If they sold you faulty product you should get your full refund, it's not like motherboards can't be DOA. Ausies have strong consumer protection laws if I remember correctly.
The issue is I have no proof and the only anecdote is after installing it and removing it the pins were damage. So the blame is on me even if I didn't actually do it wrong. It's kinda shitty but whatever.
Then just lie and say it arrived like that. What are you, moronic? Why would you be honest about that? The retailer is marking this shit up a ridiculous amount, they won't feel a refund while you will definitely feel a new mobo in your bank account.
are you really not meant to touch it at all? The pressure would've been even less than spreading thermal paste on it which you can do with a spatula or attaching the cpu cooler
I know, it's just weird. I know intel does that too now.
CPU pins somehow seem more durable, you can even manually bend them back and they will work just fine. I dunno, maybe I'm talking out of my ass.
it's keyed and can only go in one way, it has the arrow in the corner too, it's basically the same as intel sockets, which are the 7 systems i've built in the exact same manner previously
Yeah, and RE4 was the first time I actually thought that FSR2 Performance (at 4K) looks good. In earlier games I had to use at least Balanced not to look glitchy in motion.
And yeah, later I also switched to DLSS.
Meanwhile interlaced gives less stable image in motion.
Apparently the best implementation of FSR2 is... No Man's Sky for Switch. Hello Games took the original code of FSR and improved it, and now it works very well despite the extremely low internal resolution. See: https://youtu.be/sbiXpDmJq14?feature=shared&t=107
Why the frick it doesn't work like that by default, only AMD knows. But at least it's safe to say there's ample room for improvement, fingers crossed for FSR 3.5 or whatever they are going to call it.
The code AMD provides is open-source. So anyone can take it and do whatever they want with it (Nintendo used FSR1 in TotK for example. FSR1 has also been implemented is a few emulators. A few FSR3 mods have been popping up recently, and they work on both AMD and Nvidia cards).
But if the open-source license -requires- those developers to share their modifications, I have no idea. Probably not?
>needs a 1070 just to target 30fps *with drops* at 1080p
That is a 7 year old graphics card. The fact it's still rated to run this game at 1080p is something you should be grateful for.
I love that they show shit like this, but in game you would literally never see them as pawns or players. I want black people to play the cool game too and make what they would empathize with, but when they keep putting them everywhere in promo material it is forced as frick.
I'm still on a 1070, but I'm gonna upgrade in the next month or 2 when the nvidia 40 super series gets announced, not that I mean I'll definitely be buying one of those, but I'm waiting to see what the price and spec improvements are, and if they aren't good enough I'll probably just call it a day and buy a 7800xt or maybe a 4070 if they drop the price of those as part of the announcement, just so I can have better RT support and DLSS
I'm hoping the announcement just causes prices of other GPU's to drop to stay competitive
In a perfect world the 4070STi WRX would cost 500$-600$ max. It's what the 4070 should've been from the start.
Unfortunately it will probably cost more like 800 if not more.
This thread has prompted me to go back to dragons dogma. Seems my gamesave has vanished so frick it, starting from scratch again. Surely the gamer won't turn into a horrible grind amirite?
if you have to grind at all in dogma you're playing the game like a moron
I've seen some people complain about doing no damage due to the flat damage scaling (which IS stupid) but literally just buy and 1 star enhance new armor and weapons when you can. It's not that difficult.
However you will need to gear yourself and the pawn more efficiently when you tackle BBI, but before you do that at least buy some buff periapts from Fournival, please
Not in this generation, but you could for some time during 8th gen, there's no way these days to get something cheaper or even at the same price that matches a console performance.
A little bit of everything, the inflation it's also being used as a reason, but mostly the mining shit that ruined GPU prices for quite some time, and it's only now that we're slowly recovering.
Honestly most components are in a good place, it's just GPUs that are bit overpriced.
5 months ago
Anonymous
Chinks are now bulk buying GPUs to remove the chips and hook them up to AI model training rigs. The crypto shit has died down but it's simply been replaced.
>Scalpers? Miners?
Basically yeah.
A few years back cryptogays figured out that graphics cards have good technology for cryptomining. Once they realized they can turn a profit by sticking 20 GPUs into a single farm to speed up the process, a lot of people started trying this and all the GPUs were bought out and put to crypto work. So if you wanted to buy a graphics card you needed to shell out 2-3x times the normal price.
Then the worst possible thing happened - instead of saying "frick this shit, i'll buy once prices go back to normal" most people just... kept buying. At those ridiculous prices. That did immeasurable damage to the market because it essentially told GPU makers and vendors that they don't need to price their products sensibly, we'll buy anything no matter the price.
So here we are. The threat of "we won't buy your graphics cards if they're poor value" has been proven to be bark with 0 bite behind it, and now GPU vendors can suck us dry with insulting prices because they're fully aware that won't stop people from buying.
but then why would anyone want console performance at all let alone console as a platform at all?
unreal engine and nvidia are colluding to make game run like shit on consoles because consoles have amd gpus in them
this would be a great lawsuit to have start up because it's extremely anti-competition
The DDA port is excellent, and recent CapCom games have all been fine on PC, including DmC V. I see no reason for the DD2 PC version to be any different besides maybe Denuvo eating frames.
Anything above a r5600 will be enough. Denuvo doesn't impact performance nearly as much as people would have you believe. The FPS and frametime graph between the Doom Eternal I pirated and the Doom Eternal I later bought was the exact same. Only way it could reduce your performance is if you have an old ass computer.
I hate you morons so much. I'm not saying you should be doing that. I'm saying you should stop complaining your gpu and cpu that is older than even the ps4 pro is not adequate enough to run current generation games. Weren't you homosexuals shitting on the lack of current gen console exclusives because of last gen console versions and PC ports? Well now we've got current gen only games finally coming out and NOW the sour grapes are here right on time.
Stop being poor and upgrade or don't complain.
is this place really so poorgay infested that people are legit seething their decade old hardware can't stuff at 1080p60fps anymore? get a job you worms or stick to older games
It's a picture from the first game, and the II can be easily missed from the thumbnail.
Is that your excuse for writing the game's title wrong?
You stupid frick.
It isn't you blind moron, it's from a trailer form the second game. Nobody else ITT made the same mistake you did, you're completely at fault and a fricking idiot.
>buy a 4070 >almost all the new games that have any decent use for a potent gpu dont actually look better than many games released almost a decade ago and just have shit optimization >morons defend it because they can get 1440p@60fps / 1080p@120fps on their gpu that should have been capable of 1440p@160+fps
why are games so gay and stupid
I am using a 3070_ and if DD2 runs like shit I will walk out and buy a better card.
Literally the only game in a half-decade I will be willing to do this for. Capcom, don't let me down please.
I amassed a ton of cash and spent it all on portcrystal forgeries.
every time i play i spend some tim in the rift gifting a ton of pawns across all sorts of levels a portcrystal forgery, mostly the lower level ones
i am extremely amused by the thought of someone carrying a bunch of heavy portcrystals around only to be confused and then realise these cant actually be used
i have a computer with a 7900xtx, bought my brother one with an rtx4090 for his graduation
everything runs smooth like butter
the sad sacks who still run gtx970s and whine about how games are all bad now (because they can't run them at playable settings) are worse than consoleBlack folk in terms of holding technology back
dont worry jensen specifically paid for monster hunter wilds. for whatever reason monster hunter favors nvidia. rise has dlss and mhw is gameworks approved.
>Is high framerate fricking with the physics a normal thing?
It's very, very common. It also fricks with enemy AI a lot. In MH Rise, high framerates would break some cosmetic physics stuff (your dog's tail would spin around like a propeller for example), create wonky interactions with knockback from attacks, and increased monster attack tracking and aggression. MH World had framerates affecting attack speed, hitstop, tick rate of multi-hit attacks and the fire rate of guns.
VERY common.
Tieing physics to fps is very common dev practice and is specially common with japanese devs
From software games for example have physics and game mechanics tied to FPS and depended on community mods to "fix" that kind of thing.
In Unity terms, there's Update, which is a single frame and is supposed to be used for game controls, and Fixed Update, which can be described as "physics frame", as that's what it's mainly supposed to be used for, or game logic.
Fixed Update is executed every fixed amount of time. It must be executed even if actual frame wasn't finished. In Source engine terms, fixed update would be associated with tick rate.
Tick rate 66 means there are 66 logic/physics refreshes every second. The reason it can work just fine with high refresh rates is that values are interpolated between frames. So yeah, what you're seeing is actually the past, but it works well enough.
The games that break physics with high FPS are a rarity nowadays. Even Bethesda games play along with high framerate well now. It's only above 300+ fps that things go haywire but why would you ever have 300+fps?
How old is your rig? Apart from my GPU and an additional SSD I installed my old shitbox is over 10 years old. Served me pretty well up until now, hope I can get the same millage out of my new build.
Went back to my years old DD:DI save and have been farming Daimon for lv3 assassin and mk gear. At the point where I just speedrun him with periapts and blast arrows, but goddamn the AI was occasionally RNG when I tried to kill him with Thousand Kisses and periapt because sometimes Daimon would just stand around doing nothing while other times he would constantly do his flame blast and lighting if I tried to climb his face.
Already knew when it was announced and no footage was yet released that my shitbox PC wasn't gonna cut it, thankfully DD is meant to be played with a controller so I'll just play it on xbox.
>Suddenly everyone is doomposting and calling Dragons Dogma of all things shit.
I don't pay attention to release dates anymore. I can always tell something is coming out soon when Ganker starts pissing their pants about it.
>everyone is doomposting
it's one guy, mostly pretending to be a discord personal army but in actuality is just samegayging to create a siege mentality among DDfrens and keep the threads going until release
>muh raytracing bad
I wonder if any zoomers remember how changing to 32bit colors tanked your performance. Guess that was a meme too and all games should go back to 16 bit.
I'll have you know I play with the Z buffer disabled as it is a scam feature with disproportionate performance cost for virtually NO increase in visual quality.
The performance and the fact that devs don't usually build a game's visuals around it. That said, something like Metro Enhanced looks great with ray tracing because it was made for it specifically, and the game runs pretty well.
It's a meme because the average person still has a gtx 1060 that'll shit itself if it tries to raytrace. It's bad and a gimmick because it's not accessible. Just like FG was bad and a gimmick back when a 4090 was the only card that could do it, but now that FSR3 does the same shit and a lot more cards do it and a mod exists to make the 2000 series and 3000 series do it too, now suddenly it's not bad anymore. Tech opinions are mainly driven by sour grapes mindset.
lol the 2160i is actually 1920x2160 temporally upscaled to 3840x2160 using crapcom checkboard rendering
I guess they didn't want to scare morons off by putting actual 4k requirement in, kek
>My 8 year old purchase should give me 1080/60 at max settings for AT LEAST 10 years.
imagine if you tried saying this shit in the late 90s or early 2000s
you homies have no clue how quickly hardware got outdated back in the day, nowadays it's much better
Are you all zoomers? This has literally never been true at any point in PC gaming history. Hell it was arguably worse in the past since PC gaming has stagnated a lot. The 2000s felt like everything was obsolete within 1-2 years.
>Are you all zoomers?
Probably all third world zoomers. Remember when Crysis came out? Everyone was salivating over it and getting excited for the future. No one was b***hing and moaning that their mid-tier PC couldn't play it at max settings.
>mfw I paid 479.99 for a 1070 when it came out and bought a 6700k shortly there after, when I used to upgrade every year on GPUs >mfw thinking I'd never have to upgrade again
Guess it's only indieshit and old games for me cause I'm gonna push a decade on this thing.
I got too comfortable gaming on PC when we were still getting 360 ports.
I don't care about gaytracing or pore technology or one billion polygon rocks or volumetric fart clouds, none of that trash does SHIT for the actual game.
I intentionally play everything at low-medium settings with all post-processing turned off, my GPU fans don't even bother spinning for how little I push it.
I think PS5 Pro will be pointless this gen. It'll still rely on upscaling and if the game won't be 60 fps on regular PS5, it won't be 60 fps on Pro either.
its the most honest system requirements of any pc game right now. minimum requirements is re4r's recommended. recommended specs are literally a ps5. steam page even says it recommends controller over kb+m because the devs KNOW.
I'm in the exact same boat. It's frustrating because going from 3070 to 4070 seems idiotic, but waitgays would have you wait forever.
I have the money but that doesn't mean I want to be wasteful for little benefit in everything.
>3070
You might want to hold on... I mean that's not a slow card, its only problem is the small amount of VRAM, until that really gives you problems I think you can probably use it for a while still
I think the only problem I have is that I can't play most modern games at 4K unless I turn on DLSS. Not something I care about really, if I'm being honest 2K may be just enough for me, and I barely notice the jump from 2K to 4K.
To be fair the 3070 was barely a 1440p card when it released, nowadays it's a 1080p card, if you've got a 4k monitor and want to do native 4k, you're gonna have to get something much faster (and expensive)
5 months ago
Anonymous
>nowadays it's a 1080p card
really? I don't play many modern games, but the ones I play can be run at 2K with no issues. What games have you made drop the resolution to 1080? I can't see myself going back to that. >if you've got a 4k monitor
2K monitor, 4K TV. The TV being placed kinda far from me might be the reason why I think 2K is just fine.
5 months ago
Anonymous
you can still use a 3070 as a 1440p card
there's one simple trick you can do
it's a top secret
doctors hate this one simple trick!
but don't tell anyone i told you
ready?
here it comes: lower the game's texture size from ultra to high. -33% VRAM usage for no visual impact that's all you need to do to combat the supposed VRAM problem of 8-10gb cards i've yet to see anyone point me to a game where textures look demonstrably inferior when reducing them from ultra to high
5 months ago
Anonymous
Usually going from Ultra to High means you'll be just load fewer HQ textures at a any given time, so you might lower res textures for a bit until the higher res textures load in, it's not ideal, but yes, but it's not a big deal either...
5 months ago
Anonymous
This. I'm currently playing Cyberpunk on high settings 1440p, no DLSS and ray tracing though. Constant 60FPS.
its a great gpu but the pricing is just AWFUL. nvidia can only get away with it because their fanboys eat wiener for breakfast and you have pc building noobs who fell for the "AMD SUCKS" meme despite using AMD CPUs that are more fickle than GPUs. nearly every 7950x3d fails sha3.
I have a b450 motherboard with 2060 and 2600x. Should I just upgrade my gpu and cpu for the current motherboard, or get a new motherboard for future proofing? I can afford it, it's just the only games i'm looking at are dd2 and wilds
Why are you expecting to play the latest AAA game with an 8 year old card and zero hiccups?
This was never doable, even in 2016 when this shit was brand new.
Destruction Derby 2? I'm sure it could run just fine. Who knows what the frick you're trying to say since you're too lazy to type out the title of the game.
minimum is a 1070
recommended is PS5 specs
if you want raytracing then you need beefier specs
then i'll just buy the PS5 version i guess
>gaytracing
Hot take anon! Keep it up!
Maybe throw in a reddit next time, or trannie tracing
I'm not against Ray Tracing in principle, but Capcom's implementation so far has been shitty and pointless.for the most part.
Mostly reflections on marble, which DD2 won't even have for 99.9% of environments, too.
tbh most RT implementations in most games are shit, and that will never change because consoles are the primary market and consoles will never play well with RT
there's like a maximum of 5 games in total where you can look at RT and say "yeah that's worth turning on"
Full on pathtracing looks good in Metro and Cyberpunk, and Control also had good enough reflections, simply because most of the game was glass or shiny surfaces.
Obviously it only makes sense if you have a top of the line GPU and you have extra FPS to spare anyway.
I will categorically refuse to buy any game that my 1070 can’t run. Graphics have not improved significantly in the past six years or so.
Buy a new gpu already, goddamn.
>consoles are holding pc gaming back!!
They are though
if they werent you wouldn't be able to run shit on that old ass garbage GPU, and you'd still whine and moan, and the devs would have to optmize it for people like you.
>Optimizing the game is a bad thing
>games are being held back by consoles because devs have to make the games run on weak hardware
>but my weaker than a PS5 gpu should run it!
You make zero sense, you pick one or the other, if they go to town with the graphics and dont hold themselves back because of a PS5, your shit GPU wouldnt be able to handle it either, if you could, the ps5 could, and the games would have been held back.
Except no one would cry about a game pushing graphical boundries not being able to be played on 7 year old hardware.
You obviously missed the cyberpunk threads.
you can't fool me. that's driver.
No it isn't. It is cyberpunk.
cyberpunk looks very SOVLFVL
no, thats midtown madness 2 you fricking moron
Holy shit. Cyberpunk at max settings looks like THAT?
PCbros, what's your excuse for that?
Read the filename, this is clearly PS4 capture :^)
They did with Alan Wake 2 also, recommended specs are a 4090, and the game looks horrible on the PS5 so they clearly don't care.
>if they werent you wouldn't be able to run shit on that old ass garbage GPU
Consoles seem to do so just fine.
and you can also categorically frick off
the 1070 was a shit card even at release, its trash throw it away and then throw yourself away because you're an idiot clinging to hardware like a console tard
>the 1070 was a shit card even at release
Wasn't it literally the second best? 1070, then 1080. what else was there?
The 1080 was shit too. Anybody with half a brain bought a 1080 Ti that generation and skipped the 2000 series beta test entirely.
The 1080Ti came out a year later. The 1080 was garbage for the price, but the 1060 and 1070 were pretty good... at MSRP.
The RX 480 was a 1060 competitor, either one was fine.
zoomers on this board weren't even old enough to buy PC hardware in 2016 let alone 2010. They don't remember we were even having shortages back then to miners in 2016 because they'd buy all the GPUs the second they became available. Because of the miners I was able to sell my R9 280x in late 2017 for 20 dollars MORE than I bought it new in 2014. It looks like that train is long gone now though.
You do realize the irony of your post yeah?
>Using a 7 year old upper midrange gpu
...Anon... I-
According to Gaymer Nexus, PS5 equivalent GPU is in between GTX 1060 to GTX 1080 while Digital Foundry claims that it's RTX 2070 S.
I trust the former so you should do fine.
Nope, they just tested shit port of DMC5 which ran at some stupid internal resolution on PS5 even if you set the output to 1080p.
This is where the
>PS5 GPU is just GTX 1060
meme come from.
>According to Gaymer Nexus, PS5 equivalent GPU is in between GTX 1060 to GTX 1080 while Digital Foundry claims that it's RTX 2070 S.
Direct hardware comparisons don't really paint the full picture because the set hardware of consoles allows for greater optimization since the programmers know exactly what the game is running on and playtest it on that hardware. It's similar to how you can't really compare phone specs to PC specs due to phones having to to deal with battery and heat more
It's a 6600XT-equivalent, so 2070-ish is about right.
But also what
says is true, in some games you need a 6700XT (2080Ti-ish) to match how the game runs on PS5. It really depends on how garbage the PC port is.
Its a 6700 non xt this isnt a debate since thats literally what it is, Gaymers Nexus is dogshit and genuinely more deceptive than that cuck Linus
It’s a Navi 12 GPU with RT cores bolted on. It has the cores, TMUs and rops of a RX 5700 with a frequency boost. It’s closer to a 5700 XT in performance.
Why did you upgrade to the 1070?
Why didn't you stop playing games your 8800GT couldn't run?
Why go that far even? Why not stop at GeForce 4?
>le futureproofing tard
What do you mean, hair looks way better now.
>hair looks way better now
Someone post the Alan Wake 2 hair with TAA off.
Not him. I installed Anal Wake 2 but I didn't bother to play it, game without any AA looks bad. There is just too much foliage and everything keeps shimmering all the time.
I bet hair will simply look like shit without blur coming from TAA.
Is FSR the one that makes everything look blurry when the camera moves? Games dropping MSAA was a mistake.
>Is FSR the one that makes everything look blurry when the camera moves?
That would be TAA, FSR (and DLSS) try to make things better and sharper, but they're not that good at it, specially FSR.
Doesn't matter if you use FSR, TAA, DLSS textures will get blurred in the process. Those techniques hate sharp contrast.
MSAA would not work with amount of post processing effects modern games use. Static images don't show how blurry the image becomes when you move the camera, but the majority of homosexuals got used to awful image quality.
I could tweak the sharpness of the image with driver or reshade but it would only help a little.
homie cyberpunk can have full path travel lighting frick you dumbass
I was like this last year with my 980Ti. After upgrading this year to a 3080Ti I'm kicking myself for holding off for so long
based
the only reason to upgrade would be for RTX and raytracing is a meme so there really is no excuse for the shit tier optimization games have since as you say, games look like shit these days
you've got some mouth breathing morons clowning on you, because they miss the point
games today don't look good enough to warrant the hardware they require, and with most triple A games being absolute trash fires you don't really miss out either way
not him but I own a 3080 and he's right
sure you do
do tell, what games have released in the past 3 years that would require a GPU more powerful than the 1070 that is actually WORTH PLAYING?
any game of a genre you dont even play that wants to run better than a xbox one x you moron since that console is literally a 1070.
Not an answer.
good job dodging the question instead of outing your shit taste in video games :^)
Good job being a poorgay with 7 years old xbox one x
none of them of course
no game after [moment my graphics card stopped performing well] is worth playing 🙂
Why do pc child’s insist consoles are holding back the industry, then never want to upgrade their shitty 970, all the while blaming bad optimisation?
>Ganker is the normalgay board
>bunch of normalgays pretending that consumers having 5 year old hardware is why the industry is behind
You're not only based for not being a moron who spends 200+ on games every year, but also making morons who thinks games would be good if everyone could run Starfield seethe
If you want games to do more then you need better hardware. Good luck running Google maps on your flip phone, troglodyte
That's fine, but keep in mind it's about to be an 8 year old card. That's like expecting a GTX 260 to be able to play Doom 2016
>mfw it can't
>NOOOOO WHY IS MY ALMOST 8 YEAR OLD CARD FALLING OFF!!!
Are people not embarrassed saying shit like this?
>I refuse to buy games I can't play
Yeah no fricking shit, moron.
meanwhile I'm playing vermintide 2 on my gtx 750 ti and I will never upgrade because that's the only game I need to play
I only care about booba so your post pleases me.
used 3060ti (not mined) is $300 you fricking Black person..
I'd get an RX 6700 XT for that $300 instead.
>amd
go get it idiot
Good goy, make sure you get a GeForce RTX™ video card for the optimal gaming experience, powered by amazing RT effects and DLSS.
the game is mostly CPU limited
what makes you say that?
It's Capcom's first 9th gen only game
Gameplay footage showed a lot of physics, like foliage being affected by shockwaves
Dragon's Dogma ran extremely shit on target consoles
Dragon's Dogma was 30 fps even on PS4, when plenty of 7th gen remasters ran at 1080p 60 fps on PS4.
Capcom only shows requirements for 30 fps. Even for recommended settings, they use 2160i 30 fps as a baseline, instead of 1080p 60 fps like for Resident Evil games, including recent 4.
Monster Hunter World ran like absolute shit, which was Capcom's the only other open world game on "high end platforms".
Ryzen 3600, which is for 30 fps in this game, is similar in performance to consoles and it's still a relatively good CPU.
>Dragon's Dogma ran extremely shit on target consoles
I think people who haven't played it on either PS3 or 360 won't get just how bad it was. PC port gave DD second wind and is the version to play without a doubt.
>I think people who haven't played it on either PS3 or 360 won't get just how bad it was.
It ran like shit by today's standards but keep in mind this was a time when 24fps was considered acceptable
i mean shit i played all my 360 games on a crt via component and i had run that game through 3 times perfectly fine. People on console only just didn't care about fps standards back then
DD ran like shit. The PS3 version dipped as low as 20fps regularly. 360 didn't dip as low, but had screen tearing and pop-in. People argued over which was worse.
>20fps
Using magic cannon in bbi on ps3 would get you all the way down below 10 fps
the game is getting an amd ryzen bundle instead of radeon graphics bundle.
Ganker leak before capcom leaks said the developers had to choose between less enemies on screen or worse physics so yeh cpu bound.
>Ganker leak
It was 2 years before the game was announced and 1 year before capcom got hacked
>I am in a position that allows me to know certain details regarding the next Dragon's Dogma.
>I am pleased to notee that some of you have kept track of what I've wrote in the past, but one post seems to have slipped through the cracks (maybe deleted or not posted perhaps in error). Anyway, I've updated it and am posting it here for you for better clarity
>Almost all of what I wrote last time is confirmed except for the end-game coop, it's still happening but downgraded to a lower priority (most likely as update post-launch).
>Let's start from some technical info: there have been some hiccups in the last months. The team is more than satisfied with what has been achieved with the RE Engine, especially in the dungeons (magics are amazing but for now there is a very limited ray tracing, some sources are baked in a smart way, especially a new kind of fake tone mapping).However, problems have occurred in large open areas with a lot of trees/vegetation (some fights have literally brought dev-kits to their knees, even using extra power from upgraded pc stations). Currently the team is forced to choose whether to have more than 10+ enemies on screen or use sorceries with very complex animations like the tidal wave or 溶岩の海(?). The developers are still consulting with the RE guys occasionally, working to optimize the performance as much as possible.
>Ganker leak
You mean the "leak" that was complete bullshit?
i'm very worried the english won't have the charm of the first game going by what we've heard so far
capcoms insistence on movie type quality in games might really hurt the game
>7 year old gpu
I forget that the majority of the master race are slavs and brown people.
>I forget that the majority of the master race are slavs and brown people.
Half the world is either Indian, African, or Chinese.
Maybe the world ending wouldn't be so bad after all.
>tfw 4080
not my problem
>also denuvo
Frick crapcom.
does denuvo even affect the performance anymore?
No, it's brown subhumans salty they can't pirate it unless the schizo cracks it.
You're a massive goyim
Hopefully not, but i'm still skeptical about it
Yeah, you'll especially see it if games are CPU-bound because it's eating the cpu to function.
Capcom has had Denuvo and removing them a few months later to all their PC games for years. The fact that people still act shocked and appalled every single time is amazing.
>Bro, the sun rose again today! This is fricking HORSE shit
>Oh my God, it's setting a few hours later!
Because not everyone buy games on release? The last CAPCOM game i bought was RE2 and MH:W before that. None of those had Denuvo. I don't buy games with Denuvo. If people are buying it while it still has it then they should stop so we can erase this bloat software out of existence.
>RE2 and MH:W before that. None of those had Denuvo
they did though, they removed it later like they always do
you bought denuvo games and you didn't even realize proving it's irrelevant
Are you genuinely moronic?
I buy games i really want without Denuvo at full price. Examples being BG3 and Rogue Trader.
Capcom does not get the full price purchase. And I'm trying to be revolutionary here, i don't support Denuvo. And people that do are a blight.
you literally just said you bought RE2 and MHW, both games that released with denuvo
I didn't bought them when they had Denuvo. They were games without Denuvo and at discount when i bought them. Not hard to grasp.
you were talking about buying games new and blanket statement said RE2 and MHW didn't have denuvo as if they never did
homie, you bought games that had Denuvo at one point, meaning you fricking caved in after how much you talked up. You also act like if Capcom is devastated at the fact that you bought their games at a reduced price. They always put their games on sale, they don't care. Total units sold is what they're ultimately after.
The fact that you morons scream DENUVO!!! after everyone of their PC release is stupid. You know they support it. It has never been a secret.
No problem then
Yes and that's a problem it can cut between 5 and 10 fps
Not really.
Games are essentially thousand and thousands of lines of code constantly executing. Denuvo is like a few extra lines of code. People can point at this and say SO IT AFFECTS PERFORMANCE, SEE?! but it's the same as throwing a stone in a river and saying it's affected the river's flow.
There were rare cases when it did hamper performance or loading times a few years ago, but 90% of the time whining about denuvo performance is just sour grapes cope from people who don't have the money.
In my experience, no. But gays will cry about it anyway.
>Denuvo
Shan't be playing
>not having at least a 2060
ngmi baka
you can get 4060 for like 300 bucks
the sys reqs look very forgiving, sounds too good to be true after the absolute state of aaa this year
>Needs a 1070 for 1080p with drops
Has me more worried that the game is going to look like trash, reminder that a 1070 is a 7 year old card
7 years is basically fricking nothing, we've been sinking into diminishing returns for like a decade now. Only gaytracing makes any real difference (if implemented well), and DD2 doesn't use it.
there are hardware suggestions for rt on the dd2 game page
Right. But it's listed under additional notes with 2080Ti for both minimum and recommended, so I doubt it's going to be used globally.
you can already see what the game looks like, an upgraded dragon's dogma which is all we needed
Will it even run on Deck?
Doubt it, but hopefully it gets decent performance on the Ally
you should be happy, it'll look just like any anime and that spidervers shit
>mfw 1070 dates to just 2016
No way. I thought it was way over ten years.
>My PS2 specs can't run this PS5 game
FRICKING MONKEY DEVS REEEEE
Here is hoping a 4090 can do 4k120. Don't need shitty gaytracing anyway.
You VILL buy ze 4090
You VILL have a gpu bigger than a new born
But when will it get ported to the Switch 2?
I just hope my 3070 and Ryzen 7 3700x can run 1440p 60FPS med-high.
has there been any news for cross platform pawns?
>tfw bought a 6750xt just yesterday
Feels good baby. Budget gaming has never been better.
>Minimum:
>Requires a 64-bit processor and operating system
>OS: Windows 10 (64 bit)/Windows 11 (64 bit)
REEE WHY WON'T IT RUN ON MY 15 YEAR OLD OS THIS IS BULLSHIT
>Processor: Intel Core i5 10600 / AMD Ryzen 5 3600
REEEE MY 2500K IS PLENTY ENOUGH AVX IS A FRICKING HACK
>Memory: 16 GB RAM
REEEE NOBODY NEEDS 16GB IN CURRENT YEAR. 4GB IS FINE FRICKING moronic CAPCOM
>Graphics: NVIDIA GeForce GTX 1070 / AMD Radeon RX 5500 XT with 8GB VRAM
REEEEE VRAM IS A MEME MY 2014 GTX 970 STILL MAXES OUT EVERYTHING I SWEAR ON JENSEN THIS IS BULLSHIT
>people are still excited after the denuvo reveal
lol
lmao
there has not been a single game with denuvo that was good
you are as delusional as the people who thought starfield was going to be good
I?M LOSING MY FRICKING MIND
so has it been confirmed we're not getting purple and pink vocations alongside trickster?
They haven't confirmed shit, just trust the plan
the only thing that has been confirmed is that Mystic Knight is not coming back
Frick, there goes my machine gun.
Ignore him he's a lying homosexual trying to bait.
It wouldn't be the first game to leave important stuff like more vocations for the players to discover themselves when the game's out
Yeah but this is Capcom, they always spoil everything about their games before release, especially something as big as playable classes
That'd be like having a secret weapon class in Monster Hunter or a secret character in Street Fighter, that's just not something they do
i forget, was this a leak or someone's guesses?
Looks like wishful thinking OC
it's a reddit prediction post
mystic knight has already been deconfirmed. We've seen fighter using mystic knight exclusive weapons (maces).
That doesn't mean squat, we're still in the guessing phase
>itsuno has been adamant about wanting to make all the classes more distinct
>to the point of removing knives and staves from magik archers, separating strider into thief and archer, etc.
taking this into account explain to me why he would decide to give maces(the only exclusive weapon to mystic knights) to fighters. you know that the only answer that makes sense is that mystic knight is gone.
What if MK gets something like magick swords and maces instead?
>what if it gets a new weapon
at that point it would be a different class entirety
Not really, magick shields and skills were his bread and butter keeping him special.
Having a special category of swords or maces that fit him better on the magic side of things would only make sense to differentiate him more.
Now if we see his cannon on some other class, that would be true confirmation in my eyes.
Because vocations share weapons that were previously vocation-distinct in DD1
"archer" is the bow class, it gets shortbows and longbows, and strider & ranger skills
It's not impossible that "fighter" is the sword + board class, and mystic knight returns as just "fighter" with a magick shield
>I do think this is unlikely, because otherwise magick archer could have come under archer, but maybe you only progress to magick archer after levelling archer for a while or something
cool I can't believe i missed that
mace fighter sounds tight as frick
>We're probably only going to get the press demo and that's it until release.
FRICK.
I want to be 100% unspoiled, but god I want to know how many vocations there are. I don't believe in the lmao9 posters, but I just need to know, truly.
I will feel so vindicated in just a few months. Patience.
>30fps on console
>Denuvo on PC
Dead on arrival.
Gonna build a new 1500€ PC or PS5 Pro for this game. I will likely miss the launch anyways and get it for the next Christmas.
4080 here, so I'm not worried about that.
What I'm worried about is how those leaked Capcom diversity slides are going to affect this game.
>Low hardware requirements
No problem
>Denuvo
I ain't playing it.
it's going to run like shit because it's jammed packed with fricking denuvo. if you buy this game you're a massive homosexual.
>Denuvo
I sleep.
Anyone buying it is part of the problem. And i say this as a DD fan.
>Anyone buying it is part of the problem. And i say this as a DD fan.
I would eat a Saudi prince's shit to get DD2.
dude
poorgays are so obnoxious
>$70
Dead game.
If you still have that GPU nowadays that's your problem.
I have literally all my hopes riding on the nvidia GPU refresh next month. If the prices don't drop it's actually over.
>If the prices don't drop it's actually over.
They won't.
Just do what everyone else does and get a 7800XT already
I looked at that card during black friday, the discounts were up to 5% off. I laughed and closed the browser.
>the discount on this already cheap card are low, therefore I will just wait until the much more expensive card drops and buy that
???
Ah but the expensive one is green, the colour of money so it MUST be better.
>expecting deep discounts on a card that just came out
you should be worried about CPU
I got a 7900xt for 750 upgraded from a 2070. Nvidia brand loyalty has really damaged the GPU market IMO, that and AMD drivers being shit until the 6000 series.
>game needs a 7 year old card at minimum
Yes, as it should
i dont play new releases but not very optimism toward capcom but my gtx 970 & i5 4690k yet to trip on any game i play vsync on ultra mod skyrim, destiny 2 etc 60-55fps mwh max setting 60-55 fps any optimized game worth one salt no perfomance issue
I'm ready.
Upgraded from a 1070 too after 20 years of Nvidia exclusivity, so you should stop being a homosexual, OP.
I ran this game at 60fps 1080p on an igpu while waiting for my graphics card to arrive.
These threads always make me so happy.
It feels great to not be poor and listen to the desperate cries of the plebs, knowing that I will never have to deal with their struggles.
Am I fricked?
Should be fine at 1080p
The recommended specs are still set for 30fps. Why the frick would someone go for 2160p before targeting 60fps?!
I'd rather go down one step to 720p before looking to play at 30fps. These specs tell me nothing.
>Why the frick would someone go for 2160p before targeting 60fps?!
Because consoles advertise as 4k capable, therefore that's the new "standard"
Why do you think all this "huur duur modern cards can't run shit" even started? It's because everyone now targets 4k 120fps.
>Why the frick would someone go for 2160p before targeting 60fps?!
>These specs tell me nothing.
Because the CPU is an issue here, not GPU.
You people got too spoiled by 8th gen that you can't comprehend CPU bottleneck, now that consoles actually have competitive CPU.
>CPU is the bottleneck, so we won't list higher CPU requirements for the recommended
It doesn't make any fricking sense. Same for their low GPU requirement
Ryzen 3600 is still relatively powerful. Posting Ryzen 7800X3D as recommended CPU for 60 fps would only discourage people from buying the game.
Again, doesn't make any fricking sense.
You can list R5 5600
You can list a 3080
>You can list R5 5600
You heavily overestimate the power difference between 3600 and 5600.
Even 7600 is just about 55% faster than 3600.
It's clearly GPU limitation, 2080 isn't even twice as fast as 1070.
>It's clearly GPU limitation
Feel free to explain in detail how.
They literally recommend 2080 for 2160i, which is exactly twice the pixel count of 1080p. It would be just enough for 1080p 60 fps.
>But muh 4k
You know exactly why. Cucksoles will run at those settings more or less. But it's strange because every RE Engine game I've tried ran rather well. And this game isn't even using the new engine.
They had no issues listing 1080p 60 fps as recommended target for Resident Evil 4.
It's clear that CPU is the bottleneck.
>But it's strange because every RE Engine game I've tried ran rather well.
All of them were PS4 games. Even Resident Evil 4, despite what that one guy claims.
Dragon's Dogma 2 is Capcom's first 9th gen only game and it's clear they aim at 30 fps.
But it is a cpu limitation. Why would you guys think Dragon's Dogma 2 would not thrash the cpu then gpu when its trying to be an extremely dynamic combat sandbox type-ass game.
Doubt it, only games that struggles to keep 60FPS on that kind of CPU are Scatfield and jedi survivor.
Ryzen 3600 owner here, Ready or Not ran like shit on the CPU side, even on the lowest settings, when GPU was running at like 50% utilization.
Any milsim wannabe Squad clone (not that I like them, to the contrary, I'd rather Red Orchestra clone than Squad clone) is guaranteed to run like shit with a lot of players.
I'm planning on building new PC specifically on the CPU side, hopefully 8800X3D, while waiting for proper RTX 5000 series.
I didn't play it but it sounds like bad CPU utilization with reliance on single thread.
There is difference between high requirements and just trash optimization. But for high player count shooters CPU performance was always bad for sure.
Dragons dogma is also open world and theyve gotten rid of traversal loading screens. Stop ignoring the March of time and progress and upgrade
R5 3600 is plenty new for a cpu
I'm on 5600 and I'm waiting for Zen5X3D, I don't want to upgrade every gen for 25% uplift.
I would rather have loading screens maybe not as many as in scatfield lol than constant traversal stutter.
JS's CPU hunger is insane but the #800x3d chips run that game very well. I stuttered like hell on my r3600 but on a 5800x3d it's smooth sailing.
It stutters even on 7800X3D, there are certain levels where performance drops for no reason. This is mostly the case if you enable RT.
Shader compilation stutters and no amount of hardware can be thrown at it to overcome the problem.
Feels good to not play or buy EA published garbage.
I'm on 5800x3d + 4070 and the only place that kind of stutters to me is the dust planet's village and the droid camp near it.
>RT
It's not worth using. The reflections and RTGI is very barebones and ununoptimized. It looks like 3% better, the baseline graphics are already good enough.
that's not 2160p, that's 2160 interlaced.
Same pixel count, different display method.
No it’s not, it’s half the pixel count
Would you say that 480i and 480p is the same pixel count?
No it's not, it's a sort of checkerboard rendering method. It renders something like every second line of 2 frames and blends them together. It's basically 2k.
It's Capcom's own checkerboard rendering tech for RE Engine. I don't mean they invented interlacing or whatever, but it's one of their display options, not unlike a DLSS/FSR toggle.
Who the frick even uses interlaced these days!?
Capcom.
Do people don't look up game settings nowadays?
>Do people don't look up game settings nowadays?
I wouldn't expect a modern game from a reasonable sized developer to use a display method not used for over a decade.
Any temporal upscaling method is essentially interpolation.
>interpolation
meant to say interlaced
Recommended spec aren't even targeting 2160p but 2160i, this is 1/2 of the resolution.
Basically, the requirements are a scam.
It's the same shit, except the recommended are now 1080p 60fps, which is interlaced 2016 at 30.
Close enough but only when it comes to GPU, CPU load will be much higher at 1080p60.
>CPU load will be much higher at 1080p60
Not really. It's clearly GPU bound.
Dude, how is it not higher CPU load to run 60FPS than to run 30FPS?
CPU load doesn't change much if you change the resolution and keep the same framerate.
>how is it not higher CPU load to run 60FPS than to run 30FPS?
Because the CPU can already deliver those frames, but the GPU is limiting it from doing so.
Yes, technically you will se a slightly larger load on the CPU, but the point is that the 3600 is more than enough for the task, which is why only the GPU requirement changes.
You don't know that.
This game can do stable 30FPS on Cucksole CPU but it can't hold 60FPS for example.
R5 3600 is nearly equal to current gen CPUs.
If they targeted 30FPS from the start they didn't bother to optimize CPU performance.
Were you trying to say 2160?
If you don't have something over a 1070 or a 9th gen console that's kind of on you at that point. Probably gonna opt for console anyways since PC releases seem to be more shit than ever at launch now.
when can we expect the next marketing push bros?
Apparently early January according to IGN
I will play it in 2025 then, when I will be able to buy myself a new pc.
I will play it at 22fps on my 6600K and 8gb 1070
?t=23
Basically the 2160i is only 1080p pixel count, but half the framerate, which explains the steeper GPU requirement while having no real affect on the CPU.
homosexual.
2160i = 3840 x 1080 = 4,147,200 pixels
1080p = 1920 x 1080 = 2,073,600 pixels
>3840 x 1080
How do you figure that?
>what is interlacing
Go away zoomer.
RE engine supports checkerboard rendering and it looks awful.
Wow, rude, homosexual.
Could've just said that 2160p is 4 times 1080p, but due to it only drawing half the pixels on the Y axis, you still need to draw the full X.
Anyway, point still stands.
Higher GPU to do that, same CPU because the 3600 can handle it just fine.
>Higher GPU to do that, same CPU because the 3600 can handle it just fine.
This is just your assumption without any evidence.
You don't know how that game runs, it might be awfully optimized right now.
Before game launches it's all speculation.
>This is just your assumption without any evidence.
The evidence being the CPU is unchanged between minimum and recommended.
You keep panicking if you want though, not my problem.
It’s not CRT interlaced anon
It doesn’t scan a single line
Still should be double the 1080p pixel count, since a true 2160p is 4 times that.
Also illustrates how much of a meme 4K is.
1440p already looks way better than 1080p, but it's only 1.6 times the pixel count.
You start hitting massive diminishing returns on 4k.
This is assuming you can stomach no anti-aliasing at lower resolutions than 4k. Have you played any AAA game in the past 6 years? They all use TAA which will blur the frick out of 1080p and to a lesser degree 1440p. At 4k TAA is only a boon to clean up remaining edges
Is it though?
Isn’t it just two alternating 1920x1080p checkerboard patterns?
You need to remember that 4k is basically 4 1920x1080p screens.
So it's 4 screens alternating, so 2 normal 1080p screens.
It's just checkerboard rendering, just "temporally merged" interlacing.
interlacing draws every second scanline per refresh, which would halve the horizontal and vertical resolution
Is the fact that a 3070 only has 8gb of vram a concern?
Should I got for a 10 gb 3080 instead?
If you're gonna upgrade at least go for 12GB, it should last for the current console gen at least. Remember that the 1080Ti had 11GB.
Are you suggesting a 12gb 3060? Or go all the way to a 12gb 3080?
Because that shit is expensive
Don't ask my why there's only 8gb 3070 cards
Are you on 1080p? If yes, then you can pick 8gb card. If you are 1440p+, then 12gb is the minimum.
The 10gb 3080 sucks. Get the 12gb if anything. That said, they're really expensive now compared to a little over a year ago, so you're better off getting a 4070 or 4070ti.
If I can run RE4 remake in 60fps will I be able to run this in at least 30fps?
>If I can run RE4 remake in 60fps will I be able to run this in at least 30fps?
It's impossible to say because this is Capcom's first next gen only game and we don't know how well the RE engine scales
Its over for me
>have a 3060Ti
>still on a ryzen 1600X
am i fricked?
You should get a 5800X3D while you still can. You will have stutters out of the ass even if you are GPU bound.
4060 should hack it at 1080P, right?
You will be fine.
>tfw 2060 super bought from chinese, needs to be undervolted to not crash any games
How did you undervolt it? Have the same card and it seemed like it was only possible with the non super version.
You sure it's not you shit PSU?
I don't give a shit about raytracing or 1080p, I just want to be able to play the game.
1080ti chads, we keep on winning
>singleplayer MMO
I will be skipping and hoping that anons get rused by marketers into buying the game day 1. Nothing more amusing than seeing hundreds of cope and sneed posts over some easily-visible slop.
7800xt or a rtx 4070? I've never used AMD before but those are about as much as I'm willing to spend. No I won't just wait until 2025 for nvidia do further price gouging, my 1070 is already outdated.
Black person, wait until February nGreedia is refreshing some of the cards.
4070 might get slightly cheaper or maybe 4070Super won't be any more expensive doubt that
You don't have to buy right now.
Whatever you can get cheaper honestly.
Do you want RT on? 4070
What resolution are you going to play at? Are you willing to risk it that 12GB will be enough for 1440p for the next 5-6 years, judging by you sitting on the 1070 still. Probably going to be enough for 1080p, probably.
I got a 7800XT for 7700XT money, so I don't think I could've gotten a better deal.
I hope my rig upgrade for DD2 will help me for MH Wilds.
Can't believe Japanese games are what's making me upgrade my PC.
Sounds like the only games worth upgrading for
And Capcom no less
They’re really killing it with AAA games
I will run it on my ryzen 1600af and a gtx 1660 super, idc if i have to play it at 720p but i wont update yet.
>game isn't out
>on an engine that has proven to be amazingly optimized
Doomsayers are so fricking stupid
Anons don';t play games, this is a known fact. Being positive about a game means either potentially playing it (and as established, that is verboten) or being a corpo shill (likely unpaid). Now NOT talking about vidya is inconceivable so that leaves being a raging doom poster.
True but it is also Capcom's first next gen only game as well as being a big open world game with lots of physics interactions.
>lots of physics interactions.
speculation
We haven't seen anything to support that
We saw a clip of the party smashing a rock wall and causing a flood but sure, it might look the part more than having actual physics.
we've seen bridges collpsing, boulders rolling down hills, bombs blowing open creeks causing a rushing river to knock over a large foe etc.
besdies, Dragon's Dogma 2 is the first RE Engine title that isn't releasing on last gen consoles, so obviously the hardware floor requirment will be increased on PC
>speculation
You're a moron. They've said it and shown it multiple times. Look at this webm. The fricking table shakes when he lands near it. There are even shockwaves/impacts from physics objects colliding near each other.
Oh wow, physics that HL2 did 2 decades ago, so hard for the modern CPUs.
Fricking switch can pull that shit off in Zelda.
Zelda also isn't shooting at photogrammetry realism, neither is HL2
Absolutely does not matter for physics
it matters for general pressure on the hardware you moron
Keep moving the goalpost.
You have no idea what you're talking abut.
I accept your concession
lmao seethe
>still no argument
Google what photogrammetry even means, dipshit.
Until then, keep seething
>moron thinks that only the CPU is responsible for physics
LOL LMAO
What physics? The game's physics are on par with 00s games.
Keep doomposting moron . I know my CPU can handle it just fine.
>momentum and impact based shockwaves
HL2 and BotW do not do that.
>linear PS4 games with little interactivity vs physics focused open world PS5 game made for 30 fps cap
They're all the same!
Its gonna run fine and you know it, now have a nice day.
It'll run fine on modern hardware. This game is not targeting a 2013 console OR your PC hardware from like 6 years ago because the PS5 didn't exist yet
Thanks for agreeing with me.
Good evening Arisen 😀
>2060 that I picked up just for $200
Upgrade pajeets, it’s not that expensive
Anon the 2060 is basically a 1070
I think the game will be really good.
PANTHER
PEOPLE
MILKERS
Yeah pawn waifus are still a thing.
I don't like the addition of cat people and will mod them out as soon as the inevitable mod to do so is a thing
>addition
Cathomies and elves were supposed to be added in DD1, just look at the concept arts
I'm fine with elves, I just think the cat people look really moronic. They should have made a unique beast race.
Based Itsuno
That cat is fat
>Give Gift
Oh baby
>pretty kitty
I fricking hate your games Todd but I can't be mad at you, my brother
Bless Itsuno
BIG FAT CAT TATS
FAT
CAT
TATS
>sharp pointy bits perfectly placed to stab you in the boobs
IMMERSION RUINED, THANKS TODD
I loathe poorgay PC players who use hardware weaker than current gen consoles and act like they're smart for doing so
your 1070 and 6700K belong in a landfill, you homosexuals
The PS5 is equivalent to a 1070 tho
damn, that's crazy given how heaps of PS5 games have gaytracing which would utterly cripple a 1070
truly fascinating
>posts single slide from a single Gamers Nexus video in regard to a single game
wow!
Raytracing a 30 FPS 1080i image with checkerboarding isn't really that impressive
so show me it on your 1070 in that case
Why do PC gamers expect their ancient toasters to be able to run new games well?
Who the fricks still on the 10 series cards
Ganker is a third world colony full of dregs living off of welfare and still using the PC they got when they were 16 y/o
erm, fortnite pros sweety
Imagine if asiaticmoot range banned all of india
a man can dream
Yes, rangeban india...kekekeke
Works on my machine 🙂
What's the best 2000 series card in terms of price performance?
Why the frick is the 2070 $350 cheaper than the 2080?
>buy 7800x3d for this
>didn't check cpu socket prior to install
>install in the same manner as all other systems I've built (extremely slowly and gently lower the cpu into the socket in the keyed direction)
>gently confirm it's in place by lightly moving it in place to confirm it's in propertly
>latch it
>no display out
>rebuild like twice
>still nothing
>clean it all up and put parts in box
>notice like 3 bent pins on oppositing corners of the motherboard
There's no way I bent them installing it in the correct orientation in the same manner as I have the other half a dozen systems I've built, right?
I hope you didn't buy from shit retailer like Newegg.
Good luck with your RMA.
Left side 1/4 up-ish and top right-ish and 1/4 up from the bottom in the middleish on the right side. Nah I'm Australian, but I'm still fricked and it's an out of warranty repair. I was quoted 50-60 bucks or 110 if they replace the entire socket. Beat a $450 motherboard but it's shit.
If they sold you faulty product you should get your full refund, it's not like motherboards can't be DOA. Ausies have strong consumer protection laws if I remember correctly.
The issue is I have no proof and the only anecdote is after installing it and removing it the pins were damage. So the blame is on me even if I didn't actually do it wrong. It's kinda shitty but whatever.
Then just lie and say it arrived like that. What are you, moronic? Why would you be honest about that? The retailer is marking this shit up a ridiculous amount, they won't feel a refund while you will definitely feel a new mobo in your bank account.
>gently confirm it's in place by lightly moving it in place to confirm it's in propertly
you fricked it
are you really not meant to touch it at all? The pressure would've been even less than spreading thermal paste on it which you can do with a spatula or attaching the cpu cooler
The way you tell it makes it sound like you've installed it in the wrong socket.
Kinda weird having the pins on the motherboard.
AM5 is pins on the MOBO instead of CPU
I know, it's just weird. I know intel does that too now.
CPU pins somehow seem more durable, you can even manually bend them back and they will work just fine. I dunno, maybe I'm talking out of my ass.
it's keyed and can only go in one way, it has the arrow in the corner too, it's basically the same as intel sockets, which are the 7 systems i've built in the exact same manner previously
I have a 4090 idc, what I care about is if it still has the worst way to calculate damage in gaming like 1 did.
It's puzzling how they're still using that interlaced rendering, when FSR2 does better job at lower resolution and at better performance.
RE4R had both options available, but I just used a mod to use DLSS instead
Yeah, and RE4 was the first time I actually thought that FSR2 Performance (at 4K) looks good. In earlier games I had to use at least Balanced not to look glitchy in motion.
And yeah, later I also switched to DLSS.
Meanwhile interlaced gives less stable image in motion.
Apparently the best implementation of FSR2 is... No Man's Sky for Switch. Hello Games took the original code of FSR and improved it, and now it works very well despite the extremely low internal resolution. See: https://youtu.be/sbiXpDmJq14?feature=shared&t=107
Why the frick it doesn't work like that by default, only AMD knows. But at least it's safe to say there's ample room for improvement, fingers crossed for FSR 3.5 or whatever they are going to call it.
Isn't FSR open source? Can they release their version or is it on a game by game basis?
The code AMD provides is open-source. So anyone can take it and do whatever they want with it (Nintendo used FSR1 in TotK for example. FSR1 has also been implemented is a few emulators. A few FSR3 mods have been popping up recently, and they work on both AMD and Nvidia cards).
But if the open-source license -requires- those developers to share their modifications, I have no idea. Probably not?
i really hope they keep the black brit characters to a minimum, i don't need a bunch of wyll's and KSI's in my fricking dragon's dogma
>3060
>*huff huff huff*
"Im still in this shit"
>*hacks blood*
>needs a 1070 just to target 30fps *with drops* at 1080p
That is a 7 year old graphics card. The fact it's still rated to run this game at 1080p is something you should be grateful for.
Its just idiots who refuse to turn their settings down when they cant run something.
>ayo arisen, les go KILL sum muhfukkas, for REAL
I love that they show shit like this, but in game you would literally never see them as pawns or players. I want black people to play the cool game too and make what they would empathize with, but when they keep putting them everywhere in promo material it is forced as frick.
this, the only black pawns i have ever seen are the wacky abominations
It's funny how every thread that uses this image in its OP is always bait, every time.
More like the only DD2 threads still going are bait threads because no new info for a while
>tfw 1660
I'm still on a 1070, but I'm gonna upgrade in the next month or 2 when the nvidia 40 super series gets announced, not that I mean I'll definitely be buying one of those, but I'm waiting to see what the price and spec improvements are, and if they aren't good enough I'll probably just call it a day and buy a 7800xt or maybe a 4070 if they drop the price of those as part of the announcement, just so I can have better RT support and DLSS
I'm hoping the announcement just causes prices of other GPU's to drop to stay competitive
In a perfect world the 4070STi WRX would cost 500$-600$ max. It's what the 4070 should've been from the start.
Unfortunately it will probably cost more like 800 if not more.
My 10400f/1660 combo will flawlessly run this at 30fps, 900p after 6 months of patches and I will have a great time
This thread has prompted me to go back to dragons dogma. Seems my gamesave has vanished so frick it, starting from scratch again. Surely the gamer won't turn into a horrible grind amirite?
Just use the eternal ferrystone and skip all the grind.
if you have to grind at all in dogma you're playing the game like a moron
I've seen some people complain about doing no damage due to the flat damage scaling (which IS stupid) but literally just buy and 1 star enhance new armor and weapons when you can. It's not that difficult.
However you will need to gear yourself and the pawn more efficiently when you tackle BBI, but before you do that at least buy some buff periapts from Fournival, please
>he doesnt grind goblins with 2 rings of perseverance to get max vocation so the fun can begin
NGMI
Just use the save editor.
>tfw the 2080ti is starting to show it's age
It's been a good ride, brothers...
I had to get a 2070 anyway because my previous gpu was a 660 and I could barely play ps4 games
Same card as me. I'm holding out for 5090/5080
I'm thinking about just splurging on the 4080. It seems superior in every way to the 3000 series.
Why are pcgays so poor I thought you guys were the master race?
GPUs alone can be 3x more expensive than consoles.
but I thought you could build a PC easily for cheaper than consoles to run console games?
Not in this generation, but you could for some time during 8th gen, there's no way these days to get something cheaper or even at the same price that matches a console performance.
Why is that? Scalpers? Miners? AI homosexualry?
A little bit of everything, the inflation it's also being used as a reason, but mostly the mining shit that ruined GPU prices for quite some time, and it's only now that we're slowly recovering.
Honestly most components are in a good place, it's just GPUs that are bit overpriced.
Chinks are now bulk buying GPUs to remove the chips and hook them up to AI model training rigs. The crypto shit has died down but it's simply been replaced.
>Scalpers? Miners?
Basically yeah.
A few years back cryptogays figured out that graphics cards have good technology for cryptomining. Once they realized they can turn a profit by sticking 20 GPUs into a single farm to speed up the process, a lot of people started trying this and all the GPUs were bought out and put to crypto work. So if you wanted to buy a graphics card you needed to shell out 2-3x times the normal price.
Then the worst possible thing happened - instead of saying "frick this shit, i'll buy once prices go back to normal" most people just... kept buying. At those ridiculous prices. That did immeasurable damage to the market because it essentially told GPU makers and vendors that they don't need to price their products sensibly, we'll buy anything no matter the price.
So here we are. The threat of "we won't buy your graphics cards if they're poor value" has been proven to be bark with 0 bite behind it, and now GPU vendors can suck us dry with insulting prices because they're fully aware that won't stop people from buying.
but then why would anyone want console performance at all let alone console as a platform at all?
unreal engine and nvidia are colluding to make game run like shit on consoles because consoles have amd gpus in them
this would be a great lawsuit to have start up because it's extremely anti-competition
>unreal engine and nvidia are colluding to make game run like shit on consoles because consoles have amd gpus in them
kek
ok and? I thought you were the master race just buy a better one poor turd world homosexual
I guarantee that not only do I make more money than you ever will in your entire life but my penis is also 2x larger than yours.
You also lie on the internet so congrats I guess surprised you can even afford internet access.
Just found a cheap new copy of the first game for Switch. Give me the sliders for a cute e-girl pawn
Thanks friend
Nice, thank you too
I'm not a furhomosexual, but my first arisen and pawns will be kots.
Kots are based.
I'm gonna brine every furry pawn I see.
A tough choice - the bing chungus body (for all the mana reserves y'know) or the ebony amazon.
>playing on PC
kek, enjoy your shitty port
The DDA port is excellent, and recent CapCom games have all been fine on PC, including DmC V. I see no reason for the DD2 PC version to be any different besides maybe Denuvo eating frames.
>the cope
lmao
you have nothing, sönygger
Capcom's pc ports have been solid for almost a decade now, newfriend
?
What's a CPU that will safeguard my frames' anus against the Denuvo monster. Ryzen 5 7600x gonna be enough?
Anything above a r5600 will be enough. Denuvo doesn't impact performance nearly as much as people would have you believe. The FPS and frametime graph between the Doom Eternal I pirated and the Doom Eternal I later bought was the exact same. Only way it could reduce your performance is if you have an old ass computer.
>looks more or less like dd1
It's unacceptable really. Now you have Denuvo on top of that to lower the framerate even more.
>just buy a new gpu every other year
lmao people with money are more moronic than trannies at this point
I hate you morons so much. I'm not saying you should be doing that. I'm saying you should stop complaining your gpu and cpu that is older than even the ps4 pro is not adequate enough to run current generation games. Weren't you homosexuals shitting on the lack of current gen console exclusives because of last gen console versions and PC ports? Well now we've got current gen only games finally coming out and NOW the sour grapes are here right on time.
Stop being poor and upgrade or don't complain.
>rtx 3080 and 1440p 165 hz
just this game fricking overpriced piece of shit gpu
is this place really so poorgay infested that people are legit seething their decade old hardware can't stuff at 1080p60fps anymore? get a job you worms or stick to older games
I have an RX560 (4GB) and Dragon's Dogma ran perfectly for me (60fps).
You're lying.
the topic is DD 2 morono
pay attention
Write a 2 next time moron, can't believe people don't proofread their threads, disgusting.
The OP image says Dragon's Dogma 2, moron.
It's a picture from the first game, and the II can be easily missed from the thumbnail.
Is that your excuse for writing the game's title wrong?
You stupid frick.
It isn't you blind moron, it's from a trailer form the second game. Nobody else ITT made the same mistake you did, you're completely at fault and a fricking idiot.
Super funny that the "PC Master Race" is now getting their hardware outpaced by consoles.
Just get it on the console. Problem solved.
>rtx 2060
I'm sure It'll be fine
>buy a 4070
>almost all the new games that have any decent use for a potent gpu dont actually look better than many games released almost a decade ago and just have shit optimization
>morons defend it because they can get 1440p@60fps / 1080p@120fps on their gpu that should have been capable of 1440p@160+fps
why are games so gay and stupid
>all the games
meant to say, specifically the ones worth playing
you bought a 4070. by default everything looks blurry on your screen because you never run your monitors native resolution. AMD is always sharper.
What the hell are you talking about?
I am using a 3070_ and if DD2 runs like shit I will walk out and buy a better card.
Literally the only game in a half-decade I will be willing to do this for. Capcom, don't let me down please.
It'll run fine, just don't get too wienery with VRAM usage, you've only got 8gb.
Yeah. I wish I understood how much that 8GB matters. For the future, I suppose.
i know it's my own fault for opening this thread but reading Ganker's attempts to be an authority on tech is always vomit inducing
I have a RTX 2070 which is a thousand times better than a 1070 so I'm good
I amassed a ton of cash and spent it all on portcrystal forgeries.
every time i play i spend some tim in the rift gifting a ton of pawns across all sorts of levels a portcrystal forgery, mostly the lower level ones
i am extremely amused by the thought of someone carrying a bunch of heavy portcrystals around only to be confused and then realise these cant actually be used
Fug, really considering upgrading from 1080ti now... 6800xt is the most palatable gpu offering I've seen in a while
*7800xt
They've got practically identical performance, in fact, they trade blows depending on games, both are fine cards though, no doubt.
7800XT is supposed to be slightly better in AI and RT
i have a computer with a 7900xtx, bought my brother one with an rtx4090 for his graduation
everything runs smooth like butter
the sad sacks who still run gtx970s and whine about how games are all bad now (because they can't run them at playable settings) are worse than consoleBlack folk in terms of holding technology back
Almost done with building a new PC, so I'm fine.
Surely there will be a demo, right?
Every Capcom game has had a demo in the last 6 years, a demo is practically guaranteed.
No, and no.
first are press demos then eventually its digital demo everyone downloads.
>tfw 2080
Should I upgrade? Do I even have an upgrade path that isn't horrendously bad value for money?
Isn't the 2080 the recommended for 4K 30 fps? You should be fine at 1080, 80 cards are still powerful.
>5800x3d & 7900xtx
>mfw
RE Engine games are also AMD sponsored now so enjoy your 300+ fps.
200fps doesnt break the physics in RE4 so you're fine.
>RE Engine games are also AMD sponsored now so enjoy your 300+ fps.
dont worry jensen specifically paid for monster hunter wilds. for whatever reason monster hunter favors nvidia. rise has dlss and mhw is gameworks approved.
>my 2060 trying to sustain 1080/30
abandon all delusions of performance
If 30fps is the target framerate, what are the odds of 60fps breaking all the physics?
Surely consoles will have a performance mode where they can run up to 60 fps? So that sounds unlikely.
Is high framerate fricking with the physics a normal thing? I've only heard about that with bethesda games.
>Is high framerate fricking with the physics a normal thing?
It's very, very common. It also fricks with enemy AI a lot. In MH Rise, high framerates would break some cosmetic physics stuff (your dog's tail would spin around like a propeller for example), create wonky interactions with knockback from attacks, and increased monster attack tracking and aggression. MH World had framerates affecting attack speed, hitstop, tick rate of multi-hit attacks and the fire rate of guns.
Try playing a Dragon Engine Yakuza game at 144fps
VERY common.
Tieing physics to fps is very common dev practice and is specially common with japanese devs
From software games for example have physics and game mechanics tied to FPS and depended on community mods to "fix" that kind of thing.
Very normal.
In Unity terms, there's Update, which is a single frame and is supposed to be used for game controls, and Fixed Update, which can be described as "physics frame", as that's what it's mainly supposed to be used for, or game logic.
Fixed Update is executed every fixed amount of time. It must be executed even if actual frame wasn't finished. In Source engine terms, fixed update would be associated with tick rate.
Tick rate 66 means there are 66 logic/physics refreshes every second. The reason it can work just fine with high refresh rates is that values are interpolated between frames. So yeah, what you're seeing is actually the past, but it works well enough.
The games that break physics with high FPS are a rarity nowadays. Even Bethesda games play along with high framerate well now. It's only above 300+ fps that things go haywire but why would you ever have 300+fps?
>but why would you ever have 300+fps?
Why, for the upcoming screen from LG that can do 4k 240hz or 1080p 480hz of course!
https://tftcentral.co.uk/news/asus-tease-rog-swift-pg32ucdp-dual-mode-oled-monitor-ahead-of-ces-2024
ITT: Poor people who refuse to upgrade their nearly decade old GPU.
How old is your rig? Apart from my GPU and an additional SSD I installed my old shitbox is over 10 years old. Served me pretty well up until now, hope I can get the same millage out of my new build.
Went back to my years old DD:DI save and have been farming Daimon for lv3 assassin and mk gear. At the point where I just speedrun him with periapts and blast arrows, but goddamn the AI was occasionally RNG when I tried to kill him with Thousand Kisses and periapt because sometimes Daimon would just stand around doing nothing while other times he would constantly do his flame blast and lighting if I tried to climb his face.
Already knew when it was announced and no footage was yet released that my shitbox PC wasn't gonna cut it, thankfully DD is meant to be played with a controller so I'll just play it on xbox.
>Suddenly everyone is doomposting and calling Dragons Dogma of all things shit.
I don't pay attention to release dates anymore. I can always tell something is coming out soon when Ganker starts pissing their pants about it.
>everyone is doomposting
it's one guy, mostly pretending to be a discord personal army but in actuality is just samegayging to create a siege mentality among DDfrens and keep the threads going until release
That's some dedicated autism but shit I believe it.
I wonder if it's all the same guy that does this for every new release.
>muh raytracing bad
I wonder if any zoomers remember how changing to 32bit colors tanked your performance. Guess that was a meme too and all games should go back to 16 bit.
I'll have you know I play with the Z buffer disabled as it is a scam feature with disproportionate performance cost for virtually NO increase in visual quality.
raytracing shadows and raytraced reflections are fricking stupid. very little games use global raytracing because its too intensive.
Bounce light is really pretty I don't know why people hate on RT.
>why people hate on RT.
"my 1070 can't run it"
>120FPS without RT
>30FPS with RT
I wonder.
Hardware just isn't ready.
Tomshardware told me when my life flashes between my eyes I cannot have lived it without raytracing. I should stop kvetching and buy the 2080ti.
The performance and the fact that devs don't usually build a game's visuals around it. That said, something like Metro Enhanced looks great with ray tracing because it was made for it specifically, and the game runs pretty well.
It's a meme because the average person still has a gtx 1060 that'll shit itself if it tries to raytrace. It's bad and a gimmick because it's not accessible. Just like FG was bad and a gimmick back when a 4090 was the only card that could do it, but now that FSR3 does the same shit and a lot more cards do it and a mod exists to make the 2000 series and 3000 series do it too, now suddenly it's not bad anymore. Tech opinions are mainly driven by sour grapes mindset.
lol the 2160i is actually 1920x2160 temporally upscaled to 3840x2160 using crapcom checkboard rendering
I guess they didn't want to scare morons off by putting actual 4k requirement in, kek
>Dragon's Dogma needs a nearly 8 year old GPU to be playable
That's pretty reasonable though.
No, its not reasonable. My 8 year old purchase should give me 1080/60 at max settings for AT LEAST 10 years.
That has never been true and is unlikely to ever come true so.....
>My 8 year old purchase should give me 1080/60 at max settings for AT LEAST 10 years.
imagine if you tried saying this shit in the late 90s or early 2000s
you homies have no clue how quickly hardware got outdated back in the day, nowadays it's much better
Are you all zoomers? This has literally never been true at any point in PC gaming history. Hell it was arguably worse in the past since PC gaming has stagnated a lot. The 2000s felt like everything was obsolete within 1-2 years.
>Are you all zoomers?
Probably all third world zoomers. Remember when Crysis came out? Everyone was salivating over it and getting excited for the future. No one was b***hing and moaning that their mid-tier PC couldn't play it at max settings.
>30FPS
>Playable
Pick one.
I'm not complaining about hardware requirements, I'm not a total poorgay.
For a game like DD? Yeah its playable. This isn't Apex.
Playable according to every developer and consoletard
30 FPS is basically the definition of playable.
>it'll play like the ps5 version
damn that sucks.
also
>mfw I paid 479.99 for a 1070 when it came out and bought a 6700k shortly there after, when I used to upgrade every year on GPUs
>mfw thinking I'd never have to upgrade again
Guess it's only indieshit and old games for me cause I'm gonna push a decade on this thing.
I got too comfortable gaming on PC when we were still getting 360 ports.
I have a GTX 1660 Ti...
My impoverished brother.
Same but DD2 is finally giving me the justification to upgrade, ascending to a 7800 XT soon
>3080ti + i5 10600k
I think I'll be good
>CPUlet
>I think I'll be good
LMAOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOO
>just bought a 3060 ti.
Will I be fine?? I can just overclock right???
I don't care about gaytracing or pore technology or one billion polygon rocks or volumetric fart clouds, none of that trash does SHIT for the actual game.
I intentionally play everything at low-medium settings with all post-processing turned off, my GPU fans don't even bother spinning for how little I push it.
I'm worried too! Worried about DD2 being the absolute peak of vidya and everything else feeling like shit afterwards.
if the ps5 pro is landing in 2024 that's going to be a pretty tempting buy given the performance-price ratio.
I was hoping to ride out my current PC until TES VI which i assume will be another couple years so the ps5pro could be a pitstop along the way.
I think PS5 Pro will be pointless this gen. It'll still rely on upscaling and if the game won't be 60 fps on regular PS5, it won't be 60 fps on Pro either.
It'll very likely have a performance mode for 60fps on PS5 like most games that will just run at a lower res than upscaled4k.
I have a 2060 overclock and ZERO fear
Yeah, I'm probably going to upgrade before it comes out. I've gotten too used to playing games at 120+ fps.
>gtx 1080
will i make it?
>7800 XT
>Ryzen 5 7600
Let DD2 be your excuse to finally upgrade Gankerros
Nice combo.
I'll be alright
Will I be able to run the game off a chromebook
Yes, on GeForce Now.
don't capcom normally exaggerate the requirements?
Nope, if anything the requirements they show on $team are very "optimistic"
its the most honest system requirements of any pc game right now. minimum requirements is re4r's recommended. recommended specs are literally a ps5. steam page even says it recommends controller over kb+m because the devs KNOW.
just looked up the 4080
it's about 1300€
the frick
Get a 4070, or an AMD 7800 XT.
>a 4070
okay that one's pricing makes more sense. Thing is I currently have a 3070 so maybe it's not worth the update.
I'm in the exact same boat. It's frustrating because going from 3070 to 4070 seems idiotic, but waitgays would have you wait forever.
I have the money but that doesn't mean I want to be wasteful for little benefit in everything.
>3070
You might want to hold on... I mean that's not a slow card, its only problem is the small amount of VRAM, until that really gives you problems I think you can probably use it for a while still
I think the only problem I have is that I can't play most modern games at 4K unless I turn on DLSS. Not something I care about really, if I'm being honest 2K may be just enough for me, and I barely notice the jump from 2K to 4K.
To be fair the 3070 was barely a 1440p card when it released, nowadays it's a 1080p card, if you've got a 4k monitor and want to do native 4k, you're gonna have to get something much faster (and expensive)
>nowadays it's a 1080p card
really? I don't play many modern games, but the ones I play can be run at 2K with no issues. What games have you made drop the resolution to 1080? I can't see myself going back to that.
>if you've got a 4k monitor
2K monitor, 4K TV. The TV being placed kinda far from me might be the reason why I think 2K is just fine.
you can still use a 3070 as a 1440p card
there's one simple trick you can do
it's a top secret
doctors hate this one simple trick!
but don't tell anyone i told you
ready?
here it comes:
lower the game's texture size from ultra to high. -33% VRAM usage for no visual impact
that's all you need to do to combat the supposed VRAM problem of 8-10gb cards
i've yet to see anyone point me to a game where textures look demonstrably inferior when reducing them from ultra to high
Usually going from Ultra to High means you'll be just load fewer HQ textures at a any given time, so you might lower res textures for a bit until the higher res textures load in, it's not ideal, but yes, but it's not a big deal either...
This. I'm currently playing Cyberpunk on high settings 1440p, no DLSS and ray tracing though. Constant 60FPS.
The 4080 exists to upsell you to the 4090 or to prove you are the greatest nvidia fanboy.
its a great gpu but the pricing is just AWFUL. nvidia can only get away with it because their fanboys eat wiener for breakfast and you have pc building noobs who fell for the "AMD SUCKS" meme despite using AMD CPUs that are more fickle than GPUs. nearly every 7950x3d fails sha3.
I have a b450 motherboard with 2060 and 2600x. Should I just upgrade my gpu and cpu for the current motherboard, or get a new motherboard for future proofing? I can afford it, it's just the only games i'm looking at are dd2 and wilds
No highscepter
no buy
1070? That's a <$100 GPU, moron.
It's also like 7 years old, and it was mid, slightly high end back then... It's really silly to expect such a card to max out shit these days.
>2070
>1080p 144hz
I can't stop winning
*whining
imagine being such a little b***h that you cry when frames drop in your big boy toys
A 1070 is fricking old, you dipshit
It's barely 8 years old...
Why are you expecting to play the latest AAA game with an 8 year old card and zero hiccups?
This was never doable, even in 2016 when this shit was brand new.
Tell me why DD2 can't run on my i5 2500k + gtx 960 system, pls explain
Destruction Derby 2? I'm sure it could run just fine. Who knows what the frick you're trying to say since you're too lazy to type out the title of the game.
He clearly meant Darkest Dungeon 2. I'm sure it runs ok as well.
Pls tell me why Dragon's Dogma 2™ can't run on my fx 6300 + gtx 750 ti rig
posting this leak from a few years ago. interesting read.