MSI GTX 1070 GAMING X OVERUCKED MORE BY 150/400 i5 6600k 4.5ghz (No bottlenecks)
Still dips below 60 frames in assassins creed syndicate? who said this was the ultimate card for playing games at 1440p at 60fps stable ultra at all games was a liar
I'm only playing at 1080p/144hz and i get dips to 55 fps. Looks like the 1070 is another meme card
i'd like to see a video with a 1070 never dipping below 60fps in assass in carriages and shit at 1080p. they dont exist. and no one even has MSAA enabled.
but admittedly the game does look good.
Zachary Sanders
It's Assassin's Creed, what did you expect.
Hunter Cox
nope, I'm getting around 80-90% usage and my GPU is at 99% usage constantly. I do feel silly not future proofing myself a little with an i7 tho. I really do feel like the bottlenecks will be coming soon. Overclocking or not. Something in my setup cant me into 4.6ghz or more
Ayden Lewis
Your shit i5 4-core processor is bottlenecking. Why the fuck would you pair a 1000 series card with an i5?
Ryder Stewart
What settings are you playing at? Reason I ask is Syndicate is pretty badly optimized for PC and even top of the line stuff has issues with it.
Hell, I tried Syndicate with SLI GTX 1080 and it still had performance issues maxed out.
William Hughes
I have an MSI GTX 1070 Armor OC with an i7-6700k processor and I'm able to run everything at 1440p on ultra at 60fps as a minimum, including Witcher 3. Maybe you should upgrade your processor you fucking scrub.
Connor Evans
i'm monitoring my usages while gaming dumbass, my cpu isn't bottlenecking it.
everything turned up except I'm using FXAA cause MSAA would just be hopeless
Mason Morgan
>i5 6600k 4.5ghz (No bottlenecks) >gets bottlenecked
Whatever makes you sleep at night, OP.
Michael Price
>he can't actually detect when he's getting bottlenecked by shit.
Connor Mitchell
>cpu at 55% usage >gpu 100% >ur batlenacking XD yes, let me get the i7 so i can enjoy my extra cherrypicked games that utilize it.
Bentley Ramirez
WItcher 3 isn't even demanding on the CPU at all. Run that shit all day at 90-100 frames
David Richardson
This
I run a 4770K @ 4.5GHz with OP's exact fucking card, MSI 1070 gaming x.
I set graphics to medium high and I get 120 fps just fine in most games that aren't dogshit ports from consoles.
Get gud OP
Hunter Edwards
>medium >high
Kevin Miller
Your processor, or your drivers are the weakest link. Roll back to a stable driver instead of using the newest memeware Nvidia shits out because it never runs properly.
Last stable one for everyone not using a 10 series is 368.22, but because you're using the newest meme card, you might need 368.95 so it doesn't fuck up your audio.
That processor really is piss-weak though. Especially for the game you're trying to run with it. It's more in that Ubisoft can't make fucking games right, but you're still wrong and you got a shitty processor.
Jordan Campbell
thanks for the advice, but to be honest I don't really give a shit about ass creed. It was just funny seeing it drop below 60 and on some youtube benchmarking videos too. Not gonna roll back cause I just wanna play dishonored 2 when it comes out and this is just shit I pirated to test my wifes boyfriend bought me
Ayden Mitchell
>Falling for the "You only need an i5" meme >In 2016
Newsflash faggot. It's not 2012 anymore. The i5 meme is on fumes. More and more developers are beginning to take advantage of hyperthreading. It's time to stop the i5 meme.
Lucas Rodriguez
From what I remember when I had my GTX 980 Ti, Ambient Occlusion and Shadow Quality were major ones that caused performance issues. I'd recommend lowering each of them down.
Xavier Nguyen
>nope, I'm getting around 80-90% usage
Do you really think the lazy ports you play know how to properly use your shitty 4 cores i5? Do you really think all those 4 cores needs to reach 100% usage before bottlenecking?
You either stop playing (and paying for) lazy ports or buy an i7. Simple as that.
Brayden Russell
Just because your CPU is not at 100% load doesn't mean it isn't bottlenecking, you imbeciles.
Jason Nguyen
I have no idea what this post even means haha
Dylan Peterson
>i5 (no bottlenecks)
Aaron Adams
>le 144hz meme
simply eric OP
Carson Robinson
Assassins Creed Syndicate is TERRIBLE optimized, not the fault of your card. Try something like Devil May Cry 4 or Metal Gear Solid Swordyourass, you will easily go for 1440p/144hz, because they are good ports.
You fell for the PC meme, I hope you enjoy Mobas and MMORPGs
Jackson Cruz
>It's a Sup Forums tech support episode hahahahha
Eli Rodriguez
i'm playing other games jsut fnie at 120+hz, honestly. and I have those games and yeah, they're great.
to be perfectly honest, i was surprised it was stable at 65-76fps while on foot. (syndicate)
Jeremiah Rivera
>he didnt have $100 spare for a i7
Joshua Nguyen
I did actually and I read/bunch of reviews saying it literally doesnt matter if you have an i7 for gaming. I have an i7 at work, but I'm too lazy to bring that shit home and reassemble shit.
Josiah Brooks
>144Hz You fucktard, that means your GPU has to render twice as many frames than a 75Hz monitor, that's why you get less frames you piece of shit
Adrian Ortiz
Except it does matter. Hyperthreading is a thing, and developers are using it now.
Carter Wood
not him, but do you have even 1 example?
John Martin
>1440p you fell for the meme
Mason Kelly
t. 1080p shitters
Gabriel Davis
rather you need to make sure your computer can render 144 fps on a game or else 144Hz is pointless, if you get less fps it's probably a bug or bottleneck somewhere, lower it to 120Hz.
Caleb Peterson
>a bug or bottleneck somewhere
Or a shit port.
Lincoln Ward
Battlefield 1.
Battlefront EA.
Dylan Nguyen
got any source? genuinely curious about this topic (even if I don't like Battlefield games at the best of times)
Tyler Price
if the cpu was bottlenecking performance, wouldnt the gpu drop too?
Charles Smith
he's just bullshitting user
Isaac Miller
You're playing Assassin's Creed on PC.
Benjamin Evans
>who said this was the ultimate card for playing games at 1440p at 60fps stable ultra at all games was a liar well yeah because the best 1440p card is the 1080