2023 was the year that GPUs stood still::A new GPU generation did very little to change the speed you get for your money.
But boy did it change the price you have to pay for it.
You guys think I should upgrade my Voodoo 3 card? No one is joining my quake server anymore anyway
Come play Unreal with us then hehehe
Given technological progress and efficiency improvements I would argue that 2023 is the year the gpu ran backwards. We’ve been in a rut since 2020… and arguably since the 2018 crypto explosion.
Nah 2022 it was running backwards far more. 2023 was a slight recovery but still worse than 2021.
I wanted to upgrade my 1060 for the longest time for something like the 3080. But during to demand and prices hikes, I waited… 40 series got released and the prices stayed high.
So I just gave up, I got a steam deck and PS5 instead.
A lot of people did this. The GPU market for gaming might have actually shrunk. You would think Nvidia would panic but due to AI chip demand their stock is at an ATH and no company changes course or reevaluates and what they’re doing when shareholders are lining up to suck their dicks, so…no end in sight. Meanwhile AMD doesn’t seem to want to even try to make a play for market share.
Technically AMD does have more market share when you think about all the devices has AMD in them like Playstation, Xbox, steam deck and other handhelds.
But yeah Nvidia doesn’t care about gaming anymore, If I had to pick a GPU today, I would pick AMD because Nvidia 6-8 VRAM isn’t enough and AMD is better on linux.
If you want to do any game streaming though (e.g. on Sunshine/moonlight), Nvidia is still miles ahead.
I recently experimented with both of those on AWS and they are completely not usable atm. At least not over WAN and with gpu mounted to a device you don’t have compete control over.
Does streaming work any better over LAN?
Nvidia game stream is incredibly robust over the internet. Nothing else even comes close. The latency is incredibly low and the video quality is awesome with almost no compression artifacts like competitors often suffer from. A buddy and I used to stream our home PCs to the office when it was slow and even with both of us streaming at the same time the performance was great. If you weren’t playing a twitch shooter you could honestly hardly even tell it was streamed. This was over a meager 100mbps connection too.
The next closest alternative is Parsec which can manage very low latency, at the expense of significant compression artifacts if your connection isn’t rock solid or your CPU isn’t the fastest.
Steam link streaming is a very distant 3rd, and I actually found that critical components of many games simply did not work. Like for example Unity games where you adjust your camera by pushing and holding a button while dragging the mouse would just not work.
What are some issues AMD is having there? The sunshine pages show both AMD and Intel support now is I assumed they were gtg
The issue is down to encoding performance, Nvidia performs a LOT better with comparable GPUs.
With that said, h265 is okay from what I’ve seen, but any devices you’re streaming to that use h264 and even a 1060 will stream better than a 6750xt etc
Still rocking a 1080. I don’t see a big enough reason to upgrade yet. I mostly play PC games on my steam deck anyways. I thought starfield was going to give me a reason. Cyberpunk before that. I’m finally playing cyberpunk but the advanced haptics on PS5 sold me on going that route over my PC.
I just “upgraded” from a GTX 1080 to an RTX 4060 Ti 16Gb, but only because I was building a PC for my boyfriend and gave him the 1080. I’m really not seeing a noticeable difference in frame rate on 1440p.
Yeah I keep waiting for a good deal to retire my 1080ti.
Guess I could go for a 3060 or something but 4 series will probably leave my old CPU behind.
1080 gang rise up.
But seriously, my 1080 does fine for most things, and I have a 2k 144hz monitor. It’s JUST starting to show its age as I can’t blast everything on high/ultra anymore and have to turn down the biggest fps guzzling settings.
CP77, at least before the upgrade (haven’t checked since then) ran perfectly… acceptable on my 4G 5500 XT. Back when I bought it (just before the price hikes) it was the “RX 590 performance but less watts and RDNA” option, the RX 590 hit the market in 2017. And I’m quite sure that people still rocking it are, well, still rocking it. Developers might be using newer and fancier features but I’ll expect they’ll continue to support that class of cards for quite some while, you don’t want to lose out on millions of sales because millions don’t want to pay for overpriced GPUs. Allthewhile you can get perfectly fine graphics with those cards, if you look back pretty much all 201x titles hold up well nowadays.
Due to ML workloads I’ve been eyeing the Arc (cheapest way to get 16G and it’s got some oomph) but honestly so far I couldn’t get myself to buy an Intel product that isn’t a NIC, would break a life-long streak. A system RAM upgrade is definitely in the pipeline, though, DDR4 has gotten quite cheap. It’s gotten to a point where I’d recommend 64G simply because 32G sticks are the cheapest per GB (and you probably have two memory controllers).
As someone who upgraded from a 2016 GPU to a 2023 one I was completely fine with this. Prices finally came down and I got the best card 2023 offered me, which may not have been impressive for this generation but was incredible from what I came from.
And how much did you pay for the 2016 card, what range was it in, and what is the new card’s cost and range?
Overal, gpus have been a major ripoff, despite these upgrades giving good performance boosts
I believe about $300 for an AMD RX480 (great card and still going strong). This time I had a bit more money and wanted something more powerful. I went with the AMD 7800 XT Nitro ($550) which I got on release day. Sure it’s not top of the line but it has played pretty much everything I throw at it with all settings set to max and still maintaining 60fps or above. I have an UW monitor with its max resolution being 5120x1440 which is what most games will play at and everything still plays fine. It’s almost crazy to me that this card would be considered mid range.
7800XT Nitro gang rise up! Come from a 1080 and it’s been a leap!
intel GPUs definitely won out for what you get for the money
That’s not a sentence I’m used to seeing
I’ve been very happy with my Arc A770, it works great on Linux and performs well for what I paid for it.
Have you tried ML workloads, differently put: How is compatibility with stuff that expects CUDA/ROCm? Because the A770 is certainly the absolutely cheapest way to get 16G nowadays.
No, I don’t use any ML stuff or really anything that uses GPU compute at all. I just use it for gaming and other 3D applications.
I’m so glad that Intel has stepped into the GPU space, even if their cards are weaker. More competition will hopefully light a fire under NVidia to get their shit together.
I finally upgraded my GTX970 to a used RTX 3080 for 300€. The difference at least for me for the same 300€ was insane.
I just don’t see the point in upgrading every new release anyway, or even buying the most expensive one. I’ve had my gigabyte Rx 570 for several years and I can play Baldurs Gate 3 full settings with no issues. Maybe I haven’t tasted 120 fps but I’m just happy I can play modern games. When it comes time to get a new graphics card, which may be soon since I am planning to build my wife’s PC, maybe then I’ll see what’s going on with the higher end ones. Maybe I’m just a broke ass though.
Ya the problem I landed in was not anticipating how hard it would be to push my new monitor. Ultra wide 2.5k resolution with 144Hz. I can’t do cyberpunk full res more than 60fps, and that’s with dlss enabled and not all settings at max.
2070s
I had to buy 3070 ti at scalped price. Ended up paying £700 for it. I hate myself for it but the prices didn’t shift for months after and my gtx 1080 kicked the bucket. No way in hell am I buying anything this gen. My wife’s 1080 is going for now, maybe we’ll get 5080 if it’s not a rip off.
Its nvidia, its always a ripoff :p
Especially now when gaming GPUs are an afterthought for them.
Thats only nvidia though. Amd seems to still be trying to compete with nvidia some way or another
I wouldn’t say so, they also seem to have abandoned the gaming segment and nowadays are playing more or less ball with NVIDIA while trying to improve their AI stack so that they can get a higher chunk of the data centre business.
I don’t think that’s true at all. Let’s go back a while.
We had Polaris, a mid range 2016 architecture that was sold for years as a mid range then low end card.
They also had the Vega cards, which were compute-focussed and not particularly great at gaming.
Following that, they had the 5700 series. Decent gaming cards.
After that, the 6000s series. Right up there with Nvidia, and taking into consideration the die size, performance, and comparatively generous VRAM, you could argue they were the better gaming cards, despite losing in RT.
7000s series is pretty much like the 6000 except slightly further behind the 4090, albeit for half the real-world price due to AI demand bringing the already crazy 4090 prices even higher.
Idk to me it seems AMD is more competitive in gaming now than they have been for a long time.
Absolutely, AMD is very focused on Datacenter/AI now. They just presented their next gen AI system MI300X which made AMD stock go up significantly, and on the CPU side their server CPU Epyc is where the big money is at.
That said AMD is still into gaming hardware because they work with both Sony and Microsoft on making new consoles, what we get on the desktop from AMD, is probably mostly derived from that on the GPU side.Also the steam deck! They really got themselves in the console market this gen!
Yes, that was a very impressive win. Intel/Nvidia has usually been the preferred solution when power efficiency is important.
But now AMD is competing well in that segment too.
Especially now when gaming GPUs are an afterthought for them.
NVIDIA fucking sucks. But I do a lot of modeling in blender and holy damn do I want that RTX.
You and me both, bro-in-pathtracing
So how about the 2½ years from 2016 to 2018 between Nvidia GFX 1080ti and RTX 2080?
I think the headline should say A Year not THE year.What’s everyone’s recommendation for a cheap AMD GPU to use with Linux? I was looking recently at a Radeon RX 580, I know there are much better cards out there but the prices are about double (£350-400 instead of £180). I’d mostly be using it to play games like the remastered Rome Total War.
There are some used options e.g. 5700 XT-s are really cheap because many of them were mining card. For new cards there aren’t many options RX 6600 has relatively good value, but it’s only worth it if efficiency or features like hw video codecs are important for you.
Is there any issue with buying a card that was previously used for mining?
When you say RX 6600 do you mean that one specifically or the range including 6600XT etc? I don’t have a good handle on what the real world differences between the variants are.
Is there any issue with buying a card that was previously used for mining?
If used by a home user who didn’t know what they were doing they might have run it hotter for much longer than a typical gamer so the thermal paste might need a redo.
If used by some miner doing it even quasi-professionally or as a side-gig I’d much prefer it over a 2nd hand card from any typical gamer (most miners) they’ve kept the voltage/temps low and taken care of it far better than a gamer who might be power cycling regularly and definitely thermal cycling even more regularly.
No, there isn’t any more risk buying a mining card than any other used card. In both cases you should use a platform/marketplace with buyer protection options. Maybe one additional step is checking the VBIOS when testing.
The non XT is the best value of the 6600 family but depending on local pricing the 6600XT, 6650XT and even the 7600 could make sense. Just keep in mind that these are the same performance class. Some charts show the mentioned GPUs.
6600XTs seem to be going for around £200, often £180 even (used, eBay).
If you’d prefer new, you can get a 6650XT for £240. A 6650XT will be 6% faster than a 6600XT.
It’s double the performance of a 580, uses less power, will be supported longer, etc.
That’s a decent price, might have to go for one of those instead… Thanks for the information!
Been waiting for a good deal to replace my rx480 in my sister’s rig. I think they announced rx400/500/vega GPUs will only get security driver updates now and only for a while, I assume that applies to Linux too. RX580 will play many games at 1080p 60fpd but not the modern demanding ones (maybe not even at low settings).
Rumors say nextgen AMD isn’t targeting high end, maybe we have another 480 price to performance king 🤞. Then again, with AI as the new crypto, who can say.
I upgraded from an RX 480 to an RTX 3060 a few days ago. Crazy difference, especially in compute
Well, you are going from AMD to Nvidia, so there is a significant upgrade just in that. When I did my switch, I swore never to go back to AMD Gpu’s. But also going to a much more modern card than an almost 8 year old one would make anyone’s rig feel better. Glad you have a good card now!
This is the best summary I could come up with:
The performance gains were small, and a drop from 12GB to 8GB of RAM isn’t the direction we prefer to see things move, but it was still a slightly faster and more efficient card at around the same price.
In all, 2023 wasn’t the worst time to buy a $300 GPU; that dubious honor belongs to the depths of 2021, when you’d be lucky to snag a GTX 1650 for that price.
But these numbers were only possible in games that supported these GPUs’ newest software gimmick, DLSS Frame Generation (FG).
The technology is impressive when it works, and it’s been successful enough to spawn hardware-agnostic imitators like the AMD-backed FSR 3 and an alternate implementation from Intel that’s still in early stages.
And DLSS FG also adds a bit of latency, though this can be offset with latency-reducing technologies like Nvidia Reflex.
But to put it front-and-center in comparisons with previous-generation graphics cards is, at best, painting an overly rosy picture of what upgraders can actually expect.
The original article contains 787 words, the summary contains 168 words. Saved 79%. I’m a bot and I’m open source!