A developer tells this anecdote from a project from the PS2 era. They showed a pretty late build of the game to their publisher, very few weeks before the game had to be ready to begin the distribution process, and the FPS appeared at a corner of the screen, never quite falling below 30FPS, but often making important jumps. The people from the publishing company said that “Everything about the game looks fine, except for the FPS. 30 FPS is unacceptable, and we cannot publish it if you can’t reach a consistent 60 FPS”.
You don’t need to know much about development to understand that making such a demand weeks before the launch of a medium-sized project is asking the impossible, and so did this dev team. In the end, they changed the function that took the real calculation of the FPS so that it returned a fraction of the difference between 60 and the real FPS, so the next time the publisher took a look at the game, the FPS counter would always show a value between 58 and 60, even though they didn’t have the time to really optimize the game. The publisher didn’t notice the deception and the game was a commercial success.
The game was called “Tetris” btw
deleted by creator
30 is unplayable? I guess I haven’t played all these games then.
deleted by creator
Did you not see the part where it said “PS2 era”? The turning rates of the player/camera in any console first person shooter from those times are downright sluggish by mouse/keyboard standards, but the games were also designed around that slower pace.
deleted by creator
Then why tf are you even commenting?? Some people smh
deleted by creator
Meh. 60 is enough for me. I didn’t notice 144 being that much better than 60.
30 can fuck right off though.I can go down to 30 probably a bit lower as long as it is consistant, that is the most important part.
It can also have a bit to do with me powering through Watch Dogs at 1 fram per second in some parts. You never notice how good 25-30 is until your frames starts camping in the singel digits.
I remember playing Assassins Creed II on pc with a 9500GT and getting sub 20fps constantly to the point I had to wait for character animations to catch up with the dialogue so the next person could talk. Halfway through the game I upgraded to a GTX 560 and was astounded that everything was in sync and oh so smooth. I always remember that when I start getting annoyed I can’t get over 90fps in a game. As long as it’s playable!
It pretty much only makes a difference in FPS games where you’re constantly switching back and forth between crosshairs focus and peripheral vision flick reactions. At 144Hz, motion blur between frames is largely eliminated, so you have more accurate flicks and your vision at the crosshairs is much sharper.
FPS, and also just anything in general where the camera is panning quickly, such as character-centered 2d games
You can’t play racing games if you believe that. I’d far rather play an FPS at 30 than a racing game at 60. Low frame rates can give me motion sickness at high camera speeds.
I can mostly notice the difference in first person shooters. In most other things, 60 is plenty.
All you FPS kids are just doing the new version of “eww that game has 2d graphics; polygons or bust!!” from the PlayStation era.
Yes, progress is cool and good, but no it’s not the end-all be-all and no not every game has to have bleeding edge FPS to be good.
Like, we’re literally already done this shit guys; can’t we just learn from the past?
My brother or sister in pixels, this is not the same. I’m not a graphics snob. I still play pixelated, barely discernible nonsense games. When I updated from 30 to 144, it was a whole new world. Now even 60 can feel sluggish. This is not a graphical fidelity argument. It’s input and response time and motion perception. Open your mind, man. Accept the frames.
And that matters for certain games, a lot. But it doesn’t functionally matter at all for others. Same as the transition to polygons. My point, which I thought I stated clearly, was not “FPS BAD!!”, it was “FPS generally good, but stop acting like it’s the single most important factor in modern gaming.”
Simply put, if everything was 144fps then it would be easier on the eyes and motions would feel more natural. Even if it’s just navigating menus in a pixel style game.
Real life has infinite frames per second. In a world where high fps gaming becomes the norm, a low 24 fps game could be a great art style and win awards for its ‘bold art direction’.
Real life has infinite frames per second.
Not really. Real life is as many FPS as your eyes can perceive, which is about 60 (though it can vary somewhat between people). See: https://www.healthline.com/health/human-eye-fps#how-many-fps-do-people-see
That article states people can perceive images as rapidly as once every 13 milliseconds, which they math out to 75 fps, 25% higher than 60.
Looking at the study itself, they were testing whether participants could pick out a picture that displayed for 13-80 ms when “masked” by other brief pictures, with a focus on whether it made a difference if the participant was told what image they were looking for before or after seeing the images. What they found was that participants could pick out the image as low as the 13 ms mark (albeit with less accuracy) and could generally do so better if told what to look for beforehand.
What this tells me is that your source has nothing to say about anything over 75 fps. It also was testing in a fundamentally different environment than a video game, where your brain will constantly expect an image similar to and stemming from the image before it rather than seeing a completely different image. If you were to draw conclusions based on the study despite the differences, what the study would suggest is that knowing what to look for, as your brain does gaming, would make you better able to pick out individual frames. This makes me want to think that your source does not support your assertion, and that in a game you could perceive frame rates higher than 75 fps at a minimum.
From my own knowledge, there’s also a fundamental difference between perceiving reality and computer screens in the form of motion blur. Objects moving in real time will leave a faint blur behind when perceiving it that your brain can use to fill in any blanks it may have missed, making reality appear smoother than it is. For an example of this wobble a pencil back and forth to make it “bend.” Movies filmed at 24 fps capture this minute motion blur as they film which makes it easier for our brains to watch them despite the lower frame rate. Real time rendered video games do not have this effect, as there are no after images to fill in the blanks (unless you turn on motion blur, which doesn’t do a good job emulating this).
This means video games need to compensate, and the best way to do that is more frames per second so your brain doesn’t need to fill in the blanks with the motion blur it’s used to seeing in the real world. You’ll obviously get diminishing returns from the same increase, but there will still be returns.
Yeah, as much as I can give a shit about ray tracing or better shadows or whatever, as a budget gamer, frame rate is really fucking me up. I have a very low end PC so 60 is basically max. Moving back to 30 on the PS4 honestly feels like I’m playing PS2. I had the [mis]fortune of hanging out at a friends house and playing his PC rig with a 40 series card, 240hz monitor, etc, and suffice it to say it took a few days before I could get back to playing on my shit without everything feeling broken.
Now even 60 can feel sluggish.
That’s more or less the placebo effect at work, though. Most people cannot see “faster” than 60FPS; the only actual upside of running higher FPS rate is that you don’t go below 60 in case the game starts to lag for whatever reason. Now, you may be one of the few who actually see perceive changes better than normal, but for the vast majority, it’s more or less just placebo.
That’s more or less the placebo effect at work, though. Most people cannot see “faster” than 60FPS;
You can literally see the difference between 60 and 144 when moving the cursor or a window on your desktop. What are you on about
That’s just wrong. I couldn’t go back to my 60Hz phone after getting a 120Hz new one. It’s far from placebo, and saying otherwise is demonstrably false.
bleeding edge FPS
Do you think 60 FPS just got released or something
The discussion is about 144
144 FPS isn’t even bleeding edge, there are monitors with refresh rates higher than that.
One of most insufferable aspects of video game culture (pc gaming in particular) other than the relentless toxic masculinity from insecure nerds is an obsessive focus on having powerful hardware and shitting on people who think they are getting a good experience when they don’t have good hardware.
The point is to own a computer that other people don’t have so you can play a game and get an experience other people don’t have, the point isn’t to celebrate a diversity of gaming experiences and value accessibility for those without the money for a nice computer. It really doesn’t matter if these people are intending to do this consciously or not, this is a story as old as time. It is the same exact bullshit as guitar people who only think special exotic or vintage guitars are beautiful, claim to absolutely love guitar but never once in their life have stopped to think about how much more beautiful it is that any random chump can get an objectively wonderful sounding guitar for a couple of hundred dollars than it is that they own some stupid special edition guitar with a magic paint job that cost as much as my shitty car.
Good thing these people don’t fully dictate the flow of all of video game development, but they will never ever learn because this is the kind of pattern that arises not from conscious intention but rather from people uninterested in critically examining their own motivations.
It is the same damn nauseating thing with photography too….
I’d rather not get motion sickness thanks.
it depends on if its a good 30 or not. if inputs are quick and responsive, and the framerate stays at 30 then its fine. but if my device is struggling to run the game and its stuttering and unresponsive then its awful
sm64 comes to mind as the best 30fps experience ive had, and i am spoiled rotten on high refresh rate games
I get it’s a meme but I usually play on 144fps and when I go back to 60 fps I literally don’t notice a difference even down to like 40-45 I barely see much difference. 30 is noticeable and a bit shit, but my eyes get used to it after like 30 mins, so not a big deal.
For me frame time inconsistency is the most noticeable. FPS, as long as it’s consistent 40 and above is fine.
I will notice the difference in fluidity of motion but a large frame time difference destroys the experience.
I think it takes significantly more mental extrapolation between frames and general adjustment to your eyes not receiving frames at as quick of a rate but if the frame rate is fairly stable the human brain adapts.
The brains visual processing is so powerful that the difference between 30fps and 144fps on paper is much smaller in reality, especially if your brain has already learned the muscle memory of “upscaling” a low framerate to work with its perception of a 3D environment.
Competitively, for games like arena shooters or rocket league the frame rate is real but for most games it is a matter of the smoothness occurring on the physical monitor screen or it occurring on some level of mental image processing. What someone sees who has let the mental skill of processing a lower frame rate atrophy is a temporary sensation like putting on colored glasses for awhile, then taking them off and seeing everything washed out in a particular color. Weird, uncomfortable, but temporary.
The real problem is inconsistent framerate where the clock your brain has gotten used to receiving new visual information with the arrival of each new frame of visual information is slow enough to be on the edge of perception but keeps speeding up or slowing down chaotically. Your brain can’t just train a static layer of mental image processing to smoothen that out. Time almost feels like it is slowing up and speeding down a little, and it becomes emotionally discouraging that every time something fun happens the framerate dips and reality becomes choppier.
It seems to be dependent on the game for me. Some just seem like they should move smoother (like Akham Asylum on PC)
Yeah no, a game I regularly play just had 120Hz support added, and I’m never changing that back. I once even tried editing the .ini config just to change the framerate after coming back to it from a 120Hz game. It just is night and day, both in the input lag and the smoothness of the image
🤓
My monitor is 240hz and counter strike runs about 400fps max settings at 1080p on my new system. It was absolutely insane playing for the first time coming from my steam deck running 30-40fps on lowest settings.
Wait until you experience 500
Every time I get a taste for something better, things get more expensive… I’m going to avoid trying anything higher than 144 for a while.
It’s for the best that you do that.
Sincerely, someone who “had” to buy an RTX 4080 after buying a new 200 Hz monitor.
Oh. Fuck…you got the Odyssey G9 as well?
In order for me to even taste the sweet potential of that monitor, I’m having to build a whole new computer. I’m dreading it.
The best advice I can give you is to turn off the FPS counter. If the game feels like it’s stuttering, turn down the quality. If it feels fine during gameplay, don’t fuck with it, and under no circumstances should you enable an FPS counter or frame timing graph.
If you’re anything like me and you do enable the FPS counter or frame timing graphs, you’ll spend more time optimizing performance than actually enjoying the game…
LOL I like how they mention one spec of the monitor and you instantly know which one they bought.
I just built a new PC with a Ryzen 7800X3D CPU and a Radeon RX 7900 XTX GPU, loving it so far. 400fps with Counter-Strike 2 at max settings 1080p
Wow, you went all out on that. What chipset? X670, X670E, B650, or B650E?
B650, wasn’t planing on OC and figured better to put the money to the CPU and GPU
Great pick. The X670 and E variants are insanely overpriced for what they provide. People don’t need PCIE 5 when there aren’t even any non-SSD components that use it.
I made an awful mistake of getting a 1440p OLED instead of the 540 and now a TN panel is going to be very difficult to get used to.
People always go on and on about framerate but I’d take 4k at 60fps over 1080p at 144fps any day. I never really noticed a differance over 60fps. But the resolution makes a massive difference.
We are opposite people (and that’s fine).
1440p, stable 144. The dream
For me it depends on the game. A menu game from Paradox like Crusader Kings? 4k 60fps. A competitive shooter? Ideally the max resolution (for greater pinpoint accuracy) and 144fps, but between the two I’d want maximum fps for the reaction speed and responsiveness. A pretty game like Ori and the Will of the Wisps? Crank the graphics up and I’m happy with 60fps.
They both make a difference. 1080p is console TV gaming minimum resolution. A good PC can do much better. I like my ultra-widescreen 144Hz monitor more than a plain boring 16x9 high-res monitor.
But even games at 3440x1440 @ 144Hz don’t look as good as full-on VR at 120Hz and up. VR gaming for much time will make gaming on a monitor feel obsolete.
Me who play Minecraft at 20 fps because I installed a modpack called FTB :)
I’ve been there lmao
Lately tho I’ve been using fabric mods and damn is it optimized
Like it’s frustrating cause forge has a much much larger mod selection But it just so slowww
Who plays games at 30fps? I’m fairly sure 60 is industry standard now no?
Cries in nintendo
Baldurs gate 3 split screen is 30fps
For games like that it doesn’t matter so much.
If Call of Duty ran at 30 they’d never hear the last of it.
If you’re lucky 😭
Oh fair enough, great example thanks
Hello Doom 1/2 and its 35 FPS by default!
Even most console games run at 60 now, with an option to turn on some RT graphical wankery and run at 30.
I often turn it on to see what it looks like, and then decide it’s not worth it. Ratchet and Clank actually played decently at 30, and one of the Ghostwire Tokyo options allowed you to have RT and decent framerates with a minor hit to resolution.
Gsync/Freesync/VRR is a game changer for lower end hardware, because then all those dips below 60 get smoothed out to an even 45 or so. I’ve spent a lot less time fucking about with setting on PC since getting a monitor that supports that.
Meanwhile I pull out my old CRT slap in my N64 and play just like I would as a kid.
Y’all’s obsessions with graphics and frames is weird to me.
No hate, just confusion.
Bro a N64 plays most games at 60fps
Screen tearing is the real enemy
CRTs can easily push >120 fps. That’s why the super sweaty CS esports players still used them well after the first LCD panels came out.
same, gameplay matters more :)
I still don’t get the obsession over 144hz. I saw a big jump from 30 to 60 and I also see the jump from 60 to 144, but I’m happy with 60. 30 genuinely feels off to me though, but even then I can get used to it in like 10mins.
You really start to notice it in fps games where quick reaction time matters
deleted by creator
the human eye cant see more than 15 fps you’re lying
Actually a study proved that our eyes can’t see any faster than 1
Nope. Nobody is lying, you’re just wrong :)
Can’t your eye only see like 30 frames per second in the center?
Our eyes and brains don’t perceive still images or movement in the same way as a computer. There is no simple analogy between our perception and computer graphics.
I’ve read that some things can be perceived at 1000 fps. IIRC, it was a single white frame shown for 1ms between black frames. Of course most things you won’t be able to perceive at that speed, but it certainly isn’t as simple as 30 fps!
The human brain evolved to recognize threats in the wilderness.
We see movement and patterns very well because early hominid predators were very fast and camouflaged, so seeing the patterns of their fur and being able to react to sudden movements meant those early people didn’t die.
But evolution doesn’t optimize. Things only evolve up to the point where something lives long enough to reproduce. Maybe over extremely long time spans things will improve if they help find mates, but that is all evolution does.
Your brain perceives things fast enough for you not to get eaten by a tiger. How fast is that? Who the fuck knows.
All the being said, I like higher HZ monitors. I feel like I can perceive motion and react to things more quickly if the frame rate is higher. The smoother something looks, the more likely I feel that I can detect something like part of a character model rounding a corner. But no digital computer is ever going to have analog “frame times”, so any refresh rate you think feels comfortable is probably fine.
No. Iirc around 200hz is where you get diminishing returns.
No. This is something console fanboys used to spread up when pc gamets showed off with their +30fps games
That’s what I’ve heard. but also, the frequency of electricity in the USA is 60 Hz because Tesla found after experimentation that that’s the frequency where you don’t notice a lightbulb flickering anymore. Since the lightbulb flickers 120 times per second at 60 Hz, you could assume that a lower framerate than 120 fps is noticable.
Technically yes, but the more fluid the video is 8n the first place, the fewer gaps your brain has to fill in. On 30 fps you can see the moving image just fine, but your brain is always assembling the pieces and ignoring the gaps. The higher framerates reduce the number of gaps and makes a surprising difference in how smooth something looks in motion.
So you just explained why it’s actually a no.
Also most monitors only go up to 60fps, and even if you have a fancy monitor that does, your OS probably doesn’t bother to go higher than 60 anyways. Even if the game itself says the fps is higher, it just doesn’t know that your pc/monitor isnt actually bothering to render all the frames…
my man, just because you’ve never seen the refresh rate option in the monitor settings doesn’t mean it hasn’t been there since basically forever
This is blatantly false.
Windows will do whatever frame rate the EDID reports the display as being capable of. It won’t do it by default, but it’s just a simple change in the settings application.
Macs support higher than 60 Hz displays these days, with some of the laptops even having a built-in one. They call it by some stupid marketing name, but it’s a 120 Hz display.
Linux requires more tinkering with modelines and is complicated by the fact that you might either be running X or Wayland, but it’s supported as well.
To add on to this. There are phones coming out now with 90+hz screens. They are noticably smoother than the 60hz ones. My current phone does 120hz.
Yeah the OS can and will shove out frames as fast as the hardware can support them
Wayland picks up my 155, 144, and 60 hz monitors and sets them to the correct refresh rate on it’s own nowadays, so it’s even more painless.
You just won the award for stupidest comment in the whole commentsection. That is just completely false and makes no sense in any way. Your computer doesn’t just skip calculations its told do do. Where did you even get this idea lmao
What the fuck? Help us understand: which OS doesn’t limit fps and what do you see when you check frame rates with your own eyes?
That was true before high framerate monitors were a thing, which was around 10+ years ago…
no it wasn’t true back then either, CRTs have been doing 100hz and more decades ago and it was very much supported by OSes and games