• Justin@lemmy.jlh.name
    link
    fedilink
    English
    arrow-up
    14
    arrow-down
    2
    ·
    7 months ago

    1000 Hz seems to be close to the limit human of human vision, since we stop seeing motion blur above 1000hz. Seems like a good endpoint for display technology.

    • Tetsuo@jlai.lu
      link
      fedilink
      English
      arrow-up
      16
      arrow-down
      2
      ·
      7 months ago

      On the other hand I heard so many times the same argument for 144hz, 165hz and then 240hz…

      Now 1000Hz is much higher frequency but in term of frame time it’s not that far. I wouldn’t be surprised if some people could successfully tell the difference in a “blind” test.

      • Justin@lemmy.jlh.name
        link
        fedilink
        English
        arrow-up
        5
        ·
        7 months ago

        I remember people talking about 1000hz being the holy grail for vr headsets, though so it seems like there’s more consensus on 1000 Hz being a good limit. Frame time is just the inverse of hz.

        But yeah ive personally only used 144hz, I think I could see a difference with 204hz, but I’m not sure if I’d be able to discern 480 or 1000hz outside of maybe VR.

        • wia
          link
          fedilink
          English
          arrow-up
          9
          ·
          7 months ago

          There kinda isn’t really any definitive science that indicates a specific frame rate that the eye can perceive.

          There are studies however that show ranges from 30 to 90hz, and studies that show that human perception can detect flicker at up to 500hz even.

          The issue is that nothing that happens in the real world is synchronized with what you perceive. So filling in with more Hz means there are more chance for you to actually perceive the thing.

          To complicate matters further, our brains do a lot of filling in for us, and our eyes and brains can still perceive things you aren’t consciously perceiving yourself. So again more frames is always nice.

          Here are some sources

          Canadian Centre for Occupational Health and Safety. (2020). Lighting Ergonomics - Light Flicker.
          https://www.ccohs.ca/oshanswers/ergonomics/lighting_flicker.html

          Davis J, et al. (2015). Humans perceive flicker artifacts at 500 Hz.
          https://doi.org/10.1038/srep07861

          Mills M. (2020). How Many Frames per Second (FPS) the Human Eye Can See.
          https://itigic.com/how-many-frames-per-second-fps-human-eye-can-see/

          • Justin@lemmy.jlh.name
            link
            fedilink
            English
            arrow-up
            1
            ·
            edit-2
            7 months ago

            Sure, eyes dont have a “global frame refresh” like computers do. That’s why we can tell the difference between 24hz and 60hz video. Every eye cell is excited independently and continuously.

            Still, there’s a physical limit for frame time where 99% of humans wouldn’t notice a full screen flash 99.9% of the time. Being able to shake your head around with a 1000hz vr headset and not perceive and motion blur from sample and hold seems pretty close to that limit.

            • Dultas@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              ·
              7 months ago

              That’s still going to depend on the frame rate of the output source. If the source is only 200fps having 1000hz isn’t going to matter a whole hell of a lot.

          • tal@lemmy.today
            link
            fedilink
            English
            arrow-up
            2
            arrow-down
            2
            ·
            edit-2
            7 months ago

            If you’re using a game that renders each frame at an instant in time, and the aim is to get a better approximation of true motion blur to your eye, the theoretical maximum for getting smoother motion blur is gonna be when the thing is moving at one pixel a second, which is higher than the rate at which we can distinguish between individual images. Well, okay, maybe a bit more, since you could hypothetically have sub-pixel resolution.

            But point is, more rendered frames does buy you something even past the point that they’re not individually distinguishable, unless the game’s rendering engine can render perfectly-accurate motion blur itself.

  • adam_y@lemmy.world
    link
    fedilink
    English
    arrow-up
    13
    arrow-down
    2
    ·
    7 months ago

    Best gaming of my life was on a SNES coupled to a janky sanyo portable.

      • Nibodhika@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        1
        ·
        7 months ago

        The title doesn’t ask if it’s useful, it asks if it’s required, considering that no one NEEDS a display to begin with, a 1000Hz display by definition cannot be NEEDED.

        I will likely get one eventually, just like I have a 165Hz display now, but do I NEED it? Absolutely no.

  • AutoTL;DR@lemmings.worldB
    link
    fedilink
    English
    arrow-up
    3
    ·
    7 months ago

    This is the best summary I could come up with:


    Now, the high frame rate experts at Blur Busters bring word of a 4K, 1,000 Hz prototype screen being shown off by Chinese panel maker TCL CSOT at the manufacturer-focused DisplayWeek 2024 conference.

    And while recent advancements in pixel-flipping times have enabled TCL’s LCD prototype, Blur Busters estimates that 1,000 Hz OLED displays could be commercialized as soon as 2027.

    The apparent, impending breaking of the four-digit refresh rate threshold got us thinking: Are we finally approaching a point of diminishing returns in monitor-makers’ long-running battle of the Hz?

    But even faster refresh rates could help with the apparent sharpness of extremely fast objects on Ultra HD displays—think of a mouse pointer or video game crosshair that can move across the roughly 4,000 horizontal pixels on a 4K display in a single second.

    Nvidia data shows the RTX 4090 generating 600+ fps frame rates on aging-but-still-popular games like Rainbow Six Siege and Fortnite at “High” settings and a full 1440p resolution.

    Factor in a few more generations of graphics card upgrades—not to mention frame-generation technology that could offer 10 reprojected frames for every key frame—and waiting a single millisecond for your monitor to show a new image might not be totally ridiculous.


    The original article contains 628 words, the summary contains 202 words. Saved 68%. I’m a bot and I’m open source!

  • Okami@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    ·
    7 months ago

    Nope. I can’t even tell the difference between 30Hz and 60Hz unless they’re actually running side by side.

  • Fushuan [he/him]@lemm.ee
    link
    fedilink
    English
    arrow-up
    3
    arrow-down
    2
    ·
    edit-2
    7 months ago

    I don’t! I barely notice 120Hz so I just run at 60, my GPU loves me for it.

    Imagine running a 4K 1000Hz screen, and needing 66 times more computing power to render all those pixels than my wee 1080p 60Hz screen where I see stuff fucking fine.

  • mindbleach@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    1
    ·
    7 months ago

    1000 Hz is going to be the end. Microsoft’s experiments, umpteen years ago, said that’s about the limit of human perception. Not “it looks smooth,” but “it looks indistinguishable from reality because our wetware can’t discern events much finer than that.” Mmmaybe you go a little above that, to avoid two-millisecond events, but beyond that there is literally no point. Shutterglasses, I guess. Niche silliness.

    More importantly - any framerate looks amazing. Same deal as G-Sync / FreeSync: frames appear onscreen the moment they are ready. There’s no stutter at 29 Hz, or at 239 Hz.

  • bstix@feddit.dk
    link
    fedilink
    English
    arrow-up
    2
    arrow-down
    5
    ·
    7 months ago

    I doubt any game logic is going to be that fast ever. It might look better, but it won’t make much difference in how players or even the game itself can react to events in the game. Perhaps it’ll make VR more comfortable or something.

    • DarkThoughts@fedia.io
      link
      fedilink
      arrow-up
      5
      arrow-down
      1
      ·
      7 months ago

      Game logic runs independently from what your monitor can display. So it’s really just a question on what effect it has on the player itself. Maybe for VR there’s an argument to be made, although I feel 1000 Hz still sounds like complete overkill even in that area. But I’m gonna call bullshit on people who claim to be able to tell the difference of such high rates.

      • burgersc12@mander.xyz
        link
        fedilink
        English
        arrow-up
        3
        ·
        edit-2
        7 months ago

        Game logic does not always run independent of the framerate. Look at Fallout 4, if you run it at more than 60fps the dialogue literally overlaps itself.

        • Belgdore@lemm.ee
          link
          fedilink
          English
          arrow-up
          4
          ·
          7 months ago

          That’s because Bethesda is bad at making games not because there is an intrinsic need for the game logic to be tied to frame rate.

          • burgersc12@mander.xyz
            link
            fedilink
            English
            arrow-up
            1
            ·
            7 months ago

            Used to be very common, but even Switch games today lock the framerate to 30/60fps or else it runs at 2x the speed it should

        • DarkThoughts@fedia.io
          link
          fedilink
          arrow-up
          1
          arrow-down
          1
          ·
          7 months ago

          I didn’t say framerate, I said from what your monitor can display. FPS and Hz are not synonymous.

            • DarkThoughts@fedia.io
              link
              fedilink
              arrow-up
              1
              arrow-down
              1
              ·
              7 months ago

              You’ve got it the wrong way around. People play very high FPS games on (comparatively) lower Hz monitors. This has been common practice in competitive pvp shooters for decades.

              • burgersc12@mander.xyz
                link
                fedilink
                English
                arrow-up
                1
                ·
                edit-2
                7 months ago

                This is my point. A 1000Hz screen would, most likely, be played at as close to 1000 fps as possible. I am not sure why you think i have it the wrong way when it is you.

                • DarkThoughts@fedia.io
                  link
                  fedilink
                  arrow-up
                  1
                  arrow-down
                  1
                  ·
                  7 months ago

                  Your point was that game logic doesn’t run independently from your framerate, trying to refute my comment saying that game logic runs independently from your monitor. You’re clearly severely confused about the topic at hand.

      • tal@lemmy.today
        link
        fedilink
        English
        arrow-up
        3
        ·
        edit-2
        7 months ago

        For rendered stuff, it typically does make for smoother motion, even at rates much higher than the eye can see, because of motion blur.

        So, recorded video works fine at relatively low bitrates…but the camera is also set up to record a relatively-long exposure, something like a thirtieth of a second, and you see the scene averaged over that time. Your brain can see motion blur and interpret that usefully, to know that there is motion happening.

        Rendered 3D game images typically do not work like that. You see a series of perfectly-sharp images at instants in time. So your brain doesn’t get the nice smooth motion blur to work with.

        But if your computer renders and displays the intermediate images, then your eye can work with that nice smooth blur.

        It’s probably possible to compute a motion-blur more efficiently than rendering a lot of intermediate frames, get at least some kind of approximation of true motion blur, and some games do that, but brute-force rendering of more frames is simple for s developer and accurate. Plus, any game that can support a high frame rate can do it, even if it doesn’t have some kind of faux motion blur approximation.

        I have a 165Hz monitor. When moving my mouse cursor around, I can definitely see independent images of the cursor.

        EDIT: That being said, you could probably get a pretty good approximation by rendering and combining multiple frames on the card and only pushing a lower frame rate out to the monitor – that is, you only really need beefy rendering hardware, not a fancy monitor or cable, to get pretty close. I suppose that in theory, a compositor could do that. I don’t know if someone’s already done that or not.