• Malix@sopuli.xyz
      link
      fedilink
      arrow-up
      28
      ·
      edit-2
      9 months ago

      Hastily read around in the related issue-threads and seems like on it’s own the vm.max_map_count doesn’t do much… as long as apps behave. It’s some sort of “guard rail” which prevents processes of getting too many “maps”. Still kinda unclear what these maps are and what happens is a process gets to have excessive amounts.

      That said: https://access.redhat.com/solutions/99913

      According to kernel-doc/Documentation/sysctl/vm.txt:

      • This file contains the maximum number of memory map areas a process may have. Memory map areas are used as a side-effect of calling malloc, directly by mmap and mprotect, and also when loading shared libraries.
      • While most applications need less than a thousand maps, certain programs, particularly malloc debuggers, may consume lots of them, e.g., up to one or two maps per allocation.
      • The default value is 65530.
      • Lowering the value can lead to problematic application behavior because the system will return out of memory errors when a process reaches the limit. The upside of lowering this limit is that it can free up lowmem for other kernel uses.
      • Raising the limit may increase the memory consumption on the server. There is no immediate consumption of the memory, as this will be used only when the software requests, but it can allow a larger application footprint on the server.

      So, on the risk of higher memory usage, application can go wroom-wroom? That’s my takeaway from this.

      edit: ofc. I pasted the wrong link first. derrr.

      edit: Suse’s documentation has some info about the effects of this setting: https://www.suse.com/support/kb/doc/?id=000016692

      • psycho_driver@lemmy.world
        link
        fedilink
        arrow-up
        15
        ·
        edit-2
        9 months ago

        The default value is 65530

        Just checked and the steam deck has it set to 2147483642. My gentoo systems are 65530.

        • Malix@sopuli.xyz
          link
          fedilink
          arrow-up
          9
          arrow-down
          1
          ·
          9 months ago

          On one hand, I’d assume Valve knows what they’re doing, but also setting the value that high seems like it’s effectively removing the guardrail alltogether. Is that safe, also what is the worst that can happen if an app starts using maps in the billions?

          • sugar_in_your_tea@sh.itjust.works
            link
            fedilink
            arrow-up
            2
            ·
            9 months ago

            The whole point is to prevent one process from using too much memory. The whole point of the Steam Deck is to have one process use all the memory.

            So it makes sense to keep it relatively low for servers where runaway memory use is a bug that should crash the process, but not in a gaming scenario where high memory usage is absolutely expected.

      • sugar_in_your_tea@sh.itjust.works
        link
        fedilink
        arrow-up
        2
        ·
        9 months ago

        My read is that it matters for servers where a large number of allocations could indicate a bug/denial of service, so it’s better to crash the process.

        That’s not relevant on a gaming system, since you want one process to be able to use all the resources.

    • unexposedhazard
      link
      fedilink
      arrow-up
      11
      ·
      9 months ago

      It changed it for playing “the finals” some weeks ago to fix a crash. I havent had any issues with my system since then so it really might just be some value that never changed because nothing needed it.

    • umbrella@lemmy.ml
      link
      fedilink
      arrow-up
      4
      ·
      9 months ago

      also want to know, i increased this value by a lot for gaming and have been using it ever since with no visible repercussions.