Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Collapse
Brand Logo

agnos.is Forums

  1. Home
  2. Ask Lemmy
  3. What's a good local and free LLM model for Windows?

What's a good local and free LLM model for Windows?

Scheduled Pinned Locked Moved Ask Lemmy
asklemmy
7 Posts 5 Posters 0 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • C This user is from outside of this forum
    C This user is from outside of this forum
    [email protected]
    wrote on last edited by [email protected]
    #1
    This post did not contain any content.
    toes@ani.socialT T 2 Replies Last reply
    7
    • C [email protected]
      This post did not contain any content.
      toes@ani.socialT This user is from outside of this forum
      toes@ani.socialT This user is from outside of this forum
      [email protected]
      wrote on last edited by
      #2

      The OS isn't as important as the hardware being used.

      AMD, Nvidia or Intel GPU?

      How much RAM & vram are you working with?

      What's your CPU?

      Generally speaking I would suggest koboldcpp with gemma3.

      https://github.com/LostRuins/koboldcpp?tab=readme-ov-file#windows-usage-precompiled-binary-recommended

      https://huggingface.co/mlabonne/gemma-3-27b-it-abliterated-GGUF/blob/main/gemma-3-27b-it-abliterated.q6_k.gguf

      occultist8128@infosec.pubO 1 Reply Last reply
      3
      • C [email protected]
        This post did not contain any content.
        T This user is from outside of this forum
        T This user is from outside of this forum
        [email protected]
        wrote on last edited by
        #3

        Use the app Jan, it's a good one and has a place where it recommends models for your hardware

        1 Reply Last reply
        0
        • toes@ani.socialT [email protected]

          The OS isn't as important as the hardware being used.

          AMD, Nvidia or Intel GPU?

          How much RAM & vram are you working with?

          What's your CPU?

          Generally speaking I would suggest koboldcpp with gemma3.

          https://github.com/LostRuins/koboldcpp?tab=readme-ov-file#windows-usage-precompiled-binary-recommended

          https://huggingface.co/mlabonne/gemma-3-27b-it-abliterated-GGUF/blob/main/gemma-3-27b-it-abliterated.q6_k.gguf

          occultist8128@infosec.pubO This user is from outside of this forum
          occultist8128@infosec.pubO This user is from outside of this forum
          [email protected]
          wrote on last edited by
          #4

          What's the minimum requirements for running it?

          toes@ani.socialT 1 Reply Last reply
          2
          • occultist8128@infosec.pubO [email protected]

            What's the minimum requirements for running it?

            toes@ani.socialT This user is from outside of this forum
            toes@ani.socialT This user is from outside of this forum
            [email protected]
            wrote on last edited by
            #5

            Lots of RAM and a good cpu, benefits from cores. if you're comfortable with it being on the slow side.

            There's other versions of that model optimized for lower vram conditions too.

            But for better performance 8GB of vram minimum.

            S 1 Reply Last reply
            1
            • toes@ani.socialT [email protected]

              Lots of RAM and a good cpu, benefits from cores. if you're comfortable with it being on the slow side.

              There's other versions of that model optimized for lower vram conditions too.

              But for better performance 8GB of vram minimum.

              S This user is from outside of this forum
              S This user is from outside of this forum
              [email protected]
              wrote on last edited by [email protected]
              #6

              Do you have a recommendation for Nvidia RTX 3070ti 8GB, ryzen 5600x +16GB DDR4? Does it even make sense to use it? Last time I tried the results were petty underwhelming.

              toes@ani.socialT 1 Reply Last reply
              1
              • S [email protected]

                Do you have a recommendation for Nvidia RTX 3070ti 8GB, ryzen 5600x +16GB DDR4? Does it even make sense to use it? Last time I tried the results were petty underwhelming.

                toes@ani.socialT This user is from outside of this forum
                toes@ani.socialT This user is from outside of this forum
                [email protected]
                wrote on last edited by
                #7

                Try it with this model, using Q4_K_S version.

                https://huggingface.co/bartowski/mlabonne_gemma-3-12b-it-abliterated-GGUF

                You'll probably need to play with the context window size until you get an acceptable level of performance. (Likely 4096)

                Ideally you'd have more RAM, but I want to say this smaller model should work. Koboldcpp will try to use both your GPU and CPU to run the model.

                1 Reply Last reply
                1
                Reply
                • Reply as topic
                Log in to reply
                • Oldest to Newest
                • Newest to Oldest
                • Most Votes


                • Login

                • Login or register to search.
                • First post
                  Last post
                0
                • Categories
                • Recent
                • Tags
                • Popular
                • World
                • Users
                • Groups