Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Collapse
Brand Logo

agnos.is Forums

  1. Home
  2. LocalLLaMA
  3. [April 2025] Which model are you using?

[April 2025] Which model are you using?

Scheduled Pinned Locked Moved LocalLLaMA
localllama
14 Posts 7 Posters 0 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • B [email protected]

    I'm using this one because before they ceased the Open LLM Leaderboard, it was the highest rated 14B model that can run on a single GPU with 10GB VRAM.

    E This user is from outside of this forum
    E This user is from outside of this forum
    [email protected]
    wrote on last edited by
    #4

    Newbie here. I'm not sure if the documentation tells me if it can run with ollama. If I understand correctly you have to build it «by hand»? I mainly use ollama/models on the official website and I'm too scared to plunge deeper into the mechanics haha.

    B 1 Reply Last reply
    1
    0
    • eyekaytee@aussie.zoneE [email protected]

      Just thinking about making this a monthly post, which model are you using? what are the positives and negatives?

      E This user is from outside of this forum
      E This user is from outside of this forum
      [email protected]
      wrote on last edited by
      #5

      I find that for the purpose of my projects (narrative building, tabletop rpg simulation) gemma3:14b (with low temperature) works perfectly to create consistent psychological overviews.

      1 Reply Last reply
      1
      0
      • E [email protected]

        Newbie here. I'm not sure if the documentation tells me if it can run with ollama. If I understand correctly you have to build it «by hand»? I mainly use ollama/models on the official website and I'm too scared to plunge deeper into the mechanics haha.

        B This user is from outside of this forum
        B This user is from outside of this forum
        [email protected]
        wrote on last edited by
        #6

        Not for GGUF converted models.

        Just run the following command in ollama

        ollama run hf.co/wanlige/li-14b-v0.4-Q4_K_M-GGUF
        

        I also recommend you set up Open WebUI as your front-end instead of staying in a terminal.

        Also, look into "abliterated" uncensored models. There's plenty of them on hf that are already converted for you

        E 1 Reply Last reply
        1
        0
        • B [email protected]

          Not for GGUF converted models.

          Just run the following command in ollama

          ollama run hf.co/wanlige/li-14b-v0.4-Q4_K_M-GGUF
          

          I also recommend you set up Open WebUI as your front-end instead of staying in a terminal.

          Also, look into "abliterated" uncensored models. There's plenty of them on hf that are already converted for you

          E This user is from outside of this forum
          E This user is from outside of this forum
          [email protected]
          wrote on last edited by
          #7

          That's awesome thank you.

          1 Reply Last reply
          1
          0
          • eyekaytee@aussie.zoneE [email protected]

            Just thinking about making this a monthly post, which model are you using? what are the positives and negatives?

            smokeydope@lemmy.worldS This user is from outside of this forum
            smokeydope@lemmy.worldS This user is from outside of this forum
            [email protected]
            wrote on last edited by
            #8

            I have been using deephermes daily. I think CoT reasoning is so awesome and such a game changer! It really helps the model give better answers especially for hard logical problems. But I don't want it all the time especially on an already slow model. Being able to turn it on and off wirhout switching models is awesome. Mistral 24b deephermes is relatively uncensored, powerful and not painfully slow on my hardware. a high quant of llama 3.1 8b deephermes is able to fit entirely on my 8gb vram.

            1 Reply Last reply
            1
            0
            • eyekaytee@aussie.zoneE [email protected]

              Just thinking about making this a monthly post, which model are you using? what are the positives and negatives?

              A This user is from outside of this forum
              A This user is from outside of this forum
              [email protected]
              wrote on last edited by
              #9

              QWQ-32B for most questions, llama-3.1-8B for agents. I'm looking for new models to replace them though, especially the agent one.

              Want to test the new GLM models, but I'd rather wait for llama.cpp to definitely fix the bugs with them first.

              W eyekaytee@aussie.zoneE 2 Replies Last reply
              1
              0
              • eyekaytee@aussie.zoneE [email protected]

                Just thinking about making this a monthly post, which model are you using? what are the positives and negatives?

                W This user is from outside of this forum
                W This user is from outside of this forum
                [email protected]
                wrote on last edited by
                #10

                Fallen Gemma. The writing style is really good and it can keep relatively persistent personalities. On the other hand it's stupid af compared to other recent models and even the vanilla Gemma 3.

                1 Reply Last reply
                1
                0
                • A [email protected]

                  QWQ-32B for most questions, llama-3.1-8B for agents. I'm looking for new models to replace them though, especially the agent one.

                  Want to test the new GLM models, but I'd rather wait for llama.cpp to definitely fix the bugs with them first.

                  W This user is from outside of this forum
                  W This user is from outside of this forum
                  [email protected]
                  wrote on last edited by
                  #11

                  GLM? I feel like every other day there is a new abbreviation 😞

                  1 Reply Last reply
                  1
                  0
                  • opticalmoose@discuss.tchncs.deO [email protected]

                    I mainly use Llama-3-8B abliterated for everyday questions, and DeepSeek-Coder-V2-Lite for programming/Linux stuff.

                    eyekaytee@aussie.zoneE This user is from outside of this forum
                    eyekaytee@aussie.zoneE This user is from outside of this forum
                    [email protected]
                    wrote on last edited by
                    #12

                    Using DeepSeek-Coder-V2-Lite now, it's awesome!

                    1 Reply Last reply
                    1
                    0
                    • A [email protected]

                      QWQ-32B for most questions, llama-3.1-8B for agents. I'm looking for new models to replace them though, especially the agent one.

                      Want to test the new GLM models, but I'd rather wait for llama.cpp to definitely fix the bugs with them first.

                      eyekaytee@aussie.zoneE This user is from outside of this forum
                      eyekaytee@aussie.zoneE This user is from outside of this forum
                      [email protected]
                      wrote on last edited by
                      #13

                      Want to test the new GLM models

                      Which models are you referring to? These: https://github.com/THUDM/GLM-4 ?

                      A 1 Reply Last reply
                      1
                      0
                      • eyekaytee@aussie.zoneE [email protected]

                        Want to test the new GLM models

                        Which models are you referring to? These: https://github.com/THUDM/GLM-4 ?

                        A This user is from outside of this forum
                        A This user is from outside of this forum
                        [email protected]
                        wrote on last edited by
                        #14

                        That's the ones, the 0414 release.

                        1 Reply Last reply
                        1
                        0
                        • System shared this topic on
                        Reply
                        • Reply as topic
                        Log in to reply
                        • Oldest to Newest
                        • Newest to Oldest
                        • Most Votes


                        • Login

                        • Login or register to search.
                        • First post
                          Last post
                        0
                        • Categories
                        • Recent
                        • Tags
                        • Popular
                        • World
                        • Users
                        • Groups