Smorty [she/her]

I’m a person who tends to program stuff in Godot and also likes to look at clouds. Sometimes they look really spicy!
image of Biene Maja
image of cutie fluttershy

  • 22 Posts
  • 91 Comments
Joined 3 years ago
cake
Cake day: July 5th, 2023

help-circle




  • i totally agree… with everything. 6GB really is smol and, cuz imma crazy person, i currently try and optimize everything for llama3.2 3B Q4 model so people with even less GB VRAM can use it. i really like the idea of people just having some smollm laying around on their pc and devs being able to use it.

    i really should probably opt for APIs, you’re right. the only API I ever used was Cohere, cuz yea their CR+ model is real nice. but i still wanna use smol models for a smol price if any. imma have a look at the APIs you listed. Never heard of Kobold Horde and Samba so i’ll have a look at those… or i go for the lazy route and chose depseek cuz it’s apparently unreasonably cheap for SOTA perf. so eh…

    also yes! Lemmy really does seem anti AI, and i’m fine with that. i just say yeah companies use it in obviously dum ways but the tech is super interesting which is a reasonable argument i think.

    so yes, local llm go! i wanna get that new top amd gpu once that gets announced. so i’ll be able to run those spicy 32B models. for now i’ll just stick with 8B and 3B cuz they work quick and kinda do what i want.











  • I actually never heard of this saying, but I just looked up. Woah, that’s really a phrase they use internally, hm? Crazy.

    And it does accurately describe what they try to do here. It can’t really work like that, since many people use GNU/Linux specifically because it’s not owned by EvilCo™. But they could probably take over some part of the server-hosting business like this. And that is a scary thought.

    Imagine, they could make it super easy to deploy things by incoperating premade docker containers into their UI thingy. That’s - like - real bad.


  • Small addition:
    Now that VR works essentially perfectly on GNU/Linux, even on Wayland with Gnome and an nvidia GPU, I have now stopped dualbooting for occasional VR Chat and Beat Saber (which are VR games).

    In my opition, when looking away from online games with anticheat, Microsofts Office and adobes whatever software, there is no reason to use Winblows anymore.

    The amount of configuration GNU/Linux gives me is truly empowering, running any scripts I want using shortcuts being a big one for me.

    Some shortcuts I use daily
    Super+E -> Nautilus (obvious)
    Super+W -> Firefox
    Super+Y -> Youtube
    Super+C -> Local LLM chat
    Super+G -> Launch Godot
    Generally vim navigation
    











  • I’m even more excited for running 8B models at the speed of 1B! Laughably fast ok-quality generations in JSON format would be crazy useful.

    Also yeah, that 7B on mobile was not the best example. Again, probably 1B to 3B is the sweetspot for mobile (I’m running Qwen2.5 0.5B on my phone and it works tel real for simple JSON)

    EDIT: And imagine the context lengths we would be ablentonrun on our GPUs at home! What a time to be alive.