• 31 Posts
  • 2.75K Comments
Joined 1 year ago
cake
Cake day: March 22nd, 2024

help-circle
  • “Easy Presets” are a huge draw for users.

    I’ve seen (non gaming) frameworks live or die by how well they work turnkey, out of the box with zero config edits other than the absolute bare minimum to function. Even if configuration literally takes like half an hour or something and the framework has huge performance gains over another, that first impression is a massive turn off to many.

    It’s… not that people are lazy, but they’re human. Attention is finite. If realistic lighting isn’t good in Godot by default, then it needs a big red intro button that says “Click here for realistic lighting!”






  • “Christian Mission US Border crossings and family flipping”

    “Instructional Video: Her first time with a 7yo model”

    “How to counter the pro-consent advocacy?”

    WTF

    “HR899 Terminate Department of Education. Lifelong goal achieved”

    "Rural mass proverty and children that are cheaper than Vodka (2025 planning)

    (Mod Pinned) Mass social manipulation has worked in America. The Second American Revolution has commenced. THE COUNTRY IS OURS

    The claim:

    Video was uploaded to the internet by a Jane Doe who said she had found it on her fathers computer and recorded the screen.


    EDIT: The archive link has the full video, but I can’t find the source poking through endchan (or wherever it was supposedly uploaded before that).


    EDIT2: Struggling to find any source (how do I search/sort imageboards by OP post date?), but for any sleuthers out there, FYI the 750MB video’s encode date is April first, at least according to the metadata:

    Encoded date : 2025-04-01 00:57:33 UTC

    Tagged date : 2025-04-01 01:27:42 UTC

    And this metadata:

    “Title”:“Core Media Video”

    Suggest it is a raw screengrab from a Mac. Could be edited metadata, but still.











  • Random thing, I did not get a notification for this comment, I stumbled upon it. This happens all the time, and it makes me wonder how many replies I miss…

    I don’t run A3B specifically, but for Qwen3 32B Instruct I put something like “vary your prose; avoid repetitive vocabulary and sentence structure” in the system prompt, run at least 0.5 DRY, and maybe some dynamic sampler like mirostat if supported. Too much regular rep penalty makes it dumb, unfortunately.

    But I have much better luck with base model derived models. Look up the finetunes you tried, and see if they were trained from A3B instruct or base. Qwen3 Instruct is pretty overtuned.



  • , especially since something like a Mixture of Experts model could be split down to base models and loaded/unloaded as necessary.

    It doesn’t work that way. All MoE experts are ‘interleaved’ and you need all of them loaded at once, for every token. Some API servers can hotswap whole models, but its not fast, and rarely done since LLMs are pretty ‘generalized’ and tend to serve requests in parallel on API servers.

    The closest to what you’re thinking of is LoRAX (which basically hot-swaps Loras efficiently). But it needs an extremely specialized runtime derived from its associated paper, hence people tend to not use it since it doesn’t support quantization and some other features as well: https://github.com/predibase/lorax

    There is a good case for pure data processing, yeah… But it has little integration with LLMs themselves, especially with the API servers generally handling tokenizers/prompt formatting.

    But, all of its components need to be localized

    They already are! Local LLM tooling and engines are great and super powerful compared to ChatGPT (which offers no caching, no raw completion, primitive sampling, hidden thinking, and so on).