• 18 Posts
  • 1.25K Comments
Joined 2 years ago
cake
Cake day: March 22nd, 2024

help-circle
  • brucethemoose@lemmy.worldtoScience Memes@mander.xyzCause and Effect
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    28 minutes ago

    I think the flip side of this is Facebook or wherever the link was pushed to your in-laws (which is what I’d guess happened) feels… empowering. Those apps are literally optimized, with billions of dollars (and extensive science, especially psychology), to validate folk’s views in the pursuit of keeping them clicking. Their world’s telling them they’re right; of course your retort will feel offensive and wrong.

    They’re in a trap.

    And I still see lot of scientists posit ‘why is this happening?’ unironically on Twitter or something, which really frustrates me.


  • Yeah at some point folks get too old to change. You can try to get them off Fox News at least, but it just depends. Sometimes you can’t. Sometimes they really can’t even learn a new kind of remote/cable or whatever.

    Younger folks are deep down the influencer rabbit hole, and I don’t know what to do about that (other than pushing Fediverse) since Big Tech has such an iron grip on everything. I mean, the future is definitely cyberpunk.


  • This is not how some Trump voters I know think at all; it’s not how they think of Trump. Bring up any of these points and they’d think you’re totally paranoid.

    I witnessed a group of old guys watching the movie ‘One Battle After Another’ scratch their heads and have literally, honestly no idea what it was about.

    It’s because they’re in the habit of watching Fox News or other ‘old’ media. Trump’s just some Republican like they’ve always voted for with a bunch of hysterical finger pointers, and any Democrat is basically Jimmy Carter or Hillary Clinton.




  • Yeah, the APIs are super cheap. It doesn’t make a ton of sense unless you already have the GPU lying around.

    With the right settings, GLM will actually work fine in 16GB, 12GB, or even 11GB VRAM + 128GB RAM. I can even make a custom quant if you want, since I already got that set up. 24 GB just gives it a bit of ‘breathing room’ for longer context and relaxed quantization for the dense parts of the model.

    GLM Air will work on basically any modernish Nvidia GPU + like 26GB of free RAM. Its dense part is really small.

    But to be clear, you have to get into the weeds to run them efficiently this way. There’s no simple ollama run here.


  • brucethemoose@lemmy.worldtoTechnology@lemmy.zipDeepSeek-V3.2 released
    link
    fedilink
    English
    arrow-up
    7
    arrow-down
    1
    ·
    edit-2
    3 days ago

    Deepseek is only bad via the chat app, and whatever prefilter (or finetune?) they censor it with.

    The model itself (via API or run locally) isn’t too bad, especially with a system prompt or completion syntax to squash refusals. Obviously there are CCP mandated gaps (which you can just add in via context), but it’s not as tankie as you’d think.



  • brucethemoose@lemmy.worldtoTechnology@lemmy.zipDeepSeek-V3.2 released
    link
    fedilink
    English
    arrow-up
    3
    arrow-down
    1
    ·
    edit-2
    3 days ago

    With sparse attention, very interesting. It seems GQA is a thing of the past.

    I especially love Deepseek’s ‘public research’ aspect: they trained this and Terminus the same way, so the attention schemes are (more-or-less) directly comparable. That’s awesome.

    GLM 4.6 is reportedly about to drop too. Which is great, as 4.5 is without a doubt my daily driver now.


  • Prompt formatting (and the system prompt) is a huge thing, especially with models trained for ‘tool use’ a specific way, so be sure to keep that in mind. For example, if you want a long chain of steps, be sure to explicitly ask (though Qwen is uses its thinking block quite gratuitously).

    I find GLM 4.5’s default formatting to be really good though: be sure to give that a shot. It’s also awesome because the full 350B model (with some degredation) is locally runnable on a 128GB RAM + 24GB VRAM gaming rig, and the ‘Air’ version is quite fast and accurate on lesser hardware.

    Local hosting, if you can swing it, is particularly nice because the calls are literally free, and promt ingestion is cached, so you can batch them and spam the heck out of them for testing and such.







  • It’s a matter of opportunity cost. Why give Ukraine an expensive air superiority fighter that needs precious resources to operate, and is vulnerable to Russian SAMs, when they could have the equivalent cost in missiles instead?

    The Saab Gripen carries more explicit reconiassance/air-to-ground loadout, it’s designed to be cheap to maintain, it’s more flexible with short runaways and such.


    …Hence my uncertainty. I am not military, so I don’t really understand the need of a focused air superiority fighter like the Typhoon, as opposed to more SAMs + aircraft that can do A2A like the Gripen.