• 7 Posts
  • 2.08K Comments
Joined 2 years ago
cake
Cake day: March 22nd, 2024

help-circle


  • This is proof of why OpenAI’s… opaqueness is so dangerous.

    Chat LLMs tend to treat everything like an exam question or essay prompt, as a direct consequence of how the base models are finetuned. The hand is like a pivot point in a physic problem. But more importantly:

    • The chat context is sort of their whole world. Again, due to training format. So they tend to stubbornly adhere to what has already been said, and have no real means of self correction.

    • While we have no idea what OpenAI actually does, in basically every other open model, the vision component is trained separately from pure text input. Point being these models are alright at the very specific set of vision tasks they’re trained for, but the “coupling” of image input to the bulk of the LLM is very weak. The reasoning they can do over text does not carry over well.


    Point I’m trying to make is the biggest lie of Altman is pitching ChatGPT as a general intelligence… It’s not. It’s a dumb, narrow tool, like a drill with specific changeable bits. But they package and market it like it’s “smart”, which is a big fat lie.

    Go to any of the smaller AI vendors/models (like Minimax, with a new model today) and they do the opposite of this. They show specific uses in specific harnesses, and hyper optimize for that.


  • Be careful about assertions like that, though. Misinformation taken as fact is posted all the time… just yesterday, I looked up some picture everyone was harping on in a comment with hundreds of upvotes, and it was totally fake. Probably AI generated. No one even checked.

    And when I’ve pointed this out in the past, no one cared. People only prune information if it doesn’t validate their beliefs.

    I’m not saying the Factorio dev isn’t a jerk, necessarily, just that I don’t trust Fediverse commenters to have an ounce of information hygiene.


  • No. You ghost her. And you get everyone you can to ghost her.

    Don’t feed the trolls. Anything else is just getting her engagement, even so much as mentioning the issue.

    Also:

    for some LGBTQ+ people to shoot videos telling how HP helped them to come out of hiding and reveal their true self to the world.

    Thousands (millions?) of people have already done this. Hence the HP fanfic community is notorious for heavy LGBTQ+ plots, which is partially why her position is so ridiculous.



  • Corporate, for now.

    Thing is, once they’re out there, they’re free utilities, and they can’t be taken back.

    Also, they don’t really need to aggressively scrape the internet. There are many good public datasets now, and the Chinese are already making excellent use of synthetic dataset generation on (relative) shoestring budgets. Also, several nations and other large organizations are already funding open model efforts, but they just haven’t had the opportunity to catch up yet.


  • That’s pretty much what local ML is.

    If open weights LLMs take off, and business users realize they can just finetune tiny specialized models for stuff, OpenAI is toast. All of Big Tech’s bets are. It’s why they keep fanning the “AGI” lie, and why they’re pushing for regulation so hard, why they’re shoving LLMs where they just don’t fit and harping on safety.


  • To illustrate what I mean more clearly, look at the top comments/replies for the NASA Artemis posts, as an example.

    …It’s basically all conspiracy theorists, and government skeptics.

    Twitter’s focusing the Artemis posts on them because it’s what they want to see, and most engaging for them.

    In the EFF’s case, I’m not just talking about Musk’s influence. The algorithm will only show the EFF to users who would be highly engaged by it. E.g., angry skeptics who wouldn’t be swayed by the EFF anyway, or fans who already agree with the EFF. It’s literally not going to show the EFF to people who need to see it, as Twitter’s metrics would show it as unengaging.


    This is the “false image” I keep trying to dispel. Twitter is less and less an “even spread” of exposure like people think it is, like it sort of used to be, more-and-more a hyper focused bubble of what you want to hear, and only what you want to hear. All the changes Musk is making are amplifying that. Maybe that’s fine for some orgs, but there’s no point in the EFF staying in that kind of environment, regardless of ethics.



  • brucethemoose@lemmy.worldtoScience Memes@mander.xyzFake News
    link
    fedilink
    English
    arrow-up
    38
    ·
    edit-2
    9 days ago

    So sick of seeing confidently incorrect people opining, using historical examples, when they have never before cracked open a history book and have no idea of the context.

    This has always been the case through history.

    The issue is Twitter boosts them over less engaging experts. The new problem is the medium. Twitter is not a fair forum, and these takes trend deliberately.

    …And I think its really important for scientists (or anyone who believes in science) to recognize that. With all due respect, I do not understand, with everything that’s happened, why they still keep using Twitter.








  • They seem to have held back the “big” locally runnable model.

    It’s also kinda conservative/old, architecture wise: 16-bit weights, sliding window attention interleaved with global attention. No MTP, no QAT (yet), no tightly integrated vision, no hybrid mamba like Qwen/Deepseek, nothing weird like that. It’s especially glaring since we know Google is using an exotic architecture for Gemini, and has basically infinite resources for experimentation.

    It also feels kinda “deep fried” like GPT-OSS to me, see: https://github.com/ikawrakow/ik_llama.cpp/issues/1572

    it is acting crazy. it can’t do anything without the proper chat template, or it goes crazy.


    IMO it’s not very interesting, especially with so many other models that run really well on desktops.


  • it’s a form of private journalism, private opinion, and private art

    But without any of the liability hazard.

    This is my issue: the big platforms having their cake and eating it. In one breath, they claim to be little open-platform garage startups that can’t possibly be responsible for the content of their users; they’re just a utility. They need protection from Congress. In another breath, they’re the stewards of generations and children, the only ones responsible enough to tame the internet’s criminality. All while making trillions.

    They want to be “private content” protected from the government? Fine. Treat them like it, legally.