• 0 Posts
  • 15 Comments
Joined 1 year ago
cake
Cake day: June 8th, 2023

help-circle






  • Prior to the API fiasco, Reddit Inc had demonstrated a pattern of promising changes to the mods which they failed to deliver timely if at all. They’ve acknowledged this pattern, promised to do better, then failed to deliver time and again. That part isn’t new.

    Then the API changes were announced and the Reddit community gave Reddit Inc the loudest and most decisive rebuke they ever have. That was the feedback conversation. And Reddit Inc went forward with their plan unchanged. No concessions were made. No concerns were addressed or alleviated. Reddit Inc was informed of what this decision would break and they went ahead and broke it anyway.

    As a former mod, there is nothing left to discuss. There is no reason to believe Reddit Inc will act on anything that doesn’t agree with what they’ve already decided to do. I’m not going back to that kind of abusive relationship. They had their chance to listen to feedback and made it clear that they won’t.




  • Agreed. I find Bing chat is really good when I know almost nothing about what I’m searching, or when I know a whole lot about what I’m searching. Like in your example, if I know exactly what I need but can’t remember its name Bing will read all the spammy beginners’ guides for me and get the answer. And on the opposite end, if I’m looking to buy a gift in a hobby I don’t remotely understand Bing does a pretty good job of holding my hand through the search process.

    Weirdly, medium knowledge questions seem to still do better as a basic Google search. If I need to fix an appliance I’ve fixed before, but it’s been a long time so I really need a full walkthrough, the first few results on Google are faster than waiting for Bing to talk through it.


  • I use the term “autocomplete on steroids” because it gets across a vaguely accurate idea of what an LLM is and how it works to people who are thinking of it like sci-fi movie AI. Sorry if it came across that was my whole reason for considering them not intelligent.

    LLMs do seem to pass a lot of intelligence tests we’ve come up with. Talking with one for the first time is a really uncanny experience, it’s a totally different thing than the old voice assistants. But they also consistently fail at tasks that would indicate an understanding of a topic. They produce good looking equations, but the math underneath doesn’t make sense. They hallucinate facts that don’t fit with the rest of what they themselves are saying, but look similar to the way right answers are written and defended. They produce really convincing responses, but when they fail they betray some really basic failures to understand what they’re saying.

    I feel that LLMs are brute-forcing the tests people designed to measure intelligence. They can pass the bar exam, but they also contain thousands of successful bar exams to consult and millions of bits of text to glue those answers together with. But if you ask the LLM to actually do the job of a lawyer, they start producing all kinds of garbage that sounds good but doesn’t stand up to scrutiny when someone looks up the hallucinated case references.


  • Part of the problem is that AI research likes to use terminology that sounds like what people do, when that’s not what the AI actually does.

    Large language models are not intelligent in any sense. They are autocomplete on steroids. This is a computer program that was fed a book someone wrote, then mathematically tweaked to be able to guess the next word in a sentence in a way that resembles that book. That’s all it does. It does not think or learn in any sense we’d apply to a human.

    To me, LLMs sound like a massive plagiarism engine, and I think they should need to get a license from the authors whose works they used to make the LLM under whatever terms that author wants to give, just like a publisher needs to get permission to print a copy of the work. But copyright law has no easy “bright line” for what counts and what doesn’t. So the courts will have to decide whether what the AI “creates” is similar enough to the original works to count as a violation, or if the AI and its results are transformative enough to count as something new.




  • Mastodon is very active after you start following enough people and hashtags to populate your feed. It’s a bit rough to get started though: no algorithm means no content (or very random content in the local/federated feeds) until you build it up for yourself. But once you hit critical mass, I’ve found it a much nicer experience than I ever got on Twitter.