Taalas HC1: 17,000 tokens/sec on Llama 3.1 8B vs Nvidia H200’s 233 tokens/sec. 73x faster at one-tenth the power. Each chip runs ONE model, hardwired into the transistors.

  • FaceDeer@fedia.io
    link
    fedilink
    arrow-up
    2
    arrow-down
    1
    ·
    11 hours ago

    Why doesn’t it work in those contexts? It’s better than nothing in those contexts too. I’d rather have a car with onboard intelligence to take over than an uncontrolled one.

    I think you’re letting the perfect be the enemy of the good, here. There are plenty of situations where you don’t need a robot to behave perfectly. People don’t behave perfectly.

    • MagicShel@lemmy.zip
      link
      fedilink
      English
      arrow-up
      3
      ·
      edit-2
      11 hours ago

      No, it doesn’t work in this context because false positive is worse than nothing. False negative is better than nothing. Zero sum. Obviously it depends where you set the threshold of false positive and false negative. I imagined a very simple scenario the first time.

      If even only .001% of the time, you’re going to cause a shit load accidents. You’re going to average a car slamming on the breaks for no reason like every… 2 minutes would be .12, 20 would be 1.2, 200 would be 12% 800 would be 48%, so you’re going to have every car slam on their breaks every 12-15 hours of drive time. That would be an absolute mess.

      • FaceDeer@fedia.io
        link
        fedilink
        arrow-up
        1
        ·
        11 hours ago

        I have no idea what you’re thinking the scenario is here. The alternative is an uncontrolled car, I think I’d rather it had at least some brains behind the decisions it’s making.

        • MagicShel@lemmy.zip
          link
          fedilink
          English
          arrow-up
          4
          ·
          11 hours ago

          How does it decide the car is uncontrolled? That’s a failure scenario, too.

          I’m not even sure what you’re arguing. I said from the get go that there are niche cases where AI is nothing but positive. You seem to be arguing that there are a bunch more cases. Fine. Maybe the niche is slightly less thin and narrow than I think. Cool.

          • XLE@piefed.social
            link
            fedilink
            English
            arrow-up
            3
            ·
            4 hours ago

            Facedeer is just a pro-AI concern troll from Reddit.

            He kicked off his part of the thread by complaining about people, and then speculating that maybe this chip could do a thing without any evidence.

            • MagicShel@lemmy.zip
              link
              fedilink
              English
              arrow-up
              2
              ·
              25 minutes ago

              I’m middle of the road on AI. I think it has uses. I also think this technology is a dead end (i.e. this is not going to lead to AGI) and had people understood from the start the limitations of it, investment would’ve been more modest and cautious. Is a great technology. You can do cool things with it. But it will never be able to significantly replace humans. However it may be really painful watching the investor class wrestle with that reality.

              I think the chip does have uses and I think building it even with today’s models would last a long time. But the number of scenarios where it is unequivocally better than nothing is smaller than AI bros (I draw a line between an enthusiast like myself and a bro who is all in and won’t hear reason) want to think.

              Last point. In theory this chip is great. Based on my reading this is a substitute for an H100 — a data center GPU (APU?). This isn’t going into smart mines or drones and probably not cars. Not without more development. So while there is potential here, none of these use cases are practical. This is a way for OAI or whomever to run their current models just the way they are for cheaper but with a hardware cost to upgrade. This isn’t going to matter for the rest of us for a while.

          • FaceDeer@fedia.io
            link
            fedilink
            arrow-up
            1
            ·
            11 hours ago

            When the regular controller of the car - be it human, another AI, whatever - isn’t sending control signals, then the onboard controller knows that the car is uncontrolled. Of course it’s a “failure scenario”, I’m suggesting that this chip would be ideal for picking up when that sort of thing happens. The alternative is to just fall over.

            I, too, am not sure what you’re arguing. I suggested that a low-power high-speed AI chip like this would be ideal for putting in robots, which have power constraints and aren’t always in reliable contact with outside controllers. That’s a very broad “niche” indeed. I don’t know what all this landmine stuff or probabilities of brake-slamming is all about or how it relates to what I suggested.

            • MagicShel@lemmy.zip
              link
              fedilink
              English
              arrow-up
              1
              ·
              edit-2
              11 hours ago

              My scenario was a safety device that prevented cars from hitting pedestrians. You’re stuck on this autonomous self control in the event of loss of human control and it seems like you’re interpreting what I’m saying in that context, which I wasn’t. I presented a scenario when it’s a good idea and one when it isn’t. Nothing to do with your autonomous control scenario.

              But let’s see. If you’ve got a done that can fly itself for a few seconds or minutes if it loses signal, simply loitering waiting for control to continue, or maybe continuing on a flight path until it is out of jamming range. Alternative is uncontrolled crash, possibility of avoiding that is nothing but upside, whether it’s 10% or 90% success. It’s a good example of the type of scenario I was describing with the smart mine.

              I wasn’t trying to address your scenario because it already falls into the niche I was describing. I was trying to demonstrate how to consider scenarios where AI is good vs ones where it has an unacceptable tradeoff. Where the consequences of failure don’t outweigh the benefits when it gets it right.

              So I think we were talking past each other, and if my communication was unclear then I apologize. In my defense, it’s 2AM here.