• RickC137@lemmy.world
    link
    fedilink
    arrow-up
    0
    arrow-down
    4
    ·
    18 days ago

    I am not a fan of Tesla/Elon but are you sure that no human driver would fall for this?

    • jj4211@lemmy.world
      link
      fedilink
      arrow-up
      2
      ·
      18 days ago

      Lets assume that a human driver would fall for it, for sale of argument.

      Would that make it a good idea to potentially run over a kid just because a human would have to, when we have a decent option to do better than human senses?

      • RickC137@lemmy.world
        link
        fedilink
        arrow-up
        0
        arrow-down
        1
        ·
        18 days ago

        What makes you assume that a vision based system performs worse than the average human? Or that it can’t be 20 times safer?

        I think the main reason to go vision-only is the software complexity of merging mixed sensor data. Radar or Lidar alone also have their limitations.

        I wish it was a different company or that Musk would sell Tesla. But I think they are the closest to reaching full autonomy. Let’s see how it goes when FSD launches this year.

        • jj4211@lemmy.world
          link
          fedilink
          arrow-up
          0
          ·
          18 days ago

          Somehow other car companies are managing to merge data from multiple sources fine. Tesla even used to do it, but stopped to shave a few dollars in their costs.

          In terms of assuming there would be safety concerns, well this video clearly demonstrates that adding lidar avoids three scenarios, at least two of them realistic. As I said my standard is not “human driver” but safest options as demonstrated.

          • RickC137@lemmy.world
            link
            fedilink
            arrow-up
            0
            arrow-down
            1
            ·
            17 days ago

            Which other system can drive autonomous in potentially any environment without relying on map data?

            If merging data from different sensors increases complexity by factor 5, it’s just not worth it.

            • jj4211@lemmy.world
              link
              fedilink
              arrow-up
              0
              ·
              17 days ago

              One, I don’t know if ‘autonomous no matter what’ is an important enough goal versus ADAS, but for another, the gold standard in the industry except Tesla is vehicle mounted LIDAR, with investments to bring down the tech price.

              Merging data from different sources was never claimed by anyone to be too hard a problem, again, even Tesla used to and decided to downgrade their capabilities for cost. “It’s just not worth it” is a strange take on a video demonstrating quite clearly the better data from LIDAR than you can possibly get from cameras and the benefit of avoiding collisions, collisions that kill thousands a year. Even the relatively “won’t turn on unless things are perfect” autopilot has killed quite a few people, and incurred hundreds of accidents beyond that.

              • RickC137@lemmy.world
                link
                fedilink
                arrow-up
                0
                arrow-down
                1
                ·
                17 days ago

                Autopilot is not FSD and I bet many of the deaths were caused by inattentive drivers.

                Which other system has a similar architecture and similar potential?

    • ThePunnyMan@lemm.ee
      link
      fedilink
      arrow-up
      1
      ·
      18 days ago

      Part of the problem is the question of who is at fault if an autonomous car crashes. If a human falls for this and crashes, it’s their fault. They are responsible for their damages and the damages caused by their negligence. We expect a human driver to be able to handle any road hazards. If a self driving car crashes who’s fault is it? Tesla? They say their self driving is a beta test so drivers must remain attentive at all times. The human passenger? Most people would expect a self driving car would drive itself. If it crashes, I would expect the people that made the faulty software to be at fault, but they are doing everything they can to shift the blame off of themselves. If a self driving car crashes, they expect the owner to eat the cost.

      • RickC137@lemmy.world
        link
        fedilink
        arrow-up
        0
        arrow-down
        1
        ·
        17 days ago

        As soon as we have hard data from real world use and FSD is safer than the average human, it would be unethical to not solve the regulatory and legal issues and apply it on a larger scale to save human lives.

        If a human driver causes a crash, the insurance pays. Why shouldn’t they if a computer caused the crash, which drives safer overall, if only by let’s say 10%.

    • undeffeined@lemmy.ml
      link
      fedilink
      arrow-up
      1
      ·
      18 days ago

      The road runner thing seems a bit far fetched yeah. But there were also tests with heavy rain and fog which were not passed by Tesla.

      • RickC137@lemmy.world
        link
        fedilink
        arrow-up
        0
        ·
        18 days ago

        Should be fine if the car reduces speed to account for the conditions. Just like a human driver does.

      • Ghostalmedia@lemmy.worldOPM
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        1
        ·
        18 days ago

        The road runner thing isn’t far fetched. Teslas have a track record of t-boning semi trucks in overcast conditions, where the sky matches the color of the truck’s container.

    • TheSealStartedIt@feddit.org
      link
      fedilink
      arrow-up
      1
      ·
      17 days ago

      That is a completely legitimate question. That you are downvoted says a lot about the current state of Lemmy. Don’t get me wrong, I’m all for the Musk hate, but it looks like a nuanced discussion on topics where Nazi-Elon is involved is currently not possibe.