The reason I ask, is I’ve been seeing alot of news and cases of Tesla’s self driving acting up and being a point of contention. But back in 2016-17 my ex’s uncle and aunt got a Model X when they first dropped and they “auto-drove” us like 50 miles without any noticeable issue.

Was i just gambling my life or has the tech somehow gotten worse?

  • Sanctus@lemmy.world
    link
    fedilink
    English
    arrow-up
    30
    ·
    4 days ago

    The tech never “evolved” in the first place. It was a deformity from the start. LIDAR was developed for this reason. But Tesla uses cheap ass cameras that try to interpret what it is seeing through the visual data. I’m guessing with my layman’s knowledge this is why they veer at semi trucks. Because the technology itself is based on a shitty premise.

    • BertramDitore@lemm.ee
      link
      fedilink
      English
      arrow-up
      19
      ·
      4 days ago

      In general I think you’re right about the tech just being shitty, but a slight correction: LiDAR was not developed for self-driving, it’s just a relevant application of the technology. LiDAR has been around for quite a while, and was initially best known as a remote sensing technology. It is effective at remote sensing because it can penetrate certain solid materials, most importantly foliage. So when an aerial LiDAR dataset is collected for a forested area, since the light can penetrate through most of the foliage, one can essentially ‘delete’ the vegetation from the resulting point cloud, leaving a bare earth model, which is a very close approximation of the landscape’s actual topography if there had been no trees. This can be especially valuable for archaeological research, as foliage is often a significant obstacle for accurately mapping large sites, or even finding them in the first place.

      All of that to say, yeah, self-driving buzz made LiDAR well known as tech, but it wasn’t developed for that purpose.

      • Sanctus@lemmy.world
        link
        fedilink
        English
        arrow-up
        5
        ·
        4 days ago

        Thanks, I did not know that and it is honestly a way cooler application. Foliage X Ray vision.

    • RaoulDook@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      4 days ago

      In a non-Tesla car I’ve driven, there was an autopilot cruise control mode that just used cameras. In practice it only works out well if you’re driving long distance on a highway with low traffic. It’s still nice to have (much better than having no autopilot cruise mode) but I don’t trust it around multiple lanes of other cars doing unpredictable shit. Also quits working in the rain when the cameras are obscured.

  • FuglyDuck@lemmy.world
    link
    fedilink
    English
    arrow-up
    28
    ·
    4 days ago

    Tesla’s marketing has consistently lied to you about the state of their “full self driving”.

    It’s not. It never was. This is a perfect example of “fake it ‘till you make it”

  • gusgalarnyk@lemmy.world
    link
    fedilink
    arrow-up
    7
    ·
    4 days ago

    I appreciate these comments saying the tech hasn’t degraded and it’s been standstill, or that it was never great in the first place, all of which is true but I would like to interject my own Model 3 experience. When we first bought the Tesla in 2019 the self driving functionality on the highway felt safe and functional in nominal conditions. When we sold the Tesla 2 years ago (2022) the self driving felt noticably more finicky. It struggled to switch lanes, recognize when lanes started and ended, and had noticably more issues with maintaining proper speed and distance with other cars.

    It probably wasn’t significantly more dangerous, but it felt like it was. What was a feature we used for the first year or two without much complaint turned into something we never used and our driving time when down in that third year not up so it wasn’t exposure time I don’t think.

    • ContrarianTrail@lemm.ee
      link
      fedilink
      English
      arrow-up
      1
      ·
      3 days ago

      That’s Autopilot though, not FSD. Untill not too long ago it was still switching to Autopilot on highways where as in cities it was using an entirely different system. Nowdays the neural nets based system is used on highways too assuming you have FSD subscribtion. If not, then it’ll keep using Autopilot which is less advanced system.

  • NeoNachtwaechter@lemmy.world
    link
    fedilink
    arrow-up
    9
    ·
    4 days ago

    Was i just gambling my life or has the tech somehow gotten worse?

    We can safely assume that tech has evolved to the better during this time.

    But it is still too dangerous. It should not be allowed on public roads yet.

    • Drivebyhaiku@lemmy.world
      link
      fedilink
      arrow-up
      1
      ·
      3 days ago

      That the tech has evolved to be better actually is an assumption. The novel data problem hasn’t been meaningfully addressed really at all so mostly we assume that progress has been made… but it’s not meaningful progress. The promises being made for future capability is mostly pretty stale hype that hasn’t changed year to year with a lot of the targets remaining unchanged. We are getting more data on where specifically and how it’s failing, which is something, but overall it appears to be a plateau of non-linear progress with different updates being sometimes less safe than newer ones.

      That actually safe self driving cars might be decades away however is antithetical to the hype run marketing campaigns that are working overtime to put up smoke and mirrors around the issue.

  • bitchkat@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    ·
    3 days ago

    That car had AP1 which used hardware developed by MobileEye. Tesla doest like licensing software so they ditched that for a homegrown solution.

  • ch00f@lemmy.world
    link
    fedilink
    arrow-up
    6
    ·
    4 days ago

    The first Model X has Autopilot 1 which was a system designed by Mobileye. Tesla’s relationship with Mobileye fell apart and they replaced it with an Nvidia based system in 2017(?). It was really really bad at the start as they were essentially starting from scratch. This system also used 8 cameras instead of the original 1.

    Then Tesla released AP hardware 3 which was a custom-built silicon chip designed specifically for self-driving which also enabled proper navigation of surface streets in addition to the just highway lanekeeping offered in AP1. This broadened scope of actually dealing with turns and traffic from multiple angles is probably where the reputation of it being dangerous has come from.

    My HW3 enabled Model 3 does make mistakes, though it’s rarely anything like hitting a pedestrian or running off the road. Most of my issues are with navigational errors. If the GPS gets messed up in the tunnel, it’ll suddenly decide to take an exit that it isn’t supposed to, or it’ll get in the left lane to pass someone 1/4 mile from a right-exit.

  • thesohoriots@lemmy.world
    link
    fedilink
    English
    arrow-up
    5
    ·
    4 days ago

    There are also many more Teslas on the road, and the “full self driving” incidents are more widely reported on since the new ownership likes to overpromise and vastly underdeliver. Other commenters have already addressed the tech side, but a few years ago, the Tesla-specific FSD was found to be active right up until a split second before some prolific collisions with emergency vehicles, leading to speculation on liability. Tesla aside, I think it’s just laziness on the part of drivers used to FSD doing the menial tasks of driving.

  • NevermindNoMind@lemmy.world
    link
    fedilink
    arrow-up
    5
    ·
    4 days ago

    Just a guess, but it’s probably a combination of two things. First, if we say a self driving car is going to hit an edge case it can’t resolve once in every, say, 100,000 miles, the number of Tesla’s and other self driving cars on the roads now means more miles driven more frequently which means those edge cases are going to occur more frequently. Second, people are becoming over reliant on self driving - they are (incorrectly ) trusting it more and paying less attention, meaning less chance of human intervention when those edge cases occur. So probably the self driving is overall better, but the number of accidents overall is increasing.

  • ContrarianTrail@lemm.ee
    link
    fedilink
    English
    arrow-up
    4
    ·
    edit-2
    4 days ago

    Tesla recently switched from human-written code to full neural networks with their FSD system, which resulted in a significant leap in how reliable and human-like it became. Nowadays, it’s safe to assume it will perform better than the average human driver on most trips, but it’s still not reliable enough to be blindly trusted. You still need to sit there, paying attention to make sure it doesn’t do anything stupid. It’s, by far the most advanced ‘self-driving’ system available on a vehicle you can buy.

    The reason it’s constantly in the news is that anything related to Elon, AI, or self-driving is guaranteed to get clicks in today’s mediascape. It’s not a flawless system by any means but it’s also not as bad as people make it to be.

  • WastedJobe@feddit.org
    link
    fedilink
    arrow-up
    3
    ·
    4 days ago

    A guy on YT used to (or still does idk) upload entire FSD trips after every update. I haven’t watched any lately but the trend was that the software was improving with some regressions every now and then, which were fixed in the next updates.