Lake H. “Trey” Lytal III, partner at the West Palm Beach Law firm of Lytal, Reiter, Smith, Ivey & Fronrath, continues to investigate Tesla crashes around the world after a deadly crash in Palm Beach County. Lytal said, “How many more lives need to be lost? How many more loved ones need to be catastrophically injured for the company (Tesla) to do the right thing and fix this problem?” 

Lytal and his team have been pushing the automaker to admit their heavily marketed “autopilot” system is defective and a danger to all on the roadways. Lytal represents the Banner Family of Palm Beach County after a semi-tractor trailer pulled out in front of Jeremy Banner’s Tesla Model 3 on March 1, 2019. With eerie similarities to other recent crashes, the “autopilot” system was at fault for the deadly crash. Lytal said, “At some point, they (Tesla) have to acknowledge the technology they currently have on the roadway is unsafe to all of us.” He continued, “It’s turning into a weekly occurrence. We are seeing accident after accident.”

Since filing the lawsuit on behalf of the Banner family, the international press has featured Lytal, Reiter, Smith, Ivey & Fronrath.  U.S. news organizations like ABC News and Bloomberg reported on the lawsuit, but also many news agencies across Europe and China featured the news reports in multiple languages. The wide range of press coverage shows that this is not a local or national issue. This is a global safety problem that Lytal and his team hope to force Tesla to correct before the loss of more lives.

If a Tesla crash has impacted you or someone you know, please contact Trey Lytal at 561-820-2248 or email 

ABC News: Tesla sued for ‘defective’ Autopilot in wrongful death suit of Florida driver who crashed into tractor trailer:

The family of a 50-year-old Tesla driver who died when his car collided with a semitrailer in Florida, is suing the automaker for its allegedly “defective” Autopilot and safety features.

Lawyers for Jeremy Banner filed a wrongful death lawsuit in Palm Beach County on Thursday. Banner died on March 1, 2019, when his 2018 Tesla Model S crashed into a semitrailer crossing its path on a Florida highway, shearing off the roof. He had turned on autopilot 10 seconds before the crash, according to an National Transportation Safety Board (NTSB) investigation.

“We’re not just talking about the consequences of this effect to the Banner family, which are horrific,” the Banner family’s attorney Lake H. Lytal III told reporters at a press conference on Thursday afternoon. The point is “to open people’s eyes and make them realize that these products are defective. We’ve got hundreds of thousands of Tesla vehicles on our roadways right now. And this is what’s happening.”

In an email to ABC News, Tesla referred to its May statement about Banner’s accident.

“Shortly following the accident, we informed the National Highway Traffic Safety Administration and the National Transportation Safety Board that the vehicle’s logs showed that Autopilot was first engaged by the driver just 10 seconds prior to the accident, and then the driver immediately removed his hands from the wheel,” a Tesla spokesperson said in the statement. “Autopilot had not been used at any other time during that drive. We are deeply saddened by this accident and our thoughts are with everyone affected by this tragedy.”

In the lawsuit, Lytal referred to a similar accident in 2016 involving a Tesla S on autopilot crashing into a semitrailer on the same Florida highway, killing driver Joshua Brown.

PHOTO: A Tesla Model 3 with extensive roof damage photographed at a tow yard after Jeremy Banner crashed while in autopilot mode in Palm Beach County, Florida, on March 1, 2019.
via NTSB
A Tesla Model 3 with extensive roof damage photographed at a tow yard after Jeremy Banner crashed while in autopilot mode in Palm Beach County, Florida, on March 1, 2019.more +

“I’m going to take you through our accident, and it just kind of lets everybody know how similar it is to the failure that happened in 2016, in the same product defect that they dealt with in 2016. The same problems are happening right now,” Lytal said.

In Banner’s final moments, neither the system nor the driver stopped the vehicle, which was traveling at about 68 miles per hour at the moment of impact, NTSB investigators said. His hands were not on the wheel in the final eight seconds before the crash, according to their report.

The lawsuit says Banner believed his Model 3 “was safer than a human-operated vehicle because Defendant, Tesla claimed superiority regarding the vehicle’s autopilot system, including Tesla’s ‘full self-driving capability,’ Tesla’s ‘traffic-aware cruise control,’ Tesla’s ‘auto steer lane-keeping assistance’ and other safety-related components” would “prevent fatal injury resulting from driving into obstacles and/or vehicles in the path of the subject Tesla vehicle,” the lawsuit said.

Tesla “specifically knew that its product was defective and would not properly and safely avoid impacting other vehicles and obstacles in its path,” the lawsuit claimed, adding that the company “had specific knowledge of numerous prior incidents and accidents in which its safety systems on Tesla vehicles completely failed causing significant property damage, severe injury and catastrophic death to its occupants.”

In addition to Brown’s 2016 crash, Tesla has been under fire for several fatal crashes that occurred while cars were in Autopilot mode, including: a Jan. 22, 2019 crash in which a Tesla collided into a fire truck in Culver City, Los Angeles, a fatal March 23, 2018 crash involved two other cars in Mountainview, California, a May 8, 2018 fatal crash in Ft. Lauderdale, Florida, a May 11, 2018 crash in Utah, a May 29,2018 crash in Laguna Beach, California, an Oct. 12, 2018 crash on the Florida Turnpike.

In its statement to ABC News, Tesla said its cars are safe when Autopilot is used properly.

“Tesla drivers have logged more than one billion miles with Autopilot engaged, and our data shows that, when used properly by an attentive driver who is prepared to take control at all times, drivers supported by Autopilot are safer than those operating without assistance. For the past three quarters, we have released quarterly safety data directly from our vehicles which demonstrates that,” the Tesla spokesperson said.

In June the Insurance Institute for Highway Safety (IIHS) released a study revealing how the names manufacturers use – particularly ‘autopilot’ – “can send the wrong message to drivers regarding how attentive they should be.”

The survey found that the phrase ‘autopilot’ “was associated with the highest likelihood that drivers believed a behavior was safe while in operation, for every behavior measured, compared with other system names.”

ABC News’ Mina Kaji contributed to this report.

Bloomberg: Hyperdrive Tesla Sued by Family of Florida Man Killed in Autopilot Crash

Tesla Inc. was sued for the second time in three months by the family of a car owner who was killed in a crash while using the driver-assistance system Autopilot.

Jeremy Banner, 50, died when the Model 3 sedan he was driving failed to brake or steer to avoid a semi trailer that ran a stop sign on a Florida highway in March, according to the lawsuit, which also names the driver of the semi as a defendant. Banner had engaged the Autopilot system about 10 seconds before the collision.

Representatives for Tesla didn’t immediately respond to a request for comment on the suit, which alleges that the company knew its product was defective.

The National Transportation Safety Board issued a preliminary report on the crash in May and said data from the vehicle showed Autopilot was active at the time of the incident. The preliminary data indicated that neither the driver nor the Autopilot system executed evasive maneuvers.

Banner is survived by his wife and three children.

“We’re not just talking about the consequences of this defect to the Banner family, which is horrific,” Trey Lytal, a lawyer for the family, said during a press conference. “These products are defective.”

Lytal compared Banner’s crash with the one involving Joshua Brown, a Tesla Model S owner who died in a similar tractor trailer collision in 2016. The family of Walter Huang, an Apple Inc. engineer who died in a Model X last year in Mountain View, California, sued the company in May.

The case is Banner v. Tesla, 15th Judicial Circuit, Palm Beach County, Florida.

Business Insider: ‘We cannot have technology and sales take over safety’: Tesla is being sued again for a deadly Autopilot crash

Tesla’s been hit with another lawsuit from the family of a man who died when his Tesla crashed while in Autopilot mode.

The family of Jeremy Beren Banner announced through their lawyer Trey Lytal on Thursday that they are suing the company for wrongful death.

Banner was 50 years old when his Tesla Model 3 collided with a tractor-trailer on March 1 at 68 miles per hour. The car travelled 1,690 feet after the collision, and its roof was torn off.

In its report about the crash, the National Transportation Safety Board (NTSB) said Banner had engaged Autopilot roughly 10 seconds before the collision, and when he crashed “the vehicle did not detect the driver’s hands on the steering wheel.”

Following the NTSB’s findings, Tesla said that after activating Autopilot, Banner “immediately removed his hands from the wheel,” which goes against the instructions it gives drivers.

Read more: Elon Musk says a “massive effort” is required to get Tesla driverless cars to “99.9999%” safety

When questioned about whether Banner had his hands on the wheel at the time of the accident, Lytal said Tesla owners often receive alerts from their car to put their hands on the wheel when they already have their hands on the wheel.

“Just because their sensors didn’t sense that Mr Banner’s hands weren’t on the wheel doesn’t mean they were not on the wheel,” he said. Lytal also said that Tesla’s sales language around Autopilot promises “full self-driving capable car.”

“We are so far from that technology. I get it, his [Elon Musk’s] company is under stress to sell and profit… but we cannot have technology and sales take over safety,” Lytal explained.

He added that Tesla possesses video footage from the inside of the car at the time of the accident, which the family will gain access to during the lawsuit.

Banner’s family isn’t the first to sue Tesla over a fatal Autopilot accident. The family of Apple engineer Walter Huang, who died when his Tesla crashed into a highway barrier while on Autopilot in March 2018, launched a suit against the company in May, alleging the car was “defective in its design.”

Tesla did not immediately respond to Business Insider’s request for comment.

Isobel Asher Hamilton

Business Insider

Sun Sentinel: Autopilot failed to keep Tesla from sliding under semitruck at 68 mph, lawsuit claims

A Tesla car, running on Autopilot, skidded 1,600 feet after sliding under a semitruck at 68 mph, shearing off its top and killing its driver, according to a lawyer who is suing the carmaker.

The crash in west Delray Beach happened four months ago when a tractor-trailer pulled out in front of a bright red Tesla Model 3 driven by 50-year-old Jeremy Banner.

The Autopilot system failed, according to a lawsuit Banner’s family filed Thursday in Palm Beach County. The system should have braked or swerved to avoid the semitruck, Trey Lytal, the family’s attorney, said at a news conference.

About 10 seconds before the crash, Banner engaged the Autopilot system, according to a preliminary report from the National Transportation Safety Board.

Less than eight seconds before the collision, his hands weren’t detected on the steering wheel, which would have prompted warnings from the car’s automated system, investigators found.

The car traveled about the length of five football fields after the collision, Lytal said.

Banner’s family sued Tesla, trucking company First Fleet and semitruck driver Richard Keith Wood. The family — Banner’s wife, Kim, and three kids — seek more than $15,000 in damages, which is the limit to file a civil suit in Palm Beach County, for the death of their husband and father. It’ll be up to a jury to decide how much the damages actually amount to, but it will almost certainly be in the millions, Lytal said.

“My family is devastated due to the untimely and tragic death of a loving husband and father,” a statement from the family said. “It is difficult to discuss and relive what happened to Jeremy at this time. Our family has faith in the legal system that justice will be done and those responsible for his death will be held accountable.”

Tesla representatives did not respond to the lawsuit and referred to the company’s statement in May about the crash.

In it, Tesla said Banner used the Autopilot in the 10 seconds before the crash but not any other time. “Our data shows that, when used properly by an attentive driver who is prepared to take control at all times, drivers supported by Autopilot are safer than those operating without assistance,” the statement says.

Trey Lytal, of the law firm of Lytal, Reiter, Smith, Ivey & Fronrath, speaks during a news conference regarding a wrongful death lawsuit filed on behalf of Jeremy Banner’s family in West Palm Beach, Aug. 1, 2019. (John McCall / South Florida Sun Sentinel/South Florida Sun Sentinel)

Lytal said Tesla falsely advertised the Autopilot system, marketing it as self-driving technology that would “eliminate the risk of harm or injury to the vehicle operator caused by other vehicles or obstacles” while in auto pilot mode.

Because of that, Banner “reasonably believed” the Model 3 was safer than a human-operated vehicle, the lawsuit said.

“We cannot have technology and sales take over safety,” Lytal said. “Safety should be the first priority.

“I hope through this lawsuit that we correct this problem. Otherwise it won’t stop. The goal of all of this is to open people’s eyes.”

Banner, who lived in Lake Worth, was on his way to work in Boca Raton as a computer software programmer, Lytal said.

The crash occurred March 1 on State Road 7, near Pero Family Farms just north of Atlantic Avenue.

According to the preliminary report, Banner was driving south on State Road 7, where the speed limit is 55 mph, when the tractor-trailer pulled out in front of the Tesla, attempting to cross the southbound lanes and turn left to go north. Surveillance videos and forward-facing video from the Tesla show the truck slowed and blocked the Tesla’s path, the report said.

The Tesla drove beneath the trailer at 68 mph, and the roof was sheared off, killing Banner.

The crash is eerily similar to another one involving a Tesla in 2016 near Gainesville. Joshua Brown, 40, of Canton, Ohio, was traveling in a Tesla Model S on a divided highway and using the Autopilot system when he was killed.

Neither Brown nor the car braked for a tractor-trailer, which had turned left in front of the Tesla and was crossing its path. Brown’s Tesla also went beneath the trailer and its roof was torn off.

The NTSB, in a 2017 report, wrote that design limitations of the Autopilot system played a major role in the fatality, the first known one in which a vehicle operated on a highway under semi-autonomous control systems.

The agency said that Tesla told Model S owners that Autopilot should be used only on limited-access highways, primarily interstates. The report said that despite upgrades to the system, Tesla did not incorporate protections against use of the system on other types of roads.

The NTSB found that the Model S cameras and radar weren’t capable of detecting a vehicle turning into its path. Rather, the systems are designed to detect vehicles they are following to prevent rear-end collisions.

Tesla has said that Autopilot and automatic emergency braking are driver-assist systems and that drivers are told in the owner’s manual that they must monitor the road and be ready to take control.

In January 2017, the National Highway Traffic Safety Administration, which is the second federal agency investigating the crash that killed Banner, ended an investigation into the Brown crash, finding that Tesla’s Autopilot system had no safety defects.

But the agency warned automakers and drivers not to treat the semi-autonomous driving systems as if they could drive themselves. Semi-autonomous systems vary in capabilities, and Tesla’s system can keep a car centered in its lane, brake to stop from hitting things and change lanes when activated by the driver.


WPTV: Family of Tesla driver killed in crash in suburban Delray Beach sues automaker

The National Transportation Safety Board said a 2018 Tesla Model 3, driven by 50-year-old Jeremy Banner, was in autopilot mode when it collided with a semi truck in the 14000 block of S.R. 7 on March 1.

NTSB investigators said Banner turned on the autopilot feature about 10 seconds before the crash, and the autopilot did not execute any evasive maneuvers to avoid the crash.

On Thursday, attorneys representing Banner’s family held a news conference about the family’s wrongful death against Tesla, the semi truck driver, and the trucking company he worked for.

The law firm of Lytal, Reiter, Smith, Ivey & Fronrath said the autopilot system on Banner’s Tesla was defective, which resulted in his death.

“There’s no question at all it was defective,” said attorney Trey Lytal. “It did not work properly. In fact, it did not work at all.”

Lytal said there have been numerous cases around the country of the autopilot systems in Tesla vehicles not working correctly, causing accidents and injuries.

“We cannot have technology and sales take over safety,” said Lytal.

The Tesla was traveling 68 miles per hour at the time of the crash, according to the NTSB.

The Palm Beach County Sheriff’s Office said the semi truck pulled into the path of the Tesla, and the Tesla’s roof was sheared off as it passed underneath the semi. Banner died at the scene.

Tesla said in a statement that Banner did not use autopilot at any other time during the drive before the crash. Vehicle logs showed he took his hands off the steering wheel immediately after activating autopilot, the statement said.

Tesla also said it’s saddened by the crash and that drivers have traveled more than 1 billion miles while using autopilot.

“When used properly by an attentive driver who is prepared to take control at all times, drivers supported by Autopilot are safer than those operating without assistance,” the company said in a statement.

“I hope that we can correct this problem,” said Lytal, who added his law firm has not been in contact with Tesla.

The NTSB said its investigation is ongoing.

WPTV and the Associated Press contributed to this report.

International Headline by Italy Daily: Tesla ha denunciato per difettoso pilota automatico in tuta morte ingiusta di Florida driver

La famiglia di un 50-anno-vecchio driver Tesla che è morto quando la sua auto si è scontrato con un autoarticolato in Florida, ha citato in giudizio la casa automobilistica per il suo presunto “difettoso” pilota automatico e caratteristiche di sicurezza.

Avvocati per Jeremy Banner presentato una querela per omissione di soccorso nella Contea di Palm Beach, il giovedì. Banner morto il 1 Marzo, 2019, quando il suo 2018 Tesla Model S si è schiantato in un rimorchio che attraversa il suo percorso su una Florida highway, trancia il tetto. Aveva acceso autopilota 10 secondi prima dell’incidente, secondo la National Transportation Safety Board (NTSB) indagine.

“non stiamo parlando solo le conseguenze di questo effetto per il Banner di famiglia, che sono orribile,” la Bandiera di famiglia, avvocato Lago di H. Lytal III ha detto ai giornalisti in una conferenza stampa, nel pomeriggio di giovedì. Il punto è “aprire gli occhi della gente e fare loro capire che questi prodotti sono difettosi. Abbiamo centinaia di migliaia di Tesla veicoli sulle nostre strade ora. E questo è ciò che sta accadendo.”

In una e-mail a ABC News, Tesla di cui all’Maggio istruzione Banner incidente.

Interessati a Tesla?

Aggiungi Tesla come un interesse per rimanere aggiornato sulle ultime Tesla news, video e analisi da ABC News. Tesla Aggiungere Interesse

“Poco dopo l’incidente, abbiamo informato la National Highway Traffic Safety Administration e il National Transportation Safety Board, che il veicolo registri hanno mostrato che il pilota automatico è stato il primo impegnato con il driver di soli 10 secondi prima dell’incidente, e quindi il driver immediatamente tolto le mani dal volante,” una Tesla ha detto un portavoce dell’istruzione. Il “pilota automatico non era stato utilizzato in qualsiasi altro momento durante tale unità. Siamo profondamente addolorati per questo incidente e che i nostri pensieri sono con tutti colpiti da questa tragedia”.

Nella querela, Lytal di cui per un incidente simile nel 2016 coinvolge una Tesla S con il pilota automatico si schianta contro un autoarticolato sulla stessa Florida highway, uccidendo driver Joshua Brown.

via NTSB UNA Tesla Model 3, con un ampio tetto a danno fotografato in un rimorchio cantiere dopo Jeremy Banner caduto mentre in modalità autopilota nella Contea di Palm Beach, in Florida, il 1 ° Marzo 2019.

“sto andando a prendere l’utente attraverso il nostro caso, e solo tipo di permette a tutti di sapere come simile è il fallimento che è successo nel 2016, del prodotto stesso difetto che hanno affrontato nel 2016. Stessi problemi che stanno accadendo proprio ora,” Lytal detto.

(di PIÙ: Tesla pilota automatico è stato impegnato prima dell’incidente mortale: NTSB )

Nel Banner momenti finali, né il sistema né il conducente fermato il veicolo, che viaggiava a circa 68 miglia all’ora al momento dell’impatto, NTSB i ricercatori hanno detto. Le sue mani non erano sulla ruota in finale otto secondi prima dell’incidente, secondo la loro relazione.

La causa dice Banner creduto che il suo Modello 3 “è più sicuro di un uomo-operated vehicle perché Convenuto, Tesla sosteneva la superiorità per quanto riguarda il veicolo, il sistema di pilota automatico, tra cui Tesla ‘pieno di auto-capacità di guida,’ Tesla ‘di traffico-conoscenza cruise control,’Tesla ‘auto orientare al mantenimento della corsia di assistenza’ e gli altri i componenti di sicurezza” sarebbe “prevenire le lesioni mortali dovuti alla guida in ostacoli e/o veicoli nel percorso del soggetto Tesla veicolo”, la querela ha detto.

(di PIÙ: Driver Mortali Tesla Crash Condiviso Autopilota Video su YouTube)

Tesla “in particolare, sapeva che il suo prodotto era difettoso e non correttamente e in modo sicuro di evitare di impatto con altri veicoli e gli ostacoli nel suo percorso,” l’azione legale ha sostenuto, aggiungendo che la società “ha avuto conoscenza specifica di numerosi incidenti in cui i suoi sistemi di sicurezza su veicoli Tesla completamente fallito causando notevoli danni alla proprietà, lesioni gravi e catastrofico morte per i suoi occupanti.”

oltre a Brown 2016 crash, Tesla è stato sotto il fuoco per diversi incidenti mortali che si sono verificati mentre le auto erano in Autopilota modalità, tra cui: un Gen. 22, 2019 incidente nel quale una Tesla si è scontrato in un camion dei pompieri in Culver City, Los Angeles, una fatale 23 Marzo 2018 incidente ha coinvolto altre due auto in Mountainview, California, 8 Maggio, 2018 incidente mortale in Ft. Lauderdale, Florida, 11 Maggio, 2018 crash in Utah, a Può 29,2018 incidente in Laguna Beach, in California, un Oct. 12, 2018 incidente sulla Florida Turnpike.

(di PIÙ: Tesla Respinge Consumer Reports’ Critica del pilota automatico Caratteristiche)

Nella sua dichiarazione alla ABC News, Tesla ha detto la sua auto sono al sicuro quando l’Autopilota è usato correttamente.

“Tesla driver hanno registrato più di un miliardo di miglia con il pilota automatico inserito, i nostri dati mostrano che, se utilizzato correttamente, un servizio attento e driver, che è pronto a prendere il controllo in ogni momento, i driver supportati da pilota automatico sono più sicuri di quelli di funzionamento senza assistenza. Per gli ultimi tre trimestri, abbiamo rilasciato trimestrale di dati di sicurezza direttamente dal nostro veicoli che dimostra che” la Tesla ha detto un portavoce.

Nel mese di giugno l’Istituto di Assicurazione per Highway Safety (IIHS) ha rilasciato uno studio rivela come i nomi dei produttori di utilizzare – in particolare “pilota automatico” – “possono inviare il messaggio sbagliato per i guidatori su quale sia l’attenzione che dovrebbe essere.”

(di PIÙ: Tesla era in modalità Autopilota durante Utah crash del driver dice)

L’indagine ha trovato che la frase “pilota automatico” “è stato associato con una più alta probabilità che i driver che si crede un comportamento era al sicuro durante il funzionamento, per ogni comportamento misurato, rispetto ad altri nomi di sistema.”

ABC News’ Mina Kaji ha contribuito a questo rapporto.

International Headline by Miet Spiegel News: Tesla verklagt für ‘defekt’ Autopilot in unrechtmäßige Tod Klage von Florida Fahrer

Die Familie des 50-Jahr-alt Tesla-Fahrer starb, als sein Auto kollidierte mit einem Sattelzug in Florida, verklagt den Autohersteller für seine angeblich “defekt” – Autopilot und Sicherheits-features. die Anwälte für Jeremy Banner reichte eine unrechtmäßige Tod Klage in Palm Beach County am Donnerstag. Banner starb am 1. März 2019, wenn seine 2018 Tesla Model S krachte in einen Sattelzug überquert seinen Weg auf einer der Florida highway, Scheren aus dem Dach. Er hatte auf autopilot 10 Sekunden vor dem crash, nach einem National Transportation Safety Board (NTSB) Untersuchung. “Wir reden nicht nur über die Folgen dieser Wirkung auf das Banner-Familie, die schreckliche,” das Banner der Familie ist Anwalt, See H. Lytal III sagte Reportern auf einer Pressekonferenz am Donnerstag Nachmittag. Der Punkt ist, “Menschen die Augen zu öffnen und machen Sie klar, dass diese Produkte defekt sind. Wir haben Hunderte von tausenden von Tesla-Fahrzeugen auf unseren Straßen-gerade jetzt. Und dies ist, was passiert.” In einer email an ABC News, Tesla bezeichnet, Kann seine Aussage über Banner ‘ s Unfall. Interesse an Tesla? Hinzufügen Tesla als ein Interesse an bleiben Sie up to date über die neuesten Tesla-news -, video -, und Analyse von ABC News. Tesla Zinsen “Kurz nach dem Unfall, informierten wir die National Highway Traffic Safety Administration und der National Transportation Safety Board, dass die Fahrzeug-Protokolle zeigte, dass der Autopilot war zunächst beschäftigt mit der der Fahrer nur 10 Sekunden vor dem Unfall, und dann wird der Fahrer sofort entfernt seine Hände von dem lenkrad”, ein Tesla-Sprecher sagte in der Erklärung. “Autopilot hatte nicht verwendet wurde, zu einem anderen Zeitpunkt während, der Fahrt. Wir sind tief betroffen von diesem Unfall, und unsere Gedanken sind bei allen betroffenen dieser Tragödie.” In der Klage, Lytal gemäß einem ähnlichen Unfall im Jahr 2016 mit einem Tesla S auf autopilot Absturz in einem Sattelzug auf der gleichen Florida highway, Tötung Treiber Joshua Brown. über NTSB Ein Tesla Model 3 mit umfangreichen Dach-Schaden fotografiert an einer tow Hof nach Jeremy Banner stürzte im autopilot-Modus in Palm Beach County, Florida, am 1. März 2019. “ich werde nehmen Sie mit unserer Unfall -, und es nur irgendwie lässt jeden wissen, wie ähnlich es ist der Fehler, der passierte, in 2016, in der gleichen Mangel an der Ware, die Sie behandelt im Jahr 2016. Die gleichen Probleme sind, gerade jetzt,” Lytal sagte. (MEHR: teslas Autopilot beschäftigt war, vor dem tödlichen Absturz: NTSB ) Banner In die letzten Momente, weder das system, noch der Fahrer hielt das Fahrzeug, das unterwegs war, bei etwa 68 Meilen pro Stunde an den moment des Aufpralls, NTSB Ermittler sagte. Seine Hände waren nicht auf dem Rad in den letzten acht Sekunden vor dem Absturz, gemäß deren Bericht. Die Klage, sagt Banner glaubte, sein Modell 3 “sicherer war, als ein Mensch-das Fahrzeug betrieben wird, weil die Beklagte, Tesla behauptete überlegenheit in Bezug auf das Fahrzeug autopilot-system, einschließlich Tesla ‘s” full self-driving-Fähigkeit,’ Tesla ‘s” – Verkehr-bewusst, Tempomat,’Tesla ‘s’ auto lenken lane keeping assistance’ und anderen sicherheitsrelevanten Komponenten” würde “verhindern schwerwiegende Verletzungen, die sich aus der Fahrt in Hindernisse und/oder Fahrzeuge, die in den Pfad der das Thema Tesla-Fahrzeug”, die Klage gesagt. (MEHR dazu: Fahrer, die in Tödliche Tesla Absturz Geteilt Autopilot-Videos auf YouTube) Tesla “konkret wusste, dass Ihr Produkt defekt war und nicht richtig und sicher zu vermeiden, dabei andere Fahrzeuge und Hindernisse in den Weg,” die Klage behauptet, hinzu, dass das Unternehmen “bereits spezifische Kenntnisse zahlreicher vorheriger Vorfälle und Unfälle, bei denen die Sicherheit der Systeme auf Tesla-Fahrzeuge komplett fehlgeschlagen verursachen erhebliche Sachschäden, schwere Verletzungen und katastrophalen Tod seiner Insassen.” zusätzlich zu den Browns 2016 crash, Tesla hat unter Feuer für mehrere tödliche Abstürze, die auftraten, während Autos waren im Autopilot-Modus, einschließlich: eine Jan. 22, 2019 crash, in denen ein Tesla kollidierte in einem Feuer-LKW in Culver City, Los Angeles, eine fatale 23. März 2018 Absturz waren zwei andere Autos in Mountainview, Kalifornien, USA, 8. Mai 2018 tödlichen Absturz in Ft. Lauderdale, Florida, ein 11. Mai 2018 Absturz in Utah, eine Kann 29,2018 crash in Laguna Beach, California, Oct. 12, 2018 Absturz auf den Florida Turnpike. (MEHR: Tesla Rebuffs Consumer Reports’ Kritik der AutoPilot Funktionen) In einer Erklärung an ABC News, Tesla sagte, seine Autos sicher sind, wenn Autopilot verwendet wird, richtig. “Tesla Fahrer haben sich angemeldet, mehr als eine Milliarde Kilometer mit Autopilot engagiert, und unsere Daten zeigen, dass, wenn richtig eingesetzt, die von einem aufmerksamen Fahrer, der bereit ist, die Kontrolle zu allen Zeiten Treiber unterstützt vom Autopiloten sind sicherer als diejenigen, die ohne Hilfe. Für die letzten drei Viertel haben wir veröffentlicht quartalsweise Daten zur Sicherheit direkt aus unseren Fahrzeugen, die zeigt, dass” der Tesla-Sprecher sagte. Im Juni das Insurance Institute for Highway Safety (IIHS) veröffentlichte eine Studie enthüllt, wie die Namen der Hersteller verwenden – besonders “autopilot” – “kann die falsche Nachricht an die Fahrer in Bezug auf, wie aufmerksam Sie sein sollten.” (MEHR: Tesla war im Autopilot-Modus, während Utah abstürzt, sagt Fahrer) Die Umfrage ergab, dass der Satz “autopilot” – “war verbunden mit der höchsten Wahrscheinlichkeit, dass die Fahrer glaubten, dass ein Verhalten sicher war, während er in Betrieb ist, für jedes Verhalten gemessen, verglichen mit anderen system Namen.” ABC News’ Mina Kaji zu diesem Bericht beigetragen.

Miet Spiegel News