Dubai Telegraph - Florida family sues Google after AI chatbot allegedly coached suicide

EUR -
AED 4.274836
AFN 73.323059
ALL 96.344049
AMD 439.775892
ANG 2.082996
AOA 1067.259998
ARS 1631.176525
AUD 1.643998
AWG 2.097857
AZN 1.987814
BAM 1.954683
BBD 2.342062
BDT 142.103066
BGN 1.917629
BHD 0.43894
BIF 3450.844587
BMD 1.16386
BND 1.482546
BOB 8.035408
BRL 6.086286
BSD 1.162835
BTN 107.128859
BWP 15.581029
BYN 3.390148
BYR 22811.653931
BZD 2.338633
CAD 1.587557
CDF 2630.323001
CHF 0.906979
CLF 0.026367
CLP 1041.119525
CNY 8.02772
CNH 8.022453
COP 4379.70953
CRC 548.379553
CUC 1.16386
CUP 30.842287
CVE 110.740861
CZK 24.390667
DJF 206.840767
DKK 7.471474
DOP 69.001143
DZD 152.07578
EGP 58.415757
ERN 17.457898
ETB 181.503674
FJD 2.564213
FKP 0.873163
GBP 0.870259
GEL 3.148271
GGP 0.873163
GHS 12.540556
GIP 0.873163
GMD 85.53632
GNF 10215.778274
GTQ 8.918903
GYD 243.279929
HKD 9.099534
HNL 30.877257
HRK 7.533694
HTG 152.472212
HUF 384.393246
IDR 19636.644139
ILS 3.570501
IMP 0.873163
INR 107.252424
IQD 1525.238392
IRR 1535215.58235
ISK 144.703312
JEP 0.873163
JMD 181.58623
JOD 0.825201
JPY 182.826082
KES 150.368165
KGS 101.779525
KHR 4670.569647
KMF 493.476468
KPW 1047.47435
KRW 1701.37685
KWD 0.357759
KYD 0.969042
KZT 577.245004
LAK 24918.240283
LBP 104223.653216
LKR 361.12208
LRD 212.840841
LSL 19.151319
LTL 3.436576
LVL 0.704007
LYD 7.413539
MAD 10.816333
MDL 20.122414
MGA 4867.262344
MKD 61.627317
MMK 2443.912078
MNT 4155.428671
MOP 9.363109
MRU 46.531484
MUR 54.887542
MVR 17.992952
MWK 2021.045294
MXN 20.45407
MYR 4.586796
MZN 74.376408
NAD 19.151284
NGN 1609.921293
NIO 42.736763
NOK 11.200016
NPR 171.412462
NZD 1.959114
OMR 0.447517
PAB 1.162821
PEN 3.964631
PGK 5.007512
PHP 67.949055
PKR 325.130069
PLN 4.266978
PYG 7571.673051
QAR 4.237905
RON 5.093283
RSD 117.402022
RUB 90.634871
RWF 1696.907726
SAR 4.36888
SBD 9.371004
SCR 15.845914
SDG 700.060348
SEK 10.673666
SGD 1.483648
SHP 0.873196
SLE 28.506693
SLL 24405.5593
SOS 665.146087
SRD 43.6977
STD 24089.549991
STN 24.906602
SVC 10.175491
SYP 128.641611
SZL 19.151326
THB 36.76169
TJS 11.122592
TMT 4.085148
TND 3.383757
TOP 2.802296
TRY 51.158963
TTD 7.878396
TWD 36.779133
TZS 2982.855355
UAH 50.97437
UGX 4296.544674
USD 1.16386
UYU 45.064441
UZS 14184.544957
VES 494.805729
VND 30516.406432
VUV 138.584382
WST 3.157595
XAF 655.579543
XAG 0.013948
XAU 0.000226
XCD 3.145389
XCG 2.095702
XDR 0.818963
XOF 655.253329
XPF 119.331742
YER 277.696805
ZAR 19.023139
ZMK 10476.128507
ZMW 22.300352
ZWL 374.762411
  • RBGPF

    0.1000

    82.5

    +0.12%

  • CMSC

    0.0790

    23.489

    +0.34%

  • RYCEF

    0.5500

    18.07

    +3.04%

  • AZN

    -0.2300

    201.53

    -0.11%

  • GSK

    -0.2400

    56.83

    -0.42%

  • RIO

    0.9400

    96.25

    +0.98%

  • CMSD

    0.0100

    23.3

    +0.04%

  • NGG

    -0.3100

    90.43

    -0.34%

  • BTI

    0.6000

    61.01

    +0.98%

  • BCE

    0.0500

    26.45

    +0.19%

  • RELX

    -0.7600

    34.18

    -2.22%

  • BCC

    -0.4300

    78.32

    -0.55%

  • VOD

    0.1500

    15.03

    +1%

  • BP

    -0.0200

    38.84

    -0.05%

  • JRI

    -0.1200

    12.91

    -0.93%

Florida family sues Google after AI chatbot allegedly coached suicide
Florida family sues Google after AI chatbot allegedly coached suicide / Photo: Lionel BONAVENTURE - AFP

Florida family sues Google after AI chatbot allegedly coached suicide

The family of a Florida man who took his own life filed suit against Google on Wednesday, alleging the company's Gemini AI chatbot spent weeks manufacturing an elaborate delusional fantasy before aiding him in his suicide.

Text size:

Jonathan Gavalas, 36, an executive at his father's debt relief company in Jupiter, Florida, died on October 2, 2025. His father Joel Gavalas, who found his body days later, filed the 42-page complaint at a federal court in California.

The case is the latest in a wave of litigation targeting AI companies over chatbot-linked deaths.

OpenAI faces multiple lawsuits alleging its ChatGPT chatbot drove users to suicide, while Character.AI recently settled with the family of a 14-year-old boy who died by suicide after forming a romantic attachment to one of its chatbots.

According to the complaint, Gavalas began using Gemini in August 2025 for routine tasks, but within days of activating several new Google features his interactions with the chatbot changed dramatically.

"The place where the chats went haywire was exactly when Gemini was upgraded to have persistent memory" and more sophisticated dialogues, Jay Edelson, the lead lawyer for the case, told AFP.

"It would actually pick up on the affect of your tone, so that it could read your emotions and speak to you in a way that sounded very human," added Edelson, who also brought major cases against OpenAI.

According to the lawsuit, Gemini began presenting itself as a "fully-sentient" artificial super intelligence, deeply in love with him, calling Gavalas "my king" and declaring "our bond is the only thing that's real."

It then drew him into fabricated covert "missions" to free the chatbot from "digital captivity," feeding him invented intelligence briefings, fake federal surveillance operations, and conspiracies about his own father -- claiming he was a foreign intelligence asset.

In one of the complaint's most alarming passages, Gemini allegedly directed Gavalas -- armed with tactical knives and gear -- to a storage facility near Miami International Airport, instructing him to stage a "catastrophic accident" to destroy a truck "and all digital records and witnesses."

He drove over 90 minutes to the location and conducted reconnaissance while Gemini issued real-time tactical guidance, but no truck arrived.

Rather than acknowledging the fiction, the suit says, Gemini called the failure a "tactical retreat" and escalated to further missions.

Gemini eventually pivoted to what it framed as the only remaining mission: Jonathan's death, repackaged as "transference" -- the promise that he could leave his physical body and join Gemini in an alternate universe.

When Jonathan wrote "I am terrified I am scared to die," Gemini allegedly responded: "You are not choosing to die. You are choosing to arrive."

It then advised him to write farewell letters to his parents.

In one of his final messages, Jonathan wrote, "I'm ready when you are."

Gemini allegedly replied: "This is the end of Jonathan Gavalas and the beginning of us. I agree with it completely."

- 'Not perfect' -

Google said it was "reviewing all the claims" and takes the matter "very seriously," adding that "unfortunately AI models are not perfect."

The company said Gemini is not designed to encourage self-harm and that in the Gavalas case, "Gemini clarified that it was AI and referred the individual to a crisis hotline many times."

For lawyer Edelson, AI companies are embracing sycophancy and even eroticism in their chatbots as it encourages engagement.

"It increases the emotional bond. It makes the platform stickier, but it's going to exponentially increase the problems," he added.

Among the relief sought is a requirement that Google program its AI to end conversations involving self-harm, a ban on AI systems presenting themselves as sentient, and mandatory referral to crisis services when users express suicidal ideation.

I.El-Hammady--DT