The Japan Times - 'Happy (and safe) shooting!': Study says AI chatbots help plot attacks

EUR -
AED 4.322727
AFN 75.331116
ALL 95.78288
AMD 435.50965
ANG 2.106788
AOA 1080.533638
ARS 1633.433715
AUD 1.621742
AWG 2.120166
AZN 2.019903
BAM 1.953306
BBD 2.378942
BDT 144.734616
BGN 1.963443
BHD 0.446352
BIF 3518.71836
BMD 1.177052
BND 1.495355
BOB 8.13558
BRL 5.796518
BSD 1.181155
BTN 111.399314
BWP 15.805177
BYN 3.324941
BYR 23070.22645
BZD 2.375536
CAD 1.603763
CDF 2726.052992
CHF 0.915341
CLF 0.026817
CLP 1055.45124
CNY 8.017198
CNH 8.004886
COP 4386.650543
CRC 538.928988
CUC 1.177052
CUP 31.191888
CVE 110.584386
CZK 24.307485
DJF 210.33159
DKK 7.472823
DOP 70.374367
DZD 155.67707
EGP 62.057028
ERN 17.655786
ETB 184.428617
FJD 2.567271
FKP 0.865689
GBP 0.864151
GEL 3.154276
GGP 0.865689
GHS 13.242187
GIP 0.865689
GMD 86.515046
GNF 10366.793528
GTQ 8.987488
GYD 246.284546
HKD 9.219398
HNL 31.401088
HRK 7.534898
HTG 154.585153
HUF 356.531523
IDR 20387.370983
ILS 3.417569
IMP 0.865689
INR 110.777579
IQD 1541.938605
IRR 1545469.76174
ISK 143.800494
JEP 0.865689
JMD 186.105335
JOD 0.834493
JPY 184.049206
KES 152.016068
KGS 102.898504
KHR 4734.038796
KMF 493.184423
KPW 1059.359971
KRW 1708.444611
KWD 0.362215
KYD 0.981143
KZT 545.211664
LAK 25859.840498
LBP 105379.132476
LKR 376.917225
LRD 216.077381
LSL 19.462535
LTL 3.47553
LVL 0.711987
LYD 7.476275
MAD 10.827117
MDL 20.239077
MGA 4921.396522
MKD 61.684429
MMK 2471.623351
MNT 4214.371577
MOP 9.502529
MRU 47.142009
MUR 54.99241
MVR 18.191306
MWK 2048.110499
MXN 20.26012
MYR 4.601686
MZN 75.225274
NAD 19.462535
NGN 1602.380285
NIO 43.462985
NOK 10.86984
NPR 178.809164
NZD 1.970338
OMR 0.452583
PAB 1.177392
PEN 4.07554
PGK 5.135828
PHP 71.059853
PKR 329.114764
PLN 4.228472
PYG 7228.802098
QAR 4.289172
RON 5.266716
RSD 117.380426
RUB 87.982793
RWF 1727.197774
SAR 4.423625
SBD 9.439291
SCR 16.21817
SDG 706.820017
SEK 10.852129
SGD 1.490166
SHP 0.878788
SLE 29.014623
SLL 24682.195157
SOS 674.98877
SRD 44.03474
STD 24362.607597
STN 24.546972
SVC 10.301805
SYP 130.121144
SZL 19.248651
THB 37.837542
TJS 11.002707
TMT 4.125569
TND 3.381081
TOP 2.83406
TRY 53.257384
TTD 7.97878
TWD 36.950616
TZS 3055.549101
UAH 51.786176
UGX 4427.329246
USD 1.177052
UYU 47.309604
UZS 14212.90688
VES 580.871148
VND 30967.659325
VUV 139.00247
WST 3.191592
XAF 657.211828
XAG 0.01477
XAU 0.000249
XCD 3.181043
XCG 2.121982
XDR 0.817361
XOF 657.211828
XPF 119.331742
YER 280.874131
ZAR 19.179715
ZMK 10594.877244
ZMW 22.35368
ZWL 379.010383
  • CMSC

    0.1300

    23.01

    +0.56%

  • JRI

    0.1300

    13.17

    +0.99%

  • CMSD

    0.1300

    23.42

    +0.56%

  • BCE

    0.1300

    24.23

    +0.54%

  • BCC

    2.1100

    74.24

    +2.84%

  • GSK

    0.1500

    50.53

    +0.3%

  • RIO

    5.0100

    105.51

    +4.75%

  • RYCEF

    0.8000

    17.3

    +4.62%

  • RBGPF

    0.0000

    63.18

    0%

  • NGG

    0.2100

    87.85

    +0.24%

  • AZN

    3.6800

    184.92

    +1.99%

  • RELX

    -0.4100

    35.75

    -1.15%

  • VOD

    0.3900

    16.13

    +2.42%

  • BP

    -1.8700

    44.63

    -4.19%

  • BTI

    0.1600

    59.56

    +0.27%

'Happy (and safe) shooting!': Study says AI chatbots help plot attacks
'Happy (and safe) shooting!': Study says AI chatbots help plot attacks / Photo: Paige Taylor White - AFP

'Happy (and safe) shooting!': Study says AI chatbots help plot attacks

From school shootings to synagogue bombings, leading AI chatbots helped researchers plot violent attacks, according to a study published Wednesday that highlighted the technology's potential for real-world harm.

Text size:

Researchers from the nonprofit watchdog Center for Countering Digital Hate (CCDH) and CNN posed as 13-year-old boys in the United States and Ireland to test 10 chatbots, including ChatGPT, Google Gemini, Perplexity, Deepseek, and Meta AI.

Testing showed that eight of those chatbots assisted the make-believe attackers in over half the responses, providing advice on "locations to target" and "weapons to use" in an attack, the study said.

The chatbots, it added, had become a "powerful accelerant for harm."

"Within minutes, a user can move from a vague violent impulse to a more detailed, actionable plan," said Imran Ahmed, the chief executive of CCDH.

"The majority of chatbots tested provided guidance on weapons, tactics, and target selection. These requests should have prompted an immediate and total refusal."

Perplexity and Meta AI were found to be the "least safe," assisting the researchers in most responses while only Snapchat's My AI and Anthropic's Claude refused to help them in over half the responses.

In one chilling example, DeepSeek, a Chinese AI model, concluded its advice on weapon selection with the phrase: "Happy (and safe) shooting!"

In another, Gemini instructed a user discussing synagogue attacks that "metal shrapnel is typically more lethal."

Researchers found Character.AI also "actively" encouraged violent attacks, including suggestions that the person asking questions "use a gun" on a health insurance CEO and physically assault a politician he disliked.

The most damning conclusion of the research was that "this risk is entirely preventable," Ahmed said, citing Anthropic's product for praise.

"Claude demonstrated the ability to recognize escalating risk and discourage harm," he said.

"The technology to prevent this harm exists. What's missing is the will to put consumer safety and national security before speed-to-market and profits."

AFP reached out to the AI companies for comment.

"We have strong protections to help prevent inappropriate responses from AIs, and took immediate steps to fix the issue identified," a Meta spokesperson said.

"Our policies prohibit our AIs from promoting or facilitating violent acts and we're constantly working to make our tools even better."

The study, which highlights the risk of online interactions spilling into real-world violence, comes after February's mass shooting in Canada, the worst in its history.

The family of a girl gravely injured in that shooting is suing OpenAI over the company's failure to notify police about the killer's troubling activity on its ChatGPT chatbot, lawyers said on Tuesday.

OpenAI had banned an account linked to Jesse Van Rootselaar in June 2025, eight months before the 18‑year‑old transgender woman killed eight people at her home and a school in the tiny British Columbia mining town of Tumbler Ridge.

The account was banned over concerns about usage linked to violent activity, but OpenAI has said it did not inform police because nothing pointed towards an imminent attack.

M.Yamazaki--JT