The Japan Times - AI-enhanced images of real events distort view of Mideast war

EUR -
AED 4.302619
AFN 72.638318
ALL 95.603094
AMD 431.878807
ANG 2.097662
AOA 1075.507446
ARS 1630.8359
AUD 1.615579
AWG 2.110304
AZN 1.99945
BAM 1.956238
BBD 2.359669
BDT 143.812209
BGN 1.95644
BHD 0.441978
BIF 3486.028541
BMD 1.171577
BND 1.490921
BOB 8.096055
BRL 5.888817
BSD 1.171582
BTN 112.066143
BWP 15.783006
BYN 3.264603
BYR 22962.916957
BZD 2.356308
CAD 1.60594
CDF 2625.505158
CHF 0.91581
CLF 0.026408
CLP 1039.329512
CNY 7.956124
CNH 7.950219
COP 4445.398123
CRC 533.328553
CUC 1.171577
CUP 31.046801
CVE 110.655135
CZK 24.327919
DJF 208.212632
DKK 7.472548
DOP 69.416143
DZD 155.118147
EGP 61.994247
ERN 17.573661
ETB 184.376952
FJD 2.560893
FKP 0.866041
GBP 0.866089
GEL 3.139424
GGP 0.866041
GHS 13.242448
GIP 0.866041
GMD 85.525666
GNF 10283.522856
GTQ 8.938002
GYD 245.111173
HKD 9.172924
HNL 31.1758
HRK 7.533714
HTG 153.009493
HUF 358.229119
IDR 20516.663355
ILS 3.410104
IMP 0.866041
INR 112.115446
IQD 1534.766388
IRR 1538281.120455
ISK 143.612268
JEP 0.866041
JMD 185.285963
JOD 0.830666
JPY 184.939933
KES 151.344328
KGS 102.454005
KHR 4699.197143
KMF 493.234395
KPW 1054.43934
KRW 1745.468735
KWD 0.361116
KYD 0.976348
KZT 549.878462
LAK 25716.123453
LBP 105150.026727
LKR 380.231651
LRD 214.57466
LSL 19.226057
LTL 3.459363
LVL 0.708675
LYD 7.410193
MAD 10.747758
MDL 20.0931
MGA 4891.33573
MKD 61.635919
MMK 2459.473576
MNT 4193.865493
MOP 9.450699
MRU 46.863218
MUR 54.84144
MVR 18.053649
MWK 2040.295627
MXN 20.113167
MYR 4.599628
MZN 74.860808
NAD 19.225688
NGN 1605.623002
NIO 43.002772
NOK 10.739627
NPR 179.312517
NZD 1.975525
OMR 0.450414
PAB 1.171602
PEN 4.016757
PGK 5.108019
PHP 71.952469
PKR 326.382702
PLN 4.2477
PYG 7164.604642
QAR 4.268647
RON 5.208363
RSD 117.382677
RUB 86.904361
RWF 1710.502998
SAR 4.402872
SBD 9.410412
SCR 16.330594
SDG 703.542135
SEK 10.926465
SGD 1.490557
SHP 0.874701
SLE 28.823398
SLL 24567.394667
SOS 669.559557
SRD 43.575646
STD 24249.286687
STN 24.89602
SVC 10.251296
SYP 129.551813
SZL 19.313411
THB 37.889169
TJS 10.971838
TMT 4.112237
TND 3.374732
TOP 2.820877
TRY 53.230856
TTD 7.948916
TWD 36.980249
TZS 3043.348516
UAH 51.5192
UGX 4393.058898
USD 1.171577
UYU 46.541218
UZS 14150.311878
VES 595.237083
VND 30868.721224
VUV 138.221382
WST 3.166467
XAF 656.120751
XAG 0.013399
XAU 0.00025
XCD 3.166246
XCG 2.111455
XDR 0.81421
XOF 654.332389
XPF 119.331742
YER 279.567675
ZAR 19.234782
ZMK 10545.588979
ZMW 22.113613
ZWL 377.247443
  • RBGPF

    -0.2100

    60.79

    -0.35%

  • BCE

    -0.0800

    24.39

    -0.33%

  • AZN

    3.1800

    187.72

    +1.69%

  • CMSD

    -0.0400

    23.56

    -0.17%

  • NGG

    -0.2600

    86.98

    -0.3%

  • GSK

    0.0900

    50.99

    +0.18%

  • CMSC

    -0.0600

    23.05

    -0.26%

  • RIO

    2.5400

    112.04

    +2.27%

  • BP

    -0.2600

    44.14

    -0.59%

  • BTI

    1.7100

    65.35

    +2.62%

  • RELX

    -1.1500

    31.62

    -3.64%

  • JRI

    -0.0100

    13.13

    -0.08%

  • BCC

    -0.9500

    66.98

    -1.42%

  • VOD

    0.4150

    15.51

    +2.68%

  • RYCEF

    -0.1700

    16.03

    -1.06%

AI-enhanced images of real events distort view of Mideast war
AI-enhanced images of real events distort view of Mideast war / Photo: AHMAD AL-RUBAYE - AFP

AI-enhanced images of real events distort view of Mideast war

The Middle East war has unleashed a torrent of AI-driven disinformation. Beyond entirely fabricated visuals, another kind of content is spreading: authentic images "enhanced" in ways that experts say are subtly distorting perceptions of what's happening on the ground.

Text size:

In one striking photo, a kneeling US pilot is confronted by a Kuwaiti local, moments after parachuting from his jet. The high-quality image was widely shared online and even published by media outlets. Yet the pilot appears to have only four fingers on each hand.

AFP fact-checkers ran the photo through AI detection tools and found it contained a SynthID, an invisible watermark meant to identify images made with Google AI. But that's not the whole story.

The situation itself appears to be genuine. A video showing the same scene began circulating on social media on March 2, and satellite imagery verified the location. It also corresponded with reports that day that Kuwait had mistakenly shot down three US warplanes.

AFP was also able to locate an earlier version of the photo on Telegram that matched the high-resolution photo exactly, except that it was blurry.

AI verification tools determined this image, which had none of the same detail in the pilot's face, was real. This suggests it may have served as the starting point for the image that returned the Google AI result.

"AI-enhancement may subtly alter textures, faces, lighting, or background details, creating an image that looks more 'real' than the original," said Evangelos Kanoulas, a professor in AI at the University of Amsterdam.

This can "strengthen a particular narrative about an event -- for example, making a protest appear more violent, making a crowd appear larger, making facial expressions more intense."

In another case, social media users shared a dramatic image of a huge blaze near Erbil airport in Iraq, after the area was targeted by Iranian strikes on March 1.

Although SynthID detection recognised the use of Google AI in the picture, it was not a total fabrication. The original version of the image shows the same scene but with a far smaller fire and smoke column, and less vivid colours.

- 'Very different story' -

Experts warned that the line between enhancement and content generation, accidentally or intentionally, was a thin one.

"Even little changes can end up telling a very different story," said James O'Brien, a professor of computer science at the University of California, Berkeley, and "could change the perception of events".

Generative artificial intelligence is also still prone to error and may "hallucinate" elements that were not in the original image, Kanoulas added.

This happened following the shooting of Alex Pretti by federal immigration agents in the US state of Minneapolis in January, when an AI-enhanced image of the incident went viral.

The image was based on a frame taken from a genuine video of the shooting, showing Pretti falling to his knees with officers beside him, one of them holding a gun to his head.

In the grainy, low-quality frame, Pretti holds an object that in reality was a phone. In the AI-treated image, some social media users wrongly saw a weapon in his hand.

As the war triggered by the US-Israeli attacks on Iran rages on, experts said that without proper labelling, AI-enhanced images further eroded the public's trust.

This kind of content was already having "a huge impact on people and their ability to trust the truth," said O'Brien.

"People start doubting authentic images as well," Kanoulas agreed.

Y.Watanabe--JT