文档详情

面向人工智能模型的安全攻击和防御策略综述.pdf

发布:2024-11-09约9.67万字共22页下载文档
文本预览下载声明

计算机研究与发展DOI:10.7544/issn1000-1239.202440449

Journal

of

Computer

Research

and

Development61(10):2627−2648,2024

面向人工智能模型的安全攻击和防御策略综述

秦臻庄添铭朱国淞周尔强丁熠耿技

(网络与数据安全四川省重点实验室(电子科技大学)成都610054)

(zhoueq@)

SurveyofSecurityAttackandDefenseStrategiesforArtificialIntelligenceModel

Qin

Zhen,

Zhuang

Tianming,

Zhu

Guosong,

Zhou

Erqiang,

Ding

Yi,

and

Geng

Ji

(NetworkandDataSecurityKeyLaboratoryofSichuanProvince(UniversityofElectronicScienceandTechnologyofChina),Chengdu

610054)

AbstractIn

recent

years,

the

rapid

development

of

artificial

intelligence

technology,

particularly

deep

learning,

has

led

to

its

widespread

application

in

various

fields

such

as

computer

vision

and

natural

language

processing.

However,

recent

research

indicates

potential

security

risks

associated

with

these

advanced

AI

models

could

compromise

their

reliability.

In

light

of

this

concern,

this

survey

delves

into

cutting-edge

research

findings

pertaining

to

security

attacks,

attack

detection,

and

defense

strategies

for

artificial

intelligence

models.

Specifically

regarding

model

security

attacks,

our

work

focuses

on

elucidating

the

principles

and

technical

status

of

adversarial

attacks,

model

inversion

attacks,

and

model

theft

attacks.

With

regards

to

model

attack

detection

methods

explored

in

this

paper,

they

include

defensive

distillation

techniques,

regularization

approaches,

outlier

detection,

robust

statistics.

As

for

model

defense

strategies

examined

in

this

study,

they

encompass

adversarial

training

measures,

model

structure

defense

mechanisms,

query

control

defenses

along

with

other

technical

means.

This

comprehensive

survey

not

only

summarizes

but

also

expands

upon

techniques

and

methodologies

rel

显示全部
相似文档