OpenAttack是一款专为文本对抗攻击设计的开源工具套件,该工具基于Python开发,可以处理文本对抗攻击的整个过程,包括预处理文本、访问目标用户模型、生成对抗示例和评估攻击模型等等。

网站建设哪家好,找创新互联建站!专注于网页设计、网站建设、微信开发、微信平台小程序开发、集团企业网站建设等服务项目。为回馈新老客户创新互联还提供了涉县免费建站欢迎大家使用!
OpenAttack支持以下几种功能:
OpenAttack的使用范围非常广,其中包括但不限于:
我们可以使用pip安装,或者克隆该项目源码来安装OpenAttack。
使用pip安装(推荐):
- pip install OpenAttack
 
克隆代码库:
- git clone https://github.com/thunlp/OpenAttack.git
 - cd OpenAttack
 - python setup.py install
 
安装完成之后,我们可以尝试运行“demo.py”来检测OpenAttack是否能够正常工作:
(1) 基础使用:使用内置攻击模型
OpenAttack内置了一些常用的文本分类模型,如LSTM和BERT,以及用于情感分析的SST和用于自然语言推理的SNLI等数据集。
以下代码段显示了如何使用基于遗传算法的攻击模型攻击SST数据集上的BERT:
- import OpenAttack as oa
 - # choose a trained victim classification model
 - victim = oa.DataManager.load("Victim.BERT.SST")
 - # choose an evaluation dataset
 - dataset = oa.DataManager.load("Dataset.SST.sample")
 - # choose Genetic as the attacker and initialize it with default parameters
 - attacker = oa.attackers.GeneticAttacker()
 - # prepare for attacking
 - attack_eval = oa.attack_evals.DefaultAttackEval(attacker, victim)
 - # launch attacks and print attack results
 - attack_eval.eval(dataset, visualize=True)
 
(2) 高级使用:攻击自定义目标用户模型
下面的代码段显示了如何使用基于遗传算法的攻击模型攻击SST上的自定义情绪分析模型:
- import OpenAttack as oa
 - import numpy as np
 - from nltk.sentiment.vader import SentimentIntensityAnalyzer
 - # configure access interface of the customized victim model
 - class MyClassifier(oa.Classifier):
 - def __init__(self):
 - self.model = SentimentIntensityAnalyzer()
 - # access to the classification probability scores with respect input sentences
 - def get_prob(self, input_):
 - rt = []
 - for sent in input_:
 - rs = self.model.polarity_scores(sent)
 - prob = rs["pos"] / (rs["neg"] + rs["pos"])
 - rt.append(np.array([1 - prob, prob]))
 - return np.array(rt)
 - # choose the costomized classifier as the victim model
 - victim = MyClassifier()
 - # choose an evaluation dataset
 - dataset = oa.DataManager.load("Dataset.SST.sample")
 - # choose Genetic as the attacker and initialize it with default parameters
 - attacker = oa.attackers.GeneticAttacker()
 - # prepare for attacking
 - attack_eval = oa.attack_evals.DefaultAttackEval(attacker, victim)
 - # launch attacks and print attack results
 - attack_eval.eval(dataset, visualize=True)
 
OpenAttack:【GitHub传送门】