textattack.attack_recipes package
Attack Recipes Package:
We provide a number of pre-built attack recipes, which correspond to attacks from the literature. To run an attack recipe from the command line, run:
textattack attack --recipe [recipe_name]
To initialize an attack in Python script, use:
<recipe name>.build(model_wrapper)
For example, attack = InputReductionFeng2018.build(model)
creates attack, an object of type Attack
with the goal function, transformation, constraints, and search method specified in that paper. This object can then be used just like any other attack; for example, by calling attack.attack_dataset
.
TextAttack supports the following attack recipes (each recipe’s documentation contains a link to the corresponding paper):
A2T (A2T: Attack for Adversarial Training Recipe)
- class textattack.attack_recipes.a2t_yoo_2021.A2TYoo2021(goal_function: GoalFunction, constraints: List[Constraint | PreTransformationConstraint], transformation: Transformation, search_method: SearchMethod, transformation_cache_size=32768, constraint_cache_size=32768)[source]
Bases:
AttackRecipe
Towards Improving Adversarial Training of NLP Models.
(Yoo et al., 2021)
https://arxiv.org/abs/2109.00544
- static build(model_wrapper, mlm=False)[source]
Build attack recipe.
- Parameters:
model_wrapper (
ModelWrapper
) – Model wrapper containing both the model and the tokenizer.mlm (
bool
, optional, defaults toFalse
) – IfTrue
, load A2T-MLM attack. Otherwise, load regular A2T attack.
- Returns:
A2T attack.
- Return type:
Attack Recipe Class
- class textattack.attack_recipes.attack_recipe.AttackRecipe(goal_function: GoalFunction, constraints: List[Constraint | PreTransformationConstraint], transformation: Transformation, search_method: SearchMethod, transformation_cache_size=32768, constraint_cache_size=32768)[source]
Bases:
Attack
,ABC
A recipe for building an NLP adversarial attack from the literature.
- abstract static build(model_wrapper, **kwargs)[source]
Creates pre-built
Attack
that correspond to attacks from the literature.- Parameters:
model_wrapper (
ModelWrapper
) –ModelWrapper
that contains the victim model and tokenizer. This is passed toGoalFunction
when constructing the attack.kwargs – Additional keyword arguments.
- Returns:
BAE (BAE: BERT-Based Adversarial Examples)
- class textattack.attack_recipes.bae_garg_2019.BAEGarg2019(goal_function: GoalFunction, constraints: List[Constraint | PreTransformationConstraint], transformation: Transformation, search_method: SearchMethod, transformation_cache_size=32768, constraint_cache_size=32768)[source]
Bases:
AttackRecipe
Siddhant Garg and Goutham Ramakrishnan, 2019.
BAE: BERT-based Adversarial Examples for Text Classification.
https://arxiv.org/pdf/2004.01970
This is “attack mode” 1 from the paper, BAE-R, word replacement.
- We present 4 attack modes for BAE based on the
R and I operations, where for each token t in S: • BAE-R: Replace token t (See Algorithm 1) • BAE-I: Insert a token to the left or right of t • BAE-R/I: Either replace token t or insert a token to the left or right of t • BAE-R+I: First replace token t, then insert a token to the left or right of t
- static build(model_wrapper)[source]
Creates pre-built
Attack
that correspond to attacks from the literature.- Parameters:
model_wrapper (
ModelWrapper
) –ModelWrapper
that contains the victim model and tokenizer. This is passed toGoalFunction
when constructing the attack.kwargs – Additional keyword arguments.
- Returns:
BERT-Attack:
(BERT-Attack: Adversarial Attack Against BERT Using BERT)
Warning
This attack is super slow (see https://github.com/QData/TextAttack/issues/586) Consider using smaller values for “max_candidates”.
- class textattack.attack_recipes.bert_attack_li_2020.BERTAttackLi2020(goal_function: GoalFunction, constraints: List[Constraint | PreTransformationConstraint], transformation: Transformation, search_method: SearchMethod, transformation_cache_size=32768, constraint_cache_size=32768)[source]
Bases:
AttackRecipe
Li, L.., Ma, R., Guo, Q., Xiangyang, X., Xipeng, Q. (2020).
BERT-ATTACK: Adversarial Attack Against BERT Using BERT
https://arxiv.org/abs/2004.09984
This is “attack mode” 1 from the paper, BAE-R, word replacement.
- static build(model_wrapper)[source]
Creates pre-built
Attack
that correspond to attacks from the literature.- Parameters:
model_wrapper (
ModelWrapper
) –ModelWrapper
that contains the victim model and tokenizer. This is passed toGoalFunction
when constructing the attack.kwargs – Additional keyword arguments.
- Returns:
CheckList:
(Beyond Accuracy: Behavioral Testing of NLP models with CheckList)
- class textattack.attack_recipes.checklist_ribeiro_2020.CheckList2020(goal_function: GoalFunction, constraints: List[Constraint | PreTransformationConstraint], transformation: Transformation, search_method: SearchMethod, transformation_cache_size=32768, constraint_cache_size=32768)[source]
Bases:
AttackRecipe
An implementation of the attack used in “Beyond Accuracy: Behavioral Testing of NLP models with CheckList”, Ribeiro et al., 2020.
This attack focuses on a number of attacks used in the Invariance Testing Method: Contraction, Extension, Changing Names, Number, Location
https://arxiv.org/abs/2005.04118
- static build(model_wrapper)[source]
Creates pre-built
Attack
that correspond to attacks from the literature.- Parameters:
model_wrapper (
ModelWrapper
) –ModelWrapper
that contains the victim model and tokenizer. This is passed toGoalFunction
when constructing the attack.kwargs – Additional keyword arguments.
- Returns:
CLARE Recipe
(Contextualized Perturbation for Textual Adversarial Attack)
- class textattack.attack_recipes.clare_li_2020.CLARE2020(goal_function: GoalFunction, constraints: List[Constraint | PreTransformationConstraint], transformation: Transformation, search_method: SearchMethod, transformation_cache_size=32768, constraint_cache_size=32768)[source]
Bases:
AttackRecipe
Li, Zhang, Peng, Chen, Brockett, Sun, Dolan.
“Contextualized Perturbation for Textual Adversarial Attack” (Li et al., 2020)
https://arxiv.org/abs/2009.07502
This method uses greedy search with replace, merge, and insertion transformations that leverage a pretrained language model. It also uses USE similarity constraint.
- static build(model_wrapper)[source]
Creates pre-built
Attack
that correspond to attacks from the literature.- Parameters:
model_wrapper (
ModelWrapper
) –ModelWrapper
that contains the victim model and tokenizer. This is passed toGoalFunction
when constructing the attack.kwargs – Additional keyword arguments.
- Returns:
DeepWordBug
(Black-box Generation of Adversarial Text Sequences to Evade Deep Learning Classifiers)
- class textattack.attack_recipes.deepwordbug_gao_2018.DeepWordBugGao2018(goal_function: GoalFunction, constraints: List[Constraint | PreTransformationConstraint], transformation: Transformation, search_method: SearchMethod, transformation_cache_size=32768, constraint_cache_size=32768)[source]
Bases:
AttackRecipe
Gao, Lanchantin, Soffa, Qi.
Black-box Generation of Adversarial Text Sequences to Evade Deep Learning Classifiers.
https://arxiv.org/abs/1801.04354
- static build(model_wrapper, use_all_transformations=True)[source]
Creates pre-built
Attack
that correspond to attacks from the literature.- Parameters:
model_wrapper (
ModelWrapper
) –ModelWrapper
that contains the victim model and tokenizer. This is passed toGoalFunction
when constructing the attack.kwargs – Additional keyword arguments.
- Returns:
Faster Alzantot Genetic Algorithm
(Certified Robustness to Adversarial Word Substitutions)
- class textattack.attack_recipes.faster_genetic_algorithm_jia_2019.FasterGeneticAlgorithmJia2019(goal_function: GoalFunction, constraints: List[Constraint | PreTransformationConstraint], transformation: Transformation, search_method: SearchMethod, transformation_cache_size=32768, constraint_cache_size=32768)[source]
Bases:
AttackRecipe
Certified Robustness to Adversarial Word Substitutions.
Robin Jia, Aditi Raghunathan, Kerem Göksel, Percy Liang (2019).
https://arxiv.org/pdf/1909.00986.pdf
- static build(model_wrapper)[source]
Creates pre-built
Attack
that correspond to attacks from the literature.- Parameters:
model_wrapper (
ModelWrapper
) –ModelWrapper
that contains the victim model and tokenizer. This is passed toGoalFunction
when constructing the attack.kwargs – Additional keyword arguments.
- Returns:
Alzantot Genetic Algorithm
(Generating Natural Language Adversarial Examples)
Warning
This attack uses a very slow language model. Consider using the faster-alzantot
recipe instead.
- class textattack.attack_recipes.genetic_algorithm_alzantot_2018.GeneticAlgorithmAlzantot2018(goal_function: GoalFunction, constraints: List[Constraint | PreTransformationConstraint], transformation: Transformation, search_method: SearchMethod, transformation_cache_size=32768, constraint_cache_size=32768)[source]
Bases:
AttackRecipe
Alzantot, M., Sharma, Y., Elgohary, A., Ho, B., Srivastava, M.B., & Chang, K. (2018).
Generating Natural Language Adversarial Examples.
https://arxiv.org/abs/1804.07998
- static build(model_wrapper)[source]
Creates pre-built
Attack
that correspond to attacks from the literature.- Parameters:
model_wrapper (
ModelWrapper
) –ModelWrapper
that contains the victim model and tokenizer. This is passed toGoalFunction
when constructing the attack.kwargs – Additional keyword arguments.
- Returns:
HotFlip
(HotFlip: White-Box Adversarial Examples for Text Classification)
- class textattack.attack_recipes.hotflip_ebrahimi_2017.HotFlipEbrahimi2017(goal_function: GoalFunction, constraints: List[Constraint | PreTransformationConstraint], transformation: Transformation, search_method: SearchMethod, transformation_cache_size=32768, constraint_cache_size=32768)[source]
Bases:
AttackRecipe
Ebrahimi, J. et al. (2017)
HotFlip: White-Box Adversarial Examples for Text Classification
https://arxiv.org/abs/1712.06751
This is a reproduction of the HotFlip word-level attack (section 5 of the paper).
- static build(model_wrapper)[source]
Creates pre-built
Attack
that correspond to attacks from the literature.- Parameters:
model_wrapper (
ModelWrapper
) –ModelWrapper
that contains the victim model and tokenizer. This is passed toGoalFunction
when constructing the attack.kwargs – Additional keyword arguments.
- Returns:
Improved Genetic Algorithm
(Natural Language Adversarial Attacks and Defenses in Word Level)
- class textattack.attack_recipes.iga_wang_2019.IGAWang2019(goal_function: GoalFunction, constraints: List[Constraint | PreTransformationConstraint], transformation: Transformation, search_method: SearchMethod, transformation_cache_size=32768, constraint_cache_size=32768)[source]
Bases:
AttackRecipe
Xiaosen Wang, Hao Jin, Kun He (2019).
Natural Language Adversarial Attack and Defense in Word Level.
http://arxiv.org/abs/1909.06723
- static build(model_wrapper)[source]
Creates pre-built
Attack
that correspond to attacks from the literature.- Parameters:
model_wrapper (
ModelWrapper
) –ModelWrapper
that contains the victim model and tokenizer. This is passed toGoalFunction
when constructing the attack.kwargs – Additional keyword arguments.
- Returns:
Input Reduction
(Pathologies of Neural Models Make Interpretations Difficult)
- class textattack.attack_recipes.input_reduction_feng_2018.InputReductionFeng2018(goal_function: GoalFunction, constraints: List[Constraint | PreTransformationConstraint], transformation: Transformation, search_method: SearchMethod, transformation_cache_size=32768, constraint_cache_size=32768)[source]
Bases:
AttackRecipe
Feng, Wallace, Grissom, Iyyer, Rodriguez, Boyd-Graber. (2018).
Pathologies of Neural Models Make Interpretations Difficult.
https://arxiv.org/abs/1804.07781
- static build(model_wrapper)[source]
Creates pre-built
Attack
that correspond to attacks from the literature.- Parameters:
model_wrapper (
ModelWrapper
) –ModelWrapper
that contains the victim model and tokenizer. This is passed toGoalFunction
when constructing the attack.kwargs – Additional keyword arguments.
- Returns:
Kuleshov2017
(Adversarial Examples for Natural Language Classification Problems)
- class textattack.attack_recipes.kuleshov_2017.Kuleshov2017(goal_function: GoalFunction, constraints: List[Constraint | PreTransformationConstraint], transformation: Transformation, search_method: SearchMethod, transformation_cache_size=32768, constraint_cache_size=32768)[source]
Bases:
AttackRecipe
Kuleshov, V. et al.
Generating Natural Language Adversarial Examples.
https://openreview.net/pdf?id=r1QZ3zbAZ.
- static build(model_wrapper)[source]
Creates pre-built
Attack
that correspond to attacks from the literature.- Parameters:
model_wrapper (
ModelWrapper
) –ModelWrapper
that contains the victim model and tokenizer. This is passed toGoalFunction
when constructing the attack.kwargs – Additional keyword arguments.
- Returns:
MORPHEUS2020
(It’s Morphin’ Time! Combating Linguistic Discrimination with Inflectional Perturbations)
- class textattack.attack_recipes.morpheus_tan_2020.MorpheusTan2020(goal_function: GoalFunction, constraints: List[Constraint | PreTransformationConstraint], transformation: Transformation, search_method: SearchMethod, transformation_cache_size=32768, constraint_cache_size=32768)[source]
Bases:
AttackRecipe
Samson Tan, Shafiq Joty, Min-Yen Kan, Richard Socher.
It’s Morphin’ Time! Combating Linguistic Discrimination with Inflectional Perturbations
https://www.aclweb.org/anthology/2020.acl-main.263/
- static build(model_wrapper)[source]
Creates pre-built
Attack
that correspond to attacks from the literature.- Parameters:
model_wrapper (
ModelWrapper
) –ModelWrapper
that contains the victim model and tokenizer. This is passed toGoalFunction
when constructing the attack.kwargs – Additional keyword arguments.
- Returns:
Pruthi2019: Combating with Robust Word Recognition
- class textattack.attack_recipes.pruthi_2019.Pruthi2019(goal_function: GoalFunction, constraints: List[Constraint | PreTransformationConstraint], transformation: Transformation, search_method: SearchMethod, transformation_cache_size=32768, constraint_cache_size=32768)[source]
Bases:
AttackRecipe
An implementation of the attack used in “Combating Adversarial Misspellings with Robust Word Recognition”, Pruthi et al., 2019.
- This attack focuses on a small number of character-level changes that simulate common typos. It combines:
Swapping neighboring characters
Deleting characters
Inserting characters
Swapping characters for adjacent keys on a QWERTY keyboard.
https://arxiv.org/abs/1905.11268
- Parameters:
model – Model to attack.
max_num_word_swaps – Maximum number of modifications to allow.
- static build(model_wrapper, max_num_word_swaps=1)[source]
Creates pre-built
Attack
that correspond to attacks from the literature.- Parameters:
model_wrapper (
ModelWrapper
) –ModelWrapper
that contains the victim model and tokenizer. This is passed toGoalFunction
when constructing the attack.kwargs – Additional keyword arguments.
- Returns:
Particle Swarm Optimization
(Word-level Textual Adversarial Attacking as Combinatorial Optimization)
- class textattack.attack_recipes.pso_zang_2020.PSOZang2020(goal_function: GoalFunction, constraints: List[Constraint | PreTransformationConstraint], transformation: Transformation, search_method: SearchMethod, transformation_cache_size=32768, constraint_cache_size=32768)[source]
Bases:
AttackRecipe
Zang, Y., Yang, C., Qi, F., Liu, Z., Zhang, M., Liu, Q., & Sun, M. (2019).
Word-level Textual Adversarial Attacking as Combinatorial Optimization.
https://www.aclweb.org/anthology/2020.acl-main.540.pdf
Methodology description quoted from the paper:
“We propose a novel word substitution-based textual attack model, which reforms both the aforementioned two steps. In the first step, we adopt a sememe-based word substitution strategy, which can generate more candidate adversarial examples with better semantic preservation. In the second step, we utilize particle swarm optimization (Eberhart and Kennedy, 1995) as the adversarial example searching algorithm.”
And “Following the settings in Alzantot et al. (2018), we set the max iteration time G to 20.”
- static build(model_wrapper)[source]
Creates pre-built
Attack
that correspond to attacks from the literature.- Parameters:
model_wrapper (
ModelWrapper
) –ModelWrapper
that contains the victim model and tokenizer. This is passed toGoalFunction
when constructing the attack.kwargs – Additional keyword arguments.
- Returns:
PWWS
(Generating Natural Language Adversarial Examples through Probability Weighted Word Saliency)
- class textattack.attack_recipes.pwws_ren_2019.PWWSRen2019(goal_function: GoalFunction, constraints: List[Constraint | PreTransformationConstraint], transformation: Transformation, search_method: SearchMethod, transformation_cache_size=32768, constraint_cache_size=32768)[source]
Bases:
AttackRecipe
An implementation of Probability Weighted Word Saliency from “Generating Natural Language Adversarial Examples through Probability Weighted Word Saliency”, Ren et al., 2019.
Words are prioritized for a synonym-swap transformation based on a combination of their saliency score and maximum word-swap effectiveness. Note that this implementation does not include the Named Entity adversarial swap from the original paper, because it requires access to the full dataset and ground truth labels in advance.
https://www.aclweb.org/anthology/P19-1103/
- static build(model_wrapper)[source]
Creates pre-built
Attack
that correspond to attacks from the literature.- Parameters:
model_wrapper (
ModelWrapper
) –ModelWrapper
that contains the victim model and tokenizer. This is passed toGoalFunction
when constructing the attack.kwargs – Additional keyword arguments.
- Returns:
Seq2Sick
(Seq2Sick: Evaluating the Robustness of Sequence-to-Sequence Models with Adversarial Examples)
- class textattack.attack_recipes.seq2sick_cheng_2018_blackbox.Seq2SickCheng2018BlackBox(goal_function: GoalFunction, constraints: List[Constraint | PreTransformationConstraint], transformation: Transformation, search_method: SearchMethod, transformation_cache_size=32768, constraint_cache_size=32768)[source]
Bases:
AttackRecipe
Cheng, Minhao, et al.
Seq2Sick: Evaluating the Robustness of Sequence-to-Sequence Models with Adversarial Examples
https://arxiv.org/abs/1803.01128
This is a greedy re-implementation of the seq2sick attack method. It does not use gradient descent.
- static build(model_wrapper, goal_function='non_overlapping')[source]
Creates pre-built
Attack
that correspond to attacks from the literature.- Parameters:
model_wrapper (
ModelWrapper
) –ModelWrapper
that contains the victim model and tokenizer. This is passed toGoalFunction
when constructing the attack.kwargs – Additional keyword arguments.
- Returns:
TextBugger
(TextBugger: Generating Adversarial Text Against Real-world Applications)
- class textattack.attack_recipes.textbugger_li_2018.TextBuggerLi2018(goal_function: GoalFunction, constraints: List[Constraint | PreTransformationConstraint], transformation: Transformation, search_method: SearchMethod, transformation_cache_size=32768, constraint_cache_size=32768)[source]
Bases:
AttackRecipe
Li, J., Ji, S., Du, T., Li, B., and Wang, T. (2018).
TextBugger: Generating Adversarial Text Against Real-world Applications.
https://arxiv.org/abs/1812.05271
- static build(model_wrapper)[source]
Creates pre-built
Attack
that correspond to attacks from the literature.- Parameters:
model_wrapper (
ModelWrapper
) –ModelWrapper
that contains the victim model and tokenizer. This is passed toGoalFunction
when constructing the attack.kwargs – Additional keyword arguments.
- Returns:
TextFooler (Is BERT Really Robust?)
A Strong Baseline for Natural Language Attack on Text Classification and Entailment)
- class textattack.attack_recipes.textfooler_jin_2019.TextFoolerJin2019(goal_function: GoalFunction, constraints: List[Constraint | PreTransformationConstraint], transformation: Transformation, search_method: SearchMethod, transformation_cache_size=32768, constraint_cache_size=32768)[source]
Bases:
AttackRecipe
Jin, D., Jin, Z., Zhou, J.T., & Szolovits, P. (2019).
Is BERT Really Robust? Natural Language Attack on Text Classification and Entailment.
https://arxiv.org/abs/1907.11932
- static build(model_wrapper)[source]
Creates pre-built
Attack
that correspond to attacks from the literature.- Parameters:
model_wrapper (
ModelWrapper
) –ModelWrapper
that contains the victim model and tokenizer. This is passed toGoalFunction
when constructing the attack.kwargs – Additional keyword arguments.
- Returns: