Spaces:
Sleeping
Sleeping
| # This is the official config file. | |
| PECCAVI_TEXT: | |
| Entailment: | |
| task: "text-classification" | |
| model: "ynie/roberta-large-snli_mnli_fever_anli_R1_R2_R3-nli" | |
| Masking: | |
| task: "fill-mask" | |
| tokenizer: "bert-base-uncased" | |
| model: "bert-base-uncased" | |
| # tokenizer: "bert-large-cased-whole-word-masking" | |
| # model: "bert-large-cased-whole-word-masking" | |
| Vocabulary: | |
| tokenizer: "bert-base-uncased" | |
| model: "bert-base-uncased" | |
| # permissible_ratio: 0.5 | |
| # tokenizer: "bert-large-cased-whole-word-masking" | |
| # model: "bert-large-cased-whole-word-masking" | |
| permissible_ratio: 1.0 | |
| Sampling: | |
| tokenizer: "bert-base-uncased" | |
| model: "bert-base-uncased" | |
| # tokenizer: "bert-large-cased-whole-word-masking" | |
| # model: "bert-large-cased-whole-word-masking" | |
| Metrics: | |
| EuclideanDistance: "sentence-transformers/all-MiniLM-L6-v2" | |
| Distortion: "gpt2" | |
| Detector: | |
| tokenizer: "bert-base-uncased" | |
| model: "bert-base-uncased" | |
| # tokenizer: "bert-large-cased-whole-word-masking" | |
| # model: "bert-large-cased-whole-word-masking" | |
| Paraphrase: | |
| tokenizer: "humarin/chatgpt_paraphraser_on_T5_base" | |
| model: "humarin/chatgpt_paraphraser_on_T5_base" | |
| num_beams: 10 | |
| num_beam_groups: 10 | |
| num_return_sequences: 10 | |
| repetition_penalty: 10.0 | |
| diversity_penalty: 3.0 | |
| no_repeat_ngram_size: 2 | |
| temperature: 0.7 | |
| max_length: 64 | |