Références :
adv_sst2
Utilisez la commande suivante pour charger cet ensemble de données dans TFDS :
ds = tfds.load('huggingface:adv_glue/adv_sst2')
- Description :
Adversarial GLUE Benchmark (AdvGLUE) is a comprehensive robustness evaluation benchmark
that focuses on the adversarial robustness evaluation of language models. It covers five
natural language understanding tasks from the famous GLUE tasks and is an adversarial
version of GLUE benchmark.
- Licence : Aucune licence connue
- Version : 1.0.0
- Divisions :
Diviser | Exemples |
---|---|
'validation' | 148 |
- Caractéristiques :
{
"sentence": {
"dtype": "string",
"id": null,
"_type": "Value"
},
"idx": {
"dtype": "int32",
"id": null,
"_type": "Value"
}
}
adv_qqp
Utilisez la commande suivante pour charger cet ensemble de données dans TFDS :
ds = tfds.load('huggingface:adv_glue/adv_qqp')
- Description :
Adversarial GLUE Benchmark (AdvGLUE) is a comprehensive robustness evaluation benchmark
that focuses on the adversarial robustness evaluation of language models. It covers five
natural language understanding tasks from the famous GLUE tasks and is an adversarial
version of GLUE benchmark.
- Licence : Aucune licence connue
- Version : 1.0.0
- Divisions :
Diviser | Exemples |
---|---|
'validation' | 78 |
- Caractéristiques :
{
"question1": {
"dtype": "string",
"id": null,
"_type": "Value"
},
"question2": {
"dtype": "string",
"id": null,
"_type": "Value"
},
"idx": {
"dtype": "int32",
"id": null,
"_type": "Value"
}
}
adv_mnli
Utilisez la commande suivante pour charger cet ensemble de données dans TFDS :
ds = tfds.load('huggingface:adv_glue/adv_mnli')
- Description :
Adversarial GLUE Benchmark (AdvGLUE) is a comprehensive robustness evaluation benchmark
that focuses on the adversarial robustness evaluation of language models. It covers five
natural language understanding tasks from the famous GLUE tasks and is an adversarial
version of GLUE benchmark.
- Licence : Aucune licence connue
- Version : 1.0.0
- Divisions :
Diviser | Exemples |
---|---|
'validation' | 121 |
- Caractéristiques :
{
"premise": {
"dtype": "string",
"id": null,
"_type": "Value"
},
"hypothesis": {
"dtype": "string",
"id": null,
"_type": "Value"
},
"idx": {
"dtype": "int32",
"id": null,
"_type": "Value"
}
}
adv_mnli_mismatched
Utilisez la commande suivante pour charger cet ensemble de données dans TFDS :
ds = tfds.load('huggingface:adv_glue/adv_mnli_mismatched')
- Description :
Adversarial GLUE Benchmark (AdvGLUE) is a comprehensive robustness evaluation benchmark
that focuses on the adversarial robustness evaluation of language models. It covers five
natural language understanding tasks from the famous GLUE tasks and is an adversarial
version of GLUE benchmark.
- Licence : Aucune licence connue
- Version : 1.0.0
- Divisions :
Diviser | Exemples |
---|---|
'validation' | 162 |
- Caractéristiques :
{
"premise": {
"dtype": "string",
"id": null,
"_type": "Value"
},
"hypothesis": {
"dtype": "string",
"id": null,
"_type": "Value"
},
"idx": {
"dtype": "int32",
"id": null,
"_type": "Value"
}
}
adv_qnli
Utilisez la commande suivante pour charger cet ensemble de données dans TFDS :
ds = tfds.load('huggingface:adv_glue/adv_qnli')
- Description :
Adversarial GLUE Benchmark (AdvGLUE) is a comprehensive robustness evaluation benchmark
that focuses on the adversarial robustness evaluation of language models. It covers five
natural language understanding tasks from the famous GLUE tasks and is an adversarial
version of GLUE benchmark.
- Licence : Aucune licence connue
- Version : 1.0.0
- Divisions :
Diviser | Exemples |
---|---|
'validation' | 148 |
- Caractéristiques :
{
"question": {
"dtype": "string",
"id": null,
"_type": "Value"
},
"sentence": {
"dtype": "string",
"id": null,
"_type": "Value"
},
"idx": {
"dtype": "int32",
"id": null,
"_type": "Value"
}
}
adv_rte
Utilisez la commande suivante pour charger cet ensemble de données dans TFDS :
ds = tfds.load('huggingface:adv_glue/adv_rte')
- Description :
Adversarial GLUE Benchmark (AdvGLUE) is a comprehensive robustness evaluation benchmark
that focuses on the adversarial robustness evaluation of language models. It covers five
natural language understanding tasks from the famous GLUE tasks and is an adversarial
version of GLUE benchmark.
- Licence : Aucune licence connue
- Version : 1.0.0
- Divisions :
Diviser | Exemples |
---|---|
'validation' | 81 |
- Caractéristiques :
{
"sentence1": {
"dtype": "string",
"id": null,
"_type": "Value"
},
"sentence2": {
"dtype": "string",
"id": null,
"_type": "Value"
},
"idx": {
"dtype": "int32",
"id": null,
"_type": "Value"
}
}