🚀 ModernBERT-large 模型
ModernBERT-large 是一个经过多任务微调的模型,在自然语言推理、零样本分类等任务中表现出色,具备强大的推理和细粒度分析能力。
🚀 快速开始
本模型是在多个自然语言推理(NLI)任务上进行多任务微调的 ModernBERT 模型,涵盖了 MNLI、ANLI、SICK、WANLI、doc - nli、LingNLI、FOLIO、FOL - NLI、LogicNLI、Label - NLI 等任务(以及下表中的所有数据集)。这是一个 “指令” 版本的模型。该模型在 Nvidia A30 GPU 上训练了 200k 步。
它在推理任务(在 ANLI 和 FOLIO 上表现优于 llama 3.1 8B Instruct)、长上下文推理、情感分析和使用新标签的零样本分类方面表现出色。
✨ 主要特性
- 强大的推理能力:在多种推理任务中表现优秀,优于部分同类模型。
- 长上下文处理:能够处理长上下文的推理任务。
- 多任务适用性:适用于情感分析和零样本分类等多种任务。
📚 详细文档
模型测试准确率
以下表格展示了模型的测试准确率。这些是同一个变压器模型搭配不同分类头的得分。通过在单一任务(如 SST)上进行微调可以进一步提高性能,但此检查点在零样本分类和自然语言推理(矛盾/蕴含/中立分类)方面表现出色。
测试名称 |
测试准确率 |
glue/mnli |
0.89 |
glue/qnli |
0.96 |
glue/rte |
0.91 |
glue/wnli |
0.64 |
glue/mrpc |
0.81 |
glue/qqp |
0.87 |
glue/cola |
0.87 |
glue/sst2 |
0.96 |
super_glue/boolq |
0.66 |
super_glue/cb |
0.86 |
super_glue/multirc |
0.9 |
super_glue/wic |
0.71 |
super_glue/axg |
1 |
anli/a1 |
0.72 |
anli/a2 |
0.54 |
anli/a3 |
0.55 |
sick/label |
0.91 |
sick/entailment_AB |
0.93 |
snli |
0.94 |
scitail/snli_format |
0.95 |
hans |
1 |
WANLI |
0.77 |
recast/recast_ner |
0.85 |
recast/recast_sentiment |
0.97 |
recast/recast_verbnet |
0.89 |
recast/recast_megaveridicality |
0.87 |
recast/recast_verbcorner |
0.87 |
recast/recast_kg_relations |
0.9 |
recast/recast_factuality |
0.95 |
recast/recast_puns |
0.98 |
probability_words_nli/reasoning_1hop |
1 |
probability_words_nli/usnli |
0.79 |
probability_words_nli/reasoning_2hop |
0.98 |
nan - nli |
0.85 |
nli_fever |
0.78 |
breaking_nli |
0.99 |
conj_nli |
0.72 |
fracas |
0.79 |
dialogue_nli |
0.94 |
mpe |
0.75 |
dnc |
0.91 |
recast_white/fnplus |
0.76 |
recast_white/sprl |
0.9 |
recast_white/dpr |
0.84 |
add_one_rte |
0.94 |
paws/labeled_final |
0.96 |
pragmeval/pdtb |
0.56 |
lex_glue/scotus |
0.58 |
lex_glue/ledgar |
0.85 |
dynasent/dynabench.dynasent.r1.all/r1 |
0.83 |
dynasent/dynabench.dynasent.r2.all/r2 |
0.76 |
cycic_classification |
0.96 |
lingnli |
0.91 |
monotonicity - entailment |
0.97 |
scinli |
0.88 |
naturallogic |
0.93 |
dynahate |
0.86 |
syntactic - augmentation - nli |
0.94 |
autotnli |
0.92 |
defeasible - nli/atomic |
0.83 |
defeasible - nli/snli |
0.8 |
help - nli |
0.96 |
nli - veridicality - transitivity |
0.99 |
lonli |
0.99 |
dadc - limit - nli |
0.79 |
folio |
0.71 |
tomi - nli |
0.54 |
puzzte |
0.59 |
temporal - nli |
0.93 |
counterfactually - augmented - snli |
0.81 |
cnli |
0.9 |
boolq - natural - perturbations |
0.72 |
equate |
0.65 |
logiqa - 2.0 - nli |
0.58 |
mindgames |
0.96 |
ConTRoL - nli |
0.66 |
logical - fallacy |
0.38 |
cladder |
0.89 |
conceptrules_v2 |
1 |
zero - shot - label - nli |
0.79 |
scone |
1 |
monli |
1 |
SpaceNLI |
1 |
propsegment/nli |
0.92 |
FLD.v2/default |
0.91 |
FLD.v2/star |
0.78 |
SDOH - NLI |
0.99 |
scifact_entailment |
0.87 |
feasibilityQA |
0.79 |
AdjectiveScaleProbe - nli |
1 |
resnli |
1 |
semantic_fragments_nli |
1 |
dataset_train_nli |
0.95 |
nlgraph |
0.97 |
ruletaker |
0.99 |
PARARULE - Plus |
1 |
logical - entailment |
0.93 |
nope |
0.56 |
LogicNLI |
0.91 |
contract - nli/contractnli_a/seg |
0.88 |
contract - nli/contractnli_b/full |
0.84 |
nli4ct_semeval2024 |
0.72 |
biosift - nli |
0.92 |
SIGA - nli |
0.57 |
FOL - nli |
0.79 |
doc - nli |
0.81 |
mctest - nli |
0.92 |
natural - language - satisfiability |
0.92 |
idioms - nli |
0.83 |
lifecycle - entailment |
0.79 |
MSciNLI |
0.84 |
hover - 3way/nli |
0.92 |
seahorse_summarization_evaluation |
0.81 |
missing - item - prediction/contrastive |
0.88 |
Pol_NLI |
0.93 |
synthetic - retrieval - NLI/count |
0.72 |
synthetic - retrieval - NLI/position |
0.9 |
synthetic - retrieval - NLI/binary |
0.92 |
babi_nli |
0.98 |
模型信息
属性 |
详情 |
库名称 |
transformers |
基础模型 |
answerdotai/ModernBERT - large |
许可证 |
apache - 2.0 |
语言 |
英语 |
任务类型 |
零样本分类 |
训练数据集 |
nyu - mll/glue、facebook/anli |
标签 |
instruct、natural - language - inference、nli |
💻 使用示例
基础用法
[ZS] 零样本分类管道
from transformers import pipeline
classifier = pipeline("zero-shot-classification",model="tasksource/ModernBERT-large-nli")
text = "one day I will see the world"
candidate_labels = ['travel', 'cooking', 'dancing']
classifier(text, candidate_labels)
此模型的 NLI 训练数据包括 label - nli,这是一个专门为提高此类零样本分类而构建的 NLI 数据集。
[NLI] 自然语言推理管道
from transformers import pipeline
pipe = pipeline("text-classification",model="tasksource/ModernBERT-large-nli")
pipe([dict(text='there is a cat',
text_pair='there is a black cat')])
高级用法
进一步微调的骨干模型
此检查点比基础版本具有更强的推理和细粒度分析能力,可用于进一步的微调。
📄 许可证
本模型使用 apache - 2.0 许可证。
📖 引用
@inproceedings{sileo-2024-tasksource,
title = "tasksource: A Large Collection of {NLP} tasks with a Structured Dataset Preprocessing Framework",
author = "Sileo, Damien",
booktitle = "Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)",
month = may,
year = "2024",
address = "Torino, Italia",
publisher = "ELRA and ICCL",
url = "https://aclanthology.org/2024.lrec-main.1361",
pages = "15655--15684",
}