Dataset Viewer
Auto-converted to Parquet Duplicate
Model_name
stringlengths
12
30
Train_size
int64
50.8k
50.8k
Test_size
int64
12.7k
12.7k
lora
listlengths
2
8
r
int64
12
64
Trainable_parameters
int64
3.59M
52.2M
Memory Allocation
float64
0.43
4.93
accuracy
float64
0.9
90.8
f1_macro
float64
0.89
90.3
f1_weighted
float64
0.9
90.8
precision
float64
0.89
90.3
recall
float64
0.9
90.2
FacebookAI/roberta-large
50,775
12,652
[ "out_proj", "value", "dense", "key", "query" ]
12
12,918,970
1.207656
90.246601
89.802879
90.260977
89.93768
89.700889
FacebookAI/xlm-roberta-large
50,775
12,652
[ "dense", "out_proj" ]
12
6,504,058
0.825117
89.977869
89.593698
89.995384
89.727738
89.497052
Qwen/Qwen3-Reranker-0.6B
50,775
12,652
[ "down_proj", "gate_proj", "o_proj", "score", "up_proj" ]
12
9,978,802
1.703594
90.0411
89.583316
90.054603
89.763457
89.43125
RUCAIBox/mvp
50,775
12,652
[ "fc2", "q_proj", "out_proj", "fc1", "dense", "v_proj", "k_proj" ]
12
17,195,578
1.770498
90.460006
90.053034
90.474028
90.179157
89.952785
answerdotai/ModernBERT-large
50,775
12,652
[ "Wi", "Wo", "classifier", "dense" ]
12
7,573,210
1.188389
90.460006
90.070314
90.474233
90.214053
89.94897
facebook/bart-large
50,775
12,652
[ "fc2", "q_proj", "out_proj", "fc1", "dense", "v_proj", "k_proj" ]
12
17,195,578
1.690195
90.436295
90.019956
90.450537
90.173706
89.895379
facebook/opt-125m
50,775
12,652
[ "fc2", "q_proj", "out_proj", "fc1", "v_proj", "k_proj", "score" ]
12
3,594,290
0.43123
89.029402
88.547062
89.039785
88.658871
88.453779
facebook/opt-350m
50,775
12,652
[ "fc1", "fc2", "out_proj", "project_in", "project_out", "score" ]
12
6,460,026
0.905674
89.464116
89.023809
89.476387
89.192131
88.899953
google-bert/bert-large-uncased
50,775
12,652
[ "classifier", "dense" ]
12
6,504,058
0.837207
89.61429
89.180331
89.623731
89.245118
89.128224
google-t5/t5-large
50,775
12,652
[ "wo", "out_proj", "q", "wi", "dense", "k", "v", "o" ]
12
34,302,010
3.54624
90.768258
90.266154
90.787784
90.340215
90.213506
Qwen/Qwen2-1.5B
50,775
12,652
[ "down_proj", "gate_proj", "o_proj", "up_proj" ]
32
8,429,568
1.617012
0.903889
0.899308
0.904411
0.896445
0.90323
Qwen/Qwen3-1.7B
50,775
12,652
[ "down_proj", "gate_proj", "o_proj", "up_proj" ]
16
29,589,504
2.122236
0.899937
0.894225
0.900583
0.891233
0.898903
facebook/opt-2.7b
50,775
12,652
[ "fc1", "fc2", "out_proj" ]
32
19,906,560
2.394766
0.899542
0.893458
0.900166
0.890395
0.898157
tiiuae/Falcon3-3B-Base
50,775
12,652
[ "down_proj", "gate_proj", "o_proj", "up_proj" ]
16
52,174,848
1.645117
0.899146
0.893445
0.899648
0.890745
0.897115
Qwen/Qwen2.5-3B-Instruct
50,775
12,652
[ "down_proj", "gate_proj", "k_proj", "o_proj", "q_proj", "up_proj", "v_proj" ]
64
12,257,280
3.259326
0.905469
0.90008
0.905995
0.897099
0.90432
Alibaba-NLP/E2Rank-4B
50,775
12,652
[ "down_proj", "gate_proj", "k_proj", "o_proj", "q_proj", "up_proj", "v_proj" ]
64
20,146,176
4.926826
0.904916
0.899534
0.905499
0.896799
0.903723
Qwen/Qwen3-Embedding-4B
50,775
12,652
[ "down_proj", "gate_proj", "o_proj", "up_proj" ]
32
29,859,840
4.167988
0.90792
0.902632
0.908473
0.899744
0.906993
README.md exists but content is empty.
Downloads last month
22