Safetensors
English
vidore
manu commited on
Commit
a571124
1 Parent(s): 8705a6e

Upload folder using huggingface_hub

Browse files
adapter_config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "scripts/configs/siglip/../../../models/siglip-so400m-patch14-384",
5
+ "bias": "none",
6
+ "fan_in_fan_out": false,
7
+ "inference_mode": false,
8
+ "init_lora_weights": "gaussian",
9
+ "layer_replication": null,
10
+ "layers_pattern": null,
11
+ "layers_to_transform": null,
12
+ "loftq_config": {},
13
+ "lora_alpha": 32,
14
+ "lora_dropout": 0.1,
15
+ "megatron_config": null,
16
+ "megatron_core": "megatron.core",
17
+ "modules_to_save": null,
18
+ "peft_type": "LORA",
19
+ "r": 32,
20
+ "rank_pattern": {},
21
+ "revision": null,
22
+ "target_modules": "(.*(text_model).*(down_proj|gate_proj|up_proj|k_proj|q_proj|v_proj|o_proj).*$)",
23
+ "task_type": "FEATURE_EXTRACTION",
24
+ "use_dora": false,
25
+ "use_rslora": false
26
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5065a2a8318651e6be28a53764c8cd651d7d3c7477859e6c506b655c575ef20b
3
+ size 11967752
git_hash.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ 83ba45daf034b63549ce091c9fc7d1b05d17381a
preprocessor_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_valid_processor_keys": [
3
+ "images",
4
+ "do_resize",
5
+ "size",
6
+ "resample",
7
+ "do_rescale",
8
+ "rescale_factor",
9
+ "do_normalize",
10
+ "image_mean",
11
+ "image_std",
12
+ "return_tensors",
13
+ "data_format",
14
+ "input_data_format",
15
+ "do_convert_rgb"
16
+ ],
17
+ "do_convert_rgb": null,
18
+ "do_normalize": true,
19
+ "do_rescale": true,
20
+ "do_resize": true,
21
+ "image_mean": [
22
+ 0.5,
23
+ 0.5,
24
+ 0.5
25
+ ],
26
+ "image_processor_type": "SiglipImageProcessor",
27
+ "image_std": [
28
+ 0.5,
29
+ 0.5,
30
+ 0.5
31
+ ],
32
+ "processor_class": "SiglipProcessor",
33
+ "resample": 3,
34
+ "rescale_factor": 0.00392156862745098,
35
+ "size": {
36
+ "height": 384,
37
+ "width": 384
38
+ }
39
+ }
results.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"validation_set": {"ndcg_at_1": 0.584, "ndcg_at_3": 0.65926, "ndcg_at_5": 0.68619, "ndcg_at_10": 0.70901, "ndcg_at_20": 0.72156, "ndcg_at_100": 0.73549, "ndcg_at_1000": 0.73906, "map_at_1": 0.584, "map_at_3": 0.641, "map_at_5": 0.6558, "map_at_10": 0.66531, "map_at_20": 0.66871, "map_at_100": 0.67057, "map_at_1000": 0.67074, "recall_at_1": 0.584, "recall_at_3": 0.712, "recall_at_5": 0.778, "recall_at_10": 0.848, "recall_at_20": 0.898, "recall_at_100": 0.974, "recall_at_1000": 1.0, "precision_at_1": 0.584, "precision_at_3": 0.23733, "precision_at_5": 0.1556, "precision_at_10": 0.0848, "precision_at_20": 0.0449, "precision_at_100": 0.00974, "precision_at_1000": 0.001, "mrr_at_1": 0.582, "mrr_at_3": 0.6406666666666665, "mrr_at_5": 0.6558666666666664, "mrr_at_10": 0.6645682539682537, "mrr_at_20": 0.6681508925285237, "mrr_at_100": 0.6700156171199655, "mrr_at_1000": 0.6701911266661673, "naucs_at_1_max": 0.6378681589019386, "naucs_at_1_std": -0.0630487762217373, "naucs_at_1_diff1": 0.8086768926652304, "naucs_at_3_max": 0.6205150293207277, "naucs_at_3_std": -0.058016610255415935, "naucs_at_3_diff1": 0.7396439845358828, "naucs_at_5_max": 0.6141894030901982, "naucs_at_5_std": -0.014446383109978518, "naucs_at_5_diff1": 0.7136142848757988, "naucs_at_10_max": 0.6607969813177121, "naucs_at_10_std": 0.06600776788929309, "naucs_at_10_diff1": 0.6807592574014549, "naucs_at_20_max": 0.6746666290474197, "naucs_at_20_std": 0.2983425414364643, "naucs_at_20_diff1": 0.7040712894033511, "naucs_at_100_max": 0.9242995566524874, "naucs_at_100_std": 0.8109961933491295, "naucs_at_100_diff1": 0.8821316857699331, "naucs_at_1000_max": 1.0, "naucs_at_1000_std": 1.0, "naucs_at_1000_diff1": 1.0}, "syntheticDocQA_energy": {"ndcg_at_1": 0.61, "ndcg_at_3": 0.71202, "ndcg_at_5": 0.73655, "ndcg_at_10": 0.75647, "ndcg_at_20": 0.76674, "ndcg_at_100": 0.776, "ndcg_at_1000": 0.77728, "map_at_1": 0.61, "map_at_3": 0.68833, "map_at_5": 0.70183, "map_at_10": 0.71037, "map_at_20": 0.71328, "map_at_100": 0.71453, "map_at_1000": 0.71457, "recall_at_1": 0.61, "recall_at_3": 0.78, "recall_at_5": 0.84, "recall_at_10": 0.9, "recall_at_20": 0.94, "recall_at_100": 0.99, "recall_at_1000": 1.0, "precision_at_1": 0.61, "precision_at_3": 0.26, "precision_at_5": 0.168, "precision_at_10": 0.09, "precision_at_20": 0.047, "precision_at_100": 0.0099, "precision_at_1000": 0.001, "mrr_at_1": 0.62, "mrr_at_3": 0.6966666666666668, "mrr_at_5": 0.7086666666666668, "mrr_at_10": 0.7173452380952382, "mrr_at_20": 0.7203019480519481, "mrr_at_100": 0.7215746292501511, "mrr_at_1000": 0.721619674295196, "naucs_at_1_max": 0.35279481856803946, "naucs_at_1_std": -0.27883613070868557, "naucs_at_1_diff1": 0.8297484395863376, "naucs_at_3_max": 0.5471653659365916, "naucs_at_3_std": -0.29320257801270416, "naucs_at_3_diff1": 0.7976877066458377, "naucs_at_5_max": 0.6855134616638254, "naucs_at_5_std": -0.0886107020127646, "naucs_at_5_diff1": 0.7807154186020162, "naucs_at_10_max": 0.7069198049590212, "naucs_at_10_std": -0.039355742296917365, "naucs_at_10_diff1": 0.7281835811247575, "naucs_at_20_max": 0.6993118234948302, "naucs_at_20_std": -0.034936196700904094, "naucs_at_20_diff1": 0.7907239819004533, "naucs_at_100_max": 0.5559705363627031, "naucs_at_100_std": -0.5634920634920583, "naucs_at_100_diff1": 0.8707175177763411, "naucs_at_1000_max": NaN, "naucs_at_1000_std": NaN, "naucs_at_1000_diff1": NaN}, "syntheticDocQA_healthcare_industry": {"ndcg_at_1": 0.72, "ndcg_at_3": 0.80678, "ndcg_at_5": 0.82313, "ndcg_at_10": 0.8298, "ndcg_at_20": 0.83718, "ndcg_at_100": 0.84426, "ndcg_at_1000": 0.84426, "map_at_1": 0.72, "map_at_3": 0.785, "map_at_5": 0.794, "map_at_10": 0.79686, "map_at_20": 0.79879, "map_at_100": 0.79968, "map_at_1000": 0.79968, "recall_at_1": 0.72, "recall_at_3": 0.87, "recall_at_5": 0.91, "recall_at_10": 0.93, "recall_at_20": 0.96, "recall_at_100": 1.0, "recall_at_1000": 1.0, "precision_at_1": 0.72, "precision_at_3": 0.29, "precision_at_5": 0.182, "precision_at_10": 0.093, "precision_at_20": 0.048, "precision_at_100": 0.01, "precision_at_1000": 0.001, "mrr_at_1": 0.72, "mrr_at_3": 0.7899999999999998, "mrr_at_5": 0.7964999999999999, "mrr_at_10": 0.7993571428571427, "mrr_at_20": 0.8012925495557073, "mrr_at_100": 0.802182456562281, "mrr_at_1000": 0.802182456562281, "naucs_at_1_max": 0.5040470632499849, "naucs_at_1_std": -0.19158272283272318, "naucs_at_1_diff1": 0.6287869335383656, "naucs_at_3_max": 0.6078961861219718, "naucs_at_3_std": -0.07498333703621549, "naucs_at_3_diff1": 0.6880716814703113, "naucs_at_5_max": 0.5082491717282229, "naucs_at_5_std": -0.21968046477850045, "naucs_at_5_diff1": 0.7400354941030779, "naucs_at_10_max": 0.7490879963213166, "naucs_at_10_std": -0.05142056822729049, "naucs_at_10_diff1": 0.7955565204805319, "naucs_at_20_max": 0.7401836826459105, "naucs_at_20_std": -0.3278478057889798, "naucs_at_20_diff1": 0.7102777281125197, "naucs_at_100_max": NaN, "naucs_at_100_std": NaN, "naucs_at_100_diff1": NaN, "naucs_at_1000_max": NaN, "naucs_at_1000_std": NaN, "naucs_at_1000_diff1": NaN}, "syntheticDocQA_artificial_intelligence_test": {"ndcg_at_1": 0.62, "ndcg_at_3": 0.72202, "ndcg_at_5": 0.74312, "ndcg_at_10": 0.756, "ndcg_at_20": 0.76626, "ndcg_at_100": 0.77815, "ndcg_at_1000": 0.78075, "map_at_1": 0.62, "map_at_3": 0.69833, "map_at_5": 0.71033, "map_at_10": 0.71561, "map_at_20": 0.7185, "map_at_100": 0.7204, "map_at_1000": 0.72052, "recall_at_1": 0.62, "recall_at_3": 0.79, "recall_at_5": 0.84, "recall_at_10": 0.88, "recall_at_20": 0.92, "recall_at_100": 0.98, "recall_at_1000": 1.0, "precision_at_1": 0.62, "precision_at_3": 0.26333, "precision_at_5": 0.168, "precision_at_10": 0.088, "precision_at_20": 0.046, "precision_at_100": 0.0098, "precision_at_1000": 0.001, "mrr_at_1": 0.62, "mrr_at_3": 0.7016666666666667, "mrr_at_5": 0.7111666666666667, "mrr_at_10": 0.7164444444444444, "mrr_at_20": 0.7193360014168839, "mrr_at_100": 0.721291521824503, "mrr_at_1000": 0.7214070180304247, "naucs_at_1_max": 0.3587976494477431, "naucs_at_1_std": -0.23265544718949024, "naucs_at_1_diff1": 0.6797309474890378, "naucs_at_3_max": 0.4794046216365683, "naucs_at_3_std": -0.22596003558804434, "naucs_at_3_diff1": 0.6751423743712268, "naucs_at_5_max": 0.5935198821796747, "naucs_at_5_std": -0.14190598919980307, "naucs_at_5_diff1": 0.605993740372159, "naucs_at_10_max": 0.7616131084950692, "naucs_at_10_std": 0.0047327394209372015, "naucs_at_10_diff1": 0.5520581470506981, "naucs_at_20_max": 0.6338702147525689, "naucs_at_20_std": -0.3121498599439731, "naucs_at_20_diff1": 0.4777756794515646, "naucs_at_100_max": 0.8692810457516348, "naucs_at_100_std": 0.9346405228758174, "naucs_at_100_diff1": -0.08116683881238204, "naucs_at_1000_max": NaN, "naucs_at_1000_std": NaN, "naucs_at_1000_diff1": NaN}, "syntheticDocQA_government_reports": {"ndcg_at_1": 0.55, "ndcg_at_3": 0.70488, "ndcg_at_5": 0.74188, "ndcg_at_10": 0.75167, "ndcg_at_20": 0.75937, "ndcg_at_100": 0.76703, "ndcg_at_1000": 0.76703, "map_at_1": 0.55, "map_at_3": 0.66833, "map_at_5": 0.68883, "map_at_10": 0.69293, "map_at_20": 0.69509, "map_at_100": 0.69622, "map_at_1000": 0.69622, "recall_at_1": 0.55, "recall_at_3": 0.81, "recall_at_5": 0.9, "recall_at_10": 0.93, "recall_at_20": 0.96, "recall_at_100": 1.0, "recall_at_1000": 1.0, "precision_at_1": 0.55, "precision_at_3": 0.27, "precision_at_5": 0.18, "precision_at_10": 0.093, "precision_at_20": 0.048, "precision_at_100": 0.01, "precision_at_1000": 0.001, "mrr_at_1": 0.55, "mrr_at_3": 0.6716666666666667, "mrr_at_5": 0.6921666666666667, "mrr_at_10": 0.6963730158730158, "mrr_at_20": 0.6985364774114775, "mrr_at_100": 0.6996883942498957, "mrr_at_1000": 0.6996883942498957, "naucs_at_1_max": 0.39614107776896795, "naucs_at_1_std": -0.05161833883311121, "naucs_at_1_diff1": 0.7058759321138007, "naucs_at_3_max": 0.5654922562655941, "naucs_at_3_std": 0.07077374404107108, "naucs_at_3_diff1": 0.5614385203195352, "naucs_at_5_max": 0.4612609702026484, "naucs_at_5_std": -0.17170868347338447, "naucs_at_5_diff1": 0.4838798534486071, "naucs_at_10_max": 0.6551324932488741, "naucs_at_10_std": -0.3175270108043195, "naucs_at_10_diff1": 0.5189682451095561, "naucs_at_20_max": 0.8611111111111085, "naucs_at_20_std": 0.455182072829132, "naucs_at_20_diff1": 0.5640091900137633, "naucs_at_100_max": NaN, "naucs_at_100_std": NaN, "naucs_at_100_diff1": NaN, "naucs_at_1000_max": NaN, "naucs_at_1000_std": NaN, "naucs_at_1000_diff1": NaN}, "infovqa_subsampled": {"ndcg_at_1": 0.59, "ndcg_at_3": 0.68714, "ndcg_at_5": 0.70521, "ndcg_at_10": 0.72425, "ndcg_at_20": 0.73174, "ndcg_at_100": 0.7445, "ndcg_at_1000": 0.75005, "map_at_1": 0.59, "map_at_3": 0.66333, "map_at_5": 0.67333, "map_at_10": 0.68136, "map_at_20": 0.68337, "map_at_100": 0.68504, "map_at_1000": 0.68527, "recall_at_1": 0.59, "recall_at_3": 0.756, "recall_at_5": 0.8, "recall_at_10": 0.858, "recall_at_20": 0.888, "recall_at_100": 0.958, "recall_at_1000": 1.0, "precision_at_1": 0.59, "precision_at_3": 0.252, "precision_at_5": 0.16, "precision_at_10": 0.0858, "precision_at_20": 0.0444, "precision_at_100": 0.00958, "precision_at_1000": 0.001, "mrr_at_1": 0.594, "mrr_at_3": 0.6649999999999998, "mrr_at_5": 0.6746999999999996, "mrr_at_10": 0.682990476190476, "mrr_at_20": 0.6849822625860548, "mrr_at_100": 0.6866517475331556, "mrr_at_1000": 0.686884696889696, "naucs_at_1_max": 0.4750876644260999, "naucs_at_1_std": -0.10045636638443373, "naucs_at_1_diff1": 0.7674943637513642, "naucs_at_3_max": 0.5415688290260486, "naucs_at_3_std": 0.034602149986765765, "naucs_at_3_diff1": 0.6498617044832607, "naucs_at_5_max": 0.5852949578747686, "naucs_at_5_std": 0.11070429570429464, "naucs_at_5_diff1": 0.6203305154498029, "naucs_at_10_max": 0.656883904234058, "naucs_at_10_std": 0.36746419337628644, "naucs_at_10_diff1": 0.6148801675321266, "naucs_at_20_max": 0.6692752886969887, "naucs_at_20_std": 0.42040381493506335, "naucs_at_20_diff1": 0.5748871495607518, "naucs_at_100_max": 0.6902135142860049, "naucs_at_100_std": 0.5294784580498858, "naucs_at_100_diff1": 0.6635405966594489, "naucs_at_1000_max": 1.0, "naucs_at_1000_std": 1.0, "naucs_at_1000_diff1": 1.0}, "docvqa_subsampled": {"ndcg_at_1": 0.238, "ndcg_at_3": 0.30526, "ndcg_at_5": 0.3291, "ndcg_at_10": 0.35895, "ndcg_at_20": 0.37686, "ndcg_at_100": 0.41193, "ndcg_at_1000": 0.44324, "map_at_1": 0.238, "map_at_3": 0.28967, "map_at_5": 0.30287, "map_at_10": 0.31525, "map_at_20": 0.32028, "map_at_100": 0.32521, "map_at_1000": 0.32642, "recall_at_1": 0.238, "recall_at_3": 0.35, "recall_at_5": 0.408, "recall_at_10": 0.5, "recall_at_20": 0.57, "recall_at_100": 0.758, "recall_at_1000": 1.0, "precision_at_1": 0.238, "precision_at_3": 0.11667, "precision_at_5": 0.0816, "precision_at_10": 0.05, "precision_at_20": 0.0285, "precision_at_100": 0.00758, "precision_at_1000": 0.001, "mrr_at_1": 0.242, "mrr_at_3": 0.29166666666666663, "mrr_at_5": 0.30486666666666673, "mrr_at_10": 0.3173301587301587, "mrr_at_20": 0.32237405155903603, "mrr_at_100": 0.3273195432373183, "mrr_at_1000": 0.3285276950292888, "naucs_at_1_max": 0.48213743667838665, "naucs_at_1_std": 0.11681943763847841, "naucs_at_1_diff1": 0.7032765364046992, "naucs_at_3_max": 0.5260569818685588, "naucs_at_3_std": 0.17266530423890825, "naucs_at_3_diff1": 0.6249454086703624, "naucs_at_5_max": 0.5778357812048011, "naucs_at_5_std": 0.21954676415006688, "naucs_at_5_diff1": 0.6118204750271781, "naucs_at_10_max": 0.5777470327417413, "naucs_at_10_std": 0.2905982678214522, "naucs_at_10_diff1": 0.5456515785012462, "naucs_at_20_max": 0.5775000025691972, "naucs_at_20_std": 0.35737144989308395, "naucs_at_20_diff1": 0.5759268371720303, "naucs_at_100_max": 0.6922983438805927, "naucs_at_100_std": 0.6006094023301695, "naucs_at_100_diff1": 0.5791109625293102, "naucs_at_1000_max": 1.0, "naucs_at_1000_std": 1.0, "naucs_at_1000_diff1": 1.0}, "arxivqa_subsampled": {"ndcg_at_1": 0.492, "ndcg_at_3": 0.566, "ndcg_at_5": 0.58476, "ndcg_at_10": 0.61329, "ndcg_at_20": 0.63126, "ndcg_at_100": 0.65766, "ndcg_at_1000": 0.662, "map_at_1": 0.492, "map_at_3": 0.548, "map_at_5": 0.5583, "map_at_10": 0.57011, "map_at_20": 0.57493, "map_at_100": 0.57845, "map_at_1000": 0.57864, "recall_at_1": 0.492, "recall_at_3": 0.618, "recall_at_5": 0.664, "recall_at_10": 0.752, "recall_at_20": 0.824, "recall_at_100": 0.968, "recall_at_1000": 1.0, "precision_at_1": 0.492, "precision_at_3": 0.206, "precision_at_5": 0.1328, "precision_at_10": 0.0752, "precision_at_20": 0.0412, "precision_at_100": 0.00968, "precision_at_1000": 0.001, "mrr_at_1": 0.488, "mrr_at_3": 0.5460000000000002, "mrr_at_5": 0.5569999999999999, "mrr_at_10": 0.5679055555555554, "mrr_at_20": 0.5727728744251963, "mrr_at_100": 0.5763984587423269, "mrr_at_1000": 0.5765952795524785, "naucs_at_1_max": 0.5051502720320145, "naucs_at_1_std": -0.1490531541155774, "naucs_at_1_diff1": 0.7430229429291968, "naucs_at_3_max": 0.5132701970287259, "naucs_at_3_std": -0.15837310747140693, "naucs_at_3_diff1": 0.6668445823582236, "naucs_at_5_max": 0.4953827950259241, "naucs_at_5_std": -0.1276612827684026, "naucs_at_5_diff1": 0.6584245957611312, "naucs_at_10_max": 0.5159804439246654, "naucs_at_10_std": -0.048523562327243905, "naucs_at_10_diff1": 0.6143019465544078, "naucs_at_20_max": 0.5575369120966851, "naucs_at_20_std": 0.021292506297229136, "naucs_at_20_diff1": 0.5977514918610547, "naucs_at_100_max": 0.4182403348844746, "naucs_at_100_std": 0.015785480859009476, "naucs_at_100_diff1": 0.41161464945440435, "naucs_at_1000_max": -0.0, "naucs_at_1000_std": 1.0, "naucs_at_1000_diff1": 1.0}, "tabfquad_subsampled": {"ndcg_at_1": 0.52143, "ndcg_at_3": 0.61169, "ndcg_at_5": 0.6266, "ndcg_at_10": 0.65692, "ndcg_at_20": 0.67762, "ndcg_at_100": 0.69891, "ndcg_at_1000": 0.69891, "map_at_1": 0.52143, "map_at_3": 0.58869, "map_at_5": 0.59708, "map_at_10": 0.60976, "map_at_20": 0.61542, "map_at_100": 0.6186, "map_at_1000": 0.6186, "recall_at_1": 0.52143, "recall_at_3": 0.67857, "recall_at_5": 0.71429, "recall_at_10": 0.80714, "recall_at_20": 0.88929, "recall_at_100": 1.0, "recall_at_1000": 1.0, "precision_at_1": 0.52143, "precision_at_3": 0.22619, "precision_at_5": 0.14286, "precision_at_10": 0.08071, "precision_at_20": 0.04446, "precision_at_100": 0.01, "precision_at_1000": 0.001, "mrr_at_1": 0.5142857142857142, "mrr_at_3": 0.5851190476190478, "mrr_at_5": 0.5935119047619049, "mrr_at_10": 0.6058503401360544, "mrr_at_20": 0.6117789065540337, "mrr_at_100": 0.6149708790374979, "mrr_at_1000": 0.6149708790374979, "naucs_at_1_max": 0.4868863923702865, "naucs_at_1_std": 0.24971538109741254, "naucs_at_1_diff1": 0.6783266182026825, "naucs_at_3_max": 0.411472501368197, "naucs_at_3_std": 0.29670255306059673, "naucs_at_3_diff1": 0.6092358151392712, "naucs_at_5_max": 0.3865818323579307, "naucs_at_5_std": 0.295637898686679, "naucs_at_5_diff1": 0.5692539432219096, "naucs_at_10_max": 0.38925956441210646, "naucs_at_10_std": 0.27932981510868304, "naucs_at_10_diff1": 0.4986876424395795, "naucs_at_20_max": 0.3320879228556591, "naucs_at_20_std": 0.37500381318446735, "naucs_at_20_diff1": 0.5213327126314864, "naucs_at_100_max": 1.0, "naucs_at_100_std": 1.0, "naucs_at_100_diff1": -0.0, "naucs_at_1000_max": NaN, "naucs_at_1000_std": NaN, "naucs_at_1000_diff1": NaN}, "tatdqa": {"ndcg_at_1": 0.20746, "ndcg_at_3": 0.27866, "ndcg_at_5": 0.30512, "ndcg_at_10": 0.33772, "ndcg_at_20": 0.36832, "ndcg_at_100": 0.41958, "ndcg_at_1000": 0.43385, "map_at_1": 0.20746, "map_at_3": 0.26127, "map_at_5": 0.27565, "map_at_10": 0.28905, "map_at_20": 0.2974, "map_at_100": 0.30431, "map_at_1000": 0.30498, "recall_at_1": 0.20746, "recall_at_3": 0.32892, "recall_at_5": 0.39447, "recall_at_10": 0.49549, "recall_at_20": 0.61696, "recall_at_100": 0.89537, "recall_at_1000": 1.0, "precision_at_1": 0.20746, "precision_at_3": 0.10964, "precision_at_5": 0.07889, "precision_at_10": 0.04955, "precision_at_20": 0.03085, "precision_at_100": 0.00895, "precision_at_1000": 0.001, "mrr_at_1": 0.20805772699939867, "mrr_at_3": 0.26037282020444974, "mrr_at_5": 0.27459410703547765, "mrr_at_10": 0.2883975698918571, "mrr_at_20": 0.2966673640468522, "mrr_at_100": 0.3036664958789319, "mrr_at_1000": 0.3043435796303338, "naucs_at_1_max": 0.09554120254812996, "naucs_at_1_std": -0.04319255805542964, "naucs_at_1_diff1": 0.4354055996024844, "naucs_at_3_max": 0.10637847767297262, "naucs_at_3_std": 0.027853236732236504, "naucs_at_3_diff1": 0.3225557159516856, "naucs_at_5_max": 0.10795770273258308, "naucs_at_5_std": 0.04531330767590303, "naucs_at_5_diff1": 0.3034161365153788, "naucs_at_10_max": 0.13728143627291428, "naucs_at_10_std": 0.08287067171489244, "naucs_at_10_diff1": 0.27920328691666213, "naucs_at_20_max": 0.15261471308494734, "naucs_at_20_std": 0.16776681693702528, "naucs_at_20_diff1": 0.24549869081668294, "naucs_at_100_max": 0.21871573037711095, "naucs_at_100_std": 0.28884476007699017, "naucs_at_100_diff1": 0.22557527469999167, "naucs_at_1000_max": NaN, "naucs_at_1000_std": NaN, "naucs_at_1000_diff1": NaN}, "shift_project": {"ndcg_at_1": 0.16, "ndcg_at_3": 0.23547, "ndcg_at_5": 0.26475, "ndcg_at_10": 0.29016, "ndcg_at_20": 0.30291, "ndcg_at_100": 0.35133, "ndcg_at_1000": 0.38271, "map_at_1": 0.16, "map_at_3": 0.21667, "map_at_5": 0.23317, "map_at_10": 0.2434, "map_at_20": 0.24696, "map_at_100": 0.2537, "map_at_1000": 0.2548, "recall_at_1": 0.16, "recall_at_3": 0.29, "recall_at_5": 0.36, "recall_at_10": 0.44, "recall_at_20": 0.49, "recall_at_100": 0.75, "recall_at_1000": 1.0, "precision_at_1": 0.16, "precision_at_3": 0.09667, "precision_at_5": 0.072, "precision_at_10": 0.044, "precision_at_20": 0.0245, "precision_at_100": 0.0075, "precision_at_1000": 0.001, "mrr_at_1": 0.15, "mrr_at_3": 0.21666666666666667, "mrr_at_5": 0.22916666666666669, "mrr_at_10": 0.24106746031746035, "mrr_at_20": 0.24484785536101328, "mrr_at_100": 0.25164910352824316, "mrr_at_1000": 0.2527711454997698, "naucs_at_1_max": 0.10005541483901749, "naucs_at_1_std": -0.05903364572130915, "naucs_at_1_diff1": 0.34747369821425966, "naucs_at_3_max": 0.024457548283731787, "naucs_at_3_std": -0.08260648599645645, "naucs_at_3_diff1": 0.3321252210834322, "naucs_at_5_max": 0.13466488497748016, "naucs_at_5_std": 0.04738098987036162, "naucs_at_5_diff1": 0.27937713181515583, "naucs_at_10_max": 0.30304011965300015, "naucs_at_10_std": 0.18848201701727887, "naucs_at_10_diff1": 0.25580418038081487, "naucs_at_20_max": 0.3214473699812928, "naucs_at_20_std": 0.20289989776001516, "naucs_at_20_diff1": 0.23865148183383278, "naucs_at_100_max": 0.39521039184904744, "naucs_at_100_std": 0.27336473050758797, "naucs_at_100_diff1": 0.2401113519052073, "naucs_at_1000_max": NaN, "naucs_at_1000_std": NaN, "naucs_at_1000_diff1": NaN}}
special_tokens_map.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eos_token": {
3
+ "content": "</s>",
4
+ "lstrip": true,
5
+ "normalized": false,
6
+ "rstrip": true,
7
+ "single_word": false
8
+ },
9
+ "pad_token": {
10
+ "content": "</s>",
11
+ "lstrip": true,
12
+ "normalized": false,
13
+ "rstrip": true,
14
+ "single_word": false
15
+ },
16
+ "unk_token": {
17
+ "content": "<unk>",
18
+ "lstrip": true,
19
+ "normalized": false,
20
+ "rstrip": true,
21
+ "single_word": false
22
+ }
23
+ }
spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e5036bed065526c3c212dfbe288752391797c4bb1a284aa18c9a0b23fcaf8ec
3
+ size 798330
tokenizer_config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "1": {
4
+ "content": "</s>",
5
+ "lstrip": true,
6
+ "normalized": false,
7
+ "rstrip": true,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "2": {
12
+ "content": "<unk>",
13
+ "lstrip": true,
14
+ "normalized": false,
15
+ "rstrip": true,
16
+ "single_word": false,
17
+ "special": true
18
+ }
19
+ },
20
+ "additional_special_tokens": [],
21
+ "clean_up_tokenization_spaces": true,
22
+ "do_lower_case": true,
23
+ "eos_token": "</s>",
24
+ "max_length": 64,
25
+ "model_input_names": [
26
+ "input_ids"
27
+ ],
28
+ "model_max_length": 64,
29
+ "pad_token": "</s>",
30
+ "processor_class": "SiglipProcessor",
31
+ "sp_model_kwargs": {},
32
+ "tokenizer_class": "SiglipTokenizer",
33
+ "unk_token": "<unk>"
34
+ }
training_config.yml ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ config:
2
+ (): custom_colbert.utils.train_custom_colbert_models.ColModelTrainingConfig
3
+ output_dir: !path ../../../models/without_tabfquad_no_pairwise/train_real_siglip_text_only
4
+ processor:
5
+ () : custom_colbert.utils.wrapper.AutoProcessorWrapper
6
+ pretrained_model_name_or_path: !path ../../../models/siglip-so400m-patch14-384
7
+ max_length: 64
8
+ model:
9
+ (): custom_colbert.utils.wrapper.AutoColModelWrapper
10
+ pretrained_model_name_or_path: !path ../../../models/siglip-so400m-patch14-384
11
+ training_objective: "biencoder_mean"
12
+ # attn_implementation: "eager"
13
+ torch_dtype: !ext torch.bfloat16
14
+ # device_map: "auto"
15
+ # quantization_config:
16
+ # (): transformers.BitsAndBytesConfig
17
+ # load_in_4bit: true
18
+ # bnb_4bit_quant_type: "nf4"
19
+ # bnb_4bit_compute_dtype: "bfloat16"
20
+ # bnb_4bit_use_double_quant: true
21
+
22
+ dataset_loading_func: !ext custom_colbert.utils.dataset_transformation.load_train_set
23
+ eval_dataset_loader: !import ../data/test_data.yaml
24
+
25
+ max_length: 64
26
+ run_train: true
27
+ run_eval: true
28
+ add_suffix: true
29
+ loss_func:
30
+ (): custom_colbert.loss.colbert_loss.BiEncoderLoss
31
+ tr_args: !import ../tr_args/default_tr_args.yaml
32
+ peft_config:
33
+ (): peft.LoraConfig
34
+ r: 32
35
+ lora_alpha: 32
36
+ lora_dropout: 0.1
37
+ init_lora_weights: "gaussian"
38
+ bias: "none"
39
+ task_type: "FEATURE_EXTRACTION"
40
+ target_modules: '(.*(text_model).*(down_proj|gate_proj|up_proj|k_proj|q_proj|v_proj|o_proj).*$)'