stevenbucaille commited on
Commit
9bf358b
·
verified ·
1 Parent(s): f63f906
Files changed (2) hide show
  1. config.json +161 -0
  2. model.safetensors +3 -0
config.json ADDED
@@ -0,0 +1,161 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation_function": "relu",
3
+ "architectures": [
4
+ "EfficientLoFTRForKeypointMatching"
5
+ ],
6
+ "attention_bias": false,
7
+ "attention_dropout": 0.0,
8
+ "batch_norm_eps": 1e-05,
9
+ "coarse_matching_border_removal": 2,
10
+ "coarse_matching_skip_softmax": false,
11
+ "coarse_matching_temperature": 0.1,
12
+ "coarse_matching_threshold": 0.2,
13
+ "embedding_size": [
14
+ 26,
15
+ 26
16
+ ],
17
+ "fine_fusion_dims": [
18
+ 256,
19
+ 128,
20
+ 64
21
+ ],
22
+ "fine_kernel_size": 8,
23
+ "fine_matching_regress_temperature": 10.0,
24
+ "fine_matching_slice_dim": 8,
25
+ "hidden_size": 256,
26
+ "initializer_range": 0.02,
27
+ "intermediate_size": 512,
28
+ "kv_aggregation_kernel_size": 4,
29
+ "kv_aggregation_stride": 4,
30
+ "mlp_activation_function": "leaky_relu",
31
+ "model_type": "efficientloftr",
32
+ "num_attention_heads": 8,
33
+ "num_attention_layers": 4,
34
+ "num_key_value_heads": 8,
35
+ "out_features": [
36
+ 64,
37
+ 64,
38
+ 128,
39
+ 256
40
+ ],
41
+ "partial_rotary_factor": 4.0,
42
+ "q_aggregation_kernel_size": 4,
43
+ "q_aggregation_stride": 4,
44
+ "rope_scaling": {
45
+ "rope_type": "default"
46
+ },
47
+ "rope_theta": 10000.0,
48
+ "stage_block_in_channels": [
49
+ [
50
+ 1
51
+ ],
52
+ [
53
+ 64,
54
+ 64
55
+ ],
56
+ [
57
+ 64,
58
+ 128,
59
+ 128,
60
+ 128
61
+ ],
62
+ [
63
+ 128,
64
+ 256,
65
+ 256,
66
+ 256,
67
+ 256,
68
+ 256,
69
+ 256,
70
+ 256,
71
+ 256,
72
+ 256,
73
+ 256,
74
+ 256,
75
+ 256,
76
+ 256
77
+ ]
78
+ ],
79
+ "stage_block_out_channels": [
80
+ [
81
+ 64
82
+ ],
83
+ [
84
+ 64,
85
+ 64
86
+ ],
87
+ [
88
+ 128,
89
+ 128,
90
+ 128,
91
+ 128
92
+ ],
93
+ [
94
+ 256,
95
+ 256,
96
+ 256,
97
+ 256,
98
+ 256,
99
+ 256,
100
+ 256,
101
+ 256,
102
+ 256,
103
+ 256,
104
+ 256,
105
+ 256,
106
+ 256,
107
+ 256
108
+ ]
109
+ ],
110
+ "stage_block_stride": [
111
+ [
112
+ 2
113
+ ],
114
+ [
115
+ 1,
116
+ 1
117
+ ],
118
+ [
119
+ 2,
120
+ 1,
121
+ 1,
122
+ 1
123
+ ],
124
+ [
125
+ 2,
126
+ 1,
127
+ 1,
128
+ 1,
129
+ 1,
130
+ 1,
131
+ 1,
132
+ 1,
133
+ 1,
134
+ 1,
135
+ 1,
136
+ 1,
137
+ 1,
138
+ 1
139
+ ]
140
+ ],
141
+ "stage_in_channels": [
142
+ 1,
143
+ 64,
144
+ 64,
145
+ 128
146
+ ],
147
+ "stage_num_blocks": [
148
+ 1,
149
+ 2,
150
+ 4,
151
+ 14
152
+ ],
153
+ "stage_stride": [
154
+ 2,
155
+ 1,
156
+ 2,
157
+ 2
158
+ ],
159
+ "torch_dtype": "float32",
160
+ "transformers_version": "4.56.0.dev0"
161
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f9fab3e87396a9de84488cb08b1cc54c82ab9ba510c737cf0a0945c6354b859
3
+ size 64263488