Add files using large-upload tool
Browse files- model_weights/common.pt +3 -0
- model_weights/metadata.json +1 -0
- model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_14_96.pt +3 -0
- model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_17_96.pt +3 -0
- model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_19_96.pt +3 -0
- model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_31_96.pt +3 -0
- model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_36_96.pt +3 -0
- model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_39_96.pt +3 -0
- model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_43_96.pt +3 -0
- model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_48_96.pt +3 -0
- model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_49_96.pt +3 -0
- model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_4_96.pt +3 -0
- model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_50_96.pt +3 -0
- model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_55_96.pt +3 -0
- model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_59_96.pt +3 -0
- model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_60_96.pt +3 -0
- model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_68_96.pt +3 -0
- model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_75_96.pt +3 -0
- model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_78_96.pt +3 -0
- model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_81_96.pt +3 -0
- model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_82_96.pt +3 -0
- model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_91_96.pt +3 -0
- model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_93_96.pt +3 -0
- model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_94_96.pt +3 -0
- model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_9_96.pt +3 -0
model_weights/common.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e4e4090fa34d96307127606cccef3ae99aedae58279e8bdf1746d44d3bf7aa47
|
3 |
+
size 860
|
model_weights/metadata.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"sharded_backend": "zarr", "sharded_backend_version": 1, "common_backend": "torch", "common_backend_version": 1}
|
model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_14_96.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9411f9dce48a1d3a9696ced7800e3e952fd83beecc0f2b6789ead864be85342e
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_17_96.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:abcbc429045a564b70b5b8e6a3aebd65d5c9033d0851500eb76c099f47d78a95
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_19_96.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f2f56efedb6748e4ff2a0e6a118ad77842e265a67f44e3f1448ce39549f7cc07
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_31_96.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7a0d4b9cdf13a1eba2ed93d85a2982360212cddc77d53c576022e3cbe189615f
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_36_96.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bf56314f019f97c0fd50b64f933a6af12207a4f46d2ff4cb03172248709e85bc
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_39_96.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3b109bc001c35c681140f7729d952690a2e524e10f7f31b66473cd9368535724
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_43_96.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3f67a153354d75feb672b47442b5b6b4e05a393ff95b9cf077bae7263f185230
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_48_96.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fb4e3686edebeef34059ccb1a24181f803157ce2fe1ad32327842e52f332f016
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_49_96.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cabdcbeee1b045eb2bc0e6250a49fd397ab59f81b3495065dd1349cc50050a52
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_4_96.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:461f0daf5250e8e0f6a0d98a3936729d35c8fe255985e7824b06c8b5680cce96
|
3 |
+
size 1836
|
model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_50_96.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:589423c313c3b3a1596305bef910e00c563ef1fcc0305777be0d8226cd391fdf
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_55_96.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:933b45a0255ffbf603631df86198c897aed451e359ea27c8dbab3a43ce6eb6eb
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_59_96.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:318589273676b17854619902806c202d0e2313dd0f9b335247e78421e3f8fee0
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_60_96.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ae1b8b2fa8a86bf8650543a16b5ce2fedfe5225dbb442fcc74bbdc8d23365d42
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_68_96.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:07033b70d349490e62e87c37b2c422207efe1e2411b575672ba110bc3c338d55
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_75_96.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:af07f6ddc756fcb79a4871a38bab569ed54f5c8e92d27ab935c1719ec6d9ad7b
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_78_96.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d3266a9256519aec23042c13c73f48061e100e831d06dddb600c039f15926b8c
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_81_96.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:37af42582c2b74450e16a71bd80e24f762d3ffa8ac4599774d2a4a7a61820e6b
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_82_96.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8d981dae8c5c3b6dad6c9f2fe25f26e4c445772ee9d9850c359b836ac8f6ae64
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_91_96.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:01048ad3a21833fd1c3279d20e19d6190d5d695c68f738e3b6401bb8b6601ed3
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_93_96.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8c9154e83763b85c469f90fbd98d7d81388deca2aa97cdf867e343d61f026055
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_94_96.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9a15a899861ca2803444e551521376d87cbe33c95038243d97b7e65c2e7ce21c
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.self_attention.linear_proj._extra_state/shard_9_96.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:aee04ee3a9964d7e6d305e5c1ad233bf6ef32b34178cda7bf7c2083c011e4e56
|
3 |
+
size 1836
|