sileod commited on
Commit
db3c618
·
verified ·
1 Parent(s): c54851c

Add new SentenceTransformer model

Browse files
README.md CHANGED
The diff for this file is too large to render. See raw diff
 
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4a1f55f4497e38a3a6452cb64d3365cabcd8ee5f8b0ebb65522ff965008c9861
3
  size 67193928
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ce4897c03f24856a772e3c2bbddf40b4c3eef45de675bfff55146c215458795
3
  size 67193928
sentence_bert_config.json CHANGED
@@ -1,4 +1,4 @@
1
  {
2
- "max_seq_length": 7999,
3
  "do_lower_case": false
4
  }
 
1
  {
2
+ "max_seq_length": 128,
3
  "do_lower_case": false
4
  }
tokenizer.json CHANGED
@@ -2,7 +2,7 @@
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
- "max_length": 7999,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
 
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
+ "max_length": 128,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
tokenizer_config.json CHANGED
@@ -937,7 +937,7 @@
937
  "input_ids",
938
  "attention_mask"
939
  ],
940
- "model_max_length": 8192,
941
  "pad_token": "[PAD]",
942
  "sep_token": "[SEP]",
943
  "tokenizer_class": "PreTrainedTokenizerFast",
 
937
  "input_ids",
938
  "attention_mask"
939
  ],
940
+ "model_max_length": 128,
941
  "pad_token": "[PAD]",
942
  "sep_token": "[SEP]",
943
  "tokenizer_class": "PreTrainedTokenizerFast",