jacobfulano commited on
Commit
2e90941
1 Parent(s): 97f26fe

Change attention_probs_dropout_prob to 0.1 so that triton FlashAttention dependencies are avoided

Browse files

See [github issue](https://github.com/mosaicml/examples/issues/441#issuecomment-1874817465) for more detail

Files changed (1) hide show
  1. config.json +1 -1
config.json CHANGED
@@ -4,7 +4,7 @@
4
  "architectures": [
5
  "BertForMaskedLM"
6
  ],
7
- "attention_probs_dropout_prob": 0.0,
8
  "auto_map": {
9
  "AutoConfig": "configuration_bert.BertConfig",
10
  "AutoModelForMaskedLM": "bert_layers.BertForMaskedLM"
 
4
  "architectures": [
5
  "BertForMaskedLM"
6
  ],
7
+ "attention_probs_dropout_prob": 0.1,
8
  "auto_map": {
9
  "AutoConfig": "configuration_bert.BertConfig",
10
  "AutoModelForMaskedLM": "bert_layers.BertForMaskedLM"