bert_config.json 520 B

12345678910111213141516171819
  1. {
  2. "attention_probs_dropout_prob": 0.1,
  3. "directionality": "bidi",
  4. "hidden_act": "gelu",
  5. "hidden_dropout_prob": 0.1,
  6. "hidden_size": 768,
  7. "initializer_range": 0.02,
  8. "intermediate_size": 3072,
  9. "max_position_embeddings": 512,
  10. "num_attention_heads": 12,
  11. "num_hidden_layers": 12,
  12. "pooler_fc_size": 768,
  13. "pooler_num_attention_heads": 12,
  14. "pooler_num_fc_layers": 3,
  15. "pooler_size_per_head": 128,
  16. "pooler_type": "first_token_transform",
  17. "type_vocab_size": 2,
  18. "vocab_size": 13685
  19. }