{ "_name_or_path": "bert_book_title_classifier/checkpoint-7785", "architectures": [ "BertForSequenceClassification" ], "attention_probs_dropout_prob": 0.1, "classifier_dropout": null, "directionality": "bidi", "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 1024, "id2label": { "0": "Calendars", "1": "Comics & Graphic Novels", "2": "Test Preparation", "3": "Mystery, Thriller & Suspense", "4": "Science Fiction & Fantasy", "5": "Romance", "6": "Humor & Entertainment", "7": "Literature & Fiction", "8": "Gay & Lesbian", "9": "Engineering & Transportation", "10": "Cookbooks, Food & Wine", "11": "Crafts, Hobbies & Home", "12": "Arts & Photography", "13": "Education & Teaching", "14": "Parenting & Relationships", "15": "Self-Help", "16": "Computers & Technology", "17": "Medical Books", "18": "Science & Math", "19": "Health, Fitness & Dieting", "20": "Business & Money", "21": "Law", "22": "Biographies & Memoirs", "23": "History", "24": "Politics & Social Sciences", "25": "Reference", "26": "Christian Books & Bibles", "27": "Religion & Spirituality", "28": "Sports & Outdoors", "29": "Teen & Young Adult", "30": "Children's Books", "31": "Travel" }, "initializer_range": 0.02, "intermediate_size": 4096, "label2id": { "Arts & Photography": 12, "Biographies & Memoirs": 22, "Business & Money": 20, "Calendars": 0, "Children's Books": 30, "Christian Books & Bibles": 26, "Comics & Graphic Novels": 1, "Computers & Technology": 16, "Cookbooks, Food & Wine": 10, "Crafts, Hobbies & Home": 11, "Education & Teaching": 13, "Engineering & Transportation": 9, "Gay & Lesbian": 8, "Health, Fitness & Dieting": 19, "History": 23, "Humor & Entertainment": 6, "Law": 21, "Literature & Fiction": 7, "Medical Books": 17, "Mystery, Thriller & Suspense": 3, "Parenting & Relationships": 14, "Politics & Social Sciences": 24, "Reference": 25, "Religion & Spirituality": 27, "Romance": 5, "Science & Math": 18, "Science Fiction & Fantasy": 4, "Self-Help": 15, "Sports & Outdoors": 28, "Teen & Young Adult": 29, "Test Preparation": 2, "Travel": 31 }, "layer_norm_eps": 1e-12, "max_position_embeddings": 512, "model_type": "bert", "num_attention_heads": 16, "num_hidden_layers": 24, "pad_token_id": 0, "pooler_fc_size": 768, "pooler_num_attention_heads": 12, "pooler_num_fc_layers": 3, "pooler_size_per_head": 128, "pooler_type": "first_token_transform", "position_embedding_type": "absolute", "problem_type": "single_label_classification", "torch_dtype": "float32", "transformers_version": "4.40.2", "type_vocab_size": 2, "use_cache": true, "vocab_size": 28996 }