| | """ |
| | Implementation borrowed from transformers package and extended to support multiple prediction heads: |
| | |
| | https://github.com/huggingface/transformers/blob/master/src/transformers/models/bert/modeling_bert.py |
| | """ |
| |
|
| | import torch |
| | from torch import nn |
| | from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss |
| |
|
| | import transformers |
| | from transformers import BertTokenizer |
| | from transformers import models |
| | from transformers.modeling_outputs import SequenceClassifierOutput |
| |
|
| | from transformers.models.bert.configuration_bert import BertConfig |
| | from transformers.models.bert.modeling_bert import ( |
| | BertPreTrainedModel, |
| | BERT_INPUTS_DOCSTRING, |
| | _TOKENIZER_FOR_DOC, |
| | _CHECKPOINT_FOR_DOC, |
| | _CONFIG_FOR_DOC, |
| | BertModel, |
| | ) |
| |
|
| | from transformers.file_utils import ( |
| | add_code_sample_docstrings, |
| | add_start_docstrings_to_model_forward, |
| | ) |
| |
|
| |
|
| | class BertForSequenceClassification(BertPreTrainedModel): |
| | def __init__(self, config, **kwargs): |
| | super().__init__(transformers.PretrainedConfig()) |
| | self.num_labels = kwargs.get("task_labels_map", {}) |
| | self.config = config |
| |
|
| | self.bert = BertModel(config) |
| | classifier_dropout = ( |
| | config.classifier_dropout |
| | if config.classifier_dropout is not None |
| | else config.hidden_dropout_prob |
| | ) |
| | self.dropout = nn.Dropout(classifier_dropout) |
| | |
| | self.classifier1 = nn.Linear( |
| | config.hidden_size, list(self.num_labels.values())[0] |
| | ) |
| | self.classifier2 = nn.Linear( |
| | config.hidden_size, list(self.num_labels.values())[1] |
| | ) |
| |
|
| | self.init_weights() |
| |
|
| | @add_start_docstrings_to_model_forward( |
| | BERT_INPUTS_DOCSTRING.format("batch_size, sequence_length") |
| | ) |
| | @add_code_sample_docstrings( |
| | tokenizer_class=_TOKENIZER_FOR_DOC, |
| | checkpoint=_CHECKPOINT_FOR_DOC, |
| | output_type=SequenceClassifierOutput, |
| | config_class=_CONFIG_FOR_DOC, |
| | ) |
| | def forward( |
| | self, |
| | input_ids=None, |
| | attention_mask=None, |
| | token_type_ids=None, |
| | position_ids=None, |
| | head_mask=None, |
| | inputs_embeds=None, |
| | labels=None, |
| | output_attentions=None, |
| | output_hidden_states=None, |
| | return_dict=None, |
| | task_name=None, |
| | ): |
| | r""" |
| | labels (:obj:`torch.LongTensor` of shape :obj:`(batch_size,)`, `optional`): |
| | Labels for computing the sequence classification/regression loss. Indices should be in :obj:`[0, ..., |
| | config.num_labels - 1]`. If :obj:`config.num_labels == 1` a regression loss is computed (Mean-Square loss), |
| | If :obj:`config.num_labels > 1` a classification loss is computed (Cross-Entropy). |
| | """ |
| | return_dict = ( |
| | return_dict if return_dict is not None else self.config.use_return_dict |
| | ) |
| |
|
| | outputs = self.bert( |
| | input_ids, |
| | attention_mask=attention_mask, |
| | token_type_ids=token_type_ids, |
| | position_ids=position_ids, |
| | head_mask=head_mask, |
| | inputs_embeds=inputs_embeds, |
| | output_attentions=output_attentions, |
| | output_hidden_states=output_hidden_states, |
| | return_dict=return_dict, |
| | ) |
| |
|
| | pooled_output = outputs[1] |
| |
|
| | pooled_output = self.dropout(pooled_output) |
| | logits = None |
| | if task_name == list(self.num_labels.keys())[0]: |
| | logits = self.classifier1(pooled_output) |
| | elif task_name == list(self.num_labels.keys())[1]: |
| | logits = self.classifier2(pooled_output) |
| |
|
| | loss = None |
| | if labels is not None: |
| | if self.config.problem_type is None: |
| | if self.num_labels[task_name] == 1: |
| | self.config.problem_type = "regression" |
| | elif self.num_labels[task_name] > 1 and ( |
| | labels.dtype == torch.long or labels.dtype == torch.int |
| | ): |
| | self.config.problem_type = "single_label_classification" |
| | else: |
| | self.config.problem_type = "multi_label_classification" |
| |
|
| | if self.config.problem_type == "regression": |
| | loss_fct = MSELoss() |
| | if self.num_labels[task_name] == 1: |
| | loss = loss_fct(logits.squeeze(), labels.squeeze()) |
| | else: |
| | loss = loss_fct(logits, labels) |
| | elif self.config.problem_type == "single_label_classification": |
| | loss_fct = CrossEntropyLoss() |
| | loss = loss_fct( |
| | logits.view(-1, self.num_labels[task_name]), labels.view(-1) |
| | ) |
| | elif self.config.problem_type == "multi_label_classification": |
| | loss_fct = BCEWithLogitsLoss() |
| | loss = loss_fct(logits, labels) |
| | if not return_dict: |
| | output = (logits,) + outputs[2:] |
| | return ((loss,) + output) if loss is not None else output |
| |
|
| | return SequenceClassifierOutput( |
| | loss=loss, |
| | logits=logits, |
| | hidden_states=outputs.hidden_states, |
| | attentions=outputs.attentions, |
| | ) |
| |
|