Setup the BERT Pretrained Model

BERT pre-trained model is a good starting point for us to do a lot of NLP task, and we can use its performance as Benchmark. In this session, we simply load the model

Load Package

the package is in the transformers arrow-up-rightlibrary

from transformers import BertForSequenceClassification

In here, we use the one for ”Sequence Classification” because the twitter data is a sequential data: a paragraph of text

Loading the model

Model = BertForSequenceClassification.from_pretrained(
    'bert-base-uncased', # the specific model we want to use, it is matching with the tokenizer we choose
    num_labels=len(label_dict), # how many classes / labels we need
    output_attentions=False, # whether the model tells us its reasoning in making that prediction
    output_hidden_states=False
)

Last updated