Package smile.llm
Class BERT
java.lang.Object
smile.llm.BERT
Bidirectional Encoder Representations from Transformers (BERT).
BERT is a deeply bidirectional, unsupervised language representation,
pre-trained using only a plain text corpus. Context-free models such as
word2vec or GloVe generate a single word embedding representation for
each word in the vocabulary, where BERT takes into account the context
for each occurrence of a given word.
BERT makes use of Transformer, an attention mechanism that learns contextual relations between words (or sub-words) in a text. In its vanilla form, Transformer includes two separate mechanisms: an encoder that reads the text input and a decoder that produces a prediction for the task. Since BERT’s goal is to generate a language model, only the encoder mechanism is necessary.
As opposed to directional models, which read the text input sequentially (left-to-right or right-to-left), the Transformer encoder reads the entire sequence of words at once. Therefore it is considered bidirectional. This characteristic allows the model to learn the context of a word based on all of its surroundings (left and right of the word).
-
Constructor Summary
-
Method Summary
-
Constructor Details
-
BERT
public BERT()
-