DashAI.back.models.DistilBertTransformer
- class DistilBertTransformer(model=None, **kwargs)[source]
Pre-trained transformer DistilBERT allowing English text classification.
DistilBERT is a small, fast, cheap and light Transformer model trained by distilling BERT base. It has 40% less parameters than bert-base-uncased, runs 60% faster while preserving over 95% of BERT’s performances as measured on the GLUE language understanding benchmark [1].
References
[1] https://huggingface.co/docs/transformers/model_doc/distilbert
- __init__(model=None, **kwargs)[source]
Initialize the transformer model.
The process includes the instantiation of the pre-trained model and the associated tokenizer.
Methods
__init__([model])Initialize the transformer model.
fit(x_train, y_train)Fine-tune the pre-trained model.
get_schema()Generates the component related Json Schema.
load(filename)Restores an instance of a model.
predict(x_pred)Predict with the fine-tuned model.
save(filename)Store an instance of a model.
tokenize_data(dataset)Tokenize the input data.
validate_and_transform(raw_data)It takes the data given by the user to initialize the model and returns it with all the objects that the model needs to work.
Attributes
COLORCOMPATIBLE_COMPONENTSDISPLAY_NAMETYPE