Greek BERT Sentence Base Uncased Embedding

Description

A Greek version of BERT pre-trained language model.

The pre-training corpora of bert-base-greek-uncased-v1 include:

  • The Greek part of Wikipedia,
  • The Greek part of European Parliament Proceedings Parallel Corpus, and
  • The Greek part of OSCAR, a cleansed version of Common Crawl.

Predicted Entities

Download Copy S3 URI

How to use

sent_embeddings = BertSentenceEmbeddings.pretrained("sent_bert_base_uncased", "el") \
      .setInputCols("sentence") \
      .setOutputCol("bert_sentence")

nlp_pipeline = Pipeline(stages=[document_assembler, sentence_detector, sent_embeddings ])
val sent_embeddings = BertSentenceEmbeddings.pretrained("sent_bert_base_uncased", "el")
      .setInputCols("sentence")
      .setOutputCol("bert_sentence")

val pipeline = new Pipeline().setStages(Array(document_assembler, sentence_detector, sent_embeddings ))
import nlu
nlu.load("el.embed_sentence.bert.base_uncased").predict("""Put your text here.""")

Model Information

Model Name: sent_bert_base_uncased
Compatibility: Spark NLP 3.2.2+
License: Open Source
Edition: Official
Input Labels: [sentence]
Output Labels: [bert_sentence]
Language: el
Case sensitive: true

Data Source

The model is imported from: https://huggingface.co/nlpaueb/bert-base-greek-uncased-v1