This repo stores the work I participated in the URP, COMETH, at RPI.
Preprocess the corpus for fine-tuning
defineCorpus(publishers, passages, documents, output)
Use 10%, 10% and 80% propotions of the corpus for Test, Validation and Training, respectively to fine-tune the pretrained bert model
fineTuneEmbedding(publishers, output)
Combine the fine-tuned bert models with Flair Transformer
combineModelFlair(publishers, documents, ignores, output)
Use Flair embedding for its Transformer
saveEmbedding(embeddingVectors, output)