BERT, mBERT, BiBERT? A Study on Contextualized Embeddings for Supervised Neural Machine Translation

Everything Is All It Takes: A Multipronged Strategy for Zero-Shot Cross-Lingual Information Extraction

Gradual Fine-Tuning for Low-Resource Domain Adaptation

Zero-Shot Cross-Lingual Dependency Parsing through Contextual Embedding Transformation

Cross-Lingual BERT Contextual Embedding Space Mapping with Isotropic and Isometric Conditions