Giter Club home page Giter Club logo

transrelation's Introduction

Transrelation - CogALex VI Submission

This is a description of our system submitted to the CogaLex VI shared task. We use a single multilingual model in order to annotate the four different test sets. The labeled test sets can be found in the folder “annotated test sets”.

Our full paper can be found here: https://www.aclweb.org/anthology/2020.cogalex-1.7/ [1]

The task website is: https://sites.google.com/site/cogalexvisharedtask/

XLM-RoBERTa

Our system makes use of the multilingual language model XLM-RoBERTa [2]. We use the implementation provided by the transformers library, which offers the XLM-RoBERTa model pretained on 100 different languages using CommonCrawl data [3]. A linear layer is added on top of the pooled output in order to allow classification. The model was trained on the training data provided for the CogALex VI shared task, with the training being performed on all three languages at the same time. Hyperparameters were finetuned on the given validation set. The best results were achieved with the following Hyperparameters

  • Optimizer: AdamW
  • Learning rate: 2e-5
  • Epsilon: 1e-8
  • Epochs: 7
  • Batch size: 32
  • Warm up steps: 0
  • Weight Decay: 0
  • Model size: Base

Results on Validation Set

With XLMR Model

Chinese

Class Precision Recall F1 Support
SYN 0.862 0.822 0.841 129
HYP 0.938 0.841 0.887 145
ANT 0.877 0.838 0.857 136
Overall 0.894 0.834 0.863

English

Class Precision Recall F1 Support
SYN 0.492 0.463 0.477 259
HYP 0.597 0.473 0.528 292
ANT 0.801 0.497 0.613 308
Overall 0.639 0.478 0.543

German

Class Precision Recall F1 Support
SYN 0.507 0.419 0.459 272
HYP 0.595 0.459 0.518 294
ANT 0.673 0.411 0.510 275
Overall 0.592 0.430 0.496

References

[1] Wachowiak, L., Lang, C., Heinisch, B., & Gromann, D. (2020, December). CogALex-VI Shared Task: Transrelation-A Robust Multilingual Language Model for Multilingual Relation Identification. In Proceedings of the Workshop on the Cognitive Aspects of the Lexicon (pp. 59-64).

[2] Conneau, A., Khandelwal, K., Goyal, N., Chaudhary, V., Wenzek, G., Guzmán, F., ... & Stoyanov, V. (2019). Unsupervised cross-lingual representation learning at scale. arXiv preprint arXiv:1911.02116.

[3] Wolf, T., Debut, L., Sanh, V., Chaumond, J., Delangue, C., Moi, A., ... & Brew, J. (2019). HuggingFace's Transformers: State-of-the-art Natural Language Processing. ArXiv, arXiv-1910.

transrelation's People

Contributors

clang88 avatar lwachowiak avatar

Stargazers

 avatar  avatar

Watchers

 avatar  avatar  avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.