Giter Club home page Giter Club logo

nlg_evaluation_metrics's Introduction

Text Similarity as An Evaluation Measure of Text Generation

Code style - black

โ“ Context

Natural Language Generation (NLG) is the process of generating human-like language by machines. One of the key challenges in evaluating the quality of generated text is to compare it with 'gold standard' references.

However, obtaining human annotations for evaluation is an expensive and time-consuming process, making it impractical for large-scale experiments. As a result, researchers have explored alternative methods for evaluating the quality of generated text.

Two families of metrics have been proposed: trained metrics and untrained metrics. While trained metrics may not generalize well to new data, untrained metrics, such as word or character-based metrics and embedding-based metrics, offer a more flexible and cost-effective solution. To assess the performance of an evaluation metric, correlation measures such as Pearson, Spearman, or Kendall tests are used, either at the text-level or system-level.

๐ŸŽฏ Objective

This project aims to benchmark the correlation of existing metrics with human scores on generation task: translation or data2text generation or story generation.

๐Ÿš€ How to use the project

  1. First, you need to clone the repository and cd into it :
git clone https://github.com/lidamsoukaina/NLG_Evaluation_Metrics.git
cd NLG_EVALUATION_METRICS
  1. Then, you need to create a virtual environment and activate it :
python3 -m venv venv
source venv/bin/activate
  1. You need to install all the requirements using the following command :
pip install -r requirements.txt
  1. [Optional] if you are using this repository in development mode, you can run the following command to set up the git hook scripts:
pre-commit install
  1. You can now run the python files in the file name folder using the following commands :
cd cluster
python3 [file name]

To test the project, you can run the test.ipynb notebook.

๐Ÿ“ Results

TODO: List the tested metrics

TODO: Describe the criteria used to evaluate the results

Metric criterion1 criterion2 criterion3
TER XX XX XX
DepthScore XX XX XX

TODO: Describe and analyse the results

๐Ÿค” What's next ?

TODO: List the next steps

๐Ÿ“š References

TODO: List the references

โœ๏ธ Authors

  • LETAIEF Maram
  • LIDAM Soukaina

nlg_evaluation_metrics's People

Contributors

maram1764 avatar lidamsoukaina avatar

Stargazers

 avatar

Watchers

 avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.