Giter Club home page Giter Club logo

autodl's People

Contributors

didayolo avatar madclam avatar meta-meta-learning avatar nehzux avatar zhengying-liu avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

autodl's Issues

Bad scoring for parent phase

The score displayed for "All datasets" phase is the score of dataset 1.

screen1
screen2

The issue can come from Codalab or from the Parent Scoring Program.

Data browser fails on audio data (inside docker)

When trying to call the data browser on an audio dataset, inside evariste/autodl:cpu-latest docker, I get the following error:

Playing audio...
Traceback (most recent call last):
  File "data_browser.py", line 357, in <module>
    main()
  File "data_browser.py", line 353, in main
    data_browser.show_examples(num_examples=num_examples, subset=subset)
  File "data_browser.py", line 306, in show_examples
    self.show_an_example(subset=subset)
  File "data_browser.py", line 296, in show_an_example
    self.show(tensor_4d, label_confidence_pairs=label_conf_pairs)
  File "data_browser.py", line 196, in show_speech
    DataBrowser.play_sound(data)
  File "data_browser.py", line 238, in play_sound
    playsound(tmp_filepath)
  File "/usr/local/lib/python3.5/dist-packages/playsound.py", line 92, in _playsoundNix
    gi.require_version('Gst', '1.0')
  File "/usr/lib/python3/dist-packages/gi/__init__.py", line 102, in require_version
    raise ValueError('Namespace %s not available' % namespace)
ValueError: Namespace Gst not available

The WAV file is written but the playsound function fails.

Big variance of initial delay

Describe the bug
This issue concerns the competitions:
https://autodl.lri.fr/competitions/3
https://autodl.lri.fr/competitions/32
When one makes the same submission, there is a chance to get very long delay. Even though the algorithm, VM configuration, ingestion + scoring, datasets etc are all the same.

The Output Log of ingestion for the submission 6272:

2019-08-23 12:22:34,493 INFO ingestion.py: ************************************************
2019-08-23 12:22:34,493 INFO ingestion.py: ******** Processing dataset Apollon ********
2019-08-23 12:22:34,493 INFO ingestion.py: ************************************************
2019-08-23 12:22:34,493 INFO ingestion.py: Reading training set and test set...
2019-08-23 12:22:34,735 INFO ingestion.py: ===== Start core part of ingestion program. Version: v20190820 =====
2019-08-23 12:22:34,735 INFO ingestion.py: Creating model...
2019-08-23 12:23:22,627 INFO ingestion.py: Begin training the model...
2019-08-23 12:23:41,877 INFO ingestion.py: Finished training the model.
2019-08-23 12:23:41,877 INFO ingestion.py: Begin testing the model by making predictions on test set...
2019-08-23 12:23:46,053 INFO ingestion.py: Finished making predictions.
2019-08-23 12:23:46,204 INFO ingestion.py: [+] 1 predictions made, time spent so far 71.47 sec
2019-08-23 12:23:46,204 INFO ingestion.py: [+] Time left 1128.53 sec

We see that making the first prediction took 72 seconds.

The Output Log of ingestion for the submission 6386:

2019-08-23 14:56:01,055 INFO ingestion.py: ************************************************
2019-08-23 14:56:01,055 INFO ingestion.py: ******** Processing dataset Apollon ********
2019-08-23 14:56:01,055 INFO ingestion.py: ************************************************
2019-08-23 14:56:01,055 INFO ingestion.py: Reading training set and test set...
2019-08-23 14:56:01,299 INFO ingestion.py: ===== Start core part of ingestion program. Version: v20190820 =====
2019-08-23 14:56:01,300 INFO ingestion.py: Creating model...
2019-08-23 14:56:07,437 INFO ingestion.py: Begin training the model...
2019-08-23 14:56:25,904 INFO ingestion.py: Finished training the model.
2019-08-23 14:56:25,904 INFO ingestion.py: Begin testing the model by making predictions on test set...
2019-08-23 14:56:30,127 INFO ingestion.py: Finished making predictions.
2019-08-23 14:56:30,273 INFO ingestion.py: [+] 1 predictions made, time spent so far 28.97 sec
2019-08-23 14:56:30,273 INFO ingestion.py: [+] Time left 1171.03 sec

We see that this time making the first prediction only took 29 seconds.

The Output Log of ingestion for the submission 6506:

2019-08-23 17:41:37,020 INFO ingestion.py: ************************************************
2019-08-23 17:41:37,021 INFO ingestion.py: ******** Processing dataset Apollon ********
2019-08-23 17:41:37,021 INFO ingestion.py: ************************************************
2019-08-23 17:41:37,021 INFO ingestion.py: Reading training set and test set...
2019-08-23 17:41:37,271 INFO ingestion.py: ===== Start core part of ingestion program. Version: v20190820 =====
2019-08-23 17:41:37,271 INFO ingestion.py: Creating model...
2019-08-23 17:41:43,085 INFO ingestion.py: Begin training the model...
2019-08-23 17:42:00,111 INFO ingestion.py: Finished training the model.
2019-08-23 17:42:00,111 INFO ingestion.py: Begin testing the model by making predictions on test set...
2019-08-23 17:42:04,084 INFO ingestion.py: Finished making predictions.
2019-08-23 17:42:04,240 INFO ingestion.py: [+] 1 predictions made, time spent so far 26.97 sec
2019-08-23 17:42:04,240 INFO ingestion.py: [+] Time left 1173.03 sec

We see that this time making the first prediction only took 27 seconds.

So the normal level should be around 28 seconds. In addition, if we look at the same submission for other datasets, this delay is common.

This variance affects the final ALC score a lot since for the above 3 submissions, the ALC scores are respectively:
0.5583, 0.6963, 0.6646

Screenshots
Comparison of different learning curves:
6271-kakaobrain-apollon
6385-kakaobrain-apollon
6505-kakaobrain-apollon

To Reproduce
Steps to reproduce the behavior:

  1. Go to https://autodl.lri.fr/competitions/32
  2. Make exact the same submissions twice
  3. Look at the difference of scores for the two submissions
    As the delay varies in a unpredictable way, it's hard to reproduce exactly the same situation.

Expected behavior
The variance should be a lot less. The normal level standard deviation of ALC score should be less than 0.03.

Desktop (please complete the following information):

  • Docker: Ubuntu

Smartphone (please complete the following information):
[Not relevant]

Additional context
None

An example issue

System information

  • OS Platform and Distribution (e.g., Linux Ubuntu 16.04): MacOS Mojave, 10.14.3
  • Did you use the Docker image evariste/autodl?: Yes
  • Python version: 3.6.8
  • CUDA/cuDNN version (optional): None
  • GPU model and memory (optional): None

Describe the current behavior
Now everything is doing great except that so few issues are created!

Describe the expected behavior
More issues created in Issues page!

Additional info
Everybody reports bugs and other things on Issues!

Auto-migration

In single phase competitions, auto-migration should be set to False.

Here are the correct settings:

PARENT PHASE
Results Scoring Only => YES
Auto migration => NO
If submission beats old score, put submission on leaderboard => NO
Ingestion program only during scoring => NO
Is parallel parent => YES

CHILDREN PHASES
Results Scoring Only => YES
Auto migration => NO
If submission beats old score, put submission on leaderboard => NO
Ingestion program only during scoring => YES
Is parallel parent => NO

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.