python run_mrqa.py --model_type=roberta-base --model_name_or_path=roberta-base --qass_head=False --tokenizer_name=roberta-base --output_dir=outputs --train_file=squad/mosaic_unite_npairs-4/squad-train-seed-42-num-examples-16.jsonl --predict_file="squad/dev.jsonl" --do_train --do_eval --cache_dir=.cache --max_seq_length=384 --doc_stride=128 --threads=4 --save_steps=50000 --per_gpu_train_batch_size=12 --per_gpu_eval_batch_size=16 --learning_rate=3e-5 --max_answer_length=10 --warmup_ratio=0.1 --min_steps=1 --num_train_epochs=1 --seed=42 --use_cache=False --evaluate_every_epoch=False --overwrite_output_dir
python run_mrqa.py --model_type=roberta-base --model_name_or_path=roberta-base --qass_head=False --tokenizer_name=roberta-base --output_dir=outputs --train_file=squad/mosaic_unite_npairs-4/squad-train-seed-42-num-examples-16.jsonl --predict_file="squad/dev.jsonl" --do_train --do_eval --cache_dir=.cache --max_seq_length=384 --doc_stride=128 --threads=4 --save_steps=50000 --per_gpu_train_batch_size=12 --per_gpu_eval_batch_size=16 --learning_rate=3e-5 --max_answer_length=10 --warmup_ratio=0.1 --min_steps=1 --num_train_epochs=1 --seed=42 --use_cache=False --evaluate_every_epoch=False --overwrite_output_dir --aug context-shuffle
python run_mrqa.py --model_type=roberta-base --model_name_or_path=roberta-base --qass_head=False --tokenizer_name=roberta-base --output_dir=outputs --train_file=squad/mosaic_unite_npairs-4/squad-train-seed-42-num-examples-16.jsonl --predict_file="squad/dev.jsonl" --do_train --do_eval --cache_dir=.cache --max_seq_length=384 --doc_stride=128 --threads=4 --save_steps=50000 --per_gpu_train_batch_size=12 --per_gpu_eval_batch_size=16 --learning_rate=3e-5 --max_answer_length=10 --warmup_ratio=0.1 --min_steps=1 --num_train_epochs=1 --seed=42 --use_cache=False --evaluate_every_epoch=False --overwrite_output_dir --aug mosaic-2-True python run_mrqa.py --model_type=roberta-base --model_name_or_path=roberta-base --qass_head=False --tokenizer_name=roberta-base --output_dir=outputs --train_file=squad/mosaic_unite_npairs-4/squad-train-seed-42-num-examples-16.jsonl --predict_file="squad/dev.jsonl" --do_train --do_eval --cache_dir=.cache --max_seq_length=384 --doc_stride=128 --threads=4 --save_steps=50000 --per_gpu_train_batch_size=12 --per_gpu_eval_batch_size=16 --learning_rate=3e-5 --max_answer_length=10 --warmup_ratio=0.1 --min_steps=1 --num_train_epochs=1 --seed=42 --use_cache=False --evaluate_every_epoch=False --overwrite_output_dir --aug context-shuffle
python run_mrqa.py --model_type=roberta-base --model_name_or_path=roberta-base --qass_head=False --tokenizer_name=roberta-base --output_dir=outputs --train_file=squad/mosaic_unite_npairs-4/squad-train-seed-42-num-examples-16.jsonl --predict_file="squad/dev.jsonl" --do_train --do_eval --cache_dir=.cache --max_seq_length=384 --doc_stride=128 --threads=4 --save_steps=50000 --per_gpu_train_batch_size=12 --per_gpu_eval_batch_size=16 --learning_rate=3e-5 --max_answer_length=10 --warmup_ratio=0.1 --min_steps=1 --num_train_epochs=1 --seed=42 --use_cache=False --evaluate_every_epoch=False --overwrite_output_dir
sbatch --gres=gpu:rtx2080:1 sh_runs/run_baseline.sh
python run_mrqa.py --model_type=roberta-base --model_name_or_path=roberta-base --qass_head=False --tokenizer_name=roberta-base --output_dir=results/single_run_test --train_file=squad/mosaic_unite_npairs-4/squad-train-seed-42-num-examples-16.jsonl --predict_file="squad/dev.jsonl" --do_train --do_eval --cache_dir=.cache --max_seq_length=384 --doc_stride=128 --threads=4 --save_steps=50000 --per_gpu_train_batch_size=12 --per_gpu_eval_batch_size=16 --learning_rate=3e-5 --max_answer_length=10 --warmup_ratio=0.1 --min_steps=1 --num_train_epochs=1 --seed=42 --use_cache=False --evaluate_every_epoch=False --overwrite_output_dir --aug concat-coherent-text
-
srun --pty --gres=gpu:rtx2080 bash
-
sbatch --time=2:0:0 --gres=gpu:rtx2080:1 sh_runs/run_baseline.sh
-
Manually
cd /cs/labs/gabis/ednussi/
source venv_splinter/bin/activate
cd splinter/finetuning
export MODEL="roberta-base"
export OUTPUT_DIR="output"
python run_mrqa.py --model_type=roberta-base --model_name_or_path=$MODEL --qass_head=False --tokenizer_name=$MODEL --output_dir=$OUTPUT_DIR --train_file="../squad/squad-train-seed-42-num-examples-16.jsonl" --predict_file="../squad/dev.jsonl" --do_train --do_eval --cache_dir=.cache --max_seq_length=384 --doc_stride=128 --threads=4 --save_steps=50000 --per_gpu_train_batch_size=12 --per_gpu_eval_batch_size=16 --learning_rate=3e-5 --max_answer_length=10 --warmup_ratio=0.1 --min_steps=200 --num_train_epochs=10 --seed=42 --use_cache=False --evaluate_every_epoch=False --overwrite_output_dir
This repository was forked on 25th April 2021, from the original splinter repo matching the "Few-Shot Question Answering by Pretraining Span Selection" Paper.
Maybe needed for spacy
export BLIS_REALLY_COMPILE=1
conda create -n splinter-env python=3.8
conda activate splinter-env
git clone [email protected]:ednussi/splinter.git
cd splinter/finetuning
pip install -r requirements.txt
mkdir mrqa_data
cd mrqa_data
curl -L https://www.dropbox.com/sh/pfg8j6yfpjltwdx/AAC8Oky0w8ZS-S3S5zSSAuQma?dl=1 > mrqa-few-shot.zip
unzip mrqa-few-shot.zip
cd ..
Note on single Titan X took ~45 minutes to train + ~1 hour to eval.
mkdir outputs
python run_mrqa.py \
--model_type=roberta-base \
--model_name_or_path=roberta-base \
--qass_head=False \
--tokenizer_name=roberta-base \
--output_dir="outputs/output32-42-test" \
--train_file="squad/squad-train-seed-42-num-examples-32.jsonl" \
--predict_file="squad/dev.jsonl" \
--do_train \
--do_eval \
--cache_dir=.cache \
--max_seq_length=384 \
--doc_stride=128 \
--threads=4 \
--save_steps=50000 \
--per_gpu_train_batch_size=12 \
--per_gpu_eval_batch_size=12 \
--learning_rate=3e-5 \
--max_answer_length=10 \
--warmup_ratio=0.1 \
--min_steps=200 \
--num_train_epochs=10 \
--seed=42 \
--use_cache=False \
--evaluate_every_epoch=False
export MODEL="roberta-base"
for i in 64 128 256
do
for j in 42 43 44 45 46
do
echo "Loop $i-$j"
python run_mrqa.py --model_type=$MODEL --model_name_or_path=$MODEL --qass_head=False --tokenizer_name=$MODEL --output_dir="outputs/output$i-$j" --train_file="squad/squad-train-seed-$j-num-examples-$i.jsonl" --predict_file="squad/dev.jsonl" --do_train --do_eval --cache_dir=.cache --max_seq_length=384 --doc_stride=128 --threads=4 --save_steps=50000 --per_gpu_train_batch_size=12 --per_gpu_eval_batch_size=12 --learning_rate=3e-5 --max_answer_length=10 --warmup_ratio=0.1 --min_steps=200 --num_train_epochs=10 --seed=$j --use_cache=False --evaluate_every_epoch=False
python run_ner.py --model_name_or_path bert-base-uncased --dataset_name wnut_17 --do_train --do_eval --warmup_ratio=0.1 --overwrite_output_dir --num_train_epochs=10 --seed 42 --max_train_samples 16 --output_dir test_res --aug mosaic-crop