Comments (6)
Hey there are some more details on mT0 fine-tuning here: #12
The config is here: #6 (comment)
from xmtf.
Thanks for reply! Will try mentioned config.
from xmtf.
Hey @Muennighoff , It's seems that I still can't get a couple of things. Would be very appreciate If you could give me a hand here.
Well, I need to finetune your model mT0-xxl (not the initial T5X-xxl), so accordingly to the manual https://github.com/google-research/t5x/blob/main/docs/usage/finetune.md I need 3 components (excluded SeqIO Task, which is clear as for now) to proceed:
- Checkpoint -- Could you please share with mT0-xxl checkpoint? In the manual all used checkpoints are the TensorFlow weights etc, but on the HuggingFace there are only PyTorch weights. So I do need either mT0-xxl checkpoint in TensorFlow or finetune the model in PyTorch (is it even possible?)
- Gin file for the model to finetune (mT0-xxl in the case) -- Could I use the default one like https://github.com/google-research/t5x/blob/main/t5x/examples/t5/mt5/xxl.gin?
- Gin file configuring finetuning process -- I write it by my own based on https://github.com/google-research/t5x/blob/main/t5x/configs/runs/finetune.gin with some overrides, right?
Please, correct me if I wrong in some points.
from xmtf.
There's a t5x ckpt here: https://huggingface.co/bigscience/mt0-t5x
I don't remember which size that model is though; I don't have the other ones anymore, maybe @adarob does
For 2. & 3., yes I think so
from xmtf.
from xmtf.
Thanks a lot, guys!
from xmtf.
Related Issues (20)
- Were the checkpoints selected based on the held-out performance or seen task performance? HOT 2
- How to convert megatron-deepspeed checkpoints to huggingface checkpoints ? HOT 4
- How to fineutne mT0 with specific down-stream data? HOT 3
- Questions on creating instruction data HOT 1
- Use Petals without sharing GPU HOT 11
- Controlled generation HOT 1
- how to convert model weights(e.g., bigscience/bloomz-560m-optimizer-states) to Hugging Face model.bin file? HOT 2
- how to repreduce bloomz-* HOT 6
- bloomz-mt universal checkpoint HOT 2
- Export mt0-xxl-mt to ONNX fails HOT 2
- Dose mt0&bloomz trained on dev, devtest datasets of Flores-200? HOT 2
- Parsing the xP3 dataset HOT 1
- Getting machine-translated prompts of xP3mt HOT 4
- P3megds URL is not available HOT 1
- Some datasets are not in xP3all HOT 4
- What is the training config? HOT 3
- I can't find the model weights that you used for experimentation. HOT 1
- Quesiton about MTFDataset HOT 1
- Why does the number of templates differ between languages? HOT 4
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from xmtf.