This repository contains the code and data for fine-tuning a Llama-7B model on a dataset containing user inputs and mid-journey prompts as responses.
User Input: midjourney prompt for a retro-futuristic spaceship with organic shapes and bright colors
![1](https://private-user-images.githubusercontent.com/71924097/261793308-87759ce9-8814-49db-8603-eae35329aa4c.jpg?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3MjE0MjM0MjMsIm5iZiI6MTcyMTQyMzEyMywicGF0aCI6Ii83MTkyNDA5Ny8yNjE3OTMzMDgtODc3NTljZTktODgxNC00OWRiLTg2MDMtZWFlMzUzMjlhYTRjLmpwZz9YLUFtei1BbGdvcml0aG09QVdTNC1ITUFDLVNIQTI1NiZYLUFtei1DcmVkZW50aWFsPUFLSUFWQ09EWUxTQTUzUFFLNFpBJTJGMjAyNDA3MTklMkZ1cy1lYXN0LTElMkZzMyUyRmF3czRfcmVxdWVzdCZYLUFtei1EYXRlPTIwMjQwNzE5VDIxMDUyM1omWC1BbXotRXhwaXJlcz0zMDAmWC1BbXotU2lnbmF0dXJlPTRmZDMyNDNkNTY0MjViNDcxYjU2ZDMxNTVlODZjNzkyNzE3NTc5YzhkM2EyMjVmMDk5ZGI4OGQ2MjQwMjJjNzkmWC1BbXotU2lnbmVkSGVhZGVycz1ob3N0JmFjdG9yX2lkPTAma2V5X2lkPTAmcmVwb19pZD0wIn0.qDnoeAz7Re_L92Vn2ufC7mTviDiZr_pZafRCtZUpnIc)
![1](https://private-user-images.githubusercontent.com/71924097/261151690-0ac7607f-b097-420e-a3a7-5fbd015bafc6.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3MjE0MjM0MjMsIm5iZiI6MTcyMTQyMzEyMywicGF0aCI6Ii83MTkyNDA5Ny8yNjExNTE2OTAtMGFjNzYwN2YtYjA5Ny00MjBlLWEzYTctNWZiZDAxNWJhZmM2LnBuZz9YLUFtei1BbGdvcml0aG09QVdTNC1ITUFDLVNIQTI1NiZYLUFtei1DcmVkZW50aWFsPUFLSUFWQ09EWUxTQTUzUFFLNFpBJTJGMjAyNDA3MTklMkZ1cy1lYXN0LTElMkZzMyUyRmF3czRfcmVxdWVzdCZYLUFtei1EYXRlPTIwMjQwNzE5VDIxMDUyM1omWC1BbXotRXhwaXJlcz0zMDAmWC1BbXotU2lnbmF0dXJlPTM5ZDgxNWJlZWMzOTJkZTNjM2U1ZjY5NTVhNWUwZjAzMzFmODdiYmJlMzM3MjE4NTlhNWVmOWNkMDI1MDgxNTMmWC1BbXotU2lnbmVkSGVhZGVycz1ob3N0JmFjdG9yX2lkPTAma2V5X2lkPTAmcmVwb19pZD0wIn0.XHenLw1Fij9Fh4LghgGmiK2p0urw_NYXOjuxuchauuI)
![3](https://private-user-images.githubusercontent.com/71924097/261152461-3b232be3-842b-4430-ab20-a510c53d582e.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3MjE0MjM0MjMsIm5iZiI6MTcyMTQyMzEyMywicGF0aCI6Ii83MTkyNDA5Ny8yNjExNTI0NjEtM2IyMzJiZTMtODQyYi00NDMwLWFiMjAtYTUxMGM1M2Q1ODJlLnBuZz9YLUFtei1BbGdvcml0aG09QVdTNC1ITUFDLVNIQTI1NiZYLUFtei1DcmVkZW50aWFsPUFLSUFWQ09EWUxTQTUzUFFLNFpBJTJGMjAyNDA3MTklMkZ1cy1lYXN0LTElMkZzMyUyRmF3czRfcmVxdWVzdCZYLUFtei1EYXRlPTIwMjQwNzE5VDIxMDUyM1omWC1BbXotRXhwaXJlcz0zMDAmWC1BbXotU2lnbmF0dXJlPTlhZmVmY2NjYzcyNWE3MGZmNzcxMjZkMWFjYTJlNjBhOGZlNGE1YzA1NzhmYjRkMDNjZDI0MzVhNTM0ZTFhY2YmWC1BbXotU2lnbmVkSGVhZGVycz1ob3N0JmFjdG9yX2lkPTAma2V5X2lkPTAmcmVwb19pZD0wIn0.dkQOKT1vL8Nic4w1jh7a8sW6P49J4NPdwoVnu6WmL2M)
![2](https://private-user-images.githubusercontent.com/71924097/261152349-fd201b68-9e7f-43bf-a4dc-edd4ffec91fd.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3MjE0MjM0MjMsIm5iZiI6MTcyMTQyMzEyMywicGF0aCI6Ii83MTkyNDA5Ny8yNjExNTIzNDktZmQyMDFiNjgtOWU3Zi00M2JmLWE0ZGMtZWRkNGZmZWM5MWZkLnBuZz9YLUFtei1BbGdvcml0aG09QVdTNC1ITUFDLVNIQTI1NiZYLUFtei1DcmVkZW50aWFsPUFLSUFWQ09EWUxTQTUzUFFLNFpBJTJGMjAyNDA3MTklMkZ1cy1lYXN0LTElMkZzMyUyRmF3czRfcmVxdWVzdCZYLUFtei1EYXRlPTIwMjQwNzE5VDIxMDUyM1omWC1BbXotRXhwaXJlcz0zMDAmWC1BbXotU2lnbmF0dXJlPTY0MDYwMjA4ODg1N2E3OWM5OTVhMWEyZGUxZWZjNDc1MWZkNDRjOTc4ZWU5MTkyMjkxOGNiOWZlZDY5MDlmZjImWC1BbXotU2lnbmVkSGVhZGVycz1ob3N0JmFjdG9yX2lkPTAma2V5X2lkPTAmcmVwb19pZD0wIn0.eOcN3utpfOhpuUpQApVettFgbSP4lwexVJoUpm91z4k)
The dataset used for fine-tuning the model is stored in a variable named fine_tuning_data
. It contains three columns: User
, Prompt
, and Text
.
The User
column contains the user inputs, while the Prompt
column contains the corresponding mid-journey prompts. The Text
column is generated using the following template: ### Instruction: <USER PROMPT> ### Response: <RESPONSE / MID-JOURNEY PROMPTS
This template adds special tokens ### Instruction:
and ### Response:
to the user input and mid-journey prompt, respectively. These special tokens are used by the model to distinguish between the instruction and response parts of the text.
The Llama-7B model is fine-tuned on the fine_tuning_data
dataset using standard techniques for fine-tuning language models. The fine-tuned model can then be used to generate mid-journey prompts given a user input.
The following bitsandbytes
quantization config was used during training:
- load_in_8bit: False
- load_in_4bit: True
- llm_int8_threshold: 6.0
- llm_int8_skip_modules: None
- llm_int8_enable_fp32_cpu_offload: False
- llm_int8_has_fp16_weight: False
- bnb_4bit_quant_type: nf4
- bnb_4bit_use_double_quant: True
- bnb_4bit_compute_dtype: bfloat16
- PEFT 0.4.0