8000 Fix the remainder of our configs by wizeng23 · Pull Request #1356 · oumi-ai/oumi · GitHub
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content

Fix the remainder of our configs #1356

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 3 commits into from
Feb 4, 2025
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion configs/examples/misc/dev_gcp_job.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@ resources:
# Multi GPU options: "A100:2", "L4:4"
accelerators: "A100:1"

disk_size: 200 # Disk size in GB
disk_size: 1000 # Disk size in GB
# region: us-central1 # Uncomment this line to only consider a specific region.

use_spot: false
Expand Down
4 changes: 1 addition & 3 deletions configs/projects/chatqa/chatqa_stage1_train.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -14,15 +14,14 @@ model:
# torch_dtype_str: "float16"
# model_max_length: 1024

# model_name: "meta-llama/Meta-Llama-3-8B-Instruct"
# model_name: "meta-llama/Llama-3.1-8B-Instruct"
# torch_dtype_str: "bfloat16"
# model_max_length: 4096

model_name: "microsoft/Phi-3-mini-4k-instruct"
torch_dtype_str: "bfloat16"
model_max_length: 4096

trust_remote_code: True
attn_implementation: "sdpa"

data:
Expand Down Expand Up @@ -50,7 +49,6 @@ training:
use_reentrant: False # So gradients flow from frozen (non FT) input embeddings

optimizer: "adamw_torch_fused"
# optimizer: "adamw_torch"
learning_rate: 0.000005
lr_scheduler_type: "cosine_with_min_lr"
lr_scheduler_kwargs:
Expand Down
1 change: 0 additions & 1 deletion configs/projects/chatqa/chatqa_stage2_train.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -80,7 +80,6 @@ training:
use_reentrant: False # So gradients flow from frozen (non FT) input embeddings

optimizer: "adamw_torch_fused"
# optimizer: "adamw_torch"
learning_rate: 0.000005
lr_scheduler_type: "cosine_with_min_lr"
lr_scheduler_kwargs:
Expand Down
2 changes: 2 additions & 0 deletions configs/recipes/llama3_1/sft/8b_full/longctx_train.yaml
8000
Original file line number Diff line number Diff line change
Expand Up @@ -47,6 +47,8 @@ data:
- dataset_name: "HuggingFaceFW/fineweb-edu"
subset: "sample-10BT"
split: "train"
dataset_kwargs:
seq_length: 32_768
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

For general info: curious how 32_768 notation is supported: is it done by YAML or Python ?

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I'm not sure exactly, copied what was above (which I think is from Oussama)

target_col: "text"
use_async_dataset: True
stream: True
Expand Down
1 change: 0 additions & 1 deletion configs/recipes/llama3_3/inference/70b_infer.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,6 @@

model:
model_name: "meta-llama/Llama-3.3-70B-Instruct"
adapter_model: null # Update for LoRA-tuned models.
model_max_length: 2048
torch_dtype_str: "bfloat16"
attn_implementation: "sdpa"
Expand Down
1 change: 0 additions & 1 deletion configs/recipes/llama3_3/inference/70b_vllm_infer.yaml
6D40
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,6 @@

model:
model_name: "meta-llama/Llama-3.3-70B-Instruct"
adapter_model: null # Update for LoRA-tuned models.
model_max_length: 2048
torch_dtype_str: "bfloat16"
attn_implementation: "sdpa"
Expand Down
4 changes: 2 additions & 2 deletions configs/recipes/phi3/dpo/nvidia_80g_train.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -24,8 +24,8 @@ training:
use_peft: true
output_dir: "output/phi3.dpo"
trainer_type: "TRL_DPO"
per_device_train_batch_size: 30 # Each batch seems to be approx. 1.8GB
gradient_accumulation_steps: 8
per_device_train_batch_size: 4
gradient_accumulation_steps: 64

# Use for debugging purposes
# max_steps: 10
Expand Down
Original file line number Diff line number Diff line change
@@ -1,5 +1,9 @@
# Quickstart Alpaca v2 eval config for SmolLM 135M Instruct.
#
# Requirements:
# - Run `pip install oumi[evaluation]` or `pip install alpaca_eval`
# - Set the env var `OPENAI_API_KEY` to your OpenAI API key.
#
# Usage:
# oumi evaluate -c configs/recipes/smollm/evaluation/135m/quickstart_alpaca_v2_eval.yaml
#
Expand All @@ -24,3 +28,5 @@ tasks:
- evaluation_platform: alpaca_eval
eval_kwargs:
version: 2.0

# engine: VLLM # Uncomment if you're running on a CUDA GPU.
4 changes: 2 additions & 2 deletions notebooks/Oumi - A Tour.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -83,7 +83,7 @@
"cell_type": "markdown",
"metadata": {},
"source": [
"**WARNING:** After the first `pip install`, you may have to restart the notebook for the package updates to take effect (Colab Menu: `Runtime` -> `Restart Session`)"
"**WARNING:** After the first `pip install`, you may have to restart the notebook for the package updates to take effect (Colab Menu: `Runtime` -> `Restart Session`)."
]
},
{
Expand Down Expand Up @@ -509,7 +509,7 @@
"name": "python",
"nbconvert_exporter": "python",
"pygments_lexer": "ipython3",
"version": "3.11.8"
"version": "3.11.11"
}
},
"nbformat": 4,
Expand Down
2 changes: 1 addition & 1 deletion notebooks/Oumi - Deploying a Job.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -70,7 +70,7 @@
"cell_type": "markdown",
"metadata": {},
"source": [
"**WARNING:** After the first `pip install`, you may have to restart the notebook for the package updates to take effect (Colab Menu: `Runtime` -> `Restart Session`)"
"**WARNING:** After the first `pip install`, you may have to restart the notebook for the package updates to take effect (Colab Menu: `Runtime` -> `Restart Session`)"
]
},
{
Expand Down
2 changes: 1 addition & 1 deletion notebooks/Oumi - Distill a Large Model.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -75,7 +75,7 @@
"cell_type": "markdown",
"metadata": {},
"source": [
"**WARNING:** After the first `pip install`, you may have to restart the notebook for the package updates to take effect (Colab Menu: `Runtime` -> `Restart Session`)"
"**WARNING:** After the first `pip install`, you may have to restart the notebook for the package updates to take effect (Colab Menu: `Runtime` -> `Restart Session`)"
]
},
{
Expand Down
2 changes: 1 addition & 1 deletion notebooks/Oumi - Finetuning Tutorial.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -68,7 +68,7 @@
"cell_type": "markdown",
"metadata": {},
"source": [
"**WARNING:** After the first `pip install`, you may have to restart the notebook for the package updates to take effect (Colab Menu: `Runtime` -> `Restart Session`)"
"**WARNING:** After the first `pip install`, you may have to restart the notebook for the package updates to take effect (Colab Menu: `Runtime` -> `Restart Session`)"
]
},
{
Expand Down
2 changes: 1 addition & 1 deletion notebooks/Oumi - Launching Jobs on Custom Clusters.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -67,7 +67,7 @@
"cell_type": "markdown",
"metadata": {},
"source": [
"**WARNING:** After the first `pip install`, you may have to restart the notebook for the package updates to take effect (Colab Menu: `Runtime` -> `Restart Session`)"
"**WARNING:** After the first `pip install`, you may have to restart the notebook for the package updates to take effect (Colab Menu: `Runtime` -> `Restart Session`)"
]
},
{
Expand Down
2 changes: 1 addition & 1 deletion notebooks/Oumi - Running Jobs Remotely.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -93,7 +93,7 @@
"cell_type": "markdown",
"metadata": {},
"source": [
"**WARNING:** After the first `pip install`, you may have to restart the notebook for the package updates to take effect (Colab Menu: `Runtime` -> `Restart Session`)"
"**WARNING:** After the first `pip install`, you may have to restart the notebook for the package updates to take effect (Colab Menu: `Runtime` -> `Restart Session`)"
]
},
{
Expand Down
2 changes: 1 addition & 1 deletion notebooks/Oumi - Training CNN on Custom Dataset.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -67,7 +67,7 @@
"cell_type": "markdown",
"metadata": {},
"source": [
"**WARNING:** After the first `pip install`, you may have to restart the notebook for the package updates to take effect (Colab Menu: `Runtime` -> `Restart Session`)"
"**WARNING:** After the first `pip install`, you may have to restart the notebook for the package updates to take effect (Colab Menu: `Runtime` -> `Restart Session`)"
]
},
{
Expand Down
2 changes: 1 addition & 1 deletion notebooks/Oumi - Using vLLM Engine for Inference.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -73,7 +73,7 @@
"cell_type": "markdown",
"metadata": {},
"source": [
"**WARNING:** After the first `pip install`, you may have to restart the notebook for the package updates to take effect (Colab Menu: `Runtime` -> `Restart Session`)"
"**WARNING:** After the first `pip install`, you may have to restart the notebook for the package updates to take effect (Colab Menu: `Runtime` -> `Restart Session`)"
]
},
{
Expand Down
2 changes: 1 addition & 1 deletion notebooks/Oumi - Vision Language Models.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -126,7 +126,7 @@
"cell_type": "markdown",
"metadata": {},
"source": [
"**WARNING:** After the first `pip install`, you may have to restart the notebook for the package updates to take effect (Colab Menu: `Runtime` -> `Restart Session`)"
"**WARNING:** After the first `pip install`, you may have to restart the notebook for the package updates to take effect (Colab Menu: `Runtime` -> `Restart Session`)"
]
},
{
Expand Down
5 changes: 2 additions & 3 deletions src/experimental/configs/projects/zephyr/evaluation/eval.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -10,11 +10,10 @@
# - Other eval configs: configs/**/evaluation/

model:
# Note: update model_name or adapter_model with your checkpoint directories
# Note: Update with your FFT or LoRA checkpoint
model_name: "mistralai/Mistral-7B-v0.1"
adapter_model: "your-checkpoint-dir"
trust_remote_code: True
#TODO currently below are not passed in LM_HARNESS
# TODO: currently below are not passed in LM_HARNESS
torch_dtype_str: "bfloat16"
model_max_length: 2048
chat_template: "zephyr"
Expand Down
0