summaryrefslogtreecommitdiff
path: root/configs
diff options
context:
space:
mode:
Diffstat (limited to 'configs')
-rw-r--r--configs/a12_init_logit_3.yaml53
-rw-r--r--configs/a13_init_logit_0.yaml53
-rw-r--r--configs/a14_init_logit_1.yaml53
3 files changed, 159 insertions, 0 deletions
diff --git a/configs/a12_init_logit_3.yaml b/configs/a12_init_logit_3.yaml
new file mode 100644
index 0000000..bf7f6db
--- /dev/null
+++ b/configs/a12_init_logit_3.yaml
@@ -0,0 +1,53 @@
+# A12 — Init logit = 3.0 (moderate, sigmoid not saturated)
+# Purpose: A starts at σ(1.5)≈0.82, gradient ~250× larger than logit=15.
+# Does predictor learn useful topology when sigmoid is not saturated?
+# Run: python scripts/train.py --config configs/a12_init_logit_3.yaml
+
+# Model
+olmo_model_id: "allenai/OLMo-2-0425-1B"
+qwen_model_id: "Qwen/Qwen3-Embedding-0.6B"
+
+# Predictor
+predictor_hidden_dim: 1024
+predictor_rank: 32
+cascading_gate_k: 5.0
+input_norm: "none"
+init_logit: 3.0
+
+# Data
+dataset: "allenai/dolma"
+dataset_name: "v1_7"
+seq_len: 1024
+batch_size: 4
+micro_batch_size: 2
+qwen_input_prefix: ""
+
+# Eval
+eval_skip: 10000
+eval_size: 50
+
+# Training — ~50M tokens = 12500 steps
+total_steps: 12500
+lr: 3e-4
+weight_decay: 0.01
+optimizer: "adamw"
+
+# Schedules — constant tau=2 (same as S2), no sparsity
+tau_init: 2.0
+tau_final: 2.0
+tau_schedule: "cosine"
+lambda_max: 0.0
+lambda_warmup_frac: 0.2
+
+# Logging
+wandb_project: "dagformer"
+wandb_run_name: "a12-init-logit-3"
+log_every: 10
+eval_every: 500
+
+# Checkpointing
+save_every: 2500
+save_dir: "checkpoints/a12/"
+
+# Hardware
+num_gpus: 1
diff --git a/configs/a13_init_logit_0.yaml b/configs/a13_init_logit_0.yaml
new file mode 100644
index 0000000..5bd356d
--- /dev/null
+++ b/configs/a13_init_logit_0.yaml
@@ -0,0 +1,53 @@
+# A13 — Init logit = 0.0 (maximum gradient, A starts at 0.5)
+# Purpose: σ(0)=0.5, maximum gradient signal. Init NLL will be bad
+# but learning space is large. Tests if predictor can learn from scratch.
+# Run: python scripts/train.py --config configs/a13_init_logit_0.yaml
+
+# Model
+olmo_model_id: "allenai/OLMo-2-0425-1B"
+qwen_model_id: "Qwen/Qwen3-Embedding-0.6B"
+
+# Predictor
+predictor_hidden_dim: 1024
+predictor_rank: 32
+cascading_gate_k: 5.0
+input_norm: "none"
+init_logit: 0.0
+
+# Data
+dataset: "allenai/dolma"
+dataset_name: "v1_7"
+seq_len: 1024
+batch_size: 4
+micro_batch_size: 2
+qwen_input_prefix: ""
+
+# Eval
+eval_skip: 10000
+eval_size: 50
+
+# Training — ~50M tokens = 12500 steps
+total_steps: 12500
+lr: 3e-4
+weight_decay: 0.01
+optimizer: "adamw"
+
+# Schedules — constant tau=2 (same as S2), no sparsity
+tau_init: 2.0
+tau_final: 2.0
+tau_schedule: "cosine"
+lambda_max: 0.0
+lambda_warmup_frac: 0.2
+
+# Logging
+wandb_project: "dagformer"
+wandb_run_name: "a13-init-logit-0"
+log_every: 10
+eval_every: 500
+
+# Checkpointing
+save_every: 2500
+save_dir: "checkpoints/a13/"
+
+# Hardware
+num_gpus: 1
diff --git a/configs/a14_init_logit_1.yaml b/configs/a14_init_logit_1.yaml
new file mode 100644
index 0000000..f3278cf
--- /dev/null
+++ b/configs/a14_init_logit_1.yaml
@@ -0,0 +1,53 @@
+# A14 — Init logit = 1.0 (compromise: A≈0.62, strong gradient, mild dense bias)
+# Purpose: σ(0.5)≈0.62, strong gradient with slight bias toward connectivity.
+# Middle ground between A12 (too high?) and A13 (no bias at all).
+# Run: python scripts/train.py --config configs/a14_init_logit_1.yaml
+
+# Model
+olmo_model_id: "allenai/OLMo-2-0425-1B"
+qwen_model_id: "Qwen/Qwen3-Embedding-0.6B"
+
+# Predictor
+predictor_hidden_dim: 1024
+predictor_rank: 32
+cascading_gate_k: 5.0
+input_norm: "none"
+init_logit: 1.0
+
+# Data
+dataset: "allenai/dolma"
+dataset_name: "v1_7"
+seq_len: 1024
+batch_size: 4
+micro_batch_size: 2
+qwen_input_prefix: ""
+
+# Eval
+eval_skip: 10000
+eval_size: 50
+
+# Training — ~50M tokens = 12500 steps
+total_steps: 12500
+lr: 3e-4
+weight_decay: 0.01
+optimizer: "adamw"
+
+# Schedules — constant tau=2 (same as S2), no sparsity
+tau_init: 2.0
+tau_final: 2.0
+tau_schedule: "cosine"
+lambda_max: 0.0
+lambda_warmup_frac: 0.2
+
+# Logging
+wandb_project: "dagformer"
+wandb_run_name: "a14-init-logit-1"
+log_every: 10
+eval_every: 500
+
+# Checkpointing
+save_every: 2500
+save_dir: "checkpoints/a14/"
+
+# Hardware
+num_gpus: 1