summaryrefslogtreecommitdiff
path: root/configs
diff options
context:
space:
mode:
authorYurenHao0426 <blackhao0426@gmail.com>2026-01-27 09:57:37 -0600
committerYurenHao0426 <blackhao0426@gmail.com>2026-01-27 09:57:37 -0600
commitdc801c07cf38b0c495686463e6ca6f871a64440e (patch)
tree599f03114775921dbc472403c701f4a3a8ea188a /configs
parente43b3f8aa36c198b95c1e46bea2eaf3893b13dc3 (diff)
Add collaborativeagents module and update gitignore
- Add collaborativeagents subproject with adapters, agents, and evaluation modules - Update .gitignore to exclude large binary files (.whl, .tar), wandb logs, and results Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com>
Diffstat (limited to 'configs')
-rw-r--r--configs/local_models.yaml39
1 files changed, 27 insertions, 12 deletions
diff --git a/configs/local_models.yaml b/configs/local_models.yaml
index 13c7fcc..8f91955 100644
--- a/configs/local_models.yaml
+++ b/configs/local_models.yaml
@@ -1,36 +1,46 @@
+# Base path for all models
+_base_path: &base /projects/bfqt/users/yurenh2/ml-projects/personalization-user-model
+
models:
llm:
# New Multi-Backend Config
qwen_1_5b:
backend: qwen
- path: models/qwen2.5-1.5b-instruct
+ path: /projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/models/qwen2.5-1.5b-instruct
device: auto
dtype: bfloat16
max_context_length: 4096
-
+
llama_8b:
backend: llama
- path: models/llama-3.1-8b-instruct
+ path: /projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/models/llama-3.1-8b-instruct
device: auto
dtype: bfloat16
max_context_length: 8192
+ # vLLM backend for high-throughput experiments
+ llama_8b_vllm:
+ backend: vllm
+ path: /projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/models/llama-3.1-8b-instruct
+ vllm_url: http://localhost:8003/v1
+ max_context_length: 8192
+
# Legacy fallback (needed if from_config is called directly without name)
hf_id: Qwen/Qwen2.5-1.5B-Instruct
- local_path: models/qwen2.5-1.5b-instruct
+ local_path: /projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/models/qwen2.5-1.5b-instruct
dtype: bfloat16
device_map: auto
preference_extractor:
# Default/Legacy
default:
- hf_id: Qwen/Qwen2.5-0.5B-Instruct
- local_path: models/qwen2.5-0.5b-instruct
- dtype: bfloat16
- device_map: auto
+ hf_id: Qwen/Qwen2.5-0.5B-Instruct
+ local_path: /projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/models/qwen2.5-0.5b-instruct
+ dtype: bfloat16
+ device_map: auto
# New SFT Extractor
qwen3_0_6b_sft:
- path: saves/qwen3-0.6b-full-sft-h200/checkpoint-4358
+ path: /projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/saves/qwen3-0.6b-full-sft-h200/checkpoint-4358
prompt_template_path: fine_tuning_prompt_template.txt
device: auto
dtype: bfloat16
@@ -38,13 +48,18 @@ models:
embedding:
qwen3:
hf_id: Qwen/Qwen3-Embedding-8B
- local_path: models/qwen3-embedding-8b
+ local_path: /projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/models/qwen3-embedding-8b
nemotron:
hf_id: nvidia/llama-embed-nemotron-8b
- local_path: models/llama-embed-nemotron-8b
+ local_path: /projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/models/llama-embed-nemotron-8b
reranker:
qwen3_8b:
hf_id: Qwen/Qwen3-Reranker-8B
- local_path: models/rerankers/qwen3-reranker-8b
+ local_path: /projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/models/rerankers/qwen3-reranker-8b
dtype: bfloat16
device_map: auto
+ bge_base:
+ hf_id: BAAI/bge-reranker-base
+ local_path: /projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/models/rerankers/bge-reranker-base
+ dtype: float16
+ device_map: auto