summaryrefslogtreecommitdiff
path: root/scripts/test_armo_15666925.err
diff options
context:
space:
mode:
authorYurenHao0426 <blackhao0426@gmail.com>2026-01-27 12:15:45 -0600
committerYurenHao0426 <blackhao0426@gmail.com>2026-01-27 12:15:45 -0600
commit680513b7771a29f27cbbb3ffb009a69a913de6f9 (patch)
treea0d60aef9ade1b2953b915f535b990c0de95e493 /scripts/test_armo_15666925.err
parentc06ec2f3b80f8968f09eb801b69237495b055ec1 (diff)
local reward model
Diffstat (limited to 'scripts/test_armo_15666925.err')
-rw-r--r--scripts/test_armo_15666925.err31
1 files changed, 31 insertions, 0 deletions
diff --git a/scripts/test_armo_15666925.err b/scripts/test_armo_15666925.err
new file mode 100644
index 0000000..4c98df7
--- /dev/null
+++ b/scripts/test_armo_15666925.err
@@ -0,0 +1,31 @@
+/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/transformers/utils/hub.py:110: FutureWarning: Using `TRANSFORMERS_CACHE` is deprecated and will be removed in v5 of Transformers. Use `HF_HOME` instead.
+ warnings.warn(
+/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/huggingface_hub/file_download.py:798: UserWarning: Not enough free disk space to download the file. The expected file size is: 0.00 MB. The target location /work/hdd/bfqt/yurenh2/huggingface_cache/transformers/models--RLHFlow--ArmoRM-Llama3-8B-v0.1/blobs only has 0.00 MB free disk space.
+ warnings.warn(
+`torch_dtype` is deprecated! Use `dtype` instead!
+/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/huggingface_hub/file_download.py:798: UserWarning: Not enough free disk space to download the file. The expected file size is: 0.01 MB. The target location /work/hdd/bfqt/yurenh2/huggingface_cache/transformers/models--RLHFlow--ArmoRM-Llama3-8B-v0.1/blobs only has 0.00 MB free disk space.
+ warnings.warn(
+A new version of the following files was downloaded from https://huggingface.co/RLHFlow/ArmoRM-Llama3-8B-v0.1:
+- modeling_custom.py
+. Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision.
+Traceback (most recent call last):
+ File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/scripts/test_armo_reward.py", line 216, in <module>
+ main()
+ File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/scripts/test_armo_reward.py", line 42, in main
+ model.load()
+ File "/projects/bfqt/users/yurenh2/ml-projects/personalization-user-model/scripts/../src/personalization/feedback/armo_reward.py", line 81, in load
+ self._model = AutoModelForSequenceClassification.from_pretrained(
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/transformers/models/auto/auto_factory.py", line 586, in from_pretrained
+ model_class = get_class_from_dynamic_module(
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/transformers/dynamic_module_utils.py", line 616, in get_class_from_dynamic_module
+ return get_class_in_module(class_name, final_module, force_reload=force_download)
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+ File "/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/transformers/dynamic_module_utils.py", line 311, in get_class_in_module
+ module_spec.loader.exec_module(module)
+ File "<frozen importlib._bootstrap_external>", line 940, in exec_module
+ File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed
+ File "/projects/bfqt/users/yurenh2/hf_cache/huggingface/modules/transformers_modules/RLHFlow/ArmoRM_hyphen_Llama3_hyphen_8B_hyphen_v0_dot_1/eb2676d20da2f2d41082289d23c59b9f7427f955/modeling_custom.py", line 9, in <module>
+ from transformers.models.llama.modeling_llama import LLAMA_INPUTS_DOCSTRING
+ImportError: cannot import name 'LLAMA_INPUTS_DOCSTRING' from 'transformers.models.llama.modeling_llama' (/u/yurenh2/miniforge3/envs/eval/lib/python3.11/site-packages/transformers/models/llama/modeling_llama.py)